Re: Problems with ceph command - Octupus - Ubuntu 16.04

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Your 'cephadm ls' output was only from one node, I assumed you just bootstrapped the first node.

The 'cephadm logs' command should provide pager-output so you can scroll or search for a specific date.

I'm not sure what caused this but "error on write" is bad. As I already wrote check the filesystems on your nodes, dmesg etc. It seems as if two of your MONs are down which would make your cluster unavailable (no quorum). Is mon3 up and running? Bringing back one of the other two MONs would bring the cluster back up.


Zitat von Emanuel Alejandro Castelli <ecastelli@xxxxxxxxxxxxxxxxx>:

From MON1, dmesg I get this:

[3348025.306195] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348033.241973] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348048.089325] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348049.209243] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348050.201209] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348052.185167] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348056.280992] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348064.216703] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348078.808431] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348079.192418] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348080.220345] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348082.232299] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348086.232103] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348094.167722] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348110.411216] libceph: mon0 192.168.14.150:6789 socket closed (con state OPEN) [3348140.245900] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348141.173884] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348142.229859] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348144.213777] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348148.437674] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348157.397327] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348170.965496] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348172.213118] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348173.205087] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348175.188934] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348179.412719] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348187.348441] libceph: mon1 192.168.14.151:6789 socket closed (con state CONNECTING) [3348201.683707] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348202.195745] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348203.187654] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348205.175585] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348209.363409] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [3348217.299298] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING)

But....from MON2 I get this:

[5242753.074620] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242761.266727] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242779.959468] libceph: mon0 192.168.14.150:6789 socket closed (con state OPEN)
[5242806.834049] libceph: mon1 192.168.14.151:6789 socket error on write
[5242808.049952] libceph: mon1 192.168.14.151:6789 socket error on write
[5242809.041947] libceph: mon1 192.168.14.151:6789 socket error on write
[5242811.057917] libceph: mon1 192.168.14.151:6789 socket error on write
[5242815.285867] libceph: mon1 192.168.14.151:6789 socket error on write
[5242824.241921] libceph: mon1 192.168.14.151:6789 socket error on write
[5242837.554174] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242838.034339] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242839.026139] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242841.010177] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242845.234101] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242853.169905] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242870.102324] libceph: mon0 192.168.14.150:6789 socket closed (con state OPEN) [5242901.041812] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242902.033763] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242903.026350] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242905.009497] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242909.233740] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242917.169724] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING)
[5242931.761103] libceph: mon1 192.168.14.151:6789 socket error on write
[5242932.049095] libceph: mon1 192.168.14.151:6789 socket error on write
[5242933.041234] libceph: mon1 192.168.14.151:6789 socket error on write
[5242935.057066] libceph: mon1 192.168.14.151:6789 socket error on write
[5242939.185290] libceph: mon1 192.168.14.151:6789 socket error on write
[5242947.120965] libceph: mon1 192.168.14.151:6789 socket error on write
[5242962.481572] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242963.025508] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242964.017160] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING) [5242966.033128] libceph: mon2 192.168.14.152:6789 socket closed (con state CONNECTING)


Saludos,



EMANUEL CASTELLI

Arquitecto de Información - Gerencia OSS

C: (+549) 116707-4107 | Interno: 1325 | T-Phone: 7510-1325 | ecastelli@xxxxxxxxxxxxxxxxx

Lavardén 157 1er piso. CABA (C1437FBC)

----- Original Message -----
From: "Emanuel Alejandro Castelli" <ecastelli@xxxxxxxxxxxxxxxxx>
To: "Eugen Block" <eblock@xxxxxx>
Cc: "ceph-users" <ceph-users@xxxxxxx>
Sent: Tuesday, October 20, 2020 10:24:31 AM
Subject: Re: Re: Problems with ceph command - Octupus - Ubuntu 16.04

I have 3 MON, I don't know why it's showing only one.

root@osswrkprbe001:~# ceph --connect-timeout 60 status
Cluster connection interrupted or timed out

cephadm logs --name mon.osswrkprbe001 --> Is there any way to go to a specific date? Because it stars from Oct 4. I want to check from Oct 16 and ahead. I suspect that something happened that day.

Also, I don't know how to troubleshoot this. I did the same (./cephadm logs --name mon.osswrkprbe002) in the second MON but it starts the logs from Sep 30. I would need to check Oct 16 also.

I would appreciate if you can help me with the troubleshooting.

Thank you.

Saludos,



EMANUEL CASTELLI

Arquitecto de Información - Gerencia OSS

C: (+549) 116707-4107 | Interno: 1325 | T-Phone: 7510-1325 | ecastelli@xxxxxxxxxxxxxxxxx

Lavardén 157 1er piso. CABA (C1437FBC)

----- Original Message -----
From: "Eugen Block" <eblock@xxxxxx>
To: "ceph-users" <ceph-users@xxxxxxx>
Sent: Tuesday, October 20, 2020 10:02:35 AM
Subject:  Re: Problems with ceph command - Octupus - Ubuntu 16.04

Your mon container seems up and running, have you tried restarting it?
You just have one mon, is that correct? Do you see anything in the logs?

cephadm logs --name mon.osswrkprbe001

How long do you wait until you hit CTRL-C? There's a
connection-timeout option for ceph commands, maybe try a higher timeout?

ceph --connect-timeout 60 status

Is the node hosting the mon showing any issues in dmesg, df -h, syslog, etc.?

Regards,
Eugen


Zitat von Emanuel Alejandro Castelli <ecastelli@xxxxxxxxxxxxxxxxx>:

Hello


I'm facing an issue with ceph. I cannot run any ceph command. It
literally hangs. I need to hit CTRL-C to get this:




^CCluster connection interrupted or timed out




This is on Ubuntu 16.04. Also, I use Graphana with Prometheus to get
information from the cluster, but now there is no data to graph. Any
clue?


BQ_BEGIN


cephadm version
BQ_END

BQ_BEGIN


INFO:cephadm:Using recent ceph image ceph/ceph:v15 ceph version
15.2.4 (7447c15c6ff58d7fce91843b705a268a1917325c) octopus (stable)
BQ_END

cephadm ls
[
    {
        "style": "cephadm:v1",
        "name": "mon.osswrkprbe001",
        "fsid": "56820176-ae5b-4e58-84a2-442b2fc03e6d",
        "systemd_unit":
"ceph-56820176-ae5b-4e58-84a2-442b2fc03e6d@mon.osswrkprbe001",
        "enabled": true,
        "state": "running",
        "container_id":
"afbe6ef76198bf05ec972e832077849d4a4438bd56f2e177aeb9b11146577baf",
        "container_image_name": "docker.io/ceph/ceph:v15.2.1",
        "container_image_id":
"bc83a388465f0568dab4501fb7684398dca8b50ca12a342a57f21815721723c2",
        "version": "15.2.1",
        "started": "2020-10-19T19:03:16.759730",
        "created": "2020-09-04T23:30:30.250336",
        "deployed": "2020-09-04T23:48:20.956277",
        "configured": "2020-09-04T23:48:22.100283"
    },
    {
        "style": "cephadm:v1",
        "name": "mgr.osswrkprbe001",
        "fsid": "56820176-ae5b-4e58-84a2-442b2fc03e6d",
        "systemd_unit":
"ceph-56820176-ae5b-4e58-84a2-442b2fc03e6d@mgr.osswrkprbe001",
        "enabled": true,
        "state": "running",
        "container_id":
"1737b2cf46310025c0ae853c3b48400320fb35b0443f6ab3ef3d6cbb10f460d8",
        "container_image_name": "docker.io/ceph/ceph:v15.2.1",
        "container_image_id":
"bc83a388465f0568dab4501fb7684398dca8b50ca12a342a57f21815721723c2",
        "version": "15.2.1",
        "started": "2020-10-19T20:43:38.329529",
        "created": "2020-09-04T23:30:31.110341",
        "deployed": "2020-09-04T23:47:41.604057",
        "configured": "2020-09-05T00:00:21.064246"
    }
]


Thank you in advance.


Saludos,



EMANUEL CASTELLI

Arquitecto de Información - Gerencia OSS

C: (+549) 116707-4107 | Interno: 1325 | T-Phone: 7510-1325 |
ecastelli@xxxxxxxxxxxxxxxxx

Lavardén 157 1er piso. CABA (C1437FBC)


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux