Re: pacific: ceph-mon services stopped after OSDs are out/down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



There's an existing tracker issue [1] that hasn't been updated since a year. The OP reported that restarting the other MONs did resolve it, have you tried that?

[1] https://tracker.ceph.com/issues/52760

Zitat von Mevludin Blazevic <mblazevic@xxxxxxxxxxxxxx>:

Its very strange. The keyring of the ceph monitor is the same as on one of the working monitor hosts. The failed mon and the working mons also have the same selinux policies and firewalld settings. The connection is also present since, all osd deamons are up on the failed ceph monitor node.

Am 13.12.2022 um 11:43 schrieb Eugen Block:
So you get "Permission denied" errors, I'm guessing either the mon keyring is not present (or wrong) or the mon directory doesn't belong to the ceph user. Can you check

ls -l /var/lib/ceph/FSID/mon.sparci-store1/

Compare the keyring file with the ones on the working mon nodes.

Zitat von Mevludin Blazevic <mblazevic@xxxxxxxxxxxxxx>:

Hi Eugen,

I assume the mon db is stored on the "OS disk". I could not find any error related lines in cephadm.log, here is what journalctl -xe tells me:

Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.392+0000 7f318e1fa700  1 mon.sparci-store1@-1(???).paxosservice(auth 251..491) refresh upgraded, format 0 -> 3 Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.397+0000 7f3179248700  1 heartbeat_map reset_timeout 'Monitor::cpu_tp thread 0x7f3179248700' had timed out after 0.000000000s Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.397+0000 7f318e1fa700  0 mon.sparci-store1@-1(probing) e5  my rank is now 1 (was -1) Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.398+0000 7f317ba4d700 -1 mon.sparci-store1@1(probing) e5 handle_auth_bad_method hmm, they didn't like 2 result (13) Permission denied Dec 13 11:24:21 sparci-store1 systemd[1]: Started Ceph mon.sparci-store1 for 8c774934-1535-11ec-973e-525400130e4f. -- Subject: Unit ceph-8c774934-1535-11ec-973e-525400130e4f@mon.sparci-store1.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit ceph-8c774934-1535-11ec-973e-525400130e4f@mon.sparci-store1.service has finished starting up.
--
-- The start-up result is done.
Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.599+0000 7f317ba4d700 -1 mon.sparci-store1@1(probing) e5 handle_auth_bad_method hmm, they didn't like 2 result (13) Permission denied Dec 13 11:24:21 sparci-store1 ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1[786211]: debug 2022-12-13T10:24:21.600+0000 7f3177a45700  0 mon.sparci-store1@1(probing) e18  removed from monmap, suicide. Dec 13 11:24:21 sparci-store1 systemd[1]: var-lib-containers-storage-overlay-2e67bce8ea3795683c4326479c7169a713e9a7630b31f25d60cd45bbd9fa56bd-merged.mount: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit var-lib-containers-storage-overlay-2e67bce8ea3795683c4326479c7169a713e9a7630b31f25d60cd45bbd9fa56bd-merged.mount has successfully entered the 'dead' state. Dec 13 11:24:21 sparci-store1 bash[786318]: Error: no container with name or ID "ceph-8c774934-1535-11ec-973e-525400130e4f-mon.sparci-store1" found: no such container Dec 13 11:24:21 sparci-store1 bash[786346]: Error: no container with name or ID "ceph-8c774934-1535-11ec-973e-525400130e4f-mon-sparci-store1" found: no such container Dec 13 11:24:21 sparci-store1 bash[786375]: Error: no container with name or ID "ceph-8c774934-1535-11ec-973e-525400130e4f-mon.sparci-store1" found: no such container Dec 13 11:24:21 sparci-store1 systemd[1]: ceph-8c774934-1535-11ec-973e-525400130e4f@mon.sparci-store1.service: Succeeded.
-- Subject: Unit succeeded
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit ceph-8c774934-1535-11ec-973e-525400130e4f@mon.sparci-store1.service has successfully entered the 'dead' state.

Regards,

Mevludin


Am 08.12.2022 um 09:30 schrieb Eugen Block:
Hi,

do the MONs use the same SAS interface? They store the mon db on local disk, so it might be related. But without any logs or more details it's just guessing.

Regards,
Eugen

Zitat von Mevludin Blazevic <mblazevic@xxxxxxxxxxxxxx>:

Hi all,

I'm running Pacific with cephadm.

After installation, ceph automatically provisoned 5 ceph monitor nodes across the cluster. After a few OSDs crashed due to a hardware related issue with the SAS interface, 3 monitor services are stopped and won't restart again. Is it related to the OSD crash problem?

Thanks,
Mevludin

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

--
Mevludin Blazevic, M.Sc.

University of Koblenz-Landau
Computing Centre (GHRKO)
Universitaetsstrasse 1
D-56070 Koblenz, Germany
Room A023
Tel: +49 261/287-1326

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

--
Mevludin Blazevic, M.Sc.

University of Koblenz-Landau
Computing Centre (GHRKO)
Universitaetsstrasse 1
D-56070 Koblenz, Germany
Room A023
Tel: +49 261/287-1326

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux