Hello list, i upgraded to Debian 10, after that i upgraded from luminous to nautilus. I restarted the mons, then the OSDs. Everything was up and healthy. After rebooting a node, only 3/10 OSD start up: -4 20.07686 host ceph03 4 hdd 2.67020 osd.4 down 1.00000 1.00000 5 hdd 1.71660 osd.5 up 1.00000 1.00000 6 hdd 1.71660 osd.6 up 1.00000 1.00000 10 hdd 2.67029 osd.10 up 1.00000 1.00000 15 hdd 2.00000 osd.15 down 1.00000 1.00000 17 hdd 1.20000 osd.17 down 1.00000 1.00000 20 hdd 1.71649 osd.20 down 1.00000 1.00000 21 hdd 2.00000 osd.21 down 1.00000 1.00000 27 hdd 1.71649 osd.27 down 1.00000 1.00000 32 hdd 2.67020 osd.32 down 1.00000 1.00000 root@ceph03:~# /usr/bin/ceph-osd -f --cluster ceph --id 32 --setuser ceph --setgroup ceph 2020-03-25 15:46:36.330 7efddde5ec80 -1 auth: unable to find a keyring on /var/lib/ceph/osd/ceph-32/keyring: (2) No such file or directory 2020-03-25 15:46:36.330 7efddde5ec80 -1 AuthRegistry(0x56531c50a140) no keyring found at /var/lib/ceph/osd/ceph-32/keyring, disabling cephx 2020-03-25 15:46:36.330 7efddde5ec80 -1 auth: unable to find a keyring on /var/lib/ceph/osd/ceph-32/keyring: (2) No such file or directory 2020-03-25 15:46:36.330 7efddde5ec80 -1 AuthRegistry(0x7ffd04120468) no keyring found at /var/lib/ceph/osd/ceph-32/keyring, disabling cephx failed to fetch mon config (--no-mon-config to skip) root@ceph03:~# df Filesystem 1K-blocks Used Available Use% Mounted on udev 24624580 0 24624580 0% /dev tmpfs 4928216 9544 4918672 1% /run /dev/sda3 47930248 5209760 40262684 12% / tmpfs 24641068 0 24641068 0% /dev/shm tmpfs 5120 0 5120 0% /run/lock tmpfs 24641068 0 24641068 0% /sys/fs/cgroup /dev/sda1 944120 144752 734192 17% /boot tmpfs 24641068 24 24641044 1% /var/lib/ceph/osd/ceph-1 tmpfs 24641068 24 24641044 1% /var/lib/ceph/osd/ceph-6 tmpfs 24641068 24 24641044 1% /var/lib/ceph/osd/ceph-5 tmpfs 24641068 24 24641044 1% /var/lib/ceph/osd/ceph-10 tmpfs 4928212 0 4928212 0% /run/user/0 root@ceph03:~# ceph-volume lvm list ====== osd.1 ======= [block] /dev/ceph-9af8fc69-cab8-4c12-b51e-5746a0f0fc51/osd-block-b4987093-4fa5-47bd-8ddc-102b98444067 block device /dev/ceph-9af8fc69-cab8-4c12-b51e-5746a0f0fc51/osd-block-b4987093-4fa5-47bd-8ddc-102b98444067 block uuid HSK6Da-elP2-CFYz-s0RH-UNiw-bey0-dVcml1 cephx lockbox secret cluster fsid 5436dd5d-83d4-4dc8-a93b-60ab5db145df cluster name ceph crush device class None encrypted 0 osd fsid b4987093-4fa5-47bd-8ddc-102b98444067 osd id 1 type block vdo 0 devices /dev/sdj ====== osd.10 ====== [block] /dev/ceph-78f2730d-7277-4d1f-8909-449b45339f80/osd-block-fa241441-1758-4b85-9799-988eee3b2b3f block device /dev/ceph-78f2730d-7277-4d1f-8909-449b45339f80/osd-block-fa241441-1758-4b85-9799-988eee3b2b3f block uuid 440fNG-guO2-l1WJ-m5cR-GUkz-ZTUd-Fcz5Ml cephx lockbox secret cluster fsid 5436dd5d-83d4-4dc8-a93b-60ab5db145df cluster name ceph crush device class None encrypted 0 osd fsid fa241441-1758-4b85-9799-988eee3b2b3f osd id 10 type block vdo 0 devices /dev/sdl ====== osd.5 ======= [block] /dev/ceph-793608ca-9dd1-4a4f-a776-c1e292127899/osd-block-112e0c75-f61b-4e50-9bb5-775bacd854af block device /dev/ceph-793608ca-9dd1-4a4f-a776-c1e292127899/osd-block-112e0c75-f61b-4e50-9bb5-775bacd854af block uuid Z6VeNx-S9sg-ZOsh-HTw9-ykTc-YBrh-qFwz5i cephx lockbox secret cluster fsid 5436dd5d-83d4-4dc8-a93b-60ab5db145df cluster name ceph crush device class None encrypted 0 osd fsid 112e0c75-f61b-4e50-9bb5-775bacd854af osd id 5 type block vdo 0 devices /dev/sdb ====== osd.6 ======= [block] /dev/ceph-4b0cee89-03f4-4853-bc1d-09e0eb772799/osd-block-35288829-c1f6-42ab-aeb0-f2915a389e48 block device /dev/ceph-4b0cee89-03f4-4853-bc1d-09e0eb772799/osd-block-35288829-c1f6-42ab-aeb0-f2915a389e48 block uuid G9qHxC-dN0b-XBes-QVss-Bzwa-7Xtw-ikksgM cephx lockbox secret cluster fsid 5436dd5d-83d4-4dc8-a93b-60ab5db145df cluster name ceph crush device class None encrypted 0 osd fsid 35288829-c1f6-42ab-aeb0-f2915a389e48 osd id 6 type block vdo 0 devices /dev/sdc I would mount it and run the osd deamon manually, but ceph-disk list seems to be gone in Nautilus. Therefore i dont know where to mount what. Any ideas on that? Cheers, Michael _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx