Following is the error: Apr 4 19:09:39 osd03 systemd[1]: ceph-osd@2.service: Scheduled restart job, restart counter is at 3. Apr 4 19:09:39 osd03 systemd[1]: Stopped Ceph object storage daemon osd.2. Apr 4 19:09:39 osd03 systemd[1]: Starting Ceph object storage daemon osd.2... Apr 4 19:09:39 osd03 systemd[1]: Started Ceph object storage daemon osd.2. Apr 4 19:09:39 osd03 ceph-osd[3272]: 2021-04-04T19:09:39.668+0430 7fb5d7819f00 -1 bluestore(/var/lib/ceph/osd/ceph-2/block) _read_bdev_label failed to open /var/lib/ceph/osd/ceph-2/block: (1) Operation not permitted Apr 4 19:09:39 osd03 ceph-osd[3272]: 2021-04-04T19:09:39.668+0430 7fb5d7819f00 -1 #033[0;31m ** ERROR: unable to open OSD superblock on /var/lib/ceph/osd/ceph-2: (2) No such file or directory#033[0m Apr 4 19:09:39 osd03 systemd[1]: ceph-osd@2.service: Main process exited, code=exited, status=1/FAILURE Apr 4 19:09:39 osd03 systemd[1]: ceph-osd@2.service: Failed with result 'exit-code'. And this the files that shows the file exists: root@osd03:/var/lib/ceph/osd/ceph-2# ls -l total 24 lrwxrwxrwx 1 ceph ceph 93 Apr 4 19:09 block -> /dev/ceph-9d37674b-a269-4239-aa9e-66a3c74df76c/osd-block-2d3ffc61-e430-4b89-bcd4-105b2df26352 -rw------- 1 ceph ceph 37 Apr 4 19:09 ceph_fsid -rw------- 1 ceph ceph 37 Apr 4 19:09 fsid -rw------- 1 ceph ceph 55 Apr 4 19:09 keyring -rw------- 1 ceph ceph 6 Apr 4 19:09 ready -rw------- 1 ceph ceph 10 Apr 4 19:09 type -rw------- 1 ceph ceph 2 Apr 4 19:09 whoami root@osd03:/var/lib/ceph/osd/ceph-2# root@osd03:/var/lib/ceph/osd/ceph-2# On Sun, Apr 4, 2021 at 7:06 PM Andrew Walker-Brown < andrew_jbrown@xxxxxxxxxxx> wrote: > And after a reboot what errors are you getting? > > Sent from my iPhone > > On 4 Apr 2021, at 15:33, Behzad Khoshbakhti <khoshbakhtib@xxxxxxxxx> > wrote: > > > I have changed the uid and gid to 167, but still no progress. > cat /etc/group | grep -i ceph > ceph:x:167: > root@osd03:~# cat /etc/passwd | grep -i ceph > ceph:x:167:167:Ceph storage service:/var/lib/ceph:/usr/sbin/nologin > > On Sun, Apr 4, 2021 at 6:47 PM Andrew Walker-Brown < > andrew_jbrown@xxxxxxxxxxx> wrote: > >> UID and guid should both be 167 I believe. >> >> Make a note of the current values and change them to 167 using usermod >> and groupmod. >> >> I had just this issue. It’s partly to do with how perms are used within >> the containers I think. >> >> I changed the values to 167 in passwd everything worked again. Symptoms >> for me were OSDs not starting and permissions/file not found errors. >> >> Sent from my iPhone >> >> On 4 Apr 2021, at 13:43, Lomayani S. Laizer <lomlaizer@xxxxxxxxx> wrote: >> >> >> Hello, >> Permissions are correct. guid/uid is 64045/64045 >> >> ls -alh >> total 32K >> drwxrwxrwt 2 ceph ceph 200 Apr 4 14:11 . >> drwxr-xr-x 8 ceph ceph 4.0K Sep 18 2018 .. >> lrwxrwxrwx 1 ceph ceph 93 Apr 4 14:11 block -> /dev/... >> -rw------- 1 ceph ceph 37 Apr 4 14:11 ceph_fsid >> -rw------- 1 ceph ceph 37 Apr 4 14:11 fsid >> -rw------- 1 ceph ceph 56 Apr 4 14:11 keyring >> -rw------- 1 ceph ceph 6 Apr 4 14:11 ready >> -rw------- 1 ceph ceph 3 Apr 4 14:11 require_osd_release >> -rw------- 1 ceph ceph 10 Apr 4 14:11 type >> -rw------- 1 ceph ceph 3 Apr 4 14:11 whoami >> >> On Sun, Apr 4, 2021 at 3:07 PM Andrew Walker-Brown < >> andrew_jbrown@xxxxxxxxxxx> wrote: >> >>> Are the file permissions correct and UID/guid in passwd both 167? >>> >>> Sent from my iPhone >>> >>> On 4 Apr 2021, at 12:29, Lomayani S. Laizer <lomlaizer@xxxxxxxxx> wrote: >>> >>> Hello, >>> >>> +1 Am facing the same problem in ubuntu after upgrade to pacific >>> >>> 2021-04-03T10:36:07.698+0300 7f9b8d075f00 -1 bluestore(/var/lib/ceph/osd/ >>> ceph-29/block) _read_bdev_label failed to open >>> /var/lib/ceph/osd/ceph-29/block: >>> (1) Operation not permitted >>> 2021-04-03T10:36:07.698+0300 7f9b8d075f00 -1 ESC[0;31m ** ERROR: unable >>> to >>> open OSD superblock on /var/lib/ceph/osd/ceph-29: (2) No such file or >>> directoryESC[0m >>> >>> On Sun, Apr 4, 2021 at 1:52 PM Behzad Khoshbakhti < >>> khoshbakhtib@xxxxxxxxx> >>> wrote: >>> >>> > It worth mentioning as I issue the following command, the Ceph OSD >>> starts >>> > and joins the cluster: >>> > /usr/bin/ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup >>> ceph >>> > >>> > >>> > >>> > On Sun, Apr 4, 2021 at 3:00 PM Behzad Khoshbakhti < >>> khoshbakhtib@xxxxxxxxx> >>> > wrote: >>> > >>> >> Hi all, >>> >> >>> >> As I have upgrade my Ceph cluster from 15.2.10 to 16.2.0, during the >>> >> manual upgrade using the precompiled packages, the OSDs was down with >>> the >>> >> following messages: >>> >> >>> >> root@osd03:/var/lib/ceph/osd/ceph-2# ceph-volume lvm activate --all >>> >> --> Activating OSD ID 2 FSID 2d3ffc61-e430-4b89-bcd4-105b2df26352 >>> >> Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 >>> >> Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph >>> > prime-osd-dir >>> >> --dev >>> >> >>> > >>> /dev/ceph-9d37674b-a269-4239-aa9e-66a3c74df76c/osd-block-2d3ffc61-e430-4b89-bcd4-105b2df26352 >>> >> --path /var/lib/ceph/osd/ceph-2 --no-mon-config >>> >> Running command: /usr/bin/ln -snf >>> >> >>> > >>> /dev/ceph-9d37674b-a269-4239-aa9e-66a3c74df76c/osd-block-2d3ffc61-e430-4b89-bcd4-105b2df26352 >>> >> /var/lib/ceph/osd/ceph-2/block >>> >> Running command: /usr/bin/chown -h ceph:ceph >>> > /var/lib/ceph/osd/ceph-2/block >>> >> Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 >>> >> Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 >>> >> Running command: /usr/bin/systemctl enable >>> >> ceph-volume@lvm-2-2d3ffc61-e430-4b89-bcd4-105b2df26352 >>> >> Running command: /usr/bin/systemctl enable --runtime ceph-osd@2 >>> >> Running command: /usr/bin/systemctl start ceph-osd@2 >>> >> --> ceph-volume lvm activate successful for osd ID: 2 >>> >> >>> >> Content of /var/log/ceph/ceph-osd.2.log >>> >> 2021-04-04T14:54:56.625+0430 7f4afbac0f00 0 set uid:gid to >>> 64045:64045 >>> >> (ceph:ceph) >>> >> 2021-04-04T14:54:56.625+0430 7f4afbac0f00 0 ceph version 16.2.0 >>> >> (0c2054e95bcd9b30fdd908a79ac1d8bbc3394442) pacific (stable), process >>> >> ceph-osd, pid 5484 >>> >> 2021-04-04T14:54:56.625+0430 7f4afbac0f00 0 pidfile_write: ignore >>> empty >>> >> --pid-file >>> >> 2021-04-04T14:54:56.625+0430 7f4afbac0f00 -1* >>> >> bluestore(/var/lib/ceph/osd/ceph-2/block) _read_bdev_label failed to >>> open >>> >> /var/lib/ceph/osd/ceph-2/block: (1) Operation not permitted* >>> >> 2021-04-04T14:54:56.625+0430 7f4afbac0f00 -1 *** ERROR: unable to >>> open >>> >> OSD superblock on /var/lib/ceph/osd/ceph-2: (2) No such file or >>> > directory* >>> >> >>> >> >>> >> root@osd03:/var/lib/ceph/osd/ceph-2# systemctl status ceph-osd@2 >>> >> â— ceph-osd@2.service - Ceph object storage daemon osd.2 >>> >> Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled; >>> >> vendor preset: enabled) >>> >> Active: failed (Result: exit-code) since Sun 2021-04-04 14:55:06 >>> >> +0430; 50s ago >>> >> Process: 5471 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh >>> >> --cluster ${CLUSTER} --id 2 (code=exited, status=0/SUCCESS) >>> >> Process: 5484 ExecStart=/usr/bin/ceph-osd -f --cluster ${CLUSTER} >>> > --id >>> >> 2 --setuser ceph --setgroup ceph (code=exited, status=1/FAILURE) >>> >> Main PID: 5484 (code=exited, status=1/FAILURE) >>> >> >>> >> Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Scheduled >>> restart >>> >> job, restart counter is at 3. >>> >> Apr 04 14:55:06 osd03 systemd[1]: Stopped Ceph object storage daemon >>> > osd.2. >>> >> Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Start request >>> >> repeated too quickly. >>> >> Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Failed with >>> result >>> >> 'exit-code'. >>> >> Apr 04 14:55:06 osd03 systemd[1]: Failed to start Ceph object storage >>> >> daemon osd.2. >>> >> root@osd03:/var/lib/ceph/osd/ceph-2# >>> >> >>> >> root@osd03:~# lsblk >>> >> NAME MAJ:MIN RM SIZE RO TYPE >>> MOUNTPOINT >>> >> fd0 2:0 1 4K 0 disk >>> >> loop0 7:0 0 55.5M 1 loop >>> >> /snap/core18/1988 >>> >> loop1 7:1 0 69.9M 1 loop >>> >> /snap/lxd/19188 >>> >> loop2 7:2 0 55.5M 1 loop >>> >> /snap/core18/1997 >>> >> loop3 7:3 0 70.4M 1 loop >>> >> /snap/lxd/19647 >>> >> loop4 7:4 0 32.3M 1 loop >>> >> /snap/snapd/11402 >>> >> loop5 7:5 0 32.3M 1 loop >>> >> /snap/snapd/11107 >>> >> sda 8:0 0 80G 0 disk >>> >> ├─sda1 8:1 0 1M 0 part >>> >> ├─sda2 8:2 0 1G 0 part /boot >>> >> └─sda3 8:3 0 79G 0 part >>> >> └─ubuntu--vg-ubuntu--lv 253:0 0 69.5G 0 lvm / >>> >> sdb 8:16 0 16G 0 disk >>> >> └─sdb1 8:17 0 16G 0 part >>> >> >>> >> >>> > >>> └─ceph--9d37674b--a269--4239--aa9e--66a3c74df76c-osd--block--2d3ffc61--e430--4 >>> >> b89--bcd4--105b2df26352 >>> >> 253:1 0 16G 0 lvm >>> >> root@osd03:~# >>> >> >>> >> root@osd03:/var/lib/ceph/osd/ceph-2# mount | grep -i ceph >>> >> tmpfs on /var/lib/ceph/osd/ceph-2 type tmpfs (rw,relatime) >>> >> root@osd03:/var/lib/ceph/osd/ceph-2# >>> >> >>> >> any help is much appreciated >>> >> -- >>> >> >>> >> Regards >>> >> Behzad Khoshbakhti >>> >> Computer Network Engineer (CCIE #58887) >>> >> >>> >> >>> > >>> > -- >>> > >>> > Regards >>> > Behzad Khoshbakhti >>> > Computer Network Engineer (CCIE #58887) >>> > +989128610474 >>> > _______________________________________________ >>> > ceph-users mailing list -- ceph-users@xxxxxxx >>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> > >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >> > > -- > > Regards > Behzad Khoshbakhti > Computer Network Engineer (CCIE #58887) > +989128610474 > -- Regards Behzad Khoshbakhti Computer Network Engineer (CCIE #58887) +989128610474 _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx