Upgrade and lost osds Operation not permitted

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



 Hi all,

As I have upgrade my Ceph cluster from 15.2.10 to 16.2.0, during the manual
upgrade using the precompiled packages, the OSDs was down with the
following messages:

root@osd03:/var/lib/ceph/osd/ceph-2# ceph-volume lvm activate --all
--> Activating OSD ID 2 FSID 2d3ffc61-e430-4b89-bcd4-105b2df26352
Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2
Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir
--dev
/dev/ceph-9d37674b-a269-4239-aa9e-66a3c74df76c/osd-block-2d3ffc61-e430-4b89-bcd4-105b2df26352
--path /var/lib/ceph/osd/ceph-2 --no-mon-config
Running command: /usr/bin/ln -snf
/dev/ceph-9d37674b-a269-4239-aa9e-66a3c74df76c/osd-block-2d3ffc61-e430-4b89-bcd4-105b2df26352
/var/lib/ceph/osd/ceph-2/block
Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-2/block
Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1
Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2
Running command: /usr/bin/systemctl enable
ceph-volume@lvm-2-2d3ffc61-e430-4b89-bcd4-105b2df26352
Running command: /usr/bin/systemctl enable --runtime ceph-osd@2
Running command: /usr/bin/systemctl start ceph-osd@2
--> ceph-volume lvm activate successful for osd ID: 2

Content of /var/log/ceph/ceph-osd.2.log
2021-04-04T14:54:56.625+0430 7f4afbac0f00  0 set uid:gid to 64045:64045
(ceph:ceph)
2021-04-04T14:54:56.625+0430 7f4afbac0f00  0 ceph version 16.2.0
(0c2054e95bcd9b30fdd908a79ac1d8bbc3394442) pacific (stable), process
ceph-osd, pid 5484
2021-04-04T14:54:56.625+0430 7f4afbac0f00  0 pidfile_write: ignore empty
--pid-file
2021-04-04T14:54:56.625+0430 7f4afbac0f00 -1*
bluestore(/var/lib/ceph/osd/ceph-2/block) _read_bdev_label failed to open
/var/lib/ceph/osd/ceph-2/block: (1) Operation not permitted*
2021-04-04T14:54:56.625+0430 7f4afbac0f00 -1  *** ERROR: unable to open OSD
superblock on /var/lib/ceph/osd/ceph-2: (2) No such file or directory*


root@osd03:/var/lib/ceph/osd/ceph-2# systemctl status ceph-osd@2
â— ceph-osd@2.service - Ceph object storage daemon osd.2
     Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled; vendor
preset: enabled)
     Active: failed (Result: exit-code) since Sun 2021-04-04 14:55:06
+0430; 50s ago
    Process: 5471 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh
--cluster ${CLUSTER} --id 2 (code=exited, status=0/SUCCESS)
    Process: 5484 ExecStart=/usr/bin/ceph-osd -f --cluster ${CLUSTER} --id
2 --setuser ceph --setgroup ceph (code=exited, status=1/FAILURE)
   Main PID: 5484 (code=exited, status=1/FAILURE)

Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Scheduled restart
job, restart counter is at 3.
Apr 04 14:55:06 osd03 systemd[1]: Stopped Ceph object storage daemon osd.2.
Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Start request
repeated too quickly.
Apr 04 14:55:06 osd03 systemd[1]: ceph-osd@2.service: Failed with result
'exit-code'.
Apr 04 14:55:06 osd03 systemd[1]: Failed to start Ceph object storage
daemon osd.2.
root@osd03:/var/lib/ceph/osd/ceph-2#

root@osd03:~# lsblk
NAME                                  MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
fd0                                     2:0    1    4K  0 disk
loop0                                   7:0    0 55.5M  1 loop
/snap/core18/1988
loop1                                   7:1    0 69.9M  1 loop
/snap/lxd/19188
loop2                                   7:2    0 55.5M  1 loop
/snap/core18/1997
loop3                                   7:3    0 70.4M  1 loop
/snap/lxd/19647
loop4                                   7:4    0 32.3M  1 loop
/snap/snapd/11402
loop5                                   7:5    0 32.3M  1 loop
/snap/snapd/11107
sda                                     8:0    0   80G  0 disk
├─sda1                                  8:1    0    1M  0 part
├─sda2                                  8:2    0    1G  0 part /boot
└─sda3                                  8:3    0   79G  0 part
  └─ubuntu--vg-ubuntu--lv             253:0    0 69.5G  0 lvm  /
sdb                                     8:16   0   16G  0 disk
└─sdb1                                  8:17   0   16G  0 part

└─ceph--9d37674b--a269--4239--aa9e--66a3c74df76c-osd--block--2d3ffc61--e430--4
     b89--bcd4--105b2df26352
                                      253:1    0   16G  0 lvm
root@osd03:~#

root@osd03:/var/lib/ceph/osd/ceph-2# mount | grep -i ceph
tmpfs on /var/lib/ceph/osd/ceph-2 type tmpfs (rw,relatime)
root@osd03:/var/lib/ceph/osd/ceph-2#

any help is much appreciated
-- 

Regards
 Behzad Khoshbakhti
 Computer Network Engineer (CCIE #58887)
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux