Re: OSDs unable to mount BlueFS after reboot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 9/15/21 21:02, Davíð Steinn Geirsson wrote:
Hi,

On Wed, Sep 15, 2021 at 08:39:11PM +0200, Stefan Kooman wrote:
On 9/15/21 18:06, Davíð Steinn Geirsson wrote:
Just realised the debug paste I sent was for OSD 5 but the other info is for
OSD 0. They are both having the same issue, but for completeness sake here
is the debug output from OSD 0:
http://paste.debian.net/1211873/

All daemons in the cluster are running ceph pacific 16.2.5.

Can you increase debug level for the OSD, i.e. ceph config set osd.0
debug_osd 20/20

And then restart the osd?

Sure, here is the output with 20/20:
https://paste.debian.net/1211886/

Only 3 lines added as far as I can tell:
2021-09-15T18:44:03.289+0000 7fce2827af00  5 object store type is bluestore
[...]
2021-09-15T18:44:05.673+0000 7fce2827af00  2 osd.0 0 init /var/lib/ceph/osd/ceph-0 (looks like hdd)
2021-09-15T18:44:05.673+0000 7fce2827af00  2 osd.0 0 journal /var/lib/ceph/osd/ceph-0/journal

I tried again with debug_osd 99/99 (the maximum) and did not see any
additional messages.

Can you access this link: https://access.redhat.com/solutions/4939871 ?

This would indicate rocksdb corruption. Can you fsck and repair the OSD?

# ceph-bluestore-tool fsck --path /var/lib/ceph/osd/ceph-0/ --debug

# ceph-bluestore-tool repair --path /var/lib/ceph/osd/ceph-0/ --debug

The post states:

If fsck and repair does not help to recover from rocksdb corruption and all PGs areactive+clean then safest way is to re-deploy affected OSD. In case some PGs are incomplete or down, Kindly contact Red Hat Ceph Support.

And I agree with that. Better not try to fix things, but let Ceph do a clean recovery.

As for the Root Cause

The rocksdb corruption on bluestore could be due to hard reboot of OSD node or block.db device medium error. The rocksdb in blueStore contains not only OMAPs and metadata for ceph objects, but also ceph objects layout on disk, entire delayed transactions, allocator free regions and others.

You might get more information with increasing debug for rocksdb / bluefs / bluestore

ceph config set osd.0 debug_rocksdb = 20/20
ceph config set osd.0 debug_bluefs = 20/20
ceph config set osd.0 debug_bluestore = 20/20

Gr. Stefan

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux