Hi Daniel,
I had the exact same issue in a (virtual) Luminous cluster without
much data in it. The root cause was that my OSDs were too small (10 GB
only) and the rocksDB also grew until manual compaction. I had
configured the small OSDs intentionally because it was never supposed
to store lots of data. Can you provide some more details like 'ceph
odf df'?
Manual compaction did help, but then I recreated the OSDs with 20 GB
each and the issue didn't occur after that.
Regards,
Eugen
Zitat von Denis Krienbühl <denis@xxxxxxx>:
Hi
On one of our Ceph clusters, some OSDs have been marked as full.
Since this is a staging cluster that does not have much data on it,
this is strange.
Looking at the full OSDs through “ceph osd df” I figured out that
the space is mostly used by metadata:
SIZE: 122 GiB
USE: 118 GiB
DATA: 2.4 GiB
META: 116 GiB
We run mimic, and for the affected OSDs we use a db device (nvme) in
addition to the primary device (hdd).
In the logs we see the following errors:
2020-05-12 17:10:26.089 7f183f604700 1 bluefs _allocate failed
to allocate 0x400000 on bdev 1, free 0x0; fallback to bdev 2
2020-05-12 17:10:27.113 7f183f604700 1
bluestore(/var/lib/ceph/osd/ceph-8) _balance_bluefs_freespace
gifting 0x180a000000~400000 to bluefs
2020-05-12 17:10:27.153 7f183f604700 1 bluefs add_block_extent
bdev 2 0x180a000000~400000
We assume it is an issue with Rocksdb, as the following call will
quickly fix the problem:
ceph daemon osd.8 compact
The question is, why is this happening? I would think that “compact"
is something that runs automatically from time to time, but I’m not
sure.
Is it on us to run this regularly?
Any pointers are welcome. I’m quite new to Ceph :)
Cheers,
Denis
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx