Hi Alexei,
just left a comment in the ticket...
Thanks,
Igor
On 7/25/2020 3:31 PM, Aleksei Zakharov wrote:
Hi all,
I wonder if someone else faced the issue described on the tracker:
https://tracker.ceph.com/issues/45519
We thought that this problem is caused by high OSD fragmentation,
until today. For now even OSDs with fragmentation rating < .3 are
affected. We don't use separate DB/WAL partition in this setup and
strings like this before failing:
2020-07-25 11:08:22.961 7f6f489d5700 1 bluefs _allocate failed to
allocate 0x33dd4c5 on bdev 1, free 0x2bc0000; fallback to bdev 2
2020-07-25 11:08:22.961 7f6f489d5700 1 bluefs _allocate unable to
allocate 0x33dd4c5 on bdev 2, free 0xffffffffffffffff; fallback to
slow device expander
look suspicious for us.
We use 4KiB bluefs and bluestore block sizes as well as store the
objects ~1KiB size and it looks like this makes the issue to be
reproduced much more frequently. But, as I can see on the tracker /
telegram channels, different people face with it from time to time,
for example: https://paste.ubuntu.com/p/GDCXDrnrtX/ (telegram link
https://t.me/ceph_users/376)
Did anyone able to identify the root cause and/or find a workaround
for it?
BTW, ceph would be a nice small-objects storage showing 300-500usec
latency if not this issue and this:
https://tracker.ceph.com/issues/45765 one.
--
Regards,
Aleksei Zakharov
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx