Re: Nautilus: BlueFS spillover

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

On 9/27/19 10:54 AM, Eugen Block wrote:
Update: I expanded all rocksDB devices, but the warnings still appear:

BLUEFS_SPILLOVER BlueFS spillover detected on 10 OSD(s)
     osd.0 spilled over 2.5 GiB metadata from 'db' device (2.4 GiB used of 30 GiB) to slow device      osd.19 spilled over 66 MiB metadata from 'db' device (818 MiB used of 15 GiB) to slow device      osd.25 spilled over 2.2 GiB metadata from 'db' device (2.6 GiB used of 30 GiB) to slow device      osd.26 spilled over 1.6 GiB metadata from 'db' device (1.9 GiB used of 30 GiB) to slow device      osd.27 spilled over 2.6 GiB metadata from 'db' device (2.5 GiB used of 30 GiB) to slow device      osd.28 spilled over 2.4 GiB metadata from 'db' device (1.3 GiB used of 30 GiB) to slow device      osd.29 spilled over 2.9 GiB metadata from 'db' device (1.7 GiB used of 30 GiB) to slow device      osd.31 spilled over 2.2 GiB metadata from 'db' device (2.7 GiB used of 30 GiB) to slow device      osd.32 spilled over 2.4 GiB metadata from 'db' device (1.7 GiB used of 30 GiB) to slow device      osd.33 spilled over 2.2 GiB metadata from 'db' device (2.0 GiB used of 30 GiB) to slow device


OSD.19 can be ignored as it's currently not in use, but the other devices have been expanded from 20 to 30 GB (following the explanations about the compaction levels).
According to the OSD logs these are the sizes we're dealing with:

Level  Size
L0     31.84
L1     183.86
L2     923.67
L3     3.62
Sum    4.74
Int    0.00


Is there any sign that these OSDs would require even larger bdev devices (300GB)? Which would not be possible with the currently used SSDs, unfortunately.


Try it with a little bit large db devices. The db devices also holds the WAL, which has a default size of 1 GB afaik. And you also need to consider gigabyte vs. gibibytes. We ran into the same problem in our setup.....


You might also want to increase the size even further, since rocksdb needs some free space during compaction. The worst case scenario is ~60 GB per device to take compaction into account. If you do not have an extreme metadata/omap data loaded workload, you won't need more capacity for these partitions.


Regards,

Burkhard

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux