> -----Original Message----- > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Mark Nelson > Sent: 10 September 2018 18:27 > To: ceph-users@xxxxxxxxxxxxxx > Subject: Re: Bluestore DB size and onode count > > On 09/10/2018 12:22 PM, Igor Fedotov wrote: > > > Hi Nick. > > > > > > On 9/10/2018 1:30 PM, Nick Fisk wrote: > >> If anybody has 5 minutes could they just clarify a couple of things > >> for me > >> > >> 1. onode count, should this be equal to the number of objects stored > >> on the OSD? > >> Through reading several posts, there seems to be a general indication > >> that this is the case, but looking at my OSD's the maths don't > >> work. > > onode_count is the number of onodes in the cache, not the total number > > of onodes at an OSD. > > Hence the difference... Ok, thanks, that makes sense. I assume there isn't actually a counter which gives you the total number of objects on an OSD then? > >> > >> Eg. > >> ceph osd df > >> ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS > >> 0 hdd 2.73679 1.00000 2802G 1347G 1454G 48.09 0.69 115 > >> > >> So 3TB OSD, roughly half full. This is pure RBD workload (no > >> snapshots or anything clever) so let's assume worse case scenario of > >> 4MB objects (Compression is on however, which would only mean more > >> objects for given size) > >> 1347000/4=~336750 expected objects > >> > >> sudo ceph daemon osd.0 perf dump | grep blue > >> "bluefs": { > >> "bluestore": { > >> "bluestore_allocated": 1437813964800, > >> "bluestore_stored": 2326118994003, > >> "bluestore_compressed": 445228558486, > >> "bluestore_compressed_allocated": 547649159168, > >> "bluestore_compressed_original": 1437773843456, > >> "bluestore_onodes": 99022, > >> "bluestore_onode_hits": 18151499, > >> "bluestore_onode_misses": 4539604, > >> "bluestore_onode_shard_hits": 10596780, > >> "bluestore_onode_shard_misses": 4632238, > >> "bluestore_extents": 896365, > >> "bluestore_blobs": 861495, > >> > >> 99022 onodes, anyone care to enlighten me? > >> > >> 2. block.db Size > >> sudo ceph daemon osd.0 perf dump | grep db > >> "db_total_bytes": 8587829248, > >> "db_used_bytes": 2375024640, > >> > >> 2.3GB=0.17% of data size. This seems a lot lower than the 1% > >> recommendation (10GB for every 1TB) or 4% given in the official docs. I > >> know that different workloads will have differing overheads and > >> potentially smaller objects. But am I understanding these figures > >> correctly as they seem dramatically lower? > > Just in case - is slow_used_bytes equal to 0? Some DB data might > > reside at slow device if spill over has happened. Which doesn't > > require full DB volume to happen - that's by RocksDB's design. > > > > And recommended numbers are a bit... speculative. So it's quite > > possible that you numbers are absolutely adequate. > > FWIW, these are the numbers I came up with after examining the SST files > generated under different workloads: > > https://protect-eu.mimecast.com/s/7e0iCJq9Bh6pZCzILpy?domain=drive.google.com > Thanks for your input Mark and Igor. Mark I can see your RBD figures aren't too far off mine, so all looks to be as expected then. > >> > >> Regards, > >> Nick > >> > >> _______________________________________________ > >> ceph-users mailing list > >> mailto:ceph-users@xxxxxxxxxxxxxx > >> https://protect-eu.mimecast.com/s/YtrdCKZVDUX8OTAS9XW?domain=lists.ceph.com > > > > _______________________________________________ > > ceph-users mailing list > > mailto:ceph-users@xxxxxxxxxxxxxx > > https://protect-eu.mimecast.com/s/YtrdCKZVDUX8OTAS9XW?domain=lists.ceph.com > > _______________________________________________ > ceph-users mailing list > mailto:ceph-users@xxxxxxxxxxxxxx > https://protect-eu.mimecast.com/s/YtrdCKZVDUX8OTAS9XW?domain=lists.ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com