Hi, Start by setting: ceph config set mgr mgr/balancer/upmap_max_deviation 1 This configures the balancer to squeeze the OSDs to within 1 PG of eachother. I'm starting to think this should be the default. Cheers, dan On Wed, Jul 28, 2021 at 9:08 AM Manuel Holtgrewe <zyklenfrei@xxxxxxxxx> wrote: > > Dear all, > > I'm running Ceph 14.2.11. I have 140 HDDs in my cluster of 4 nodes, 35 HDDs > per node. I am observing fill ratios of 66% to 70% of OSDs and then one > with 82% (see attached ceph-osd-df.txt for output of "ceph osd df"). > > Previously, I had problems with single OSDs filling up to 85% and then > everything coming to a grinding halt. Ideally, I would like to have all OSD > fill grade to be close to the mean of 67%... At the very least I need to > get the 82% OSD back into the range. > > I have upmap balancing enabled and the balancer says: > > # ceph balancer status > { > "last_optimize_duration": "0:00:00.053686", > "plans": [], > "mode": "upmap", > "active": true, > "optimize_result": "Unable to find further optimization, or pool(s) > pg_num is decreasing, or distribution is already perfect", > "last_optimize_started": "Wed Jul 28 09:03:02 2021" > } > > Creating an offline balancing plan looks like this: > > # ceph osd getmap -o om > got osdmap epoch 281708 > # osdmaptool om --upmap out.txt --upmap-pool hdd_ec --upmap-deviation 1 > --upmap-active > osdmaptool: osdmap file 'om' > writing upmap command output to: out.txt > checking for upmap cleanups > upmap, max-count 10, max deviation 1 > limiting to pools hdd_ec ([3]) > pools hdd_ec > prepared 0/10 changes > Time elapsed 0.0275739 secs > Unable to find further optimization, or distribution is already perfect > osd.0 pgs 43 > [...] > # wc -l out.txt > 0 out.txt > > Does anyone have a suggestion how to proceed getting the 82% OSD closer to > the mean fill ratio (and maybe the other OSDs as well?) > > Thanks, > Manuel > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx