>If I understand the balancer correct, it balances PGs not data. >This worked perfectly fine in your case. > >I prefer a PG count of ~100 per OSD, you are at 30. Maybe it would >help to bump the PGs. > I can remember someone writing something smart about how to increase your pg's. Lets say the rbd.ssd is not balanced properly because this is the biggest one. I am not sure if I should increase from 8 to 16. Because that would just half the data in the pg's and they probably end up on the same osd's in the same ratio as now? Or am I assuming this incorrectly? Is 16 adviced? >> >> I have straw2, balancer=on, crush-compat and it gives worst spread over >> my ssd drives (4 only) being used by only 2 pools. One of these pools >> has pg 8. Should I increase this to 16 to create a better result, or >> will it never be any better. >> >> For now I like to stick to crush-compat, so I can use a default centos7 >> kernel. >> >> Luminous 12.2.8, 3.10.0-862.14.4.el7.x86_64, CentOS Linux release >> 7.5.1804 (Core) >> >> >> >> [@c01 ~]# cat balancer-1-before.txt | egrep '^19|^20|^21|^30' >> 19 ssd 0.48000 1.00000 447GiB 164GiB 283GiB 36.79 0.93 31 >> 20 ssd 0.48000 1.00000 447GiB 136GiB 311GiB 30.49 0.77 32 >> 21 ssd 0.48000 1.00000 447GiB 215GiB 232GiB 48.02 1.22 30 >> 30 ssd 0.48000 1.00000 447GiB 151GiB 296GiB 33.72 0.86 27 >> >> [@c01 ~]# ceph osd df | egrep '^19|^20|^21|^30' >> 19 ssd 0.48000 1.00000 447GiB 157GiB 290GiB 35.18 0.87 30 >> 20 ssd 0.48000 1.00000 447GiB 125GiB 322GiB 28.00 0.69 30 >> 21 ssd 0.48000 1.00000 447GiB 245GiB 202GiB 54.71 1.35 30 >> 30 ssd 0.48000 1.00000 447GiB 217GiB 230GiB 48.46 1.20 30 >> >> [@c01 ~]# ceph osd pool ls detail | egrep 'fs_meta|rbd.ssd' >> pool 19 'fs_meta' replicated size 3 min_size 2 crush_rule 5 object_hash >> rjenkins pg_num 16 pgp_num 16 last_change 22425 lfor 0/9035 flags >> hashpspool stripe_width 0 application cephfs >> pool 54 'rbd.ssd' replicated size 3 min_size 2 crush_rule 5 object_hash >> rjenkins pg_num 8 pgp_num 8 last_change 24666 flags hashpspool >> stripe_width 0 application rbd >> >> [@c01 ~]# ceph df |egrep 'ssd|fs_meta' >> fs_meta 19 170MiB 0.07 >> 240GiB 2451382 >> fs_data.ssd 33 0B 0 >> 240GiB 0 >> rbd.ssd 54 266GiB 52.57 >> 240GiB 75902 >> fs_data.ec21.ssd 55 0B 0 >> 480GiB 0 >> > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com