If I understand the balancer correct, it balances PGs not data. This worked perfectly fine in your case. I prefer a PG count of ~100 per OSD, you are at 30. Maybe it would help to bump the PGs. Kevin Am Sa., 5. Jan. 2019 um 14:39 Uhr schrieb Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx>: > > > I have straw2, balancer=on, crush-compat and it gives worst spread over > my ssd drives (4 only) being used by only 2 pools. One of these pools > has pg 8. Should I increase this to 16 to create a better result, or > will it never be any better. > > For now I like to stick to crush-compat, so I can use a default centos7 > kernel. > > Luminous 12.2.8, 3.10.0-862.14.4.el7.x86_64, CentOS Linux release > 7.5.1804 (Core) > > > > [@c01 ~]# cat balancer-1-before.txt | egrep '^19|^20|^21|^30' > 19 ssd 0.48000 1.00000 447GiB 164GiB 283GiB 36.79 0.93 31 > 20 ssd 0.48000 1.00000 447GiB 136GiB 311GiB 30.49 0.77 32 > 21 ssd 0.48000 1.00000 447GiB 215GiB 232GiB 48.02 1.22 30 > 30 ssd 0.48000 1.00000 447GiB 151GiB 296GiB 33.72 0.86 27 > > [@c01 ~]# ceph osd df | egrep '^19|^20|^21|^30' > 19 ssd 0.48000 1.00000 447GiB 157GiB 290GiB 35.18 0.87 30 > 20 ssd 0.48000 1.00000 447GiB 125GiB 322GiB 28.00 0.69 30 > 21 ssd 0.48000 1.00000 447GiB 245GiB 202GiB 54.71 1.35 30 > 30 ssd 0.48000 1.00000 447GiB 217GiB 230GiB 48.46 1.20 30 > > [@c01 ~]# ceph osd pool ls detail | egrep 'fs_meta|rbd.ssd' > pool 19 'fs_meta' replicated size 3 min_size 2 crush_rule 5 object_hash > rjenkins pg_num 16 pgp_num 16 last_change 22425 lfor 0/9035 flags > hashpspool stripe_width 0 application cephfs > pool 54 'rbd.ssd' replicated size 3 min_size 2 crush_rule 5 object_hash > rjenkins pg_num 8 pgp_num 8 last_change 24666 flags hashpspool > stripe_width 0 application rbd > > [@c01 ~]# ceph df |egrep 'ssd|fs_meta' > fs_meta 19 170MiB 0.07 > 240GiB 2451382 > fs_data.ssd 33 0B 0 > 240GiB 0 > rbd.ssd 54 266GiB 52.57 > 240GiB 75902 > fs_data.ec21.ssd 55 0B 0 > 480GiB 0 > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com