I have straw2, balancer=on, crush-compat and it gives worst spread over my ssd drives (4 only) being used by only 2 pools. One of these pools has pg 8. Should I increase this to 16 to create a better result, or will it never be any better. For now I like to stick to crush-compat, so I can use a default centos7 kernel. Luminous 12.2.8, 3.10.0-862.14.4.el7.x86_64, CentOS Linux release 7.5.1804 (Core) [@c01 ~]# cat balancer-1-before.txt | egrep '^19|^20|^21|^30' 19 ssd 0.48000 1.00000 447GiB 164GiB 283GiB 36.79 0.93 31 20 ssd 0.48000 1.00000 447GiB 136GiB 311GiB 30.49 0.77 32 21 ssd 0.48000 1.00000 447GiB 215GiB 232GiB 48.02 1.22 30 30 ssd 0.48000 1.00000 447GiB 151GiB 296GiB 33.72 0.86 27 [@c01 ~]# ceph osd df | egrep '^19|^20|^21|^30' 19 ssd 0.48000 1.00000 447GiB 157GiB 290GiB 35.18 0.87 30 20 ssd 0.48000 1.00000 447GiB 125GiB 322GiB 28.00 0.69 30 21 ssd 0.48000 1.00000 447GiB 245GiB 202GiB 54.71 1.35 30 30 ssd 0.48000 1.00000 447GiB 217GiB 230GiB 48.46 1.20 30 [@c01 ~]# ceph osd pool ls detail | egrep 'fs_meta|rbd.ssd' pool 19 'fs_meta' replicated size 3 min_size 2 crush_rule 5 object_hash rjenkins pg_num 16 pgp_num 16 last_change 22425 lfor 0/9035 flags hashpspool stripe_width 0 application cephfs pool 54 'rbd.ssd' replicated size 3 min_size 2 crush_rule 5 object_hash rjenkins pg_num 8 pgp_num 8 last_change 24666 flags hashpspool stripe_width 0 application rbd [@c01 ~]# ceph df |egrep 'ssd|fs_meta' fs_meta 19 170MiB 0.07 240GiB 2451382 fs_data.ssd 33 0B 0 240GiB 0 rbd.ssd 54 266GiB 52.57 240GiB 75902 fs_data.ec21.ssd 55 0B 0 480GiB 0 _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com