In your situation the JJ Balancer might help. > > On 2024-11-27 17:53, Anthony D'Atri wrote: >>> Hi, >>> My Ceph cluster is out-of-balance. The amount of PG's per OSD ranges from about 50 up to 100 PG's per OSD. This is far from balanced. >> Do you have multiple CRUSH roots or device classes? Are all OSDs the same weight? > > Yes, I have 2 CRUSH roots and 2 device classes. 1 CRUSH root is using the ssd device class, another CRUSH root is using the hdd device class. I am only talking about the HDD's. > >>> My disk sizes differs from 1.6T up to 2.4T. >> Ah. The number of PG replicas as reported by `ceph osd df` should be proportional to the OSD capacities. With some of your OSDs double the size of others, it is natural that some will have double the number of PGs. >> That said, 2.4T is an unusual size for a drive. What specific storage drives are you using? Chances are that you’d benefit from bumping pg_num on at least some of your pools. > > Sorry, I was wrong. It is 1.6T and 2.2T (not 2.4T). > The 2.2T OSD's are ranging between 65-104 PG's, while the 1.6T OSD's are ranging between 47-80 PG's. > > Bumping the PG's per pool might balance it a bit, but without bumping the PG's the balancer module should balance it as well, right? _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx