I am having with the change from pg 8 to pg 16 [@c01 ceph]# ceph osd df | egrep '^ID|^19|^20|^21|^30' ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS 19 ssd 0.48000 1.00000 447GiB 161GiB 286GiB 35.91 0.84 35 20 ssd 0.48000 1.00000 447GiB 170GiB 277GiB 38.09 0.89 36 21 ssd 0.48000 1.00000 447GiB 215GiB 232GiB 48.08 1.12 36 30 ssd 0.48000 1.00000 447GiB 220GiB 227GiB 49.18 1.15 37 Coming from [@c01 ~]# ceph osd df | egrep '^19|^20|^21|^30' 19 ssd 0.48000 1.00000 447GiB 157GiB 290GiB 35.18 0.87 30 20 ssd 0.48000 1.00000 447GiB 125GiB 322GiB 28.00 0.69 30 21 ssd 0.48000 1.00000 447GiB 245GiB 202GiB 54.71 1.35 30 30 ssd 0.48000 1.00000 447GiB 217GiB 230GiB 48.46 1.20 30 I guess I should know more about the technology behind this, to appreciate this result. (I guess the pgs stay this way until more data is added when having "Error EDOM: Unable to find further optimization, ...") >> >> >If I understand the balancer correct, it balances PGs not data. >> >This worked perfectly fine in your case. >> > >> >I prefer a PG count of ~100 per OSD, you are at 30. Maybe it would >> >help to bump the PGs. >> > > >> I am not sure if I should increase from 8 to 16. Because that would just >> >> half the data in the pg's and they probably end up on the same osd's in >> the same ratio as now? Or am I assuming this incorrectly? >> Is 16 adviced? >> > >If you had only one PG (the very extremest usecase) it would always be optimally >misplaced. If you have lots, there are many more chances of ceph spreading them >correctly. There is some hashing and pseudorandom in there to screw it >up at times, >but considering what you can do with few -vs- many PGs, having many allows for >better spread than few, upto some point where the cpu handling all PGs eats more >resources than its worth. _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com