Hello, On Thu, 16 Apr 2015 00:41:29 +0200 Steffen W Sørensen wrote: > Hi, > > Successfully upgrade a small development 4x node Giant 0.87-1 cluster to > Hammer 0.94-1, each node with 6x OSD - 146GB, 19 pools, mainly 2 in > usage. Only minor thing now ceph -s complaining over too may PGs, > previously Giant had complain of too few, so various pools were bumped > up till health status was okay as before upgrading. Admit, that after > bumping PGs up in Giant we had changed pool sizes from 3 to 2 & min 1 in > fear of perf. when backfilling/recovering PGs. > That later change would have _increased_ the number of recommended PG, not decreased it. With your cluster 2048 PGs total (all pools combined!) would be the sweet spot, see: http://ceph.com/pgcalc/ It seems to me that you increased PG counts assuming that the formula is per pool. > > # ceph -s > cluster 16fe2dcf-2629-422f-a649-871deba78bcd > health HEALTH_WARN > too many PGs per OSD (1237 > max 300) > monmap e29: 3 mons at > {0=10.0.3.4:6789/0,1=10.0.3.2:6789/0,2=10.0.3.1:6789/0} election epoch > 1370, quorum 0,1,2 2,1,0 mdsmap e142: 1/1/1 up {0=2=up:active}, 1 > up:standby osdmap e3483: 24 osds: 24 up, 24 in > pgmap v3719606: 14848 pgs, 19 pools, 530 GB data, 133 kobjects > 1055 GB used, 2103 GB / 3159 GB avail > 14848 active+clean > This is an insanely high PG count for this cluster and is certain to impact performance and resource requirements (all these PGs need to peer after all). > Can we just reduce PGs again and should we decrement in minor steps one > pool at a time… > No, as per the documentation you can only increase PGs and PGPs. So your options are to totally flatten this cluster or if pools with important data exist to copy them to new, correctly sized, pools and delete all the oversized ones after that. Christian -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Global OnLine Japan/Fusion Communications http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com