Re: Reg: PG

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The extra pools are probably the data and metadata pools that are automatically created for cephfs.

http://ceph.com/pgcalc/ is a useful tool for helping to work out how many PGs your pools should have.

Rich

On 04/05/17 15:41, David Turner wrote:
I'm guessing you have more than just the 1 pool with 128 PGs in your cluster (seeing as you have 320 PGs total, I would guess 2 pools with 128 PGs and 1 pool with 64 PGs).  The combined total number of PGs for all of your pools is 320 and with only 3 OSDs and most likely replica size 3... that leaves you with too many (320) PGs per OSD.  This will not likely affect your testing, but if you want to fix the problem you will need to delete and recreate your pools with a combined lower total number of PGs.

The number of PGs is supposed to reflect how much data each pool is going to have.  If you have 1 pool that will have 75% of your cluster's data, another pool with 20%, and a third pool with 5%... then the number of PGs they have should reflect that.  Based on trying to have somewhere between 100-200 PGs per osd, and the above estimation for data distribution, you should have 128 PGs in the first pool, 32 PGs in the second, and 8 PGs in the third.  Each OSD would have 168 PGs and each PG will be roughly the same size between each pool.  If you were to add more OSDs, then you would need to increase those numbers to account for the additional OSDs to maintain the same distribution.  The above math is only for 3 OSDs.  If you had 6 OSDs, then the goal would be to have somewhere between 200-400 PGs total to maintain the same 100-200 PGs per OSD.

On Thu, May 4, 2017 at 10:24 AM psuresh <psuresh@xxxxxxxxxxxx> wrote:
Hi,

I'm running 3 osd in my test setup.   I have created PG pool with 128 as per the ceph documentation.  
But i'm getting too many PGs warning.   Can anyone clarify? why i'm getting this warning.  

Each OSD contain 240GB disk.

    cluster 9d325da2-3d87-4b6b-8cca-e52a4b65aa08
     health HEALTH_WARN
            too many PGs per OSD (320 > max 300)
     monmap e2: 3 mons at {dev-ceph-mon1:6789/0,dev-ceph-mon2:6789/0,dev-ceph-mon3:6789/0}
            election epoch 6, quorum 0,1,2 dev-ceph-mon1,dev-ceph-mon2,dev-ceph-mon3
      fsmap e40: 1/1/1 up {0=dev-ceph-mds-active=up:active}
     osdmap e356: 3 osds: 3 up, 3 in
            flags sortbitwise,require_jewel_osds
      pgmap v32407: 320 pgs, 3 pools, 27456 MB data, 220 kobjects
            100843 MB used, 735 GB / 833 GB avail
                 320 active+clean

Regards,
Suresh
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux