Would http://ceph.com/pgcalc/ help? On Mon, Jul 18, 2016 at 01:27:38PM +0700, Khang Nguyễn Nhật wrote: > Hi all, > I have a cluster consists of: 3 Monitors, 1 RGW, 1 host of 24 OSDs(2TB/OSD) and > some pool as: > ap-southeast.rgw.data.root > ap-southeast.rgw.control > ap-southeast.rgw.gc > ap-southeast.rgw.log > ap-southeast.rgw.intent-log > ap-southeast.rgw.usage > ap-southeast.rgw.users.keys > ap-southeast.rgw.users.email > ap-southeast.rgw.users.swift > ap-southeast.rgw.users.uid > ap-southeast.rgw.buckets.index > ap-southeast.rgw.buckets.data > ap-southeast.rgw.buckets.non-ec > ap-southeast.rgw.meta > In which "ap-southeast.rgw.buckets.data" is a erasure pool(k=20, m=4) and all > of the remaining pool are replicated(size=3). I've used (100*OSDs)/size to > calculate the number of PGs, e.g. 100*24/3 = 800(nearest power of 2: 1024) for > replicated pools and 100*24/24=100(nearest power of 2: 128) for erasure pool. > I'm not sure this is the best placement group number, someone can give me some > advice ? > Thank ! > SECURITY NOTE: file ~/.netrc must not be accessible by others > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Thanks, Chengwei
Attachment:
signature.asc
Description: Digital signature
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com