Hi, Pavel V. Kaygorodov wrote: > I have updated my cluster to Hammer and got a warning "too many PGs > per OSD (2240 > max 300)". I know, that there is no way to decrease > number of page groups, so I want to re-create my pools with less pg > number, move all my data to them, delete old pools and rename new > pools as the old ones. Also I want to preserve the user rights on new > pools. I have several pools with RBD images, some of them with > snapshots. > > Which is the proper way to do this? I'm not a ceph expert and I can just tell you my (little but happy) experience. ;) I had the same problem with pools of my radosgw pools ie: - the .rgw.* pools except ".rgw.buckets", and - the .users.* pools So, **warning**, it was for very tiny pools. The version of Ceph was Hammer 94.1, nodes were Ubuntu 14.04 with 3.16 kernel. These commands worked well for me: --------------------------------------------- # /!\ Before I have stopped my radosgws (ie the ceph clients of the pools). old_pool=foo new_pool=foo.new ceph osd pool create $new_pool 64 rados cppool $old_pool $new_pool ceph osd pool delete $old_pool $old_pool --yes-i-really-really-mean-it ceph osd pool rename $new_pool $old_pool # And I have restarted my radosgws. --------------------------------------------- That's all. In my case, it was very fast because the pools didn't contain very much data. And I prolong your question: is it possible to do the same process but with a pool of the cephfs? For instance, the pool metadata? If I try the commands above, I have an error with the delete command: ~# ceph osd pool delete metadata metadata --yes-i-really-really-mean-it Error EBUSY: pool 'metadata' is in use by CephFS However, I'm sure no client use the cephfs (it's a cluster for test). -- François Lafont _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com