Hi all, I need to reduce the number of PGs in a pool from 2048 to 512 and would really like to do that in a single step. I executed the set pg_num 512 command, but the PGs are not all merged. Instead I get this intermediate state: pool 13 'con-fs2-meta2' replicated size 4 min_size 2 crush_rule 3 object_hash rjenkins pg_num 2048 pgp_num 1946 pg_num_target 512 pgp_num_target 512 autoscale_mode off last_change 916710 lfor 0/0/618995 flags hashpspool,nodelete,selfmanaged_snaps max_bytes 107374182400 stripe_width 0 compression_mode none application cephfs This is really annoying, because it will not only lead to repeated redundant data movements and but I also need to rebalance this pool onto fewer OSDs, which cannot hold the 1946 PGs it will be merged to intermittently. How can I override the autoscaler interfering with admin operations in such tight corners? As you can see, we disabled autoscaler on all pools and also globally. Still, it interferes with admin commands in an unsolicited way. I would like the PG merge happen on the fly as the data moves to the new OSDs. Best regards, ================= Frank Schilder AIT Risø Campus Bygning 109, rum S14 _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx