Hello team, I have one small size ceph cluster in production having 53 SSD of 7TB each with 6 node Version:- octopus In last two days we just cleared out ~100 tb of data out of 370tb total size So there is bunch of pg with active+clean+snaptrim & snaptrim_wait state comes into the action after that one of the pool placement group number decrease by autoscaler from 2048 to 512 And autoscaler showing remaining time of 4y which is too long can you guys help me out how to fix this issue root@nvme1:~# ceph -s cluster: id: c30f5720-ca5c-11ec-b19c-f9781f61e8ec health: HEALTH_WARN noscrub,nodeep-scrub flag(s) set 34 pgs not deep-scrubbed in time services: mon: 3 daemons, quorum nvme3,nvme2,nvme1 (age 14M) mgr: nvme2.ttfore(active, since 13M), standbys: nvme3.zhnsuf osd: 53 osds: 53 up (since 2h), 53 in (since 11d); 61 remapped pgs flags noscrub,nodeep-scrub task status: data: pools: 7 pools, 2784 pgs objects: 26.37M objects, 69 TiB usage: 177 TiB used, 194 TiB / 370 TiB avail pgs: 1011732/79097316 objects misplaced (1.279%) 2723 active+clean 58 active+remapped+backfill_wait 3 active+remapped+backfilling io: client: 78 KiB/s rd, 3.8 MiB/s wr, 238 op/s rd, 314 op/s wr recovery: 17 MiB/s, 6 objects/s progress: PG autoscaler decreasing pool 8 PGs from 2048 to 512 (1h) [............................] (remaining: 4y) root@nvme1:~# Thanks Arihant Jain _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx