Hi,
just because the autoscaler doesn’t increase the pg_num doesn’t mean
you can’t increase it manually. Have you tried that?
Zitat von Daniel Persson <mailto.woden@xxxxxxxxx>:
Hi Team.
We are currently in the process of changing the size of our cache pool.
Currently it's set to 32 PGs and distributed weirdly on our OSDs. The
system has tried automatically to scale it up to 256 PGs without succeeding
and I read that cache pools are not automatically scaled so we are in the
process of scaling. Our plan is to remove the old one and create a new one
with more PGs.
I've run the pool in readproxy now for a week so most of the objects should
be available in cold storage but I want to be totally sure so we don't lose
any data.
I read in the documentation that you could remove the overlay and that
would redirect clients to cold storage.
Is a preferred strategy to remove the overlay and then run
cache-flush-evict-all to clear it and then replace or should I be fine just
to remove overlay and tiering and replace it with a new pool?
Currently we have configured it to have a write caching of 0.5 hours and
read cache of 2 days.
------
ceph osd pool set cephfs_data_cache cache_min_flush_age 1800
ceph osd pool set cephfs_data_cache cache_min_evict_age 172800
----
The cache is still 25Tb in size and would be sad to lose if we have
unwritten data.
Best regards
Daniel
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx