Thanks very much!
This will only work for replicated pools, and not
erasure-coded pools, correct?
Vlad
On 2/18/22 13:37, Matt Vandermeulen wrote:
Yep, just change the CRUSH rule:
ceph osd pool set my_cephfs_metadata_pool crush_rule
replicated_nvme
If you have a rule set called replicated_nvme, that'll set
it on the pool named my_cephfs_metadata_pool.
Of course this will cause a significant data movement.
If you need to add the rule, and want it to use host as a
failure domain:
ceph osd crush rule create-replicated replicated_nvme
default host nvme
On 2022-02-18 15:26, Vladimir Brik wrote:
Hello
Is it possible to change which device class a replicated
pool is using?
For example, if my cephfs metadata pool was configured on
creation to
use ssd device class, can I later change it to use nvme
device class?
Thanks,
Vlad
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx