It's from here: https://ceph.com/docs/v0.79/dev/cache-pool/#cache-mode In that page there are both commands On Tue, Oct 28, 2014 at 6:03 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: > On Tue, Oct 28, 2014 at 3:24 AM, Cristian Falcas > <cristi.falcas@xxxxxxxxx> wrote: >> Hello, >> >> In the documentation about creating an cache pool, you find this: >> >> "Cache mode >> >> The most important policy is the cache mode: >> >> ceph osd pool set foo-hot cache-mode writeback" >> >> But when trying to run the above command, I get an error: >> >> ceph osd pool set ssd_cache cache-mode writeback >> Invalid command: cache-mode not in >> size|min_size|crash_replay_interval|pg_num|pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool|target_max_bytes|target_max_objects|cache_target_dirty_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid >> osd pool set <poolname> >> size|min_size|crash_replay_interval|pg_num|pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool|target_max_bytes|target_max_objects|cache_target_dirty_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid >> <val> {--yes-i-really-mean-it} : set pool parameter <var> to <val> >> Error EINVAL: invalid command >> >> Is this deprecated? I'm using version 0.80. >> >> Those are the commands I used to create the cache: >> >> # Set up a read/write cache pool ssd_cache for pool images: >> ceph osd tier add images ssd_cache >> ceph osd tier cache-mode ssd_cache writeback >> # Direct all traffic for images to ssd_cache: >> ceph osd tier set-overlay images ssd_cache >> ceph osd pool set ssd_cache cache-mode writeback >> # Set the target size and enable the tiering agent for ssd_cache: >> ceph osd pool set ssd_cache hit_set_type bloom >> ceph osd pool set ssd_cache hit_set_count 1 >> ceph osd pool set ssd_cache hit_set_period 3600 # 1 hour >> ceph osd pool set ssd_cache target_max_bytes 400000000000 # 500 GB >> # will begin flushing dirty objects when 40% of the pool is dirty and >> begin evicting clean objects when we reach 80% of the target size. >> ceph osd pool set ssd_cache cache_target_dirty_ratio .4 >> ceph osd pool set ssd_cache cache_target_full_ratio .8 > > Where are you seeing the "ceph osd pool set ssd_cache cache-mode > writeback" from? You're setting that with the "ceph osd tier > cache-mode ssd_cache writeback" command. :) > -Greg _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com