Hi Eugen, Sorry for the missing information. "cached-hdd-cache" is the overlay tier of "cached-hdd" and configured in "readproxy" mode. $ ceph osd dump | grep cached-hdd pool 24 'cached-hdd' replicated size 3 min_size 2 crush_rule 1 object_hash rjenkins pg_num 512 pgp_num 512 autoscale_mode warn last_change 27242 lfor 22992/22992/22992 flags hashpspool,selfmanaged_snaps tiers 25 read_tier 25 write_tier 25 stripe_width 0 application rbd pool 25 'cached-hdd-cache' replicated size 3 min_size 1 crush_rule 3 object_hash rjenkins pg_num 128 pgp_num 128 autoscale_mode warn last_change 30376 lfor 22992/30306/30304 flags hashpspool,incomplete_clones,selfmanaged_snaps tier_of 24 cache_mode readproxy target_bytes 1099511627776 target_objects 500000 hit_set bloom{false_positive_probability: 0.05, target_size: 0, seed: 0} 1200s x4 decay_rate 20 search_last_n 1 min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0 Regs, Icy On Thu, 28 May 2020 at 18:25, Eugen Block <eblock@xxxxxx> wrote: > I don't see a cache_mode enabled on the pool, did you set one? > > > > Zitat von icy chan <icy.kf.chan@xxxxxxxxx>: > > > Hi, > > > > I had configured a cache tier with max object counts 500k. But no evict > > happens when the object counts hit the configured maximum. > > Anyone experienced this issue? What should I do? > > > > $ ceph health detail > > HEALTH_WARN 1 cache pools at or near target size > > CACHE_POOL_NEAR_FULL 1 cache pools at or near target size > > cache pool 'cached-hdd-cache' with 887.11k objects at/near target max > > 500k objects > > > > $ ceph df | grep -e "POOL\|cached-hdd" > > POOLS: > > POOL ID STORED OBJECTS USED %USED > > MAX AVAIL > > cached-hdd 24 1.4 TiB 1.52M 1.4 TiB 0.60 > > 78 TiB > > cached-hdd-cache 25 842 GiB 887.14k 842 GiB 15.97 > > 1.4 TiB > > > > $ ceph osd pool get cached-hdd-cache all > > size: 3 > > min_size: 1 > > pg_num: 128 > > pgp_num: 128 > > crush_rule: nvme-repl-rule > > hashpspool: true > > nodelete: false > > nopgchange: false > > nosizechange: false > > write_fadvise_dontneed: false > > noscrub: false > > nodeep-scrub: false > > hit_set_type: bloom > > hit_set_period: 1200 > > hit_set_count: 4 > > hit_set_fpp: 0.05 > > use_gmt_hitset: 1 > > target_max_objects: 500000 > > target_max_bytes: 1099511627776 > > cache_target_dirty_ratio: 0 > > cache_target_dirty_high_ratio: 0.7 > > cache_target_full_ratio: 0.9 > > cache_min_flush_age: 0 > > cache_min_evict_age: 0 > > min_read_recency_for_promote: 1 > > min_write_recency_for_promote: 1 > > fast_read: 0 > > hit_set_grade_decay_rate: 20 > > hit_set_search_last_n: 1 > > pg_autoscale_mode: warn > > > > > > Regs, > > Icy > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx