Re: [SOLVED] Cache tier full not evicting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Nick.   That did it!   Cache cleans it self up now.   

On Sep 14, 2015, at 11:49 , Nick Fisk <nick@xxxxxxxxxx> wrote:

Have you set the target_max_bytes? Otherwise those ratios are not relative to anything, they use the target_max_bytes as a max, not the pool size.
 
From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of deeepdish
Sent: 14 September 2015 16:27
To: ceph-users@xxxxxxxxxxxxxx
Subject:  Cache tier full not evicting
 
Hi Everyone,
 
Getting close to cracking my understanding of cache tiering, and ec pools.   Stuck on one anomaly which I do not understand — spent hours reviewing docs online, can’t seem to pin point what I’m doing wrong.   Referencing http://xo4t.mj.am/link/xo4t/no2irn4/1/4BSmK1EUshpYjOdI2VWk4g/aHR0cDovL2NlcGguY29tL2RvY3MvbWFzdGVyL3JhZG9zL29wZXJhdGlvbnMvY2FjaGUtdGllcmluZy8
 
Setup:
 
Test / PoC Lab environment (not production)
 
1x [26x OSD/MON host]
1x MON VM
 
Erasure coded pool consisting of 10 spinning OSDs  (journals on SSDs - 5:1 spinner:SSD ratio)
Cache tier consisting of 2 SSD OSDs
 
Issue:
 
Cache tier is not honoring configured thresholds.   In my particular case, I have 2 OSDs in pool ‘cache’ (140G each == 280G total pool capacity).   
 
Pool cache is configured with replica factor of 2 (size = 2, min size = 1)
 
Initially I tried the following settings:
 
ceph osd pool set cache cache_target_dirty_ratio 0.3
ceph osd pool set cache cache_target_full_ratio 0.7
ceph osd pool set cache cache_min_flush_age 1
ceph osd pool set cache cache_min_evict_age 1
 
My cache tier’s utilization hit 96%+, causing the pool to run out of capacity.
 
I realized that in a replicated pool, only 1/2 the capacity is available and made the following adjustments:
 
ceph osd pool set cache cache_target_dirty_ratio 0.1
ceph osd pool set cache cache_target_full_ratio 0.3
ceph osd pool set cache cache_min_flush_age 1
ceph osd pool set cache cache_min_evict_age 1
 
The above implies that 0.3 = 60% of replicated (2x) pool size) and 0.1 = 20% of replicated (2x) pool size.   
 
Even with above revised values, I still see the cache tier getting full.  
 
The cache tier can only be flushed / evicted by manually running the following:
 
rados -p cache cache-flush-evict-all
 
Thank you.
 
 
 
 
 
 
 
 


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux