Hello, On Fri, 2 Jun 2017 14:30:56 +0800 jiajia zhong wrote: > christian, thanks for your reply. > > 2017-06-02 11:39 GMT+08:00 Christian Balzer <chibi@xxxxxxx>: > > > On Fri, 2 Jun 2017 10:30:46 +0800 jiajia zhong wrote: > > > > > hi guys: > > > > > > Our ceph cluster is working with tier cache. > > If so, then I suppose you read all the discussions here as well and not > > only the somewhat lacking documentation? > > > > > I am running "rados -p data_cache cache-try-flush-evict-all" to evict all > > > the objects. > > Why? > > And why all of it? > > > we found that when the threshold(flush/evict) was triggered, the > performance would make us a bit upset :), so I wish to flush/evict the tier > in a spare time,eg, middle night,In this scenario,the tier could not pay > any focus on flush/evict while the great w/r operations on cephfs which we > are using. > As I said, eviction (which is basically zeroing the data in cache) has very little impact. Flushing, moving data from the cache tier to the main pool is another story. But what you're doing here is completely invalidating your cache (the eviction part), so the performance will be very bad after this as well. If you have low utilization periods, consider a cron job that lowers the dirty ratio (causing only flushes to happen) and then after a while (few minutes should do, experiment) restore the old setting. For example: --- # Preemptive Flushing before midnight 45 23 * * * root ceph osd pool set cache cache_target_dirty_ratio 0.52 # And back to normal levels 55 23 * * * root ceph osd pool set cache cache_target_dirty_ratio 0.60 --- This will of course only help if the amount of data promoted into your cache per day is small enough to fit into the flushed space. Otherwise your cluster has no other choice to start flushing when things get full. Christian > > > > > But It a bit slow > > > > > Define slow, but it has to do a LOT of work and housekeeping to do this, > > so unless your cluster is very fast (probably not, or you wouldn't > > want/need a cache tier) and idle, that's the way it is. > > > > > 1. Is there any way to speed up the evicting? > > > > > Not really, see above. > > > > > 2. Is evicting triggered by itself good enough for cluster ? > > > > > See above, WHY are you manually flushing/evicting? > > > explained above. > > > > Are you aware that flushing is the part that's very I/O intensive, while > > evicting is a very low cost/impact operation? > > > not very sure, my instinct believed those. > > > > In normal production, the various parameters that control this will do > > fine, if properly configured of course. > > > > > 3. Does the flushing and evicting slow down the whole cluster? > > > > > Of course, as any good sysadmin with the correct tools (atop, iostat, > > etc, graphing Ceph performance values with Grafana/Graphite) will be able > > to see instantly. > > actually, we are using graphite, but I could not see that instantly, lol > :(, I could only got the threshold triggered by calculating after happening. > > btw, we have cephfs to store a huge number of small files, (64T , about > 100K per file), > > > > > > > > Christian > > -- > > Christian Balzer Network/Systems Engineer > > chibi@xxxxxxx Rakuten Communications > > -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Rakuten Communications _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com