Yes (I mean yes, it's real). Ceph's tiering works by moving whole (4MB)
objects to the cache pool, updating them there (with 4K random writes?)
and evicting them back when cache is full. I.e. the bad part here is that
it can't do "write-through".
Also there are some configuration options regarding the eviction process,
you can try to tune them. But don't expect the basis to change: when the
cache pool is full, Ceph will still need to evict something from there.
Why do you want cache tiering at all? Just use `allow_ec_overwrites=true`
if you're using EC and mount your RBD/CephFS directly without a cache pool.
During one of my test i found that fio inside my VM generates 1 MiB/s
(about 150 IOPS), but `ceph -s' shows me 500 MiB/s of flushing and 280
MiB/s of evicting data. How it could be? Is it real? Do you have any
optimization policies inside CEPH to eliminate such behaviour?
--
With best regards,
Vitaliy Filippov
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com