Cache tier weirdness

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

still my test cluster with 0.94.6. 
It's a bit fuzzy, but I don't think I saw this with Firefly, but then
again that is totally broken when it comes to cache tiers (switching
between writeback and forward mode).

goat is a cache pool for rbd:
---
# ceph osd pool ls detail
pool 2 'rbd' replicated size 3 min_size 1 crush_ruleset 2 object_hash rjenkins pg_num 512 pgp_num 512 last_change 11729 lfor 11662 flags hashpspool tiers 9 read_tier 9 write_tier 9 stripe_width 0

pool 9 'goat' replicated size 1 min_size 1 crush_ruleset 3 object_hash rjenkins pg_num 128 pgp_num 128 last_change 11730 flags hashpspool,incomplete_clones tier_of 2 cache_mode writeback target_bytes 524288000 hit_set bloom{false_positive_probability: 0.05, target_size: 0, seed: 0} 3600s x1 stripe_width 0
---

Initial state is this:
---
# rados df
pool name                 KB      objects       clones     degraded      unfound           rd        rd KB           wr        wr KB
goat                      34          429            0            0           0         1051      4182046       145803     10617422
rbd                164080702        40747            0            0           0       419664     71142697      4430922    531299267
  total used       599461060        41176
  total avail     5301740284
  total space     5940328912
---

First we put some data in there with 
"rados -p rbd  bench 20 write -t 32 --no-cleanup"
which easily exceeds the target bytes of 512MB and gives us:
---
pool name                 KB      objects
goat                  356386          372 
---

For starters, that's not the number I would have expected given how this
configured:
cache_target_dirty_ratio: 0.5
cache_target_full_ratio: 0.9

Lets ignore (but not forget) that discrepancy for now.

After doing a read with "rados -p rbd  bench 20 rand -t 32" to my utter
bafflement I get:
---
pool name                 KB      objects
goat                    8226          199 
---

And after a second read it's all gone, looking at the network traffic it
all originated from the base pool nodes and got relayed through the node
hosting the cache pool:
---
pool name                 KB      objects
goat                      34          191
---

I verified that the actual objects are on the base pool with 4MB each,
while their "copies" are on the cache pool OSDs with zero length. 

Can anybody unbaffle me? ^o^

Christian
-- 
Christian Balzer        Network/Systems Engineer                
chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux