08-Mar-16 00:41, Robert LeBlanc пишет:
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256
Did you also set "target_max_bytes" to the size of the pool? That bit
us when we didn't have it set. The ratio then uses the
target_max_bytes to know when to flush.
Yes, later I set this option.
But the cluster has earned again after I add new OSD in the cache tier
pool and 'full OSD' status was dropped.
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1
On Sun, Mar 6, 2016 at 2:17 AM, Mike Almateia <mike.almateia@xxxxxxxxx> wrote:
Hello Cephers!
When my cluster hit "full ratio" settings, objects from cache pull didn't
flush to a cold storage.
1. Hit the 'full ratio':
2016-03-06 11:35:23.838401 osd.64 10.22.11.21:6824/31423 4327 : cluster
[WRN] OSD near full (90%)
2016-03-06 11:35:55.447205 osd.64 10.22.11.21:6824/31423 4329 : cluster
[WRN] OSD near full (90%)
2016-03-06 11:36:29.255815 osd.64 10.22.11.21:6824/31423 4332 : cluster
[WRN] OSD near full (90%)
2016-03-06 11:37:04.769765 osd.64 10.22.11.21:6824/31423 4333 : cluster
[WRN] OSD near full (90%)
...
2. Well, ok. Set the option 'ceph osd pool set hotec cache_target_full_ratio
0.8'.
But no one of objects didn't flush at all
3. Ok. Try flush all object manually:
[root@c1 ~]# rados -p hotec cache-flush-evict-all
rbd_data.34d1f5746d773.0000000000016ba9
4. After full day objects still in cache pool, didn't flush at all:
[root@c1 ~]# rados df
pool name KB objects clones degraded unfound
rd rd KB wr wr KB
data 0 0 0 0 0
6 4 158212 215700473
hotec 797656118 25030755 0 0 0
370599 163045649 69947951 17786794779
rbd 0 0 0 0 0
0 0 0 0
total used 2080570792 25030755
It a bug or predictable action?
--
Mike. runs!
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com