You don't need to disconnect any clients from the RADOS cluster. Tiering configuration should be transparent to Ceph clients. On Fri, Dec 16, 2016 at 5:57 PM, JiaJia Zhong <zhongjiajia@xxxxxxxxxxxx> wrote: > hi skinjo, > forgot to ask that if it's necessary to disconnect all the client before > doing set-overlay ? we didn't sweep the clients out while setting overlay > > ------------------ Original ------------------ > From: "JiaJia Zhong"<zhongjiajia@xxxxxxxxxxxx>; > Date: Wed, Dec 14, 2016 11:24 AM > To: "Shinobu Kinjo"<skinjo@xxxxxxxxxx>; > Cc: "CEPH list"<ceph-users@xxxxxxxxxxxxxx>; "ukernel"<ukernel@xxxxxxxxx>; > Subject: Re: can cache-mode be set to readproxy for > tiercachewith ceph 0.94.9 ? > > ------------------ Original ------------------ > From: "Shinobu Kinjo"<skinjo@xxxxxxxxxx>; > Date: Wed, Dec 14, 2016 10:56 AM > To: "JiaJia Zhong"<zhongjiajia@xxxxxxxxxxxx>; > Cc: "CEPH list"<ceph-users@xxxxxxxxxxxxxx>; "ukernel"<ukernel@xxxxxxxxx>; > Subject: Re: can cache-mode be set to readproxy for > tiercachewith ceph 0.94.9 ? > > >> ps: When we first met this issue, restarting the mds could cure that. (but >> that was ceph 0.94.1). > > Is this still working? > I think It's hard to reproduce the issue, the cluster woks well now. > Since you're using 0.94.9, bug(#12551) you mentioned was fixed. > > Can you do the followings to see object appear to you as ZERO size is > actually there: > # rados -p ${cache pool} ls > # rados -p ${cache pool} get ${object} /tmp/file > # ls -l /tmp/file > I did these with the ZERO file, It's the original OK file via rados get, > It's only not normal in cephfs > ------------------ Original ------------------ > From: "Shinobu Kinjo"<skinjo@xxxxxxxxxx>; > Date: Tue, Dec 13, 2016 06:21 PM > To: "JiaJia Zhong"<zhongjiajia@xxxxxxxxxxxx>; > Cc: "CEPH list"<ceph-users@xxxxxxxxxxxxxx>; "ukernel"<ukernel@xxxxxxxxx>; > Subject: Re: can cache-mode be set to readproxy for tier > cachewith ceph 0.94.9 ? > > > > On Tue, Dec 13, 2016 at 4:38 PM, JiaJia Zhong <zhongjiajia@xxxxxxxxxxxx> > wrote: >> >> hi cephers: >> we are using ceph hammer 0.94.9, yes, It's not the latest ( jewel), >> with some ssd osds for tiering, cache-mode is set to readproxy, >> everything seems to be as expected, >> but when reading some small files from cephfs, we got 0 bytes. > > > Would you be able to share: > > #1 How small is actual data? > #2 Is the symptom reproduceable with same size of different data? > #3 can you share your ceph.conf(ceph --show-config)? > >> >> >> I did some search and got the below link, >> >> http://ceph-users.ceph.narkive.com/g4wcB8ED/cephfs-with-cache-tiering-reading-files-are-filled-with-0s >> that's almost the same as what we are suffering from except the >> cache-mode in the link is writeback, ours is readproxy. >> >> that bug shall have been FIXED in 0.94.9 >> (http://tracker.ceph.com/issues/12551) >> but we still can encounter with that occasionally :( >> >> enviroment: >> - ceph: 0.94.9 >> - kernel client: 4.2.0-36-generic ( ubuntu 14.04 ) >> - any others needed ? >> >> Question: >> 1. does readproxy mode work on ceph0.94.9 ? since there are only >> writeback and readonly in the document for hammer. >> 2. any one with (Jewel or Hammer) met the same issue ? >> >> >> loop Yan, Zheng >> Quote from the link for convince. >> """ >> Hi, >> >> I am experiencing an issue with CephFS with cache tiering where the kernel >> clients are reading files filled entirely with 0s. >> >> The setup: >> ceph 0.94.3 >> create cephfs_metadata replicated pool >> create cephfs_data replicated pool >> cephfs was created on the above two pools, populated with files, then: >> create cephfs_ssd_cache replicated pool, >> then adding the tiers: >> ceph osd tier add cephfs_data cephfs_ssd_cache >> ceph osd tier cache-mode cephfs_ssd_cache writeback >> ceph osd tier set-overlay cephfs_data cephfs_ssd_cache >> >> While the cephfs_ssd_cache pool is empty, multiple kernel clients on >> different hosts open the same file (the size of the file is small, <10k) >> at >> approximately the same time. A number of the clients from the OS level see >> the entire file being empty. I can do a rados -p {cache pool} ls for the >> list of files cached, and do a rados -p {cache pool} get {object} >> /tmp/file >> and see the complete contents of the file. >> I can repeat this by setting cache-mode to forward, rados -p {cache pool} >> cache-flush-evict-all, checking no more objects in cache with rados -p >> {cache pool} ls, resetting cache-mode to writeback with an empty pool, and >> doing the multiple same file opens. >> >> Has anyone seen this issue? It seems like what may be a race condition >> where the object is not yet completely loaded into the cache pool so the >> cache pool serves out an incomplete object. >> If anyone can shed some light or any suggestions to help debug this issue, >> that would be very helpful. >> >> Thanks, >> Arthur""" >> >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com