On Tue, Feb 2, 2016 at 8:28 PM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: > No, I've never seen this issue on the Fedora stock kernels. > > So either my workflow is not triggering it on the Fedora software stack or > the issues is CentOS / RHEL - specific. I mean did you encounter this problem when using ceph-fuse on 4.3.5 kernel ? (fuse mount can also be affected by kernel bug) Regards Yan, Zheng > > Anyway I will file the ceph-fuse bug then. > > On Tue, Feb 2, 2016 at 12:43 PM, Yan, Zheng <ukernel@xxxxxxxxx> wrote: > > On Tue, Feb 2, 2016 at 5:32 PM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> > wrote: > > One of my clients is using 4.3.5-300.fc23.x86_64 (Fedora release 23) > > did you encounter this problem on client using 4.3.5 kernel? If you did, > this issue should be ceph-fuse bug. > > while all the other clients reply on 3.10.0-327.4.4.el7.x86_64 (CentOS Linux > release 7.2.1511) Should I file report a bug on the RedHat bugzilla? > > you can open a bug at http://tracker.ceph.com/projects/cephfs/issues Regards > Yan, Zheng > > On Tue, Feb 2, 2016 at 8:57 AM, Yan, Zheng <ukernel@xxxxxxxxx> wrote: On > Tue, Feb 2, 2016 at 2:27 AM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> > wrote: What version are you running on your servers and clients? Are you > using 4.1 or 4.2 kernel? https://bugzilla.kernel.org/show_bug.cgi?id=104911. > Upgrade to 4.3+ kernel or 4.1.17 kernel or 4.2.8 kernel can resolve this > issue. On the clients: ceph-fuse --version ceph version 9.2.0 > (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) MDS/OSD/MON: ceph --version ceph > version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) Exactly what > changes are you making that aren't visible? I am creating some new files in > non-root folders. What's the output of "ceph -s"? ceph -s cluster > 98d72518-6619-4b5c-b148-9a781ef13bcb health HEALTH_OK monmap e1: 1 mons at > {000-s-ragnarok=XXX.XXX.XXX.XXX:6789/0} election epoch 1, quorum 0 > 000-s-ragnarok mdsmap e576: 1/1/1 up {0=000-s-ragnarok=up:active} osdmap > e233: 16 osds: 16 up, 16 in flags sortbitwise pgmap v1927636: 1088 pgs, 2 > pools, 1907 GB data, 2428 kobjects 3844 GB used, 25949 GB / 29793 GB avail > 1088 active+clean client io 4381 B/s wr, 2 op In addition on the clients' > side I have cat /etc/fuse.conf user_allow_other auto_cache large_read > max_write = 16777216 max_read = 16777216 -Mykola On Mon, Feb 1, 2016 at 5:06 > PM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: On Monday, February 1, 2016, > Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: Hi guys, This is sort of > rebuttal. I have a CephFS deployed and mounted on a couple of clients via > ceph-fuse (due to quota support and possibility to kill the ceph-fuse > process to avoid stale mounts). So the problems is that some times the > changes made on one client are not visible on the others. It appears to me > as rather random process. The only solution is to touch a new file in any > particular folder that apparently triggers synchronization. I've been using > a kernel-side client before with no such kind of problems. So the questions > is it expected behavior of ceph-fuse? What version are you running on your > servers and clients? Exactly what changes are you making that aren't > visible? What's the output of "ceph -s"? We see bugs like this occasionally > but I can't think of any recent ones in ceph-fuse -- they're actually seen a > lot more often in the kernel client. -Greg Regards, Mykola > _______________________________________________ ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com