On Tue, Feb 2, 2016 at 5:32 PM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: > One of my clients is using > > 4.3.5-300.fc23.x86_64 (Fedora release 23) did you encounter this problem on client using 4.3.5 kernel? If you did, this issue should be ceph-fuse bug. > > while all the other clients reply on > > 3.10.0-327.4.4.el7.x86_64 (CentOS Linux release 7.2.1511) > > Should I file report a bug on the RedHat bugzilla? you can open a bug at http://tracker.ceph.com/projects/cephfs/issues Regards Yan, Zheng > > On Tue, Feb 2, 2016 at 8:57 AM, Yan, Zheng <ukernel@xxxxxxxxx> wrote: > > On Tue, Feb 2, 2016 at 2:27 AM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> > wrote: > > What version are you running on your servers and clients? > > Are you using 4.1 or 4.2 kernel? > https://bugzilla.kernel.org/show_bug.cgi?id=104911. Upgrade to 4.3+ kernel > or 4.1.17 kernel or 4.2.8 kernel can resolve this issue. > > On the clients: ceph-fuse --version ceph version 9.2.0 > (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) MDS/OSD/MON: ceph --version ceph > version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) Exactly what > changes are you making that aren't visible? I am creating some new files in > non-root folders. What's the output of "ceph -s"? ceph -s cluster > 98d72518-6619-4b5c-b148-9a781ef13bcb health HEALTH_OK monmap e1: 1 mons at > {000-s-ragnarok=XXX.XXX.XXX.XXX:6789/0} election epoch 1, quorum 0 > 000-s-ragnarok mdsmap e576: 1/1/1 up {0=000-s-ragnarok=up:active} osdmap > e233: 16 osds: 16 up, 16 in flags sortbitwise pgmap v1927636: 1088 pgs, 2 > pools, 1907 GB data, 2428 kobjects 3844 GB used, 25949 GB / 29793 GB avail > 1088 active+clean client io 4381 B/s wr, 2 op In addition on the clients' > side I have cat /etc/fuse.conf user_allow_other auto_cache large_read > max_write = 16777216 max_read = 16777216 -Mykola On Mon, Feb 1, 2016 at 5:06 > PM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: On Monday, February 1, 2016, > Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: > > Hi guys, This is sort of rebuttal. I have a CephFS deployed and mounted on a > couple of clients via ceph-fuse (due to quota support and possibility to > kill the ceph-fuse process to avoid stale mounts). So the problems is that > some times the changes made on one client are not visible on the others. It > appears to me as rather random process. The only solution is to touch a new > file in any particular folder that apparently triggers synchronization. I've > been using a kernel-side client before with no such kind of problems. So the > questions is it expected behavior of ceph-fuse? > > What version are you running on your servers and clients? Exactly what > changes are you making that aren't visible? What's the output of "ceph -s"? > We see bugs like this occasionally but I can't think of any recent ones in > ceph-fuse -- they're actually seen a lot more often in the kernel client. > -Greg > > Regards, Mykola > > _______________________________________________ ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com