No, I've never seen this issue on the Fedora stock kernels.
So either my workflow is not triggering it on the Fedora software stack or the issues is CentOS / RHEL - specific.
Anyway I will file the ceph-fuse bug then.
On Tue, Feb 2, 2016 at 12:43 PM, Yan, Zheng <ukernel@xxxxxxxxx> wrote:
On Tue, Feb 2, 2016 at 5:32 PM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote:One of my clients is using 4.3.5-300.fc23.x86_64 (Fedora release 23)did you encounter this problem on client using 4.3.5 kernel? If you did, this issue should be ceph-fuse bug.while all the other clients reply on 3.10.0-327.4.4.el7.x86_64 (CentOS Linux release 7.2.1511) Should I file report a bug on the RedHat bugzilla?you can open a bug at http://tracker.ceph.com/projects/cephfs/issues Regards Yan, ZhengOn Tue, Feb 2, 2016 at 8:57 AM, Yan, Zheng <ukernel@xxxxxxxxx> wrote: On Tue, Feb 2, 2016 at 2:27 AM, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: What version are you running on your servers and clients? Are you using 4.1 or 4.2 kernel? https://bugzilla.kernel.org/show_bug.cgi?id=104911. Upgrade to 4.3+ kernel or 4.1.17 kernel or 4.2.8 kernel can resolve this issue. On the clients: ceph-fuse --version ceph version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) MDS/OSD/MON: ceph --version ceph version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) Exactly what changes are you making that aren't visible? I am creating some new files in non-root folders. What's the output of "ceph -s"? ceph -s cluster 98d72518-6619-4b5c-b148-9a781ef13bcb health HEALTH_OK monmap e1: 1 mons at {000-s-ragnarok=XXX.XXX.XXX.XXX:6789/0} election epoch 1, quorum 0 000-s-ragnarok mdsmap e576: 1/1/1 up {0=000-s-ragnarok=up:active} osdmap e233: 16 osds: 16 up, 16 in flags sortbitwise pgmap v1927636: 1088 pgs, 2 pools, 1907 GB data, 2428 kobjects 3844 GB used, 25949 GB / 29793 GB avail 1088 active+clean client io 4381 B/s wr, 2 op In addition on the clients' side I have cat /etc/fuse.conf user_allow_other auto_cache large_read max_write = 16777216 max_read = 16777216 -Mykola On Mon, Feb 1, 2016 at 5:06 PM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: On Monday, February 1, 2016, Mykola Dvornik <mykola.dvornik@xxxxxxxxx> wrote: Hi guys, This is sort of rebuttal. I have a CephFS deployed and mounted on a couple of clients via ceph-fuse (due to quota support and possibility to kill the ceph-fuse process to avoid stale mounts). So the problems is that some times the changes made on one client are not visible on the others. It appears to me as rather random process. The only solution is to touch a new file in any particular folder that apparently triggers synchronization. I've been using a kernel-side client before with no such kind of problems. So the questions is it expected behavior of ceph-fuse? What version are you running on your servers and clients? Exactly what changes are you making that aren't visible? What's the output of "ceph -s"? We see bugs like this occasionally but I can't think of any recent ones in ceph-fuse -- they're actually seen a lot more often in the kernel client. -Greg Regards, Mykola _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com