Note that we have had issues with deadlocks when re-exporting CephFS via Samba. It appears to only occur with Mac clients, though. In some cases it has hung on a request for a high-level directory and hung that branch for all clients. On Fri, Mar 13, 2020 at 1:56 AM Konstantin Shalygin <k0ste@xxxxxxxx> wrote: > > > On 3/11/20 11:16 PM, Seth Galitzer wrote: > > I have a hybrid environment and need to share with both Linux and > > Windows clients. For my previous iterations of file storage, I > > exported nfs and samba shares directly from my monolithic file server. > > All Linux clients used nfs and all Windows clients used samba. Now > > that I've switched to ceph, things are a bit more complicated. I built > > a gateway to export nfs and samba as needed, and connect that as a > > client to my ceph cluster. > > > > After having file locking problems with kernel nfs, I made the switch > > to nfs-ganesha, which has helped immensely. For Linux clients that > > have high I/O needs, like desktops and some web servers, I connect to > > ceph directly for those shares. For all other Linux needs, I use nfs > > from the gateway. For all Windows clients (desktops and a small number > > of servers), I use samba exported from the gateway. > > > > Since my ceph cluster went live in August, I have had some kind of > > strange (to me) error at least once a week, almost always related to > > the gateway client. Last night, it was MDS_CLIENT_OLDEST_TID. Since > > we're on Spring Break at my university and not very busy, I decided to > > unmount/remount the ceph share, requiring stopping nfs and samba > > services. Stopping nfs-ganesha took a while, but it finally completed > > with no complaints from the ceph cluster. Stopping samba took longer > > and gave me MDS_SLOW_REQUEST and MDS_CLIENT_LATE_RELEASE on the mds. > > It finally finished, and I was able to unmount/remount the ceph share > > and that finally cleared all the errors. > > > > This is leading me to believe that samba on the gateway and all the > > clients attaching to that is putting a strain on the connection back > > to ceph. Which finally brings me to my question: is there a better way > > to export samba to my clients using the ceph back end? Or is this as > > good as it gets and I just have to put up with the seemingly frequent > > errors? I can live with the errors and have been able to handle them > > so far, but I know people who have much bigger clusters and many more > > clients than me (by an order of magnitude) and don't see nearly as > > many errors as I do. Which is why I'm trying to figure out what is > > special about my setup. > > > > All my ceph nodes are running latest nautilus on Centos 7 (I just > > updated last week to 14.2.8), as is the gateway host. I'm mounting > > ceph directly on the gateway (by way of the kernel using cephfs, not > > rados/rbd) to a single mount point and exporting from there. > > > > My searches so far have not turned up anything extraordinarily useful, > > so I'm asking for some guidance here. Any advice is welcome. > > You can connect to your cluster directly from userland, without kernel. > Use Samba vfs_ceph for this. > > > > k > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx