Thanks folks. > * 2.6.18 is crazy-old. Certainly. I'm hoping I only need to work with CentOS6, then I at least get 2.6.32. I wouldn't try to do RBD with these kernels, though. I would be more likely to have an Ubuntu 12.04 node mount an RBD and export with NFS. But as you mention, that has it's own issues. > Your other options are re-exporting NFS or CIFS, or trying to use ceph-fuse. ceph-fuse would only be for use with cephfs, no? Or can ceph-fuse interface to RBD or radosgw? I'm probably just slightly ahead of the curve here. I am trying to shoehorn ceph into my scenario (exporting a single filesystem to lots of cluster nodes) without using cephfs since it's not quite ready. Even though I lose parallel file access when using NFS (all traffic would go through that one NFS node), I was hoping that mounting an RBD, formatting it, and exporting it with NFS would avoid the ESTALE issues (as opposed to using cephfs, mounting via ceph-fuse, and then exporting _that_ with NFS). I actually have a phone call with inktank tomorrow. I'll ask some more questions then. - Travis On Wed, Jun 6, 2012 at 12:27 PM, Sage Weil <sage@xxxxxxxxxxx> wrote: > On Wed, 6 Jun 2012, Travis Rhoden wrote: >> Sorry for the re-post. First attempt got eaten by the mailing list >> because it wasn't plain text. My first attempt! >> >> > With CephFS not quite ready for production, what would be the >> > recommended way for sharing Ceph to older operating systems? >> > Specifically, I want to share a file system (i.e. mount a common >> > directory) on several CentOS5 and CentOS6 nodes. I don't think I can >> > mount an RBD from more than one node, so what is my best bet? >> > >> > Mount an RBD on a file "head" and export it via NFS? >> > >> > I feel like that is the only sane choice, but wanted to ask in case I >> > am missing something better. > > Two things: > > * 2.6.18 is crazy-old. It's possible the libceph and rbd bits could be > merged in, but getting the fs to work with ancient VFS will be a world of > pain. > > * NFS reexport works in basic scenarios, but should not be used in > production as there are various conditions that can lead to ESTALE. > Making it work is somewhere on the mid-term roadmap, but it'll be a while. > > Your best bet is to run cephfs on a modern kernel. But be warned, we > aren't ready to support the fs just yet. Soon! > > sage > > >> >> - Travis >> > >> > >> > On Wed, Jun 6, 2012 at 9:58 AM, Sam Zaydel <sam.zaydel@xxxxxxxxxxx> wrote: >> >> >> >> I think you will be able to build from source the client bits that you >> >> need, but I saw this from an earlier similar discussion mentioning >> >> Python version, so I would make sure that python > 2.5 is used, >> >> realistically 2.6 should be the norm today. >> >> >> >> http://comments.gmane.org/gmane.comp.file-systems.ceph.devel/5717 >> >> >> >> On Tue, Jun 5, 2012 at 6:16 PM, <Eric_YH_Chen@xxxxxxxxxx> wrote: >> >> > Dear All: >> >> > >> >> >ÿÿÿÿ My ceph cluster is installed on Ubuntu 12.04 with kernel 3.2.0. >> >> > >> >> >ÿÿÿÿ But I have a situation that I may need to mount cephfs on centos 5.7, which with kernel 2.6.18. >> >> > >> >> > (that is to say , I only need to execute mount.ceph on the centos 5.7, not the whole ceph system) >> >> > >> >> > I want to find a solution that can provide reliable and high availability share file system. >> >> > >> >> > Is it possible to do this? Or any other recommend way? Ex: export NFSÿÿ >> >> > >> >> > Thanks! >> >> > >> >> > >> >> > -- >> >> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> >> > the body of a message to majordomo@xxxxxxxxxxxxxxx >> >> > More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> >> >> >> >> >> >> -- >> >> Inked by Sam Zaydel >> >> -- >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> > >> > >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html