Dear Jeff, thanks a lot for your comment! It sounds like I hit a bug here. Would you still feel queasy if I added "option lookup-unhashed yes" to avoid dht? What I reported here was some initial testing. In the end I like to use glusterfs to provide a uniform name space for our O(20) workstations with lokal storage of 4 to 16TB. The data should be mirrored (once) for reliability and stored locally were possible for speed. I also would prefer not to glue together local disk using LVM or software raid to keep it easy to add/remove disks without having to expand a filesystem. Any hints how to set this up with glusterfs? ... Matt On Tuesday 06 July 2010 06:30:35 pm Jeff Darcy wrote: > On 07/06/2010 11:19 AM, Matthias Munnich wrote: > > Hi! > > > > I am trying to combine nufa, distribute and replicate but am running in > > to messages like > > > > ls: cannot open directory .: Stale NFS file handle > > > > When I try to list in the mounted directory. I don't use NFS at all and > > am puzzled as to what is going on. Attached you find my client config > > file. The comments marked "ok" are setups which work. However, more than > > one disk is local which let me to use 3 layers: > > 1: replicate, 2: distribute: 3: nufa > > but somehow this is not working. Does anybody spot what is wrong? > > Any help is appreciated. > > First, you can pretty much ignore the reference to NFS. It's just a bad > errno-to-string conversion. > > Second, it seems like there are several places where we treat ESTALE > specially, but only one in the I/O path where we generate it. That one > is in dht-common.c, which is shared between distribute and nufa. The > function is dht_revalidate_cbk, and the ESTALE comes from detecting that > the dht "layout" structure is inconsistent. This leads me to wonder > whether the problem has to do with the fact that distribute/nufa both > use this code and the same set of extended attributes, and might be > stepping on each other. In general, having explored in some depth how > these translators work, the idea of stacking nufa/distribute on top of > one another (or themselves) makes me a bit queasy. > > >From your volfile, it looks like you want to create files on one of two > > filesystems replicated between localhost and mentha, and look for files > created elsewhere on dahlia and salvia. Assuming the four nodes are > similar, you might want to consider using nufa with local-volume-name > set to one of the two replicated subvolumes, and let mentha use the > other replicated subvolume for the other direction. Also, you should be > able to use the localhost filesystems with just storage/posix instead of > protocol/client (I assume you must have a separate glusterfsd running > for this setup to work) which would eliminate some context switches and > another layer of translator hierarchy. See > http://www.gluster.com/community/documentation/index.php/NUFA_with_single_p > rocess for further examples and explanation, and good luck. > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >