Szabolcs, do you still face this issue? avati 2008/1/19, Székelyi Szabolcs <cc@xxxxxxxxx>: > > Anand Avati wrote: > > Szabolcs, > > I suspect it might be an issue with 2.7.2-glfs8. We are seeing similar > > issues with the 2.7.2 fuse. Please let us know if 2.7.0 works well for > you. > > Well, with fuse-2.7.0-glfs7, the same happens. > > -- > Szabolcs > > > 2008/1/17, Székelyi Szabolcs <cc@xxxxxxxxx <mailto:cc@xxxxxxxxx>>: > > > > AFR with posix-locks behaves really strange nowadays... GlusterFS is > a > > fresh TLA checkout (patch-636), FUSE is brand the new 2.7.2-glfs8. > > > > I have 4 servers with a 4-way AFR on each and features/posix-locks > > loaded just above storage/posix bricks. On each AFR, one replica is > the > > local storage, the remaining 3 are on the other 3 servers. > > > > The 4 AFR bricks are mounted on each server from 'localhost'. > > > > The machines are freshly booted. Basic FS functions (ls, copy, cat) > work > > fine. > > > > Now I run a distributed locking test using [1]. On the "master" > locker I > > get: > > > > > # /tmp/locktests -n 10 -c 3 -f /mnt/glusterfs/testfile > > > Init > > > process initalization > > > .................... > > > -------------------------------------- > > > > > > TEST : TRY TO WRITE ON A READ LOCK:========== > > > TEST : TRY TO WRITE ON A WRITE LOCK:========== > > > TEST : TRY TO READ ON A READ LOCK:========== > > > TEST : TRY TO READ ON A WRITE LOCK:========== > > > TEST : TRY TO SET A READ LOCK ON A READ LOCK: > > > > After about 5 minutes, another > > > > > RDONLY: fcntl: Transport endpoint is not connected > > > > appears, and the locking processes exit on all slave servers, the > master > > blocks. > > > > The mount point locks up. Even an `ls` from a different terminal > seems > > to block forever. > > > > You can find my server config below. Client configs are simple, just > a > > protocol/client brick from localhost. I can provide server debug > logs if > > you need. > > > > Any idea? > > > > Thanks, > > -- > > Szabolcs > > > > > > [1] > > > http://nfsv4.bullopensource.org/tools/tests_tools/locktests-net.tar.gz > > > > > > My server config (from a single node, lu1): > > > > volume data-posix > > type storage/posix > > option directory /srv/glusterfs > > end-volume > > > > volume data1 > > type features/posix-locks > > subvolumes data-posix > > end-volume > > > > volume data2 > > type protocol/client > > option transport-type tcp/client > > option remote-host lu2 > > option remote-subvolume data2 > > end-volume > > > > volume data3 > > type protocol/client > > option transport-type tcp/client > > option remote-host lu3 > > option remote-subvolume data3 > > end-volume > > > > volume data4 > > type protocol/client > > option transport-type tcp/client > > option remote-host lu4 > > option remote-subvolume data4 > > end-volume > > > > volume data-afr > > type cluster/afr > > subvolumes data1 data2 data3 data4 > > end-volume > > > > volume server > > type protocol/server > > subvolumes data1 data-afr > > option transport-type tcp/server > > option auth.ip.data1.allow 10.0.0.* > > option auth.ip.data-afr.allow 127.0.0.1 <http://127.0.0.1>, > 10.0.0.* > > end-volume > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel > -- If I traveled to the end of the rainbow As Dame Fortune did intend, Murphy would be there to tell me The pot's at the other end.