Thanks. I updated to 2.0.0 but the daemons will not start and give a very generic error that does not help 2009-05-11 12:41:56 E [glusterfsd.c:483:_xlator_graph_init] drdan0199: validating translator failed 2009-05-11 12:41:56 E [glusterfsd.c:1145:main] glusterfs: translator initialization failed. exiting Can you see something wrong in the volume file? This works fine for 2.0.0rc4 --START-- volume storage type storage/posix option directory /scratch/glusterfs/export end-volume # Required for AFR (file replication) module volume locks type features/locks subvolumes storage end-volume volume brick type performance/io-threads #option thread-count 1 option thread-count 8 subvolumes locks end-volume volume server type protocol/server subvolumes brick option transport-type tcp option auth.addr.brick.allow 10.232.* end-volume volume drdan0191 type protocol/client option transport-type tcp option remote-host drdan0191.en.desres.deshaw.com option remote-subvolume brick end-volume volume drdan0192 type protocol/client option transport-type tcp option remote-host drdan0192.en.desres.deshaw.com option remote-subvolume brick end-volume [...] volume nufa type cluster/nufa option local-volume-name `hostname -s` #subvolumes replicate1 replicate2 replicate3 replicate4 replicate5 subvolumes drdan0191 drdan0192 drdan0193 drdan0194 drdan0195 drdan0196 drdan0197 drdan0198 drdan0199 drdan0200 end-volume # This, from https://savannah.nongnu.org/bugs/?24972, does the # filesystem mounting at server start time. Like an /etc/fstab entry volume fuse type mount/fuse option direct-io-mode 1 option entry-timeout 1 #option attr-timeout 1 (not recognized in 2.0) option mountpoint /mnt/glusterfs subvolumes nufa end-volume --END-- Thanks, fds ________________________________ From: Liam Slusser [mailto:lslusser at gmail.com] Sent: Thursday, May 07, 2009 1:51 PM To: Sacerdoti, Federico Cc: gluster-users at gluster.org Subject: Re: rm -rf errors You should try upgrading to the 2.0.0 release and try again. They fixed all sorts of bugs. liam On Thu, May 7, 2009 at 8:21 AM, Sacerdoti, Federico <Federico.Sacerdoti at deshawresearch.com> wrote: Hello, I am evaluating glusterfs and have seen some strange behavior with remove. I have gluster/2.0.0rc4 setup on 10 linux nodes connected with GigE. The config is Nufa/fuse with one storage brick per server, as seen in the attached nufa.vol config file, which I use for both clients and servers. My experiment is to launch 10 parallel writers, each of whom writes 32GiB worth of data in small files (2MB) to a shared gluster-fuse mounted filesystem. The files are named uniquely per client, so each file is only written once. This worked well, and I am seeing performance close to that of native disk, even with 8-writers per node. However when I do a parallel "rm -rf writedir/" on the 10 nodes, where writedir is the directory written in by the parallel writers described above, I see strange effects. There are 69,000 UNLINK errors in the glusterfsd.log of one server, in the form shown below. This alone is not surprising as the operation is ocurring in parallel. However the remove took much longer than expected, 92min, and more surprisingly the rm command exited 0 but files remained in the writedir! I ran rm -rf writedir from a single client, and it too exited 0 but left the writedir non-empty. Is this expected? Thanks, Federico --From glusterfsd.log-- 2009-05-04 11:35:15 E [fuse-bridge.c:964:fuse_unlink_cbk] glusterfs-fuse: 5764889: UNLINK() /write.2MB.runid1.p1/5 => -1 (No such file or directory) 2009-05-04 11:35:15 E [dht-common.c:1294:dht_err_cbk] nufa: subvolume drdan0192 returned -1 (No such file or directory) 2009-05-04 11:35:15 E [fuse-bridge.c:964:fuse_unlink_cbk] glusterfs-fuse: 5764894: UNLINK() /write.2MB.runid1.p1/51 => -1 (No such file or directory) --end-- <<nufa.vol>> _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090511/57af954a/attachment-0001.htm>