On Friday 06 November 2009 04:48:08 am Vikas Gorur wrote: > Krzysztof Strasburger wrote: > > volume replicated > > type cluster/replicate > > subvolumes sub1 sub2 > > end-volume > > > > and on host2: > > > > volume replicated > > type cluster/replicate > > subvolumes sub2 sub1 > > end-volume > > > > then following (positive) side-effects should occur: > > 1. After a crash, ls -R would correctly self-heal the volume either on > > host1 or on host2 (on this one, which has the newer sobvolume as the > > first on the list). > > 2. This is probably almost invisible, but the directory-read workload > > should be more equally distributed between sub1 and sub2. > > Is this the right workaround. > > This is not a workaround. Shuffling the order of subvolumes can have > disastrous consequences. > Replicate uses the first subvolume as the lock server, and if you > shuffle the order > the two clients will use different subvolumes as lock servers. This can > cause data to be inconsistent. > > We plan to fix this known issue, however, in one of the 3.x releases. If > you need a workaround, > the correct thing to do is generate a list of all files from the second > subvolume like this: > > [root at backend2] # find /export/directory/ > filelist.txt > > Then trigger self heal on all the files from the mountpoint: > > [root at mountpoint] # cat filelist.txt | xargs stat > > This will recreate all the files. > > Vikas > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > What if you set the lock servers option to the number of sub-volumes you have? -- Benjamin Long