ok - I figured it out (at least for version 1.3 that ships with ubuntu - as long as there the files exist (even at 0 length) - on ehe new volume- it will start to sync them. I'll go to the latest release and try out a few more scenarios. (That's good enuff for me - I'm not looking for perfect just something reasonable.) -Adrian On Sun, May 3, 2009 at 12:48 AM, Adrian Terranova <aterranova at gmail.com>wrote: > crap - just realized I cut pasted the server2x (sorry about that.) > > volume client3 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 127.0.0.1 # IP address of the remote brick > option remote-port 6998 # default server port is 6996 > option remote-subvolume brick3 # name of the remote volume > end-volume > > ## Add AFR (Automatic File Replication) feature. > volume afr > type cluster/afr > subvolumes client3 client1 client2 > # option replicate *:3 > end-volume > > > [snip] > > On Sun, May 3, 2009 at 12:42 AM, Adrian Terranova <aterranova at gmail.com>wrote: > >> Hello all, >> >> I've setup AFR - and am very impressed with the product - however - when I >> do a delete of /home/export1 and /home/export2 - what needs to happen for >> autoheal to happen? (I would like to understand this in some detail before >> implementing for my home directory data (mostly - just trying to work out >> the procedure for adding / replacing a volume)- I tried remounting the >> client, and restarting the server with a couple of find variations- none >> seemed to work) Is this an artifact of my one host setup or something? >> >> New files seems to show up - but the existing files / directories don't >> seem to come back when I read them. >> >> How would I get my files back onto replaced subvolumes? >> >> --Adrian >> >> >> >> XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX >> [snip]server >> peril at mythbuntufe-desktop:/etc/glusterfs$ grep -v \^# >> glusterfs-server.vol |more >> >> >> >> volume posix1 >> type storage/posix # POSIX FS translator >> option directory /home/export1 # Export this directory >> end-volume >> >> volume brick1 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix1 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6996 # Default is 6996 >> subvolumes brick1 >> option auth.ip.brick1.allow * # access to "brick" volume >> end-volume >> >> >> >> volume posix2 >> type storage/posix # POSIX FS translator >> option directory /home/export2 # Export this directory >> end-volume >> >> volume brick2 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix2 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6997 # Default is 6996 >> subvolumes brick2 >> option auth.ip.brick2.allow * # Allow access to "brick" volume >> end-volume >> >> >> >> >> volume posix3 >> type storage/posix # POSIX FS translator >> option directory /home/export3 # Export this directory >> end-volume >> >> volume brick3 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix3 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6998 # Default is 6996 >> subvolumes brick3 >> option auth.ip.brick3.allow * # access to "brick" volume >> end-volume >> >> >> XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX >> >> [snip]client >> peril at mythbuntufe-desktop:/etc/glusterfs$ grep -v \^# >> glusterfs-server.vol |more >> >> >> >> volume posix1 >> type storage/posix # POSIX FS translator >> option directory /home/export1 # Export this directory >> end-volume >> >> volume brick1 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix1 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6996 # Default is 6996 >> subvolumes brick1 >> option auth.ip.brick1.allow * # access to "brick" volume >> end-volume >> >> >> >> volume posix2 >> type storage/posix # POSIX FS translator >> option directory /home/export2 # Export this directory >> end-volume >> >> volume brick2 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix2 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6997 # Default is 6996 >> subvolumes brick2 >> option auth.ip.brick2.allow * # Allow access to "brick" volume >> end-volume >> >> >> >> >> volume posix3 >> type storage/posix # POSIX FS translator >> option directory /home/export3 # Export this directory >> end-volume >> >> volume brick3 >> type features/posix-locks >> option mandatory on # enables mandatory locking on all files >> subvolumes posix3 >> end-volume >> >> volume server >> type protocol/server >> option transport-type tcp/server # For TCP/IP transport >> option listen-port 6998 # Default is 6996 >> subvolumes brick3 >> option auth.ip.brick3.allow * # access to "brick" volume >> end-volume >> >> >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090503/366a4335/attachment.htm>