2 servers which AFR eachother as server & client. I was pretty happy with it, I think if it could just not care about those silly simlink issues it'd be great. nice to know I can just add a brick and it'll just work. I didn't mind the bit of manual intervention since it was minimal and ultimate got the job done. but here's my config to help you understand. the configs on both servers are identical. volume home1 type storage/posix # POSIX FS translator option directory /gluster/home # Export this directory end-volume volume posix-locks-home1 type features/posix-locks option mandatory on subvolumes home1 end-volume ## Reference volume "home2" from remote server volume home2 type protocol/client # POSIX FS translator option transport-type tcp/client option remote-host ##.##.##.## # IP address of remote host option remote-subvolume posix-locks-home1 # use home1 on remote host option transport-timeout 10 # value in seconds; it should be set rel atively low end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes posix-locks-home1 option auth.addr.posix-locks-home1.allow ##.##.##.##,127.0.0.1 # Allow access to "home1" volume end-volume ### Create automatic file replication volume home type cluster/afr option read-subvolume posix-locks-home1 subvolumes posix-locks-home1 home2 # subvolumes posix-locks-home1 end-volume At 09:19 PM 10/10/2008, Krishna Srinivas wrote: >Keith, > >It was difficult to follow your mail. What is your setup? I understand >that when destination of symlinks don't exist it causes problems >during selfheal, is that right? Do you have server side afr or client >side? > >Krishna > >On Fri, Oct 10, 2008 at 4:16 PM, Keith Freedman ><freedman at freeformit.com> wrote: > > just a quick status for anyone who cares. > > > > once the find seemed to work properly for a few directories I > > remounted both servers with the proper AFR confi and they seem to be > > working just fine. auto-healing serverB as appropriate and sever1 > > getting updated when a file on serverb is updated. > > > > :) > > > > > > At 03:22 AM 10/10/2008, Keith Freedman wrote: > >>no one answered me so I'll just report my findings: > >> > >>Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 > >>Disk on serverb crashes, loose everything. re-install > >>copied over my AFR config from server1 change IP addresses as appropriate. > >> > >>using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null > >>to auto-heal them and it seems to be going just fine. > >> > >>once that's finished, I'll re-add serverB to server1's AFR config and > >>I presume it'll be fine. > >> > >>anyway, it was a minor irritation and overall the auto-healing once > >>going has been a lifesaver. > >> > >>Keith > >> > >>At 06:21 AM 10/9/2008, Keith Freedman wrote: > >> >I have 2 servers that AFR eachother. > >> > > >> >one of them suffered a drive failure and is being rebuilt. > >> > > >> >the question is. What will happen if I just mount the empty drive > >> >back as the AFR node. > >> > > >> >will it just start grabbing the data from the other server (which is > >> >exactly what I want), OR > >> >will it start deleting the data from the other server (which is > >> terribly bad). > >> > > >> >Another thought was to, on the current working server with good data, > >> >disable the remote afr node from it's config (so it's only using AFR > >> >on itself), and then leave the other machines config as is, and > turn it on. > >> >This way I can be sure that the node with the data wont go nuts and > >> >start deleting, but that updates to it will get replicated to the > >> >other machine. > >> > > >> >this particular set is running 1.4pre5 if that changes the answer. > >> > > >> >Thanks, > >> >Keith > >> > > >> > > >> >_______________________________________________ > >> >Gluster-users mailing list > >> >Gluster-users at gluster.org > >> >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >> > >> > >>_______________________________________________ > >>Gluster-users mailing list > >>Gluster-users at gluster.org > >>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >