----- Original Message ----- > From: "Anuradha Talur" <atalur@xxxxxxxxxx> > To: "songxin" <songxin_1980@xxxxxxx> > Cc: "gluster-user" <gluster-users@xxxxxxxxxxx> > Sent: Thursday, March 3, 2016 12:31:41 PM > Subject: Re: about tail command > > > > ----- Original Message ----- > > From: "songxin" <songxin_1980@xxxxxxx> > > To: "Anuradha Talur" <atalur@xxxxxxxxxx> > > Cc: "gluster-user" <gluster-users@xxxxxxxxxxx> > > Sent: Wednesday, March 2, 2016 4:09:01 PM > > Subject: Re:Re: about tail command > > > > > > > > Thank you for your reply.I have two more questions as below > > > > > > 1. the command "gluster v replace-brick " is async or sync? The replace is > > complete when the command quit ? > It is a sync command, replacing the brick finishes as the command returns. > > In one of the earlier mails I gave incomplete command for replace brick, > sorry about that. > The only replace-brick operation allowed from glusterfs 3.7.9 onwards is > 'gluster v replace-brick <volname> <hostname:src_brick> <hostname:dst_brick> > commit force'. Sorry for spamming, but there is a typo here, I meant glusterfs 3.7.0 onwards, not 3.7.9. > > 2.I run "tail -n 0" on mount point.Does it trigger the heal? > > > > > > Thanks, > > Xin > > > > > > > > > > > > > > > > At 2016-03-02 15:22:35, "Anuradha Talur" <atalur@xxxxxxxxxx> wrote: > > > > > > > > >----- Original Message ----- > > >> From: "songxin" <songxin_1980@xxxxxxx> > > >> To: "gluster-user" <gluster-users@xxxxxxxxxxx> > > >> Sent: Tuesday, March 1, 2016 7:19:23 PM > > >> Subject: about tail command > > >> > > >> Hi, > > >> > > >> recondition: > > >> A node:128.224.95.140 > > >> B node:128.224.162.255 > > >> > > >> brick on A node:/data/brick/gv0 > > >> brick on B node:/data/brick/gv0 > > >> > > >> > > >> reproduce steps: > > >> 1.gluster peer probe 128.224.162.255 (on A node) > > >> 2. gluster volume create gv0 replica 2 128.224.95.140:/data/brick/gv0 > > >> 128.224.162.255:/data/brick/gv0 force (on A node) > > >> 3.gluster volume start gv0 (on A node) > > >> 4. mount -t glusterfs 128.224.95.140:/gv0 gluster (on A node) > > >> 5.create some files (a,b,c) in dir gluster (on A node) > > >> 6.shutdown the B node > > >> 7.change the files (a,b,c) in dir gluster (on A node) > > >> 8.reboot B node > > >> 9.start glusterd on B node but glusterfsd is offline (on B node) > > >> 10. gluster volume remove-brick gv0 replica 1 > > >> 128.224.162.255:/data/brick/gv0 > > >> force (on A node) > > >> 11. gluster volume add-brick gv0 replica 2 > > >> 128.224.162.255:/data/brick/gv0 > > >> force (on A node) > > >> > > >> Now the files are not same between two brick > > >> > > >> 12." gluster volume heal gv0 info " show entry num is 0 (on A node) > > >> > > >> Now What I should do if I want to sync file(a,b,c) on two brick? > > >> > > >Currently, once you add a brick to a cluster, files won't sync > > >automatically. > > >Patch has been sent to handle this requirement. Auto-heal will be > > >available > > >soon. > > > > > >You could kill the newly added brick and perform the following operations > > >from mount > > >for the sync to start : > > >1) create a directory <dirname> > > >2) setfattr -n "user.dirname" -v "value" <dirname> > > >3) delete the directory <dirname> > > > > > >Once these steps are done, start the killed brick. self-heal-daemon will > > >heal the files. > > > > > >But, for the case you have mentioned, why are you removing brick and using > > >add-brick again? > > >Is it because you don't want to change the brick-path? > > > > > >You could use "replace-brick" command. > > >gluster v replace-brick <volname> <hostname:old-brick-path> > > ><hostname:new-brick-path> > > >Note that source and destination should be different for this command to > > >work. > > > > > >> I know the "heal full" can work , but I think the command take too long > > >> time. > > >> > > >> So I run "tail -n 1 file" to all file on A node, but some files are sync > > >> but > > >> some files are not. > > >> > > >> My question is below: > > >> 1.Why the tail can't sync all files? > > >Did you run the tail command on mount point or from the backend (bricks)? > > >If you run from bricks, sync won't happen. Was client-side healing on? > > >To check if they were on or off, run `gluster v get volname all | grep > > >self-heal`, cluster.metadata-self-heal, cluster.data-self-heal, > > >cluster.entry-self-heal should be on. > > > > > >> 2.Can the command "tail -n 1 filename" trigger selfheal, just like "ls > > >> -l > > >> filename"? > > >> > > >> Thanks, > > >> Xin > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> _______________________________________________ > > >> Gluster-users mailing list > > >> Gluster-users@xxxxxxxxxxx > > >> http://www.gluster.org/mailman/listinfo/gluster-users > > > > > >-- > > >Thanks, > > >Anuradha. > > > > -- > Thanks, > Anuradha. > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users > -- Thanks, Anuradha. _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users