On yet another related note: I noticed that - consistent with the keeping the file system attributes - the files them selves are left intact on the old brick. Once I remove the file system attributes... > # for x in `getfattr -m - /srv/sda7 2> /dev/null | tail -n +2` ; do setfattr -x $x /srv/sda7 ; done ....should I remove the old files? Or should I leave the files intact and migrate back to the original brick: > gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 start ...and then heal the volume? Eric Pretorious Truckee, CA >________________________________ > From: Eric <epretorious at yahoo.com> >To: "gluster-users at gluster.org" <gluster-users at gluster.org> >Sent: Wednesday, September 5, 2012 5:42 PM >Subject: Re: migration operations: Stopping a migration > > >On a related note: Now that the PoC has been completed, I'm not able to migrate back to the original brick because of the undocumented, save-you-from-yourself file system attribute feature: > > >> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda8 192.168.1.1:/srv/sda7 start >> /srv/sda7 or a prefix of it is already part of a volume > > > >Is there a simpler, more-direct method of migrating back to the original brick or should I wipe the file system attributes manually? I only ask because: > > >1. the long-term effects of this strategy aren't addressed in the Administration Guide AFAICT, and; > >2. though the intent of the feature has merit, it lacks elegance. e.g., the addition of a "force" attribute > >?? (like that of the commit feature) could be justified in this instance, IMHO. > > >?? > gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda8 192.168.1.1:/srv/sda7 start force >?? > Usage: volume replace-brick <VOLNAME> <BRICK> <NEW-BRICK> {start|pause|abort|status|commit [force]} > > > >Eric Pretorious >Truckee, CA > > > > >>________________________________ >> From: Eric <epretorious at yahoo.com> >>To: "gluster-users at gluster.org" <gluster-users at gluster.org> >>Sent: Wednesday, September 5, 2012 5:27 PM >>Subject: Re: migration operations: Stopping a migration >> >> >>On a hunch, I attempted the "volume replace-brick <VOLNAME> <BRICK> <NEW-BRICK> commit" command and, without much fanfare, the volume information was updated: >> >> >>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 commit >>> replace-brick commit successful >>> >>> gluster> volume info >>>? >>> Volume Name: Repositories >>> Type: Distributed-Replicate >>> Volume ID: 926262ae-2aa6-4bf7-b19e-cf674431b06c >>> Status: Started >>> Number of Bricks: 2 x 2 = 4 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 192.168.1.1:/srv/sda8 >>> Brick2: 192.168.1.2:/srv/sda7 >>> Brick3: 192.168.1.1:/srv/sdb7 >>> Brick4: 192.168.1.2:/srv/sdb7 >>> >>> gluster> volume status >>> Status of volume: Repositories >>> Gluster process??? ??? ??? ??? ??? ??? Port??? Online??? Pid >>> ------------------------------------------------------------------------------ >>> Brick 192.168.1.1:/srv/sda8??? ??? ??? ??? 24012??? Y??? 13796 >>> Brick 192.168.1.2:/srv/sda7??? ??? ??? ??? 24009??? Y??? 4946 >>> Brick 192.168.1.1:/srv/sdb7??? ??? ??? ??? 24010??? Y??? 5438 >>> Brick 192.168.1.2:/srv/sdb7??? ??? ??? ??? 24010??? Y??? 4951 >>> NFS Server on localhost??? ??? ??? ??? ??? 38467??? Y??? 13803 >>> Self-heal Daemon on localhost??? ??? ??? ??? N/A??? Y??? 13808 >>> NFS Server on 192.168.1.2??? ??? ??? ??? 38467??? Y??? 7969 >>> Self-heal Daemon on 192.168.1.2??? ??? ??? ??? N/A??? Y??? 7974 >> >> >>The XFS attributes are still intact on the old brick, however: >> >> >>> [eric at sn1 ~]$ for x in sda7 sdb7 sda8 ; do sudo getfattr -m - /srv/$x 2> /dev/null ; done >>> # file: srv/sda7 >>> trusted.afr.Repositories-client-0 >>> trusted.afr.Repositories-client-1 >>> trusted.afr.Repositories-io-threads >>> trusted.afr.Repositories-replace-brick >>> trusted.gfid >>> trusted.glusterfs.dht >>> trusted.glusterfs.volume-id >>> >>> # file: srv/sdb7 >>> trusted.afr.Repositories-client-2 >>> trusted.afr.Repositories-client-3 >>> trusted.gfid >>> trusted.glusterfs.dht >>> trusted.glusterfs.volume-id >>> >>> # file: srv/sda8 >>> trusted.afr.Repositories-io-threads >>> trusted.afr.Repositories-replace-brick >>> trusted.gfid >>> trusted.glusterfs.volume-id >> >> >> >>Is this intentional (i.e., leaving the the attributes intact)? Or functionality that has yet to be implemented? >> >> >>Eric Pretorious >>Truckee, CA >> >> >> >>>________________________________ >>> From: Eric <epretorious at yahoo.com> >>>To: "gluster-users at gluster.org" <gluster-users at gluster.org> >>>Sent: Wednesday, September 5, 2012 5:05 PM >>>Subject: migration operations: Stopping a migration >>> >>> >>>I've created a distributed replicated volume: >>> >>> >>>> gluster> volume info >>>>? >>>> Volume Name: Repositories >>>> Type: Distributed-Replicate >>>> Volume ID: 926262ae-2aa6-4bf7-b19e-cf674431b06c >>>> Status: Started >>>> Number of Bricks: 2 x 2 = 4 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: 192.168.1.1:/srv/sda7 >>>> Brick2: 192.168.1.2:/srv/sda7 >>>> Brick3: 192.168.1.1:/srv/sdb7 >>>> Brick4: 192.168.1.2:/srv/sdb7 >>> >>> >>> >>>...and begun migrating data from one brick to another as a PoC: >>> >>> >>>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 start >>>> replace-brick started successfully >>>> >>>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 status >>>> Number of files migrated = 5147?????? Current file= /centos/5.8/os/x86_64/CentOS/gnome-pilot-conduits-2.0.13-7.el5.x86_64.rpm >>>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 status >>>> Number of files migrated = 24631??????? Migration complete >>> >>> >>>After the migration is finished, though, the list of bricks is wrong: >>> >>> >>> >>>> gluster> volume heal Repositories info????????????????????????????????????????????????????????? >>>> Heal operation on volume Repositories has been successful >>>> >>>> Brick 192.168.1.1:/srv/sda7 >>>> Number of entries: 0 >>>> >>>> Brick 192.168.1.2:/srv/sda7 >>>> Number of entries: 0 >>>> >>>> Brick 192.168.1.1:/srv/sdb7 >>>> Number of entries: 0 >>>> >>>> Brick 192.168.1.2:/srv/sdb7 >>>> Number of entries: 0 >>> >>> >>>...and the XFS attributes are still intact on the old brick: >>> >>> >>>> [eric at sn1 ~]$ for x in sda7 sdb7 sda8 ; do sudo getfattr -m - /srv/$x 2> /dev/null ; done >>>> # file: srv/sda7 >>>> trusted.afr.Repositories-client-0 >>>> trusted.afr.Repositories-client-1 >>>> trusted.afr.Repositories-io-threads >>>> trusted.afr.Repositories-replace-brick >>>> trusted.gfid >>>> trusted.glusterfs.dht >>>> trusted.glusterfs.pump-path >>>> trusted.glusterfs.volume-id >>>> >>>> # file: srv/sdb7 >>>> trusted.afr.Repositories-client-2 >>>> trusted.afr.Repositories-client-3 >>>> trusted.gfid >>>> trusted.glusterfs.dht >>>> trusted.glusterfs.volume-id >>>> >>>> # file: srv/sda8 >>>> trusted.afr.Repositories-io-threads >>>> trusted.afr.Repositories-replace-brick >>>> trusted.gfid >>>> trusted.glusterfs.volume-id >>> >>> >>> >>>Have I missed a step? Or: Is this (i.e., clean-up) a bug or functionality that hasn't been implemented yet? >>> >>> >>>Eric Pretorious >>>Truckee, CA >>> >>>_______________________________________________ >>>Gluster-users mailing list >>>Gluster-users at gluster.org >>>http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >>> >>> >>_______________________________________________ >>Gluster-users mailing list >>Gluster-users at gluster.org >>http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >> >> >> >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gluster.org/pipermail/gluster-users/attachments/20120905/7ba7255d/attachment-0001.htm>