On a related note: Now that the PoC has been completed, I'm not able to migrate back to the original brick because of the undocumented, save-you-from-yourself file system attribute feature: > gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda8 192.168.1.1:/srv/sda7 start > /srv/sda7 or a prefix of it is already part of a volume Is there a simpler, more-direct method of migrating back to the original brick or should I wipe the file system attributes manually? I only ask because: 1. the long-term effects of this strategy aren't addressed in the Administration Guide AFAICT, and; 2. though the intent of the feature has merit, it lacks elegance. e.g., the addition of a "force" attribute ?? (like that of the commit feature) could be justified in this instance, IMHO. ?? > gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda8 192.168.1.1:/srv/sda7 start force ?? > Usage: volume replace-brick <VOLNAME> <BRICK> <NEW-BRICK> {start|pause|abort|status|commit [force]} Eric Pretorious Truckee, CA >________________________________ > From: Eric <epretorious at yahoo.com> >To: "gluster-users at gluster.org" <gluster-users at gluster.org> >Sent: Wednesday, September 5, 2012 5:27 PM >Subject: Re: migration operations: Stopping a migration > > >On a hunch, I attempted the "volume replace-brick <VOLNAME> <BRICK> <NEW-BRICK> commit" command and, without much fanfare, the volume information was updated: > > >> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 commit >> replace-brick commit successful >> >> gluster> volume info >>? >> Volume Name: Repositories >> Type: Distributed-Replicate >> Volume ID: 926262ae-2aa6-4bf7-b19e-cf674431b06c >> Status: Started >> Number of Bricks: 2 x 2 = 4 >> Transport-type: tcp >> Bricks: >> Brick1: 192.168.1.1:/srv/sda8 >> Brick2: 192.168.1.2:/srv/sda7 >> Brick3: 192.168.1.1:/srv/sdb7 >> Brick4: 192.168.1.2:/srv/sdb7 >> >> gluster> volume status >> Status of volume: Repositories >> Gluster process??? ??? ??? ??? ??? ??? Port??? Online??? Pid >> ------------------------------------------------------------------------------ >> Brick 192.168.1.1:/srv/sda8??? ??? ??? ??? 24012??? Y??? 13796 >> Brick 192.168.1.2:/srv/sda7??? ??? ??? ??? 24009??? Y??? 4946 >> Brick 192.168.1.1:/srv/sdb7??? ??? ??? ??? 24010??? Y??? 5438 >> Brick 192.168.1.2:/srv/sdb7??? ??? ??? ??? 24010??? Y??? 4951 >> NFS Server on localhost??? ??? ??? ??? ??? 38467??? Y??? 13803 >> Self-heal Daemon on localhost??? ??? ??? ??? N/A??? Y??? 13808 >> NFS Server on 192.168.1.2??? ??? ??? ??? 38467??? Y??? 7969 >> Self-heal Daemon on 192.168.1.2??? ??? ??? ??? N/A??? Y??? 7974 > > >The XFS attributes are still intact on the old brick, however: > > >> [eric at sn1 ~]$ for x in sda7 sdb7 sda8 ; do sudo getfattr -m - /srv/$x 2> /dev/null ; done >> # file: srv/sda7 >> trusted.afr.Repositories-client-0 >> trusted.afr.Repositories-client-1 >> trusted.afr.Repositories-io-threads >> trusted.afr.Repositories-replace-brick >> trusted.gfid >> trusted.glusterfs.dht >> trusted.glusterfs.volume-id >> >> # file: srv/sdb7 >> trusted.afr.Repositories-client-2 >> trusted.afr.Repositories-client-3 >> trusted.gfid >> trusted.glusterfs.dht >> trusted.glusterfs.volume-id >> >> # file: srv/sda8 >> trusted.afr.Repositories-io-threads >> trusted.afr.Repositories-replace-brick >> trusted.gfid >> trusted.glusterfs.volume-id > > > >Is this intentional (i.e., leaving the the attributes intact)? Or functionality that has yet to be implemented? > > >Eric Pretorious >Truckee, CA > > > >>________________________________ >> From: Eric <epretorious at yahoo.com> >>To: "gluster-users at gluster.org" <gluster-users at gluster.org> >>Sent: Wednesday, September 5, 2012 5:05 PM >>Subject: migration operations: Stopping a migration >> >> >>I've created a distributed replicated volume: >> >> >>> gluster> volume info >>>? >>> Volume Name: Repositories >>> Type: Distributed-Replicate >>> Volume ID: 926262ae-2aa6-4bf7-b19e-cf674431b06c >>> Status: Started >>> Number of Bricks: 2 x 2 = 4 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 192.168.1.1:/srv/sda7 >>> Brick2: 192.168.1.2:/srv/sda7 >>> Brick3: 192.168.1.1:/srv/sdb7 >>> Brick4: 192.168.1.2:/srv/sdb7 >> >> >> >>...and begun migrating data from one brick to another as a PoC: >> >> >>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 start >>> replace-brick started successfully >>> >>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 status >>> Number of files migrated = 5147?????? Current file= /centos/5.8/os/x86_64/CentOS/gnome-pilot-conduits-2.0.13-7.el5.x86_64.rpm >>> gluster> volume replace-brick Repositories 192.168.1.1:/srv/sda7 192.168.1.1:/srv/sda8 status >>> Number of files migrated = 24631??????? Migration complete >> >> >>After the migration is finished, though, the list of bricks is wrong: >> >> >> >>> gluster> volume heal Repositories info????????????????????????????????????????????????????????? >>> Heal operation on volume Repositories has been successful >>> >>> Brick 192.168.1.1:/srv/sda7 >>> Number of entries: 0 >>> >>> Brick 192.168.1.2:/srv/sda7 >>> Number of entries: 0 >>> >>> Brick 192.168.1.1:/srv/sdb7 >>> Number of entries: 0 >>> >>> Brick 192.168.1.2:/srv/sdb7 >>> Number of entries: 0 >> >> >>...and the XFS attributes are still intact on the old brick: >> >> >>> [eric at sn1 ~]$ for x in sda7 sdb7 sda8 ; do sudo getfattr -m - /srv/$x 2> /dev/null ; done >>> # file: srv/sda7 >>> trusted.afr.Repositories-client-0 >>> trusted.afr.Repositories-client-1 >>> trusted.afr.Repositories-io-threads >>> trusted.afr.Repositories-replace-brick >>> trusted.gfid >>> trusted.glusterfs.dht >>> trusted.glusterfs.pump-path >>> trusted.glusterfs.volume-id >>> >>> # file: srv/sdb7 >>> trusted.afr.Repositories-client-2 >>> trusted.afr.Repositories-client-3 >>> trusted.gfid >>> trusted.glusterfs.dht >>> trusted.glusterfs.volume-id >>> >>> # file: srv/sda8 >>> trusted.afr.Repositories-io-threads >>> trusted.afr.Repositories-replace-brick >>> trusted.gfid >>> trusted.glusterfs.volume-id >> >> >> >>Have I missed a step? Or: Is this (i.e., clean-up) a bug or functionality that hasn't been implemented yet? >> >> >>Eric Pretorious >>Truckee, CA >> >>_______________________________________________ >>Gluster-users mailing list >>Gluster-users at gluster.org >>http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >> >> >> >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gluster.org/pipermail/gluster-users/attachments/20120905/22ee0b26/attachment.htm>