Re: Fwd: Gluster Volume Replication using 2 AWS instances on Autoscaling

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 03/13/2014 10:39 PM, bernhard glomm wrote:

------------------------------------------------------------------------
*Ecologic Institute* 	*Bernhard Glomm*
IT Administration

Phone: 	+49 (30) 86880 134
Fax: 	+49 (30) 86880 100
Skype: 	bernhard.glomm.ecologic

Website: <http://ecologic.eu> | Video:
<http://www.youtube.com/v/hZtiK04A9Yo> | Newsletter:
<http://ecologic.eu/newsletter/subscribe> | Facebook:
<http://www.facebook.com/Ecologic.Institute> | Linkedin:
<http://www.linkedin.com/company/ecologic-institute-berlin-germany> |
Twitter: <http://twitter.com/EcologicBerlin> | YouTube:
<http://www.youtube.com/user/EcologicInstitute> | Google+:
<http://plus.google.com/113756356645020994482>
Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717
Berlin | Germany
GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.:
DE811963464
Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH
------------------------------------------------------------------------


Begin forwarded message:

*From: *bernhard glomm <bernhard.glomm@xxxxxxxxxxx
<mailto:bernhard.glomm@xxxxxxxxxxx>>
*Subject: **Re:  Gluster Volume Replication using 2 AWS
instances on Autoscaling*
*Date: *March 13, 2014 6:08:10 PM GMT+01:00
*To: *Vijay Bellur <vbellur@xxxxxxxxxx <mailto:vbellur@xxxxxxxxxx>>

??? I thought  replace-brick was not recommended at the moment
in 3.4.2 on a replica 2 volume I use successfully:

replace-brick is not recommended for data migration. commit force just performs a volume topology update and does not perform any data migration.

-Vijay


volume remove-brick <vol-name>  replica 1 <brick-name>  force
# replace the old brick with the new one, mount another disk or what
ever, than
volume add-brick <vol-name> replica 2 <brick-name> force
volume heal <vol-name>

hth

Bernhard


On Mar 13, 2014, at 5:48 PM, Vijay Bellur <vbellur@xxxxxxxxxx
<mailto:vbellur@xxxxxxxxxx>> wrote:

On 03/13/2014 09:18 AM, Alejandro Planas wrote:
Hello,

We have 2 AWS instances, 1 brick on each instance, one replicated volume
among both instances. When one of the instances fails completely and
autoscaling replaces it with a new one, we are having issues recreating
the replicated volume again.

Can anyone provide some light on the gluster commands required to
include this new replacement instance (with one brick) as a member of
the replicated volume?


You can probably use:

volume replace-brick <volname> <old-brick> <new-brick> commit force

This will remove the old-brick from the volume and bring in new-brick to
the volume. self-healing can then synchronize data to the new brick.

Regards,
Vijay

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx>
http://supercolony.gluster.org/mailman/listinfo/gluster-users




_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users





[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux