Re: Issues removing then adding a brick to a replica volume (Gluster 3.7.6)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 02/23/2016 04:34 PM, Lindsay Mathieson wrote:
On 23/02/2016 8:29 PM, Sahina Bose wrote:
Late jumping into this thread, but curious -

Is there a specific reason that you are removing and adding a brick? Will replace-brick not work for you?


Testing procedures for replacing a failed brick (disk crash etc),


The recommended way for replacing brick in a replica volume is  - replace-brick <src brick path> <destination brick path> commit force
We found that the issues related to heal that you encountered with decreasing and increasing replica count do not exist here.

In case the entire host needs to be replaced (for instance re-installing host/reformatting disks- and assuming the brick directories are same as before), here is a flow that works. Can you check if this will solve your usecase?

(Follow steps 1-4 in case host3 has been re-installed, and /var/lib/glusterd re-initialized)
  1. Stop glusterd on host being replaced (say, host3)
  2. Check gluster peer status from working node to obtain previous UUID of host3
  3. Edit gluster UUID in /var/lib/glusterd/glusterd.info on host3 to previous UUID obtained in step 2.
  4. Copy peer info from a working peer to /var/lib/glusterd/peers (without the peer UUID of node being replaced, here host3)
  5. Create and remove a tmp dir at volume mount points
  6. Restart glusterd -- heal will start and the brick on replaced node should be synced automatically.



_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux