On Wed, Apr 11, 2018 at 7:38 PM, Jose Sanchez <josesanc@xxxxxxxxxxxx> wrote:
Hi KarthikLooking at the information you have provided me, I would like to make sure that I’m running the right commands.1. gluster volume heal scratch info
If the count is non zero, trigger the heal and wait for heal info count to become zero.
2. gluster volume remove-brick scratch replica 1 gluster02ib:/gdata/brick1/scratch gluster02ib:/gdata/ forcebrick2/ scratch
3. gluster volume add-brick “#" scratch gluster02ib:/gdata/brick1/ scratch gluster02ib:/gdata/brick2/ scratchBased on the configuration I have, Brick 1 from Node A and B are tide together and Brick 2 from Node A and B are also tide together. Looking at your remove command (step #2), it seems that you want me to remove Brick 1 and 2 from Node B (gluster02ib). is that correct? I thought the data was distributed in bricks 1 between nodes A and B) and duplicated on Bricks 2 (node A and B).
Data is duplicated between bricks 1 of nodes A & B and bricks 2 of nodes A & B and data is distributed between these two pairs.
You need not always remove the bricks 1 & 2 from node B itself. The idea here is to keep one copy from both the replica pairs.
Also when I add the bricks back to gluster, do I need to specify if it is distributed or replicated?? and Do i need a configuration #?? for example on your command (Step #2) you have “replica 1” when remove bricks, do I need to do the same when adding the nodes back ?
No. You just need to erase the data on those bricks and add those bricks back to the volume. The previous remove-brick command will make the volume plain distribute. Then simply adding the bricks without specifying any "#" will expand the volume as a plain distribute volue.
Im planning on moving with this changes in few days. At this point each brick has 14tb and adding bricks 1 from node A and B, i have a total of 28tb, After doing all the process, (removing and adding bricks) I should be able to see a total of 56Tb right ?
Yes after all these you will have 56TB in total.
After adding the bricks, do volume rebalance, so that the data which were present previously, will be moved to the correct bricks.
HTH,
Karthik
ThanksJose--------------------------------- Jose SanchezSystems/Network Analyst 1Center of Advanced Research ComputingMSC 01 1190Albuquerque, NM 87131-0001575.636.4232On Apr 7, 2018, at 8:29 AM, Karthik Subrahmanya <ksubrahm@xxxxxxxxxx> wrote:Hi Jose,Thanks for providing the volume info. You have 2 subvolumes. Data is replicated within the bricks of that subvolumes.First one consisting of Node A's brick1 & Node B's brick1 and the second one consisting of Node A's brick2 and Node B's brick2.You don't have the same data on all the 4 bricks. Data are distributed between these two subvolumes.To remove the replica you can use the commandgluster volume remove-brick scratch replica 1 gluster02ib:/gdata/brick1/scratch gluster02ib:/gdata/brick2/ forcescratch So you will have one copy of data present from both the distributes.Before doing this make sure "gluster volume heal scratch info" value is zero. So copies you retain will have the correct data.After the remove-brick erase the data from the backend.Then you can expand the volume by following the steps at [1].Regards,KarthikOn Fri, Apr 6, 2018 at 11:39 PM, Jose Sanchez <josesanc@xxxxxxxxxxxx> wrote:Hi Karthikthis is our configuration, is 2x2 =4 , they are all replicated , each brick has 14tb. we have 2 nodes A and B, each one with brick 1 and 2.Node A (replicated A1 (14tb) and B1 (14tb) ) same with node B (Replicated A2 (14tb) and B2 (14tb)).Do you think we need to degrade the node first before removing it. i believe the same copy of data is on all 4 bricks, we would like to keep one of them, and add the other bricks as extra spaceThanks for your help on thisJose[root@gluster01 ~]# gluster volume info scratchVolume Name: scratchType: Distributed-ReplicateVolume ID: 23f1e4b1-b8e0-46c3-874a-58b4728ea106 Status: StartedSnapshot Count: 0Number of Bricks: 2 x 2 = 4Transport-type: tcp,rdmaBricks:Brick1: gluster01ib:/gdata/brick1/scratch Brick2: gluster02ib:/gdata/brick1/scratch Brick3: gluster01ib:/gdata/brick2/scratch Brick4: gluster02ib:/gdata/brick2/scratch Options Reconfigured:performance.readdir-ahead: onnfs.disable: on[root@gluster01 ~]# gluster volume status allStatus of volume: scratchGluster process TCP Port RDMA Port Online Pid------------------------------------------------------------ ------------------ Brick gluster01ib:/gdata/brick1/scratch 49152 49153 Y 1743 Brick gluster02ib:/gdata/brick1/scratch 49156 49157 Y 1732 Brick gluster01ib:/gdata/brick2/scratch 49154 49155 Y 1738 Brick gluster02ib:/gdata/brick2/scratch 49158 49159 Y 1733 Self-heal Daemon on localhost N/A N/A Y 1728Self-heal Daemon on gluster02ib N/A N/A Y 1726Task Status of Volume scratch------------------------------------------------------------ ------------------ There are no active volume tasks--------------------------------- Jose SanchezSystems/Network Analyst 1Center of Advanced Research ComputingMSC 01 1190Albuquerque, NM 87131-0001575.636.4232On Apr 6, 2018, at 3:49 AM, Karthik Subrahmanya <ksubrahm@xxxxxxxxxx> wrote:Hi Jose,By switching into pure distribute volume you will lose availability if something goes bad.I am guessing you have a nX2 volume.If you want to preserve one copy of the data in all the distributes, you can do that by decreasing the replica count in the remove-brick operation.If you have any inconsistency, heal them first using the "gluster volume heal <volname>" command and wait till the"gluster volume heal <volname> info" output becomes zero, before removing the bricks, so that you will have the correct data.If you do not want to preserve the data then you can directly remove the bricks.Even after removing the bricks the data will be present in the backend of the removed bricks. You have to manually erase them (both data and .glusterfs folder).See [1] for more details on remove-brick.HTH,KarthikOn Thu, Apr 5, 2018 at 8:17 PM, Jose Sanchez <josesanc@xxxxxxxxxxxx> wrote:We have a Gluster setup with 2 nodes (distributed replication) and we would like to switch it to the distributed mode. I know the data is duplicated between those nodes, what is the proper way of switching it to a distributed, we would like to double or gain the storage space on our gluster storage node. what happens with the data, do i need to erase one of the nodes?Jose--------------------------------- Jose SanchezSystems/Network AnalystCenter of Advanced Research ComputingMSC 01 1190Albuquerque, NM 87131-0001575.636.4232
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users