Other people have seen these too. Seems like some kind of deadlock issue. What commands did you run prior to running rebalance command? Can you restart /etc/init.d/glusterd on all the servers and try again? On Tue, May 24, 2011 at 10:02 AM, Karim Latouche <klatouche at digitalchocolate.com> wrote: > Hi > > Sorry i forgot to include the log message I got when running a > gluster volume rebalance > >> I [glusterd-rebalance.c:483:glusterd_rebalance_cmd_attempted_log] >> 0-glusterd: Received rebalance volume 1 on dccbackup >> [2011-05-24 03:59:23.645067] W >> [socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading >> from socket failed. Error (Transport endpoint is not connected), peer >> (127.0.0.1:1019) >> [2011-05-24 03:59:28.323938] I >> [glusterd-rebalance.c:483:glusterd_rebalance_cmd_attempted_log] 0-glusterd: >> Received rebalance volume 1 on dccbackup >> [2011-05-24 03:59:28.352874] W >> [socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading >> from socket failed. Error (Transport endpoint is not connected), peer >> (127.0.0.1:1018) > > And this is the output of the command > > gluster volume rebalance dccbackup start > starting rebalance on volume dccbackup has been unsuccessful > > Thx > > KL > >> Look in the gluster logs and see what it says. /var/log/gluster* >> >> If you intention is to migrate data then alternatively you can also do: >> >> gluster volume rebalance VOLNAME start >> >> >> On Tue, May 24, 2011 at 2:59 AM, Karim Latouche >> <klatouche at digitalchocolate.com> ?wrote: >>> >>> Hi >>> >>> Thx for your help. >>> >>> I finaly could add the last brick . >>> >>> I now have a Distributed-replicated volume >>> >>>> gluster volume info >>>> >>>> Volume Name: dccbackup >>>> Type: Distributed-Replicate >>>> Status: Started >>>> Number of Bricks: 2 x 2 = 4 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: mysqldccbackup:/export >>>> Brick2: mysqldccbackupmirror:/export >>>> Brick3: gfsbrick1:/brick1 >>>> Brick4: gfsbrick2:/brick2 >>> >>> On the client side though ?I don't see any increasing in storage size. >>> I tried to rebalance >>> >>> gluster volume rebalance ?dccbackup fix-layout start >>> >>> and got . >>> >>>> starting rebalance on volume dccbackup has been unsuccessful >>> >>> I'm kind of stuck here. >>> >>> Thank you >>> >>> KL >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>>> What is the output of: >>>> >>>> # ls /etc/glusterd/peers >>>> 39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> ?4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> ?a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> >>>> Above output is mine. >>>> >>>> If you cat each of these files, there should be one for each peer OTHER >>>> than the host that your are logged into), and hostnames each ?files, >>>> like >>>> this: >>>> # cat /etc/glusterd/peers/* >>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> state=3 >>>> hostname1=jc1letgfs15-pfs1 >>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> state=3 >>>> hostname1=jc1letgfs18-pfs1 >>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> state=3 >>>> hostname1=jc1letgfs14-pfs1 >>>> >>>> Here is a listing of all the peer files on all four nodes of my >>>> GlusterFS >>>> servers - I created this configuration in the same way you did - two >>>> servers, then add two servers: >>>> >>>> jc1letgfs14 >>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> state=3 >>>> hostname1=jc1letgfs15-pfs1 >>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> state=3 >>>> hostname1=jc1letgfs18-pfs1 >>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> state=3 >>>> hostname1=10.20.72.191 >>>> >>>> jc1letgfs15 >>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> state=3 >>>> hostname1=jc1letgfs18-pfs1 >>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> state=3 >>>> hostname1=10.20.72.191 >>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> state=3 >>>> hostname1=jc1letgfs14-pfs1 >>>> >>>> jc1letgfs17 >>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> state=3 >>>> hostname1=jc1letgfs15-pfs1 >>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>> state=3 >>>> hostname1=jc1letgfs18-pfs1 >>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> state=3 >>>> hostname1=jc1letgfs14-pfs1 >>>> >>>> jc1letgfs18 >>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>> state=3 >>>> hostname1=jc1letgfs15-pfs1 >>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>> state=3 >>>> hostname1=10.20.72.191 >>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>> state=3 >>>> hostname1=jc1letgfs14-pfs1 >>>> >>>> I hope this is of some help. >>>> >>>> James Burnash >>>> Unix Engineer >>>> >>>> >>>> -----Original Message----- >>>> From: gluster-users-bounces at gluster.org >>>> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Karim Latouche >>>> Sent: Monday, May 23, 2011 1:02 PM >>>> To: Mohit Anchlia >>>> Cc: gluster-users at gluster.org >>>> Subject: Re: Trouble to add brick >>>> >>>> hi >>>> >>>> Here is the output of gluster peer status ?command of the four nodes >>>> >>>> mysqldccbackup mysqldccbackupmirror are the two first nodes >>>> >>>> I got issues when I tried to add gfsbrick1 and two to the already >>>> working >>>> cluster >>>> >>>> >>>> >>>>> [root at gfsbrick1 ~]# gluster peer status >>>>> Number of Peers: 1 >>>>> >>>>> Hostname: 10.0.4.225 >>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>> State: Establishing Connection (Connected) >>>>> [root at gfsbrick2 ~]# gluster peer status >>>>> Number of Peers: 2 >>>>> >>>>> Hostname: 10.0.4.225 >>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> Hostname: mysqldccbackupmirror >>>>> Uuid: 0961e0d2-e2e8-497d-9c51-4b96c061f939 >>>>> State: Peer Rejected (Connected) >>>>> [root at mysqldccbackup ~]# gluster peer status >>>>> Number of Peers: 3 >>>>> >>>>> Hostname: mysqldccbackupmirror >>>>> Uuid: 0961e0d2-e2e8-497d-9c51-4b96c061f939 >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> Hostname: gfsbrick2 >>>>> Uuid: 14f07412-b2e5-4b34-8980-d180a64b720f >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> Hostname: gfsbrick1 >>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>> State: Establishing Connection (Connected) >>>> >>>>> root at mysqldccbackupmirror ~]# gluster peer status >>>>> Number of Peers: 1 >>>>> >>>>> Hostname: 10.0.4.225 >>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>> State: Peer in Cluster (Connected) >>>>> [root at mysqldccbackupmirror ~]# gluster peer status >>>>> Number of Peers: 1 >>>>> >>>>> Hostname: 10.0.4.225 >>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>> State: Peer in Cluster (Connected) >>>> >>>> Thx a lot >>>> >>>> KL >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>>> Please send the complete output of gluster peer status from both the >>>>> nodes. >>>>> >>>>> On Mon, May 23, 2011 at 9:44 AM, Karim Latouche >>>>> <klatouche at digitalchocolate.com> ? ? wrote: >>>>>> >>>>>> hi >>>>>> >>>>>> Thx for your answer. >>>>>> >>>>>> I should have mention that I already set up a proper hosts file to >>>>>> avoid >>>>>> DNS >>>>>> resolution issue but I still have same issue. >>>>>> >>>>>> The last brick is allways the one having issues to be added . I tried >>>>>> adding >>>>>> gfsbrick2 first then gfsbrick1 and the result is the same >>>>>> >>>>>> I get >>>>>> >>>>>>> Hostname: gfsbrick2 >>>>>>> Uuid: 14f07412-b2e5-4b34-8980-d180a64b720f >>>>>>> State: Peer in Cluster (Connected) >>>>>>> >>>>>>> Hostname: gfsbrick1 >>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>> State: Establishing Connection (Connected) >>>>>> >>>>>> instead of >>>>>> >>>>>> >>>>>>> Hostname: gfsbrick2 >>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>> State: Establishing Connection (Connected) >>>>>> >>>>>> It 's really strange. >>>>>> >>>>>> >>>>>> Thx >>>>>> >>>>>> KL >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> Karim, >>>>>>> ? ? ?gfsbrick1 is not able to contact gfsbrick2 yet. Generally >>>>>>> happens >>>>>>> if >>>>>>> the DNS resolutions dont happen as expected. You can map them >>>>>>> manually >>>>>>> in >>>>>>> /etc/hosts for both the machines?. >>>>>>> >>>>>>> something like: >>>>>>> on gfsbrick1 >>>>>>> <ip-of-brick2> ? ? ? gfsbrick2 >>>>>>> on gfsbrick2 >>>>>>> <ip-of-brick1> ? ? ? gfsbrick1 >>>>>>> >>>>>>> Pranith. >>>>>>> >>>>>>> ----- Original Message ----- >>>>>>> From: "Karim Latouche"<klatouche at digitalchocolate.com> >>>>>>> To: gluster-users at gluster.org >>>>>>> Sent: Monday, May 23, 2011 7:56:22 PM >>>>>>> Subject: Trouble to add brick >>>>>>> >>>>>>> Hi >>>>>>> >>>>>>> I got a strange problems >>>>>>> >>>>>>> I run a two bricks replicated volume and I tried to add 2 more bricks >>>>>>> >>>>>>> The first one went ok >>>>>>> >>>>>>>> gluster peer probe gfsbrick1 >>>>>>>> Probe successful >>>>>>> >>>>>>> It seems I can't add the second one >>>>>>> >>>>>>> I get a >>>>>>> >>>>>>>> Hostname: gfsbrick2 >>>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>>> State: Establishing Connection (Connected) >>>>>>> >>>>>>> Did anyone encounter this issue ? >>>>>>> >>>>>>> Thx >>>>>>> >>>>>>> KL >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-users mailing list >>>>>> Gluster-users at gluster.org >>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>> >>>> >>>> DISCLAIMER: >>>> This e-mail, and any attachments thereto, is intended only for use by >>>> the >>>> addressee(s) named herein and may contain legally privileged and/or >>>> confidential information. If you are not the intended recipient of this >>>> e-mail, you are hereby notified that any dissemination, distribution or >>>> copying of this e-mail, and any attachments thereto, is strictly >>>> prohibited. >>>> If you have received this in error, please immediately notify me and >>>> permanently delete the original and any copy of any e-mail and any >>>> printout >>>> thereof. E-mail transmission cannot be guaranteed to be secure or >>>> error-free. The sender therefore does not accept liability for any >>>> errors or >>>> omissions in the contents of this message which arise as a result of >>>> e-mail >>>> transmission. >>>> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, >>>> at >>>> its discretion, monitor and review the content of all e-mail >>>> communications. >>>> http://www.knight.com >>> > >