On 5/24/11 8:09 PM, Mohit Anchlia wrote: > Other people have seen these too. Seems like some kind of deadlock > issue. What commands did you run prior to running rebalance command? I tried gluster volume rebalance dccbackup fix-layout start but was unsuccessfull > Can you restart /etc/init.d/glusterd on all the servers and try again? I did and still get the same error I [glusterd-rebalance.c:483:glusterd_rebalance_cmd_attempted_log] 0-glusterd: Received rebalance volume 1 on dccbackup [2011-05-24 04:42:18.521586] W [socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading from socket failed. Error (Transport endpoint is not connected), peer (127.0.0.1:1022) KL > On Tue, May 24, 2011 at 10:02 AM, Karim Latouche > <klatouche at digitalchocolate.com> wrote: >> Hi >> >> Sorry i forgot to include the log message I got when running a >> gluster volume rebalance >> >>> I [glusterd-rebalance.c:483:glusterd_rebalance_cmd_attempted_log] >>> 0-glusterd: Received rebalance volume 1 on dccbackup >>> [2011-05-24 03:59:23.645067] W >>> [socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading >>> from socket failed. Error (Transport endpoint is not connected), peer >>> (127.0.0.1:1019) >>> [2011-05-24 03:59:28.323938] I >>> [glusterd-rebalance.c:483:glusterd_rebalance_cmd_attempted_log] 0-glusterd: >>> Received rebalance volume 1 on dccbackup >>> [2011-05-24 03:59:28.352874] W >>> [socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading >>> from socket failed. Error (Transport endpoint is not connected), peer >>> (127.0.0.1:1018) >> And this is the output of the command >> >> gluster volume rebalance dccbackup start >> starting rebalance on volume dccbackup has been unsuccessful >> >> Thx >> >> KL >> >>> Look in the gluster logs and see what it says. /var/log/gluster* >>> >>> If you intention is to migrate data then alternatively you can also do: >>> >>> gluster volume rebalance VOLNAME start >>> >>> >>> On Tue, May 24, 2011 at 2:59 AM, Karim Latouche >>> <klatouche at digitalchocolate.com> wrote: >>>> Hi >>>> >>>> Thx for your help. >>>> >>>> I finaly could add the last brick . >>>> >>>> I now have a Distributed-replicated volume >>>> >>>>> gluster volume info >>>>> >>>>> Volume Name: dccbackup >>>>> Type: Distributed-Replicate >>>>> Status: Started >>>>> Number of Bricks: 2 x 2 = 4 >>>>> Transport-type: tcp >>>>> Bricks: >>>>> Brick1: mysqldccbackup:/export >>>>> Brick2: mysqldccbackupmirror:/export >>>>> Brick3: gfsbrick1:/brick1 >>>>> Brick4: gfsbrick2:/brick2 >>>> On the client side though I don't see any increasing in storage size. >>>> I tried to rebalance >>>> >>>> gluster volume rebalance dccbackup fix-layout start >>>> >>>> and got . >>>> >>>>> starting rebalance on volume dccbackup has been unsuccessful >>>> I'm kind of stuck here. >>>> >>>> Thank you >>>> >>>> KL >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>>> What is the output of: >>>>> >>>>> # ls /etc/glusterd/peers >>>>> 39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> 4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> >>>>> Above output is mine. >>>>> >>>>> If you cat each of these files, there should be one for each peer OTHER >>>>> than the host that your are logged into), and hostnames each files, >>>>> like >>>>> this: >>>>> # cat /etc/glusterd/peers/* >>>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> state=3 >>>>> hostname1=jc1letgfs15-pfs1 >>>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> state=3 >>>>> hostname1=jc1letgfs18-pfs1 >>>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> state=3 >>>>> hostname1=jc1letgfs14-pfs1 >>>>> >>>>> Here is a listing of all the peer files on all four nodes of my >>>>> GlusterFS >>>>> servers - I created this configuration in the same way you did - two >>>>> servers, then add two servers: >>>>> >>>>> jc1letgfs14 >>>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> state=3 >>>>> hostname1=jc1letgfs15-pfs1 >>>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> state=3 >>>>> hostname1=jc1letgfs18-pfs1 >>>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> state=3 >>>>> hostname1=10.20.72.191 >>>>> >>>>> jc1letgfs15 >>>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> state=3 >>>>> hostname1=jc1letgfs18-pfs1 >>>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> state=3 >>>>> hostname1=10.20.72.191 >>>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> state=3 >>>>> hostname1=jc1letgfs14-pfs1 >>>>> >>>>> jc1letgfs17 >>>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> state=3 >>>>> hostname1=jc1letgfs15-pfs1 >>>>> /etc/glusterd/peers/4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> uuid=4583d8b7-f724-4017-b549-9aa9e27dd50f >>>>> state=3 >>>>> hostname1=jc1letgfs18-pfs1 >>>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> state=3 >>>>> hostname1=jc1letgfs14-pfs1 >>>>> >>>>> jc1letgfs18 >>>>> /etc/glusterd/peers/39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> uuid=39bd607b-b355-4e0d-80ae-ef507e0dbabe >>>>> state=3 >>>>> hostname1=jc1letgfs15-pfs1 >>>>> /etc/glusterd/peers/94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> uuid=94f28f8c-8958-435b-80e4-3bd6d4fc5156 >>>>> state=3 >>>>> hostname1=10.20.72.191 >>>>> /etc/glusterd/peers/a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> uuid=a70858c6-cac3-43fa-b743-4d2226f8a433 >>>>> state=3 >>>>> hostname1=jc1letgfs14-pfs1 >>>>> >>>>> I hope this is of some help. >>>>> >>>>> James Burnash >>>>> Unix Engineer >>>>> >>>>> >>>>> -----Original Message----- >>>>> From: gluster-users-bounces at gluster.org >>>>> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Karim Latouche >>>>> Sent: Monday, May 23, 2011 1:02 PM >>>>> To: Mohit Anchlia >>>>> Cc: gluster-users at gluster.org >>>>> Subject: Re: Trouble to add brick >>>>> >>>>> hi >>>>> >>>>> Here is the output of gluster peer status command of the four nodes >>>>> >>>>> mysqldccbackup mysqldccbackupmirror are the two first nodes >>>>> >>>>> I got issues when I tried to add gfsbrick1 and two to the already >>>>> working >>>>> cluster >>>>> >>>>> >>>>> >>>>>> [root at gfsbrick1 ~]# gluster peer status >>>>>> Number of Peers: 1 >>>>>> >>>>>> Hostname: 10.0.4.225 >>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>> State: Establishing Connection (Connected) >>>>>> [root at gfsbrick2 ~]# gluster peer status >>>>>> Number of Peers: 2 >>>>>> >>>>>> Hostname: 10.0.4.225 >>>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> Hostname: mysqldccbackupmirror >>>>>> Uuid: 0961e0d2-e2e8-497d-9c51-4b96c061f939 >>>>>> State: Peer Rejected (Connected) >>>>>> [root at mysqldccbackup ~]# gluster peer status >>>>>> Number of Peers: 3 >>>>>> >>>>>> Hostname: mysqldccbackupmirror >>>>>> Uuid: 0961e0d2-e2e8-497d-9c51-4b96c061f939 >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> Hostname: gfsbrick2 >>>>>> Uuid: 14f07412-b2e5-4b34-8980-d180a64b720f >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> Hostname: gfsbrick1 >>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>> State: Establishing Connection (Connected) >>>>>> root at mysqldccbackupmirror ~]# gluster peer status >>>>>> Number of Peers: 1 >>>>>> >>>>>> Hostname: 10.0.4.225 >>>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>>> State: Peer in Cluster (Connected) >>>>>> [root at mysqldccbackupmirror ~]# gluster peer status >>>>>> Number of Peers: 1 >>>>>> >>>>>> Hostname: 10.0.4.225 >>>>>> Uuid: 2f4a1869-8c7e-49ea-a7b1-d9b2d364e5d6 >>>>>> State: Peer in Cluster (Connected) >>>>> Thx a lot >>>>> >>>>> KL >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>> Please send the complete output of gluster peer status from both the >>>>>> nodes. >>>>>> >>>>>> On Mon, May 23, 2011 at 9:44 AM, Karim Latouche >>>>>> <klatouche at digitalchocolate.com> wrote: >>>>>>> hi >>>>>>> >>>>>>> Thx for your answer. >>>>>>> >>>>>>> I should have mention that I already set up a proper hosts file to >>>>>>> avoid >>>>>>> DNS >>>>>>> resolution issue but I still have same issue. >>>>>>> >>>>>>> The last brick is allways the one having issues to be added . I tried >>>>>>> adding >>>>>>> gfsbrick2 first then gfsbrick1 and the result is the same >>>>>>> >>>>>>> I get >>>>>>> >>>>>>>> Hostname: gfsbrick2 >>>>>>>> Uuid: 14f07412-b2e5-4b34-8980-d180a64b720f >>>>>>>> State: Peer in Cluster (Connected) >>>>>>>> >>>>>>>> Hostname: gfsbrick1 >>>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>>> State: Establishing Connection (Connected) >>>>>>> instead of >>>>>>> >>>>>>> >>>>>>>> Hostname: gfsbrick2 >>>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>>> State: Establishing Connection (Connected) >>>>>>> It 's really strange. >>>>>>> >>>>>>> >>>>>>> Thx >>>>>>> >>>>>>> KL >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>> Karim, >>>>>>>> gfsbrick1 is not able to contact gfsbrick2 yet. Generally >>>>>>>> happens >>>>>>>> if >>>>>>>> the DNS resolutions dont happen as expected. You can map them >>>>>>>> manually >>>>>>>> in >>>>>>>> /etc/hosts for both the machines?. >>>>>>>> >>>>>>>> something like: >>>>>>>> on gfsbrick1 >>>>>>>> <ip-of-brick2> gfsbrick2 >>>>>>>> on gfsbrick2 >>>>>>>> <ip-of-brick1> gfsbrick1 >>>>>>>> >>>>>>>> Pranith. >>>>>>>> >>>>>>>> ----- Original Message ----- >>>>>>>> From: "Karim Latouche"<klatouche at digitalchocolate.com> >>>>>>>> To: gluster-users at gluster.org >>>>>>>> Sent: Monday, May 23, 2011 7:56:22 PM >>>>>>>> Subject: Trouble to add brick >>>>>>>> >>>>>>>> Hi >>>>>>>> >>>>>>>> I got a strange problems >>>>>>>> >>>>>>>> I run a two bricks replicated volume and I tried to add 2 more bricks >>>>>>>> >>>>>>>> The first one went ok >>>>>>>> >>>>>>>>> gluster peer probe gfsbrick1 >>>>>>>>> Probe successful >>>>>>>> It seems I can't add the second one >>>>>>>> >>>>>>>> I get a >>>>>>>> >>>>>>>>> Hostname: gfsbrick2 >>>>>>>>> Uuid: 00000000-0000-0000-0000-000000000000 >>>>>>>>> State: Establishing Connection (Connected) >>>>>>>> Did anyone encounter this issue ? >>>>>>>> >>>>>>>> Thx >>>>>>>> >>>>>>>> KL >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Gluster-users mailing list >>>>>>>> Gluster-users at gluster.org >>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>> >>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users at gluster.org >>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>> >>>>> >>>>> DISCLAIMER: >>>>> This e-mail, and any attachments thereto, is intended only for use by >>>>> the >>>>> addressee(s) named herein and may contain legally privileged and/or >>>>> confidential information. If you are not the intended recipient of this >>>>> e-mail, you are hereby notified that any dissemination, distribution or >>>>> copying of this e-mail, and any attachments thereto, is strictly >>>>> prohibited. >>>>> If you have received this in error, please immediately notify me and >>>>> permanently delete the original and any copy of any e-mail and any >>>>> printout >>>>> thereof. E-mail transmission cannot be guaranteed to be secure or >>>>> error-free. The sender therefore does not accept liability for any >>>>> errors or >>>>> omissions in the contents of this message which arise as a result of >>>>> e-mail >>>>> transmission. >>>>> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, >>>>> at >>>>> its discretion, monitor and review the content of all e-mail >>>>> communications. >>>>> http://www.knight.com >>