ping is ok restart of gluster done : [root at node00 ~]# for i in `seq -w 1 10` ; do echo ; echo node$i ; echo ; ssh node$i "service glusterd restart";done node01 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node02 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node03 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node04 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node05 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node06 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node07 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node08 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node09 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] node10 Stopping glusterd:[ OK ] Starting glusterd:[ OK ] but same error message ... [root at node07 ~]# gluster volume add-brick cluster node05:/gluster2 node06:/gluster2 Operation failed on node05 Cheers, Laurent DOUCHY. On 9/1/11 6:54 PM, Mohit Anchlia wrote: > Can you ping node05 from node07 from where you are trying to do the > add? Also, try restarting gluster process on every node and try again. > > On Thu, Sep 1, 2011 at 9:39 AM, Laurent DOUCHY<Laurent.Douchy at unige.ch> wrote: >> see below >> >> Cheers, >> Laurent DOUCHY. >> >> >> On 9/1/11 6:01 PM, Mohit Anchlia wrote: >>> You can check few things on 5 and 6: >>> >>> 1) gluster processes are running on node5 and 6 >> yes: >> >> node05 >> >> root 4902 1 0 Aug31 ? 00:00:00 >> /opt/glusterfs/3.2.2/sbin/glusterd >> root 9626 1 0 19:55 ? 00:00:00 >> /opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol -p >> /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log >> root 9690 9686 0 20:04 ? 00:00:00 bash -c ps -edf | grep >> gluster >> root 9704 9690 0 20:04 ? 00:00:00 grep gluster >> >> node06 >> >> root 4441 1 0 Aug31 ? 00:00:00 >> /opt/glusterfs/3.2.2/sbin/glusterd >> root 9178 1 0 19:55 ? 00:00:00 >> /opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol -p >> /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log >> root 9242 9238 0 20:04 ? 00:00:00 bash -c ps -edf | grep >> gluster >> root 9256 9242 0 20:04 ? 00:00:00 grep gluster >> >>> 2) both nodes are able to see each other >> yes: >> >> [root at node05 ~]# ping node06 >> PING node06.isdc.unige.ch (129.194.168.70) 56(84) bytes of data. >> 64 bytes from node06.isdc.unige.ch (129.194.168.70): icmp_seq=1 ttl=64 >> time=0.376 ms >> >> [root at node06 ~]# ping node05 >> PING node05.isdc.unige.ch (129.194.168.69) 56(84) bytes of data. >> 64 bytes from node05.isdc.unige.ch (129.194.168.69): icmp_seq=1 ttl=64 >> time=0.337 ms >>> 3) do gluster peer status on both the nodes and see what you see >> node 5 trust node 6 and node 6 trust node 5 >> >> [root at node05 ~]# gluster peer status >> Number of Peers: 5 >> >> Hostname: node08 >> Uuid: dd114546-5b94-4a62-9301-260703bf5707 >> State: Peer in Cluster (Connected) >> >> Hostname: node06 >> Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a >> State: Peer in Cluster (Connected) >> >> Hostname: node10 >> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0 >> State: Peer in Cluster (Connected) >> >> Hostname: 129.194.168.71 >> Uuid: a35fb0a1-af35-4a04-b38a-434f68369508 >> State: Peer in Cluster (Connected) >> >> Hostname: node09 >> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04 >> State: Peer in Cluster (Connected) >> >> >> >> >> [root at node06 ~]# gluster peer status >> Number of Peers: 5 >> >> Hostname: node08 >> Uuid: dd114546-5b94-4a62-9301-260703bf5707 >> State: Peer in Cluster (Connected) >> >> Hostname: node09 >> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04 >> State: Peer in Cluster (Connected) >> >> Hostname: node05 >> Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b >> State: Peer in Cluster (Connected) >> >> Hostname: node10 >> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0 >> State: Peer in Cluster (Connected) >> >> Hostname: 129.194.168.71 >> Uuid: a35fb0a1-af35-4a04-b38a-434f68369508 >> State: Peer in Cluster (Connected) >> >> >>> 4) check iptables >> same file on each node (the installation is manage by puppet) >>> On Thu, Sep 1, 2011 at 8:57 AM, Laurent DOUCHY<Laurent.Douchy at unige.ch> >>> wrote: >>>> It works ... >>>> >>>> [root at node07 ~]# gluster volume add-brick cluster node09:/gluster3 >>>> node10:/gluster3 >>>> Add Brick successful >>>> >>>> >>>> On 9/1/11 5:39 PM, Mohit Anchlia wrote: >>>>> Can you try with node09:/gluster3 and node10:gluster3 instead? >>>>> >>>>> On Thu, Sep 1, 2011 at 2:49 AM, Laurent DOUCHY<Laurent.Douchy at unige.ch> >>>>> wrote: >>>>>> Hi, >>>>>> >>>>>> I work on the node7 so it doesn't appear on the list. >>>>>> >>>>>> I create a folder /gluster3 on node5 and node6 and try to add them to >>>>>> my >>>>>> volume but it failed with the same message :( >>>>>> >>>>>> [root at node07 ~]# gluster volume add-brick cluster node05:/gluster3 >>>>>> node06:/gluster3 >>>>>> Operation failed on node05 >>>>>> >>>>>> next step is to reinstall from scratch the node I hope I can avoid >>>>>> this. >>>>>> >>>>>> On 8/31/11 9:08 PM, Mohit Anchlia wrote: >>>>>>> I don't see node07 in above output of gluster peer status. >>>>>>> >>>>>>> Can you try to add bricks on the hosts that gluster1, gluster2? So add >>>>>>> gluster3 and see if that works. >>>>>>> >>>>>>> On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY >>>>>>> <Laurent.Douchy at unige.ch> wrote: >>>>>>>> Hi, >>>>>>>> >>>>>>>> I try to add 2 bricks or 4 bricks for the same effect. >>>>>>>> >>>>>>>> I try to reinstall gluster without success. >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Laurent DOUCHY. >>>>>>>> >>>>>>>> >>>>>>>> On 8/31/11 8:07 PM, Burnash, James wrote: >>>>>>>>> Hi Laurent. >>>>>>>>> >>>>>>>>> Since your configuration specifies replication, you must add bricks >>>>>>>>> in >>>>>>>>> the >>>>>>>>> same number as your number of replicas. >>>>>>>>> >>>>>>>>> For instance - if you have 2 replicas (most normal case), you would >>>>>>>>> need >>>>>>>>> to do something like this: >>>>>>>>> >>>>>>>>> gluster volume add-brick cluster node05:/gluster1 node06:/gluster1 >>>>>>>>> >>>>>>>>> James Burnash >>>>>>>>> Unix Engineer >>>>>>>>> Knight Capital Group >>>>>>>>> >>>>>>>>> >>>>>>>>> -----Original Message----- >>>>>>>>> From: gluster-users-bounces at gluster.org >>>>>>>>> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Laurent >>>>>>>>> DOUCHY >>>>>>>>> Sent: Wednesday, August 31, 2011 12:49 PM >>>>>>>>> To: gluster-users at gluster.org >>>>>>>>> Subject: add bricks on distributed replicated volume >>>>>>>>> failed >>>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for >>>>>>>>> gluster. >>>>>>>>> >>>>>>>>> I manage to configure a distributed and replicated volume on 4 nodes >>>>>>>>> : >>>>>>>>> >>>>>>>>> [root at node07 ~]# gluster volume info cluster >>>>>>>>> >>>>>>>>> Volume Name: cluster >>>>>>>>> Type: Distributed-Replicate >>>>>>>>> Status: Started >>>>>>>>> Number of Bricks: 4 x 2 = 8 >>>>>>>>> Transport-type: tcp >>>>>>>>> Bricks: >>>>>>>>> Brick1: node09:/gluster1 >>>>>>>>> Brick2: node10:/gluster1 >>>>>>>>> Brick3: node09:/gluster2 >>>>>>>>> Brick4: node10:/gluster2 >>>>>>>>> Brick5: node07:/gluster1 >>>>>>>>> Brick6: node08:/gluster1 >>>>>>>>> Brick7: node07:/gluster2 >>>>>>>>> Brick8: node08:/gluster2 >>>>>>>>> >>>>>>>>> But I can't add new nodes to this volume >>>>>>>>> >>>>>>>>> [root at node07 ~]# gluster peer status >>>>>>>>> Number of Peers: 5 >>>>>>>>> >>>>>>>>> Hostname: node10 >>>>>>>>> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0 >>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>> >>>>>>>>> Hostname: node08 >>>>>>>>> Uuid: dd114546-5b94-4a62-9301-260703bf5707 >>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>> >>>>>>>>> Hostname: node09 >>>>>>>>> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04 >>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>> >>>>>>>>> Hostname: node06 >>>>>>>>> Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a >>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>> >>>>>>>>> Hostname: node05 >>>>>>>>> Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b >>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>> [root at node07 ~]# gluster volume add-brick cluster node05:/gluster1 >>>>>>>>> node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed >>>>>>>>> on >>>>>>>>> node05 >>>>>>>>> >>>>>>>>> I try to detach nodes 5 and 6, restart glusterd do the probe and the >>>>>>>>> add-brick but still nothing ... >>>>>>>>> >>>>>>>>> Did some one have any idea to fix this ? >>>>>>>>> >>>>>>>>> Thanks in advance, >>>>>>>>> Laurent. >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Gluster-users mailing list >>>>>>>>> Gluster-users at gluster.org >>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>>>> >>>>>>>>> >>>>>>>>> DISCLAIMER: >>>>>>>>> This e-mail, and any attachments thereto, is intended only for use >>>>>>>>> by >>>>>>>>> the >>>>>>>>> addressee(s) named herein and may contain legally privileged and/or >>>>>>>>> confidential information. If you are not the intended recipient of >>>>>>>>> this >>>>>>>>> e-mail, you are hereby notified that any dissemination, distribution >>>>>>>>> or >>>>>>>>> copying of this e-mail, and any attachments thereto, is strictly >>>>>>>>> prohibited. >>>>>>>>> If you have received this in error, please immediately notify me and >>>>>>>>> permanently delete the original and any copy of any e-mail and any >>>>>>>>> printout >>>>>>>>> thereof. E-mail transmission cannot be guaranteed to be secure or >>>>>>>>> error-free. The sender therefore does not accept liability for any >>>>>>>>> errors or >>>>>>>>> omissions in the contents of this message which arise as a result of >>>>>>>>> e-mail >>>>>>>>> transmission. >>>>>>>>> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group >>>>>>>>> may, >>>>>>>>> at >>>>>>>>> its discretion, monitor and review the content of all e-mail >>>>>>>>> communications. >>>>>>>>> http://www.knight.com >>>>>>>>> _______________________________________________ >>>>>>>>> Gluster-users mailing list >>>>>>>>> Gluster-users at gluster.org >>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>>> _______________________________________________ >>>>>>>> Gluster-users mailing list >>>>>>>> Gluster-users at gluster.org >>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>>> > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users