kp, please find attached tar files. [root at vhead-010 ~]# ssh foo-1-private [root at foo-1-private ~]# gluster volume create foo foo-1-private:/mnt/brick Creation of volume foo has been successful. Please start the volume to access data. [root at foo-1-private ~]# gluster volume start foo Starting volume foo has been successful [root at foo-1-private ~]# gluster peer probe foo-3-private Probe successful [root at foo-1-private ~]# gluster peer status Number of Peers: 1 Hostname: foo-3-private Uuid: 0c76c6ec-a7ea-405e-9d1d-8e100e155ec3 State: Peer Rejected (Connected) [root at foo-1-private ~]# cd /etc/glusterd/vols/ [root at foo-1-private vols]# tar cf /tmp/foo-1-private.tar foo [root at foo-1-private vols]# ssh foo-3-private [root at foo-3-private ~]# cd /etc/glusterd/vols [root at foo-3-private vols]# tar cf /tmp/foo-3-private.tar foo [root at foo-3-private vols]# Thanks, Tomo (2011/08/17 15:46), krish wrote: > Tomoaki, > > Can you send the info file under /etc/glusterd/vols/<volname>/ from machines were the peer probe > command was issued and the 'peer' who is being rejected? > > thanks, > kp > > On 08/17/2011 12:10 PM, Tomoaki Sato wrote: >> Mohit >> >> (2011/08/17 13:53), Mohit Anchlia wrote: >>> Not sure. It could be because the new node doesn't have the volume >>> configs. Can you try gluster volume sync to sync the configs and see >>> if that helps? >> >> - at foo-1-private - >> gluster> volume sync foo-3-private >> please delete all the volumes before full sync >> gluster> >> -- >> >> - at foo-3-private - >> gluster> volume sync foo foo-1-private >> sync from localhost not allowed >> gluster> >> -- >> >> Question is "How should I add extra peers to existing file systems ?". >> extra peers = new nodes. >> >> Could you tell me right instructions to gluster probe new nodes after the volume starting ? >> >>> >>> Also, not sure why you are getting "Unable to find hostname: foo-3-private" >>> >> >> "Unable to find hostname: foo-3-private" was printed out on both OK and NG cases. >> "Cksums of volume foo differ. local cksum = 1403573944, remote cksum = -1413994823" was printed out on NG case only. >> >> 'Peer in Cluster', OK case: >> [2011-08-17 15:08:47.462646] I [glusterd-handler.c:602:glusterd_handle_cli_probe] 0-glusterd: Received CLI probe req foo-3-private 24007 >> [2011-08-17 15:08:47.466194] I [glusterd-handler.c:391:glusterd_friend_find] 0-glusterd: Unable to find hostname: foo-3-private >> [2011-08-17 15:08:47.466224] I [glusterd-handler.c:3120:glusterd_probe_begin] 0-glusterd: Unable to find peerinfo for host: foo-3-private (24007) >> [2011-08-17 15:08:47.469365] W [rpc-transport.c:728:rpc_transport_load] 0-rpc-transport: missing 'option transport-type'. defaulting to "socket" >> [2011-08-17 15:08:47.473671] I [glusterd-handler.c:3102:glusterd_friend_add] 0-glusterd: connect returned 0 >> [2011-08-17 15:08:47.474167] I [glusterd-handshake.c:317:glusterd_set_clnt_mgmt_program] 0-: Using Program glusterd clnt mgmt, Num (1238433), Version (1) >> [2011-08-17 15:08:47.474214] I [glusterd-utils.c:2127:glusterd_friend_find_by_hostname] 0-glusterd: Friend foo-3-private found.. state: 0 >> [2011-08-17 15:08:47.483485] I [glusterd-rpc-ops.c:364:glusterd3_1_probe_cbk] 0-glusterd: Received probe resp from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f, host: foo-3-private >> [2011-08-17 15:08:47.483516] I [glusterd-handler.c:379:glusterd_friend_find] 0-glusterd: Unable to find peer by uuid >> [2011-08-17 15:08:47.483562] I [glusterd-utils.c:2127:glusterd_friend_find_by_hostname] 0-glusterd: Friend foo-3-private found.. state: 0 >> [2011-08-17 15:08:47.483764] I [glusterd-rpc-ops.c:409:glusterd3_1_probe_cbk] 0-glusterd: Received resp to probe req >> [2011-08-17 15:08:47.484040] I [glusterd-rpc-ops.c:454:glusterd3_1_friend_add_cbk] 0-glusterd: Received ACC from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f, host: foo-3-private, port: 0 >> [2011-08-17 15:08:47.484088] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Probe Sent to Peer >> [2011-08-17 15:08:47.484153] I [glusterd-handler.c:3293:glusterd_xfer_cli_probe_resp] 0-glusterd: Responded to CLI, ret: 0 >> [2011-08-17 15:08:47.484554] I [glusterd-handler.c:2882:glusterd_handle_probe_query] 0-glusterd: Received probe from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f >> [2011-08-17 15:08:47.484585] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Accepted peer request >> [2011-08-17 15:08:47.484647] I [glusterd-handler.c:2917:glusterd_handle_probe_query] 0-glusterd: Responded to 192.168.1.129, op_ret: 0, op_errno: 0, ret: 0 >> [2011-08-17 15:08:47.485499] I [glusterd-handler.c:2614:glusterd_handle_incoming_friend_req] 0-glusterd: Received probe from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f >> [2011-08-17 15:08:47.485536] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Accepted peer request >> [2011-08-17 15:08:47.485590] I [glusterd-handler.c:3270:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to 192.168.1.129 (0), ret: 0 >> [2011-08-17 15:08:47.485713] I [glusterd-sm.c:492:glusterd_ac_send_friend_update] 0-: Added uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f, host: foo-3-private >> [2011-08-17 15:08:47.486203] I [glusterd-rpc-ops.c:636:glusterd3_1_friend_update_cbk] 0-glusterd: Received ACC from uuid: >> [2011-08-17 15:08:47.486259] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Peer in Cluster >> [2011-08-17 15:08:47.486284] I [glusterd-handler.c:2761:glusterd_handle_friend_update] 0-glusterd: Received friend update from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f >> [2011-08-17 15:08:47.486316] I [glusterd-handler.c:2806:glusterd_handle_friend_update] 0-: Received uuid: 4b5b0ecb-7d18-4ec4-90d9-0df2d392b63f, hostname:192.168.1.129 >> [2011-08-17 15:08:47.486335] I [glusterd-handler.c:2809:glusterd_handle_friend_update] 0-: Received my uuid as Friend >> >> >> 'Peer Rejected', NG case: >> [2011-08-17 15:10:19.21262] I [glusterd-handler.c:602:glusterd_handle_cli_probe] 0-glusterd: Received CLI probe req foo-3-private 24007 >> [2011-08-17 15:10:19.24605] I [glusterd-handler.c:391:glusterd_friend_find] 0-glusterd: Unable to find hostname: foo-3-private >> [2011-08-17 15:10:19.24648] I [glusterd-handler.c:3120:glusterd_probe_begin] 0-glusterd: Unable to find peerinfo for host: foo-3-private (24007) >> [2011-08-17 15:10:19.27736] W [rpc-transport.c:728:rpc_transport_load] 0-rpc-transport: missing 'option transport-type'. defaulting to "socket" >> [2011-08-17 15:10:19.32034] I [glusterd-handler.c:3102:glusterd_friend_add] 0-glusterd: connect returned 0 >> [2011-08-17 15:10:19.32389] I [glusterd-handshake.c:317:glusterd_set_clnt_mgmt_program] 0-: Using Program glusterd clnt mgmt, Num (1238433), Version (1) >> [2011-08-17 15:10:19.32426] I [glusterd-utils.c:2127:glusterd_friend_find_by_hostname] 0-glusterd: Friend foo-3-private found.. state: 0 >> [2011-08-17 15:10:19.40671] I [glusterd-rpc-ops.c:364:glusterd3_1_probe_cbk] 0-glusterd: Received probe resp from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f, host: foo-3-private >> [2011-08-17 15:10:19.40720] I [glusterd-handler.c:379:glusterd_friend_find] 0-glusterd: Unable to find peer by uuid >> [2011-08-17 15:10:19.40748] I [glusterd-utils.c:2127:glusterd_friend_find_by_hostname] 0-glusterd: Friend foo-3-private found.. state: 0 >> [2011-08-17 15:10:19.40983] I [glusterd-rpc-ops.c:409:glusterd3_1_probe_cbk] 0-glusterd: Received resp to probe req >> [2011-08-17 15:10:19.42854] I [rpc-clnt.c:696:rpc_clnt_handle_cbk] 0-rpc-clnt: recieved rpc message (XID: 0x2a, Ver: 2, Program: 52743234, ProgVers: 1, Proc: 1) from rpc-transport (management) >> [2011-08-17 15:10:19.50762] I [glusterd-rpc-ops.c:454:glusterd3_1_friend_add_cbk] 0-glusterd: Received ACC from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f, host: foo-3-private, port: 0 >> [2011-08-17 15:10:19.50794] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Probe Sent to Peer >> [2011-08-17 15:10:19.50851] I [glusterd-handler.c:3293:glusterd_xfer_cli_probe_resp] 0-glusterd: Responded to CLI, ret: 0 >> [2011-08-17 15:10:19.51413] I [glusterd-handler.c:2882:glusterd_handle_probe_query] 0-glusterd: Received probe from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f >> [2011-08-17 15:10:19.51444] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Accepted peer request >> [2011-08-17 15:10:19.51487] I [glusterd-handler.c:2917:glusterd_handle_probe_query] 0-glusterd: Responded to 192.168.1.129, op_ret: 0, op_errno: 0, ret: 0 >> [2011-08-17 15:10:19.51853] I [glusterd-handler.c:2614:glusterd_handle_incoming_friend_req] 0-glusterd: Received probe from uuid: 464b3ea0-1b2b-4683-8209-72220dcb295f >> [2011-08-17 15:10:19.51885] I [glusterd-utils.c:2088:glusterd_friend_find_by_uuid] 0-glusterd: Friend found... state: Accepted peer request >> [2011-08-17 15:10:19.51930] E [glusterd-utils.c:1407:glusterd_compare_friend_volume] 0-: Cksums of volume foo differ. local cksum = 1403573944, remote cksum = -1413994823 >> [2011-08-17 15:10:19.51975] I [glusterd-handler.c:3270:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to 192.168.1.129 (0), ret: 0 >> >> >>> On Tue, Aug 16, 2011 at 8:18 PM, Tomoaki Sato<tsato at valinux.co.jp> wrote: >>>> Mohit, >>>> >>>> let me say again. >>>> 3.1.6-1 fail to 'peer probe' after 'start volume' in my environment. >>>> case-A) peer probe foo-3-private --> Peer in Cluster >>>> >>>> <delete all configuration files and reboot all foo-X-private nodes> >>>> >>>> [root at foo-1-private ~]# gluster peer probe foo-3-private >>>> Probe successful >>>> [root at foo-1-private ~]# gluster peer status >>>> Number of Peers: 1 >>>> >>>> Hostname: foo-3-private >>>> Uuid: ef7d3c43-219a-4d13-a918-2639455cfbe7 >>>> State: Peer in Cluster (Connected) >>>> >>>> case-B) create volume then peer probe foo-3-private --> Peer in Cluster >>>> >>>> <delete all configuration files and reboot all foo-X-private nodes> >>>> >>>> [root at foo-1-private ~]# gluster volume create foo foo-1-private:/mnt/brick >>>> Creation of volume foo has been successful. Please start the volume to >>>> access data. >>>> [root at foo-1-private ~]# gluster peer probe foo-3-private >>>> Probe successful >>>> [root at foo-1-private ~]# gluster peer status >>>> Number of Peers: 1 >>>> >>>> Hostname: foo-3-private >>>> Uuid: fe44c954-4679-4389-a0e6-4c1fd4569a02 >>>> State: Peer in Cluster (Connected) >>>> >>>> case-C) start volume then peer probe foo-3-private --> Peer Rejected >>>> >>>> <delete all configuration files and reboot all foo-X-private nodes> >>>> >>>> [root at foo-1-private ~]# gluster volume create foo foo-1-private:/mnt/brick >>>> Creation of volume foo has been successful. Please start the volume to >>>> access data. >>>> [root at foo-1-private ~]# gluster volume start foo >>>> Starting volume foo has been successful >>>> [root at foo-1-private ~]# gluster peer probe foo-3-private >>>> Probe successful >>>> [root at foo-1-private ~]# gluster peer status >>>> Number of Peers: 1 >>>> >>>> Hostname: foo-3-private >>>> Uuid: bb6932e4-5bf0-4d34-872f-4a5fc1d0b6f8 >>>> State: Peer Rejected (Connected) >>>> >>>> >>>>> Can you for now put it in /etc/hosts and test? >>>> >>>> All foo-X-private hosts have no entries in /etc/hosts. >>>> All the nodes obtain IP addresses from a DHCP server and register the IP >>>> address - host name paires to a DNS server. >>>> >>>>> >>>>> also, make sure you have same version of gluster running on all the nodes. >>>> >>>> Since all three foo-X-private hosts are generated from a common VM template, >>>> same version of gluster run on all the nodes. >>>> >>>>> >>>>> What's the result of gluster peer status on node 3? >>>> >>>> [root at foo-1-private ~]# ssh foo-3-private gluster peer status >>>> Number of Peers: 1 >>>> >>>> Hostname: 192.168.1.129 >>>> Uuid: 828bcc00-14d3-4505-8b35-d0ac6ca0730a >>>> State: Peer Rejected (Connected) >>>> [root at foo-1-private ~]# >>>> >>>> >>>> Best, >>>> >>>> >>>> (2011/08/17 0:53), Mohit Anchlia wrote: >>>>> >>>>> I see this in the logs: >>>>> >>>>> [2011-08-16 11:57:05.642903] I >>>>> [glusterd-handler.c:391:glusterd_friend_find] 0-glusterd: Unable to >>>>> find hostname: foo-3-private >>>>> >>>>> Can you for now put it in /etc/hosts and test? >>>>> >>>>> also, make sure you have same version of gluster running on all the nodes. >>>>> >>>>> What's the result of gluster peer status on node 3? >>>>> >>>>> On Mon, Aug 15, 2011 at 8:18 PM, Tomoaki Sato<tsato at valinux.co.jp> wrote: >>>>>> >>>>>> Mohit >>>>>> >>>>>> I've tried same test and reproduce the 'Peer Rejected' status. >>>>>> please find config files and log files in attached taz. >>>>>> >>>>>> >>>>>> [root at vhead-010 ~]# date >>>>>> Tue Aug 16 11:55:15 JST 2011 >>>>>> [root at vhead-010 ~]# cat a.sh >>>>>> #!/bin/bash >>>>>> for i in foo-{1..3}-private >>>>>> do >>>>>> ssh ${i} service glusterd stop >>>>>> ssh ${i} 'find /etc/glusterd -type f|xargs rm -f' >>>>>> ssh ${i} rm -rf /etc/glusterd/vols/* >>>>>> ssh ${i} service iptables stop >>>>>> ssh ${i} cp /dev/null /var/log/glusterfs/nfs.log >>>>>> ssh ${i} cp /dev/null /var/log/glusterfs/bricks/mnt-brick.log >>>>>> ssh ${i} cp /dev/null /var/log/glusterfs/.cmd_log_history >>>>>> ssh ${i} cp /dev/null >>>>>> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log >>>>>> ssh ${i} service glusterd start >>>>>> ssh ${i} find /etc/glusterd >>>>>> ssh ${i} service glusterd status >>>>>> done >>>>>> [root at vhead-010 ~]# bash a.sh >>>>>> Stopping glusterd:[ OK ] >>>>>> Flushing firewall rules: [ OK ] >>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>> Unloading iptables modules: [ OK ] >>>>>> Starting glusterd:[ OK ] >>>>>> /etc/glusterd >>>>>> /etc/glusterd/glusterd.info >>>>>> /etc/glusterd/nfs >>>>>> /etc/glusterd/nfs/run >>>>>> /etc/glusterd/peers >>>>>> /etc/glusterd/vols >>>>>> glusterd (pid 15617) is running... >>>>>> Stopping glusterd:[ OK ] >>>>>> Flushing firewall rules: [ OK ] >>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>> Unloading iptables modules: [ OK ] >>>>>> Starting glusterd:[ OK ] >>>>>> /etc/glusterd >>>>>> /etc/glusterd/glusterd.info >>>>>> /etc/glusterd/nfs >>>>>> /etc/glusterd/nfs/run >>>>>> /etc/glusterd/peers >>>>>> /etc/glusterd/vols >>>>>> glusterd (pid 15147) is running... >>>>>> Stopping glusterd:[ OK ] >>>>>> Flushing firewall rules: [ OK ] >>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>> Unloading iptables modules: [ OK ] >>>>>> Starting glusterd:[ OK ] >>>>>> /etc/glusterd >>>>>> /etc/glusterd/glusterd.info >>>>>> /etc/glusterd/nfs >>>>>> /etc/glusterd/nfs/run >>>>>> /etc/glusterd/peers >>>>>> /etc/glusterd/vols >>>>>> glusterd (pid 15177) is running... >>>>>> [root at vhead-010 ~]# ssh foo-1-private >>>>>> Last login: Tue Aug 16 09:51:49 2011 from dlp.local.valinux.co.jp >>>>>> [root at localhost ~]# gluster peer probe foo-2-private >>>>>> Probe successful >>>>>> [root at localhost ~]# gluster peer status >>>>>> Number of Peers: 1 >>>>>> >>>>>> Hostname: foo-2-private >>>>>> Uuid: 20b73d9a-ede0-454f-9fbb-b0eee9ce26a3 >>>>>> State: Peer in Cluster (Connected) >>>>>> [root at localhost ~]# gluster volume create foo foo-1-private:/mnt/brick >>>>>> Creation of volume foo has been successful. Please start the volume to >>>>>> access data. >>>>>> [root at localhost ~]# gluster volume start foo >>>>>> Starting volume foo has been successful >>>>>> [root at localhost ~]# gluster volume add-brick foo foo-2-private:/mnt/brick >>>>>> Add Brick successful >>>>>> [root at localhost ~]# gluster peer probe foo-3-private >>>>>> Probe successful >>>>>> [root at localhost ~]# gluster peer status >>>>>> Number of Peers: 2 >>>>>> >>>>>> Hostname: foo-2-private >>>>>> Uuid: 20b73d9a-ede0-454f-9fbb-b0eee9ce26a3 >>>>>> State: Peer in Cluster (Connected) >>>>>> >>>>>> Hostname: foo-3-private >>>>>> Uuid: 7587ae34-9209-484a-9576-3939e061720c >>>>>> State: Peer Rejected (Connected) >>>>>> [root at localhost ~]# exit >>>>>> logout >>>>>> Connection to foo-1-private closed. >>>>>> [root at vhead-010 ~]# find foo_log_and_conf >>>>>> foo_log_and_conf >>>>>> foo_log_and_conf/foo-2-private >>>>>> foo_log_and_conf/foo-2-private/glusterd >>>>>> foo_log_and_conf/foo-2-private/glusterd/glusterd.info >>>>>> foo_log_and_conf/foo-2-private/glusterd/nfs >>>>>> foo_log_and_conf/foo-2-private/glusterd/nfs/nfs-server.vol >>>>>> foo_log_and_conf/foo-2-private/glusterd/nfs/run >>>>>> foo_log_and_conf/foo-2-private/glusterd/nfs/run/nfs.pid >>>>>> foo_log_and_conf/foo-2-private/glusterd/peers >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/peers/461f6e21-90c4-4b6c-bda8-7b99bacb2722 >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/info >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/bricks >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/bricks/foo-2-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/bricks/foo-1-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/foo.foo-2-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/cksum >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/run >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/run/foo-2-private-mnt-brick.pid >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/foo-fuse.vol >>>>>> >>>>>> foo_log_and_conf/foo-2-private/glusterd/vols/foo/foo.foo-1-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-2-private/glusterfs >>>>>> foo_log_and_conf/foo-2-private/glusterfs/nfs.log >>>>>> foo_log_and_conf/foo-2-private/glusterfs/bricks >>>>>> foo_log_and_conf/foo-2-private/glusterfs/bricks/mnt-brick.log >>>>>> foo_log_and_conf/foo-2-private/glusterfs/.cmd_log_history >>>>>> foo_log_and_conf/foo-2-private/glusterfs/etc-glusterfs-glusterd.vol.log >>>>>> foo_log_and_conf/foo-1-private >>>>>> foo_log_and_conf/foo-1-private/glusterd >>>>>> foo_log_and_conf/foo-1-private/glusterd/glusterd.info >>>>>> foo_log_and_conf/foo-1-private/glusterd/nfs >>>>>> foo_log_and_conf/foo-1-private/glusterd/nfs/nfs-server.vol >>>>>> foo_log_and_conf/foo-1-private/glusterd/nfs/run >>>>>> foo_log_and_conf/foo-1-private/glusterd/nfs/run/nfs.pid >>>>>> foo_log_and_conf/foo-1-private/glusterd/peers >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/peers/20b73d9a-ede0-454f-9fbb-b0eee9ce26a3 >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/peers/7587ae34-9209-484a-9576-3939e061720c >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/info >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/bricks >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/bricks/foo-2-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/bricks/foo-1-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/foo.foo-2-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/cksum >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/run >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/run/foo-1-private-mnt-brick.pid >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/foo-fuse.vol >>>>>> >>>>>> foo_log_and_conf/foo-1-private/glusterd/vols/foo/foo.foo-1-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-1-private/glusterfs >>>>>> foo_log_and_conf/foo-1-private/glusterfs/nfs.log >>>>>> foo_log_and_conf/foo-1-private/glusterfs/bricks >>>>>> foo_log_and_conf/foo-1-private/glusterfs/bricks/mnt-brick.log >>>>>> foo_log_and_conf/foo-1-private/glusterfs/.cmd_log_history >>>>>> foo_log_and_conf/foo-1-private/glusterfs/etc-glusterfs-glusterd.vol.log >>>>>> foo_log_and_conf/foo-3-private >>>>>> foo_log_and_conf/foo-3-private/glusterd >>>>>> foo_log_and_conf/foo-3-private/glusterd/glusterd.info >>>>>> foo_log_and_conf/foo-3-private/glusterd/nfs >>>>>> foo_log_and_conf/foo-3-private/glusterd/nfs/run >>>>>> foo_log_and_conf/foo-3-private/glusterd/peers >>>>>> >>>>>> foo_log_and_conf/foo-3-private/glusterd/peers/461f6e21-90c4-4b6c-bda8-7b99bacb2722 >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/info >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/bricks >>>>>> >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/bricks/foo-2-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/bricks/foo-1-private:-mnt-brick >>>>>> >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/foo.foo-2-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/cksum >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/foo-fuse.vol >>>>>> >>>>>> foo_log_and_conf/foo-3-private/glusterd/vols/foo/foo.foo-1-private.mnt-brick.vol >>>>>> foo_log_and_conf/foo-3-private/glusterfs >>>>>> foo_log_and_conf/foo-3-private/glusterfs/nfs.log >>>>>> foo_log_and_conf/foo-3-private/glusterfs/bricks >>>>>> foo_log_and_conf/foo-3-private/glusterfs/bricks/mnt-brick.log >>>>>> foo_log_and_conf/foo-3-private/glusterfs/.cmd_log_history >>>>>> foo_log_and_conf/foo-3-private/glusterfs/etc-glusterfs-glusterd.vol.log >>>>>> [root at vhead-010 ~]# exit >>>>>> >>>>>> Best, >>>>>> >>>>>> (2011/08/16 9:35), Mohit Anchlia wrote: >>>>>>> >>>>>>> I should have also asked you to stop and delete volume before getting >>>>>>> rid of gluster config files. Can you get rid of directories also >>>>>>> inside vols and try to restart? It's trying to look for volume files >>>>>>> that we just removed. >>>>>>> >>>>>>> Also, just disable iptables for now explicitly. >>>>>>> >>>>>>> On Mon, Aug 15, 2011 at 5:22 PM, Tomoaki Sato<tsato at valinux.co.jp> >>>>>>> wrote: >>>>>>>> >>>>>>>>> 1) run peer detach for all the servers >>>>>>>> >>>>>>>> done. >>>>>>>> >>>>>>>>> 2) from server 1 ->3 and 3->1 make sure ports are open and iptables >>>>>>>>> are turned off. >>>>>>>> >>>>>>>> done. >>>>>>>> by the way, the same test on 3.1.5-1 works fine with same environment. >>>>>>>> >>>>>>>>> 3) remove config files under /etc/glusterd >>>>>>>> >>>>>>>> please review following logs. >>>>>>>> >>>>>>>>> 4) run your tests again. >>>>>>>> >>>>>>>> I don't know why but glusterd service failed to start on all 3 hosts. >>>>>>>> >>>>>>>> [root at vhead-010 ~]# date >>>>>>>> Tue Aug 16 09:12:53 JST 2011 >>>>>>>> [root at vhead-010 ~]# cat a.sh >>>>>>>> #!/bin/bash >>>>>>>> for i in foo-{1..3}-private >>>>>>>> do >>>>>>>> ssh ${i} service glusterd stop >>>>>>>> ssh ${i} 'find /etc/glusterd -type f|xargs rm -f' >>>>>>>> ssh ${i} service iptables restart >>>>>>>> ssh ${i} iptables -vL >>>>>>>> ssh ${i} service glusterd start >>>>>>>> ssh ${i} find /etc/glusterd >>>>>>>> ssh ${i} service glusterd status >>>>>>>> done >>>>>>>> [root at vhead-010 ~]# bash a.sh >>>>>>>> Stopping glusterd:[ OK ] >>>>>>>> Flushing firewall rules: [ OK ] >>>>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>>>> Unloading iptables modules: [ OK ] >>>>>>>> Chain INPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> Starting glusterd:[ OK ] >>>>>>>> /etc/glusterd >>>>>>>> /etc/glusterd/glusterd.info >>>>>>>> /etc/glusterd/nfs >>>>>>>> /etc/glusterd/nfs/run >>>>>>>> /etc/glusterd/peers >>>>>>>> /etc/glusterd/vols >>>>>>>> /etc/glusterd/vols/foo >>>>>>>> /etc/glusterd/vols/foo/bricks >>>>>>>> /etc/glusterd/vols/foo/run >>>>>>>> glusterd is stopped >>>>>>>> Stopping glusterd:[ OK ] >>>>>>>> Flushing firewall rules: [ OK ] >>>>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>>>> Unloading iptables modules: [ OK ] >>>>>>>> Chain INPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> Starting glusterd:[ OK ] >>>>>>>> /etc/glusterd >>>>>>>> /etc/glusterd/glusterd.info >>>>>>>> /etc/glusterd/nfs >>>>>>>> /etc/glusterd/nfs/run >>>>>>>> /etc/glusterd/peers >>>>>>>> /etc/glusterd/vols >>>>>>>> /etc/glusterd/vols/foo >>>>>>>> /etc/glusterd/vols/foo/bricks >>>>>>>> /etc/glusterd/vols/foo/run >>>>>>>> glusterd is stopped >>>>>>>> Stopping glusterd:[ OK ] >>>>>>>> Flushing firewall rules: [ OK ] >>>>>>>> Setting chains to policy ACCEPT: filter [ OK ] >>>>>>>> Unloading iptables modules: [ OK ] >>>>>>>> Chain INPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> >>>>>>>> Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) >>>>>>>> pkts bytes target prot opt in out source >>>>>>>> destination >>>>>>>> Starting glusterd:[ OK ] >>>>>>>> /etc/glusterd >>>>>>>> /etc/glusterd/glusterd.info >>>>>>>> /etc/glusterd/nfs >>>>>>>> /etc/glusterd/nfs/run >>>>>>>> /etc/glusterd/peers >>>>>>>> /etc/glusterd/vols >>>>>>>> /etc/glusterd/vols/foo >>>>>>>> /etc/glusterd/vols/foo/bricks >>>>>>>> /etc/glusterd/vols/foo/run >>>>>>>> glusterd is stopped >>>>>>>> [root at vhead-010 ~]# date >>>>>>>> Tue Aug 16 09:13:20 JST 2011 >>>>>>>> [root at vhead-010 ~]# ssh foo-1-private >>>>>>>> Last login: Tue Aug 16 09:06:57 2011 from dlp.local.valinux.co.jp >>>>>>>> [root at localhost ~]# tail -20 >>>>>>>> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log >>>>>>>> ... >>>>>>>> [2011-08-16 09:13:01.85858] I [glusterd.c:304:init] 0-management: Using >>>>>>>> /etc/glusterd as working directory >>>>>>>> [2011-08-16 09:13:01.87294] E [rpc-transport.c:799:rpc_transport_load] >>>>>>>> 0-rpc-transport: >>>>>>>> /opt/glusterfs/3.1.6/lib64/glusterfs/3.1.6/rpc-transport/rdma.so: >>>>>>>> cannot >>>>>>>> open shared object file: No such file or directory >>>>>>>> [2011-08-16 09:13:01.87340] E [rpc-transport.c:803:rpc_transport_load] >>>>>>>> 0-rpc-transport: volume 'rdma.management': transport-type 'rdma' is not >>>>>>>> valid or not found on this machine >>>>>>>> [2011-08-16 09:13:01.87402] E >>>>>>>> [glusterd-store.c:654:glusterd_store_handle_retrieve] 0-glusterd: >>>>>>>> Unable >>>>>>>> to >>>>>>>> retrieve store handle for /etc/glusterd/glusterd.info, error: No such >>>>>>>> file >>>>>>>> or directory >>>>>>>> [2011-08-16 09:13:01.87422] E >>>>>>>> [glusterd-store.c:761:glusterd_retrieve_uuid] >>>>>>>> 0-: Unable to get store handle! >>>>>>>> [2011-08-16 09:13:01.87514] I [glusterd.c:95:glusterd_uuid_init] >>>>>>>> 0-glusterd: >>>>>>>> generated UUID: c0cef9f9-a79e-4189-8955-d83927db9cee >>>>>>>> [2011-08-16 09:13:01.87681] E >>>>>>>> [glusterd-store.c:654:glusterd_store_handle_retrieve] 0-glusterd: >>>>>>>> Unable >>>>>>>> to >>>>>>>> retrieve store handle for /etc/glusterd/vols/foo/info, error: No such >>>>>>>> file >>>>>>>> or directory >>>>>>>> [2011-08-16 09:13:01.87704] E >>>>>>>> [glusterd-store.c:1328:glusterd_store_retrieve_volumes] 0-: Unable to >>>>>>>> restore volume: foo >>>>>>> >>>>>>>> [2011-08-16 09:13:01.87732] E [xlator.c:843:xlator_init] 0-management: >>>>>>>> Initialization of volume 'management' failed, review your volfile again >>>>>>>> [2011-08-16 09:13:01.87751] E [graph.c:331:glusterfs_graph_init] >>>>>>>> 0-management: initializing translator failed >>>>>>>> [2011-08-16 09:13:01.87818] I [glusterfsd.c:712:cleanup_and_exit] >>>>>>>> 0-glusterfsd: shutting down >>>>>>>> [root at localhost ~]# exit >>>>>>>> >>>>>>>> Best, >>>>>>>> >>>>>>>> (2011/08/16 8:52), Mohit Anchlia wrote: >>>>>>>>> >>>>>>>>> Logs are generally in /var/log/gluster >>>>>>>>> >>>>>>>>> Since you are playing with it. I would suggest this: >>>>>>>>> >>>>>>>>> 1) run peer detach for all the servers >>>>>>>>> 2) from server 1 ->3 and 3->1 make sure ports are open and iptables >>>>>>>>> are turned off. >>>>>>>>> 3) remove config files under /etc/glusterd >>>>>>>>> 4) run your tests again. >>>>>>>>> >>>>>>>>> On Mon, Aug 15, 2011 at 4:28 PM, Tomoaki Sato<tsato at valinux.co.jp> >>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>> Thanks, Mohit >>>>>>>>>> >>>>>>>>>> (2011/08/16 8:05), Mohit Anchlia wrote: >>>>>>>>>>> >>>>>>>>>>> What's in your logs? >>>>>>>>>> >>>>>>>>>> I can obtain logs needed. could you tell me the instruction to take >>>>>>>>>> the >>>>>>>>>> logs? >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Did you have foo-3-private before in your gluster cluster ever or >>>>>>>>>>> adding this host for the first time? >>>>>>>>>> >>>>>>>>>> It was first time. >>>>>>>>>> All foo-X-private has no entries in /etc/glusterd/peers/ and >>>>>>>>>> /etc/glusterd/vols/. >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Try gluster peer detach and then remove any left over configuration >>>>>>>>>>> in >>>>>>>>>>> /etc/glusterd config directory. After that try again and see if that >>>>>>>>>>> works. >>>>>>>>>> >>>>>>>>>> [root at vhead-010 ~]# date >>>>>>>>>> Tue Aug 16 08:17:49 JST 2011 >>>>>>>>>> [root at vhead-010 ~]# cat a.sh >>>>>>>>>> #!/bin/bash >>>>>>>>>> for i in foo-{1..3}-private >>>>>>>>>> do >>>>>>>>>> ssh ${i} service glusterd stop >>>>>>>>>> ssh ${i} rm -rf /etc/glusterd/peers/* >>>>>>>>>> ssh ${i} rm -rf /etc/glusterd/vols/* >>>>>>>>>> ssh ${i} service glusterd start >>>>>>>>>> ssh ${i} find /etc/glusterd >>>>>>>>>> done >>>>>>>>>> [root at vhead-010 ~]# bash a.sh >>>>>>>>>> Stopping glusterd:[ OK ] >>>>>>>>>> Starting glusterd:[ OK ] >>>>>>>>>> /etc/glusterd >>>>>>>>>> /etc/glusterd/glusterd.info >>>>>>>>>> /etc/glusterd/nfs >>>>>>>>>> /etc/glusterd/nfs/nfs-server.vol >>>>>>>>>> /etc/glusterd/nfs/run >>>>>>>>>> /etc/glusterd/peers >>>>>>>>>> /etc/glusterd/vols >>>>>>>>>> Stopping glusterd:[ OK ] >>>>>>>>>> Starting glusterd:[ OK ] >>>>>>>>>> /etc/glusterd >>>>>>>>>> /etc/glusterd/glusterd.info >>>>>>>>>> /etc/glusterd/nfs >>>>>>>>>> /etc/glusterd/nfs/nfs-server.vol >>>>>>>>>> /etc/glusterd/nfs/run >>>>>>>>>> /etc/glusterd/peers >>>>>>>>>> /etc/glusterd/vols >>>>>>>>>> Stopping glusterd:[ OK ] >>>>>>>>>> Starting glusterd:[ OK ] >>>>>>>>>> /etc/glusterd >>>>>>>>>> /etc/glusterd/glusterd.info >>>>>>>>>> /etc/glusterd/nfs >>>>>>>>>> /etc/glusterd/nfs/nfs-server.vol >>>>>>>>>> /etc/glusterd/nfs/run >>>>>>>>>> /etc/glusterd/peers >>>>>>>>>> /etc/glusterd/vols >>>>>>>>>> [root at vhead-010 ~]# ssh foo-1-private >>>>>>>>>> [root at localhost ~]# gluster peer probe foo-2-private >>>>>>>>>> Probe successful >>>>>>>>>> [root at localhost ~]# gluster peer status >>>>>>>>>> Number of Peers: 1 >>>>>>>>>> >>>>>>>>>> Hostname: foo-2-private >>>>>>>>>> Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2 >>>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>>> [root at localhost ~]# gluster volume create foo >>>>>>>>>> foo-1-private:/mnt/brick >>>>>>>>>> Creation of volume foo has been successful. Please start the volume >>>>>>>>>> to >>>>>>>>>> access da >>>>>>>>>> ta. >>>>>>>>>> [root at localhost ~]# gluster volume start foo >>>>>>>>>> Starting volume foo has been successful >>>>>>>>>> [root at localhost ~]# gluster volume add-brick foo >>>>>>>>>> foo-2-private:/mnt/brick >>>>>>>>>> Add Brick successful >>>>>>>>>> [root at localhost ~]# gluster peer probe foo-3-private >>>>>>>>>> Probe successful >>>>>>>>>> [root at localhost ~]# gluster peer status >>>>>>>>>> Number of Peers: 2 >>>>>>>>>> >>>>>>>>>> Hostname: foo-2-private >>>>>>>>>> Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2 >>>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>>> >>>>>>>>>> Hostname: foo-3-private >>>>>>>>>> Uuid: 7fb98dac-fef7-4b33-837c-6483a767ec3e >>>>>>>>>> State: Peer Rejected (Connected) >>>>>>>>>> [root at localhost ~]# cat /var/log/glusterfs/.cmd_log_history >>>>>>>>>> ... >>>>>>>>>> [2011-08-16 08:20:28.862619] peer probe : on host >>>>>>>>>> foo-2-private:24007 >>>>>>>>>> [2011-08-16 08:20:28.912419] peer probe : on host foo-2-private:24007 >>>>>>>>>> FAILED >>>>>>>>>> [2011-08-16 08:20:58.382350] Volume create : on volname: foo >>>>>>>>>> attempted >>>>>>>>>> [2011-08-16 08:20:58.382461] Volume create : on volname: foo >>>>>>>>>> type:DEFAULT >>>>>>>>>> count: >>>>>>>>>> 1 bricks: foo-1-private:/mnt/brick >>>>>>>>>> [2011-08-16 08:20:58.384674] Volume create : on volname: foo SUCCESS >>>>>>>>>> [2011-08-16 08:21:04.831772] volume start : on volname: foo SUCCESS >>>>>>>>>> [2011-08-16 08:21:22.682292] Volume add-brick : on volname: foo >>>>>>>>>> attempted >>>>>>>>>> [2011-08-16 08:21:22.682385] Volume add-brick : volname: foo type >>>>>>>>>> DEFAULT >>>>>>>>>> count: >>>>>>>>>> 1 bricks: foo-2-private:/mnt/brick >>>>>>>>>> [2011-08-16 08:21:22.682499] Volume add-brick : on volname: foo >>>>>>>>>> SUCCESS >>>>>>>>>> [2011-08-16 08:21:39.124574] peer probe : on host >>>>>>>>>> foo-3-private:24007 >>>>>>>>>> [2011-08-16 08:21:39.135609] peer probe : on host foo-3-private:24007 >>>>>>>>>> FAILED >>>>>>>>>> >>>>>>>>>> Tomo >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Aug 15, 2011 at 3:37 PM, Tomoaki Sato<tsato at valinux.co.jp> >>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> following instructions work fine with 3.1.5-1 but with 3.1.6-1. >>>>>>>>>>>> >>>>>>>>>>>> 1. make a new file system without peers. [OK] >>>>>>>>>>>> >>>>>>>>>>>> foo-1-private# gluster volume create foo foo-1-private:/mnt/brick >>>>>>>>>>>> foo-1-private# gluster volume start foo >>>>>>>>>>>> foo-1-private# gluster peer status >>>>>>>>>>>> No peers present >>>>>>>>>>>> foo-1-private# >>>>>>>>>>>> >>>>>>>>>>>> 2. add a peer to the file system. [NG] >>>>>>>>>>>> >>>>>>>>>>>> foo-1-private# gluster peer probe foo-2-private >>>>>>>>>>>> Probe successful >>>>>>>>>>>> foo-1-private# gluster peer status >>>>>>>>>>>> Number of Peers: 1 >>>>>>>>>>>> >>>>>>>>>>>> Hostname: foo-2-private >>>>>>>>>>>> Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2 >>>>>>>>>>>> State: Peer Rejected (Connected) >>>>>>>>>>>> foo-1-private# gluster volume add-brick foo >>>>>>>>>>>> foo-2-private:/mnt/brick >>>>>>>>>>>> Host foo-2-private not connected >>>>>>>>>>>> foo-1-private# >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> following instructions work fine even with 3.1.6-1. >>>>>>>>>>>> >>>>>>>>>>>> 1. make a new file system with single peer. [OK] >>>>>>>>>>>> >>>>>>>>>>>> foo-1-private# gluster peer status >>>>>>>>>>>> No peer presents >>>>>>>>>>>> foo-1-private# gluster peer probe foo-2-private >>>>>>>>>>>> Probe successful >>>>>>>>>>>> foo-1-private# gluster peer status >>>>>>>>>>>> Number of Peers: 1 >>>>>>>>>>>> >>>>>>>>>>>> Hostname: foo-2-private >>>>>>>>>>>> Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2 >>>>>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>>>>> foo-1-private# gluster volume create foo foo-1-private:/mnt/brick >>>>>>>>>>>> Creation of volume foo has been successful. Please start the volume >>>>>>>>>>>> to >>>>>>>>>>>> access data. >>>>>>>>>>>> foo-1-private# gluster volume start foo >>>>>>>>>>>> Starting volume foo has been successful >>>>>>>>>>>> foo-1-private# gluster volume add-brick foo >>>>>>>>>>>> foo-2-private:/mnt/brick >>>>>>>>>>>> Add Brick successful >>>>>>>>>>>> foo-1-private# >>>>>>>>>>>> >>>>>>>>>>>> But ... >>>>>>>>>>>> >>>>>>>>>>>> 2. add a peer to the file system. [NG] >>>>>>>>>>>> >>>>>>>>>>>> foo-1-private# gluster peer probe foo-3-private >>>>>>>>>>>> Probe successful >>>>>>>>>>>> foo-1-private# gluster peer status >>>>>>>>>>>> Number of Peers: 2 >>>>>>>>>>>> >>>>>>>>>>>> Hostname: foo-2-private >>>>>>>>>>>> Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2 >>>>>>>>>>>> State: Peer in Cluster (Connected) >>>>>>>>>>>> >>>>>>>>>>>> Hostname: foo-3-private >>>>>>>>>>>> Uuid: 7fb98dac-fef704b33-837c-6483a767ec3e >>>>>>>>>>>> State: Peer Rejected (Connected) >>>>>>>>>>>> foo-1-private# gluster volume add-brick foo >>>>>>>>>>>> foo-3-private:/mnt/brick >>>>>>>>>>>> Host foo-3-private not connected >>>>>>>>>>>> foo-1-private# >>>>>>>>>>>> >>>>>>>>>>>> How should I add extra peers to existing file systems ? >>>>>>>>>>>> >>>>>>>>>>>> Best, >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Gluster-users mailing list >>>>>>>>>>>> Gluster-users at gluster.org >>>>>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>> >>>>>>>> >>>>>> >>>>>> >>>> >>>> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users -------------- next part -------------- A non-text attachment was scrubbed... Name: foo-1-private.tar Type: application/octet-stream Size: 10240 bytes Desc: not available URL: <http://gluster.org/pipermail/gluster-users/attachments/20110817/448b88a6/attachment-0002.obj> -------------- next part -------------- A non-text attachment was scrubbed... Name: foo-3-private.tar Type: application/octet-stream Size: 10240 bytes Desc: not available URL: <http://gluster.org/pipermail/gluster-users/attachments/20110817/448b88a6/attachment-0003.obj>