furthermore, when I stop gluster, and restartd glusterfs, in the log, I have ==> etc-glusterfs-glusterd.vol.log <== [2013-01-11 16:39:55.438506] I [glusterfsd.c:1666:main] 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.3.1 [2013-01-11 16:39:55.440098] I [glusterd.c:807:init] 0-management: Using /var/lib/glusterd as working directory [2013-01-11 16:39:55.440797] C [rdma.c:4102:gf_rdma_init] 0-rpc-transport/rdma: Failed to get IB devices [2013-01-11 16:39:55.440859] E [rdma.c:4993:init] 0-rdma.management: Failed to initialize IB Device [2013-01-11 16:39:55.440881] E [rpc-transport.c:316:rpc_transport_load] 0-rpc-transport: 'rdma' initialization failed [2013-01-11 16:39:55.440901] W [rpcsvc.c:1356:rpcsvc_transport_create] 0-rpc-service: cannot create listener, initing the transport failed [2013-01-11 16:39:55.440992] I [glusterd.c:95:glusterd_uuid_init] 0-glusterd: retrieved UUID: eece061b-1cd0-4f30-ad17-61809297aba9 [2013-01-11 16:39:56.050996] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-0 [2013-01-11 16:39:56.051041] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-1 [2013-01-11 16:39:56.235444] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-0 [2013-01-11 16:39:56.235482] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-1 [2013-01-11 16:39:56.235810] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-0 [2013-01-11 16:39:56.235831] E [glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key: brick-1 [2013-01-11 16:39:56.236277] I [rpc-clnt.c:968:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 [2013-01-11 16:39:56.236843] I [glusterd-handler.c:2227:glusterd_friend_add] 0-management: connect returned 0 [2013-01-11 16:39:56.241266] E [glusterd-store.c:2586:glusterd_resolve_all_bricks] 0-glusterd: resolve brick failed in restore [2013-01-11 16:39:56.243958] E [glusterd-utils.c:3418:glusterd_brick_start] 0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/ssl [2013-01-11 16:39:56.247827] E [glusterd-utils.c:3418:glusterd_brick_start] 0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/dist [2013-01-11 16:39:56.251832] E [glusterd-utils.c:3418:glusterd_brick_start] 0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/bucket [2013-01-11 16:39:56.258909] I [rpc-clnt.c:968:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 ==> nfs.log.1 <== [2013-01-11 16:39:56.259055] W [socket.c:410:__socket_keepalive] 0-socket: failed to set keep idle on socket 7 [2013-01-11 16:39:56.259108] W [socket.c:1876:socket_server_event_handler] 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported [2013-01-11 16:39:56.259154] W [socket.c:410:__socket_keepalive] 0-socket: failed to set keep idle on socket 8 [2013-01-11 16:39:56.259172] W [socket.c:1876:socket_server_event_handler] 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported ==> etc-glusterfs-glusterd.vol.log <== [2013-01-11 16:39:56.266390] I [rpc-clnt.c:968:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 ==> glustershd.log.1 <== [2013-01-11 16:39:56.266520] W [socket.c:410:__socket_keepalive] 0-socket: failed to set keep idle on socket 7 [2013-01-11 16:39:56.266562] W [socket.c:1876:socket_server_event_handler] 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported ==> etc-glusterfs-glusterd.vol.log <== Given volfile: +------------------------------------------------------------------------------+ 1: volume management 2: type mgmt/glusterd 3: option working-directory /var/lib/glusterd 4: option transport-type socket,rdma 5: option transport.socket.keepalive-time 10 6: option transport.socket.keepalive-interval 2 7: option transport.socket.read-fail-log off 8: end-volume +------------------------------------------------------------------------------+ ==> glustershd.log.1 <== [2013-01-11 16:39:56.266610] W [socket.c:410:__socket_keepalive] 0-socket: failed to set keep idle on socket 8 [2013-01-11 16:39:56.266624] W [socket.c:1876:socket_server_event_handler] 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported ==> etc-glusterfs-glusterd.vol.log <== [2013-01-11 16:39:56.267030] I [glusterd-handshake.c:397:glusterd_set_clnt_mgmt_program] 0-: Using Program glusterd mgmt, Num (1238433), Version (2) [2013-01-11 16:39:56.267053] I [glusterd-handshake.c:403:glusterd_set_clnt_mgmt_program] 0-: Using Program Peer mgmt, Num (1238437), Version (2) ==> nfs.log.1 <== [2013-01-11 16:39:58.148908] W [nfs.c:735:nfs_init_state] 1-nfs: /sbin/rpc.statd not found. Disabling NLM ==> etc-glusterfs-glusterd.vol.log <== [2013-01-11 16:39:58.149702] I [glusterd-handler.c:1486:glusterd_handle_incoming_friend_req] 0-glusterd: Received probe from uuid: 184a81f4-ff0f-48d6-adb8-798b98957b1a [2013-01-11 16:39:58.149818] E [glusterd-utils.c:1926:glusterd_compare_friend_volume] 0-: Cksums of volume puppet-bucket differ. local cksum = 1273524870, remote cksum = 1932840611 [2013-01-11 16:39:58.149858] I [glusterd-handler.c:2395:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to bastille.mdc (0), ret: 0 ==> nfs.log.1 <== [2013-01-11 16:39:58.179450] E [socket.c:333:__socket_server_bind] 1-socket.nfs-server: binding to failed: Address already in use [2013-01-11 16:39:58.179512] E [socket.c:336:__socket_server_bind] 1-socket.nfs-server: Port is already in use [2013-01-11 16:39:58.179535] W [rpcsvc.c:1363:rpcsvc_transport_create] 1-rpc-service: listening on transport failed [2013-01-11 16:39:58.179663] E [rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not register with portmap [2013-01-11 16:39:58.179710] E [socket.c:333:__socket_server_bind] 1-socket.nfs-server: binding to failed: Address already in use [2013-01-11 16:39:58.179727] E [socket.c:336:__socket_server_bind] 1-socket.nfs-server: Port is already in use [2013-01-11 16:39:58.179743] W [rpcsvc.c:1363:rpcsvc_transport_create] 1-rpc-service: listening on transport failed [2013-01-11 16:39:58.179815] E [rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not register with portmap [2013-01-11 16:39:58.180193] E [socket.c:333:__socket_server_bind] 1-socket.nfs-server: binding to failed: Address already in use [2013-01-11 16:39:58.180214] E [socket.c:336:__socket_server_bind] 1-socket.nfs-server: Port is already in use [2013-01-11 16:39:58.180230] W [rpcsvc.c:1363:rpcsvc_transport_create] 1-rpc-service: listening on transport failed [2013-01-11 16:39:58.180300] E [rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not register with portmap [2013-01-11 16:39:58.180319] I [nfs.c:821:init] 1-nfs: NFS service started [2013-01-11 16:39:58.186245] W [graph.c:316:_log_if_unknown_option] 1-nfs-server: option 'rpc-auth.auth-glusterfs' is not recognized [2013-01-11 16:39:58.186346] W [graph.c:316:_log_if_unknown_option] 1-nfs-server: option 'rpc-auth-allow-insecure' is not recognized [2013-01-11 16:39:58.186366] W [graph.c:316:_log_if_unknown_option] 1-nfs-server: option 'transport-type' is not recognized [2013-01-11 16:39:58.186400] I [client.c:2142:notify] 1-puppet-ssl-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.187286] I [client.c:2142:notify] 1-puppet-ssl-client-1: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.188173] I [client.c:2142:notify] 1-puppet-dist-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.189031] I [client.c:2142:notify] 1-puppet-dist-client-1: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.189703] I [client.c:2142:notify] 1-puppet-bucket-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.190559] I [client.c:2142:notify] 1-puppet-bucket-client-1: parent translators are ready, attempting connect on transport Given volfile: +------------------------------------------------------------------------------+ 1: volume puppet-bucket-client-0 2: type protocol/client 3: option remote-host sandy.mdc 4: option remote-subvolume /opt/gluster-data/snake-puppet/bucket 5: option transport-type tcp 6: end-volume 7: 8: volume puppet-bucket-client-1 9: type protocol/client 10: option remote-host irene.mdc 11: option remote-subvolume /opt/gluster-data/puppet/bucket 12: option transport-type tcp 13: end-volume 14: 15: volume puppet-bucket-replicate-0 16: type cluster/replicate 17: subvolumes puppet-bucket-client-0 puppet-bucket-client-1 18: end-volume 19: 20: volume puppet-bucket 21: type debug/io-stats 22: option latency-measurement off 23: option count-fop-hits off 24: subvolumes puppet-bucket-replicate-0 25: end-volume 26: 27: volume puppet-dist-client-0 28: type protocol/client 29: option remote-host sandy.mdc 30: option remote-subvolume /opt/gluster-data/snake-puppet/dist 31: option transport-type tcp 32: end-volume 33: 34: volume puppet-dist-client-1 35: type protocol/client 36: option remote-host irene.mdc 37: option remote-subvolume /opt/gluster-data/puppet/dist 38: option transport-type tcp 39: end-volume 40: 41: volume puppet-dist-replicate-0 42: type cluster/replicate 43: option data-self-heal-algorithm full 44: subvolumes puppet-dist-client-0 puppet-dist-client-1 45: end-volume 46: 47: volume puppet-dist 48: type debug/io-stats 49: option latency-measurement off 50: option count-fop-hits off 51: subvolumes puppet-dist-replicate-0 52: end-volume 53: 54: volume puppet-ssl-client-0 55: type protocol/client 56: option remote-host sandy.mdc 57: option remote-subvolume /opt/gluster-data/snake-puppet/ssl 58: option transport-type tcp 59: end-volume 60: 61: volume puppet-ssl-client-1 62: type protocol/client 63: option remote-host irene.mdc 64: option remote-subvolume /opt/gluster-data/puppet/ssl 65: option transport-type tcp 66: end-volume 67: 68: volume puppet-ssl-replicate-0 69: type cluster/replicate 70: option metadata-change-log on 71: option data-self-heal-algorithm full 72: subvolumes puppet-ssl-client-0 puppet-ssl-client-1 73: end-volume 74: 75: volume puppet-ssl 76: type debug/io-stats 77: option latency-measurement off 78: option count-fop-hits off 79: subvolumes puppet-ssl-replicate-0 80: end-volume 81: 82: volume nfs-server 83: type nfs/server 84: option nfs.dynamic-volumes on 85: option nfs.nlm on 86: option rpc-auth.addr.puppet-ssl.allow * 87: option nfs3.puppet-ssl.volume-id bb2ffdd5-f00c-4016-ab07-301a6ede3042 88: option rpc-auth.addr.puppet-dist.allow * 89: option nfs3.puppet-dist.volume-id 376220d6-dcdd-4f3f-9809-397046a78f5a 90: option rpc-auth.addr.puppet-bucket.allow * 91: option nfs3.puppet-bucket.volume-id 3a7e146c-7c37-41ea-baa5-5262c79b1232 92: subvolumes puppet-ssl puppet-dist puppet-bucket 93: end-volume +------------------------------------------------------------------------------+ [2013-01-11 16:39:58.191727] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-ssl-client-1: changing port to 24010 (from 0) [2013-01-11 16:39:58.191806] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-dist-client-1: changing port to 24012 (from 0) [2013-01-11 16:39:58.191844] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-bucket-client-1: changing port to 24014 (from 0) [2013-01-11 16:39:58.191881] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-ssl-client-0: changing port to 24012 (from 0) [2013-01-11 16:39:58.191974] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-dist-client-0: changing port to 24010 (from 0) [2013-01-11 16:39:58.192024] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-bucket-client-0: changing port to 24014 (from 0) ==> glustershd.log.1 <== [2013-01-11 16:39:58.381647] I [graph.c:241:gf_add_cmdline_options] 0-puppet-ssl-replicate-0: adding option 'node-uuid' for volume 'puppet-ssl-replicate-0' with value 'eece061b-1cd0-4f30-ad17-61809297aba9' [2013-01-11 16:39:58.381673] I [graph.c:241:gf_add_cmdline_options] 0-puppet-dist-replicate-0: adding option 'node-uuid' for volume 'puppet-dist-replicate-0' with value 'eece061b-1cd0-4f30-ad17-61809297aba9' [2013-01-11 16:39:58.381686] I [graph.c:241:gf_add_cmdline_options] 0-puppet-bucket-replicate-0: adding option 'node-uuid' for volume 'puppet-bucket-replicate-0' with value 'eece061b-1cd0-4f30-ad17-61809297aba9' [2013-01-11 16:39:58.390396] I [client.c:2142:notify] 1-puppet-ssl-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.391487] I [client.c:2142:notify] 1-puppet-ssl-client-1: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.392209] I [client.c:2142:notify] 1-puppet-dist-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.392995] I [client.c:2142:notify] 1-puppet-dist-client-1: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.393804] I [client.c:2142:notify] 1-puppet-bucket-client-0: parent translators are ready, attempting connect on transport [2013-01-11 16:39:58.394598] I [client.c:2142:notify] 1-puppet-bucket-client-1: parent translators are ready, attempting connect on transport Given volfile: +------------------------------------------------------------------------------+ 1: volume puppet-bucket-client-0 2: type protocol/client 3: option remote-host sandy.mdc 4: option remote-subvolume /opt/gluster-data/snake-puppet/bucket 5: option transport-type tcp 6: end-volume 7: 8: volume puppet-bucket-client-1 9: type protocol/client 10: option remote-host irene.mdc 11: option remote-subvolume /opt/gluster-data/puppet/bucket 12: option transport-type tcp 13: end-volume 14: 15: volume puppet-bucket-replicate-0 16: type cluster/replicate 17: option background-self-heal-count 0 18: option metadata-self-heal on 19: option data-self-heal on 20: option entry-self-heal on 21: option self-heal-daemon on 22: option iam-self-heal-daemon yes 23: subvolumes puppet-bucket-client-0 puppet-bucket-client-1 24: end-volume 25: 26: volume puppet-dist-client-0 27: type protocol/client 28: option remote-host sandy.mdc 29: option remote-subvolume /opt/gluster-data/snake-puppet/dist 30: option transport-type tcp 31: end-volume 32: 33: volume puppet-dist-client-1 34: type protocol/client 35: option remote-host irene.mdc 36: option remote-subvolume /opt/gluster-data/puppet/dist 37: option transport-type tcp 38: end-volume 39: 40: volume puppet-dist-replicate-0 41: type cluster/replicate 42: option background-self-heal-count 0 43: option metadata-self-heal on 44: option data-self-heal on 45: option entry-self-heal on 46: option self-heal-daemon on 47: option data-self-heal-algorithm full 48: option iam-self-heal-daemon yes 49: subvolumes puppet-dist-client-0 puppet-dist-client-1 50: end-volume 51: 52: volume puppet-ssl-client-0 53: type protocol/client 54: option remote-host sandy.mdc 55: option remote-subvolume /opt/gluster-data/snake-puppet/ssl 56: option transport-type tcp 57: end-volume 58: 59: volume puppet-ssl-client-1 60: type protocol/client 61: option remote-host irene.mdc 62: option remote-subvolume /opt/gluster-data/puppet/ssl 63: option transport-type tcp 64: end-volume 65: 66: volume puppet-ssl-replicate-0 67: type cluster/replicate 68: option background-self-heal-count 0 69: option metadata-self-heal on 70: option data-self-heal on 71: option entry-self-heal on 72: option self-heal-daemon on 73: option metadata-change-log on 74: option data-self-heal-algorithm full 75: option iam-self-heal-daemon yes 76: subvolumes puppet-ssl-client-0 puppet-ssl-client-1 77: end-volume 78: 79: volume glustershd 80: type debug/io-stats 81: subvolumes puppet-ssl-replicate-0 puppet-dist-replicate-0 puppet-bucket-replicate-0 82: end-volume +------------------------------------------------------------------------------+ [2013-01-11 16:39:58.395877] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-ssl-client-1: changing port to 24010 (from 0) [2013-01-11 16:39:58.395978] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-bucket-client-0: changing port to 24014 (from 0) [2013-01-11 16:39:58.396048] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-dist-client-1: changing port to 24012 (from 0) [2013-01-11 16:39:58.396106] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-bucket-client-1: changing port to 24014 (from 0) [2013-01-11 16:39:58.396161] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-ssl-client-0: changing port to 24012 (from 0) [2013-01-11 16:39:58.396223] I [rpc-clnt.c:1657:rpc_clnt_reconfig] 1-puppet-dist-client-0: changing port to 24010 (from 0) ==> nfs.log.1 <== [2013-01-11 16:40:02.148931] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-ssl-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.149212] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-1: Connected to 10.136.200.16:24010, attached to remote volume '/opt/gluster-data/puppet/ssl'. [2013-01-11 16:40:02.149238] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.149289] I [afr-common.c:3628:afr_notify] 1-puppet-ssl-replicate-0: Subvolume 'puppet-ssl-client-1' came back up; going online. [2013-01-11 16:40:02.149382] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-1: Server lk version = 1 [2013-01-11 16:40:02.149711] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-dist-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.149931] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-1: Connected to 10.136.200.16:24012, attached to remote volume '/opt/gluster-data/puppet/dist'. [2013-01-11 16:40:02.149951] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.149995] I [afr-common.c:3628:afr_notify] 1-puppet-dist-replicate-0: Subvolume 'puppet-dist-client-1' came back up; going online. [2013-01-11 16:40:02.150086] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-1: Server lk version = 1 [2013-01-11 16:40:02.150727] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-bucket-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.151013] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-1: Connected to 10.136.200.16:24014, attached to remote volume '/opt/gluster-data/puppet/bucket'. [2013-01-11 16:40:02.151042] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.151091] I [afr-common.c:3628:afr_notify] 1-puppet-bucket-replicate-0: Subvolume 'puppet-bucket-client-1' came back up; going online. [2013-01-11 16:40:02.151187] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-bucket-client-1: Server lk version = 1 [2013-01-11 16:40:02.151623] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-ssl-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.151924] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-0: Connected to 10.136.200.27:24012, attached to remote volume '/opt/gluster-data/snake-puppet/ssl'. [2013-01-11 16:40:02.151950] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.152166] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-0: Server lk version = 1 [2013-01-11 16:40:02.152472] I [afr-common.c:1965:afr_set_root_inode_on_first_lookup] 1-puppet-ssl-replicate-0: added root inode [2013-01-11 16:40:02.152566] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-dist-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.152807] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-0: Connected to 10.136.200.27:24010, attached to remote volume '/opt/gluster-data/snake-puppet/dist'. [2013-01-11 16:40:02.152827] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.152991] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-0: Server lk version = 1 [2013-01-11 16:40:02.153187] I [afr-common.c:1965:afr_set_root_inode_on_first_lookup] 1-puppet-dist-replicate-0: added root inode [2013-01-11 16:40:02.153403] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-bucket-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.153644] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-0: Connected to 10.136.200.27:24014, attached to remote volume '/opt/gluster-data/snake-puppet/bucket'. [2013-01-11 16:40:02.153665] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.153797] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-bucket-client-0: Server lk version = 1 [2013-01-11 16:40:02.154054] I [afr-common.c:1965:afr_set_root_inode_on_first_lookup] 1-puppet-bucket-replicate-0: added root inode ==> glustershd.log.1 <== [2013-01-11 16:40:02.381825] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-ssl-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.382098] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-1: Connected to 10.136.200.16:24010, attached to remote volume '/opt/gluster-data/puppet/ssl'. [2013-01-11 16:40:02.382119] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.382203] I [afr-common.c:3628:afr_notify] 1-puppet-ssl-replicate-0: Subvolume 'puppet-ssl-client-1' came back up; going online. [2013-01-11 16:40:02.382321] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-1: Server lk version = 1 [2013-01-11 16:40:02.382889] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-bucket-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.383190] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-0: Connected to 10.136.200.27:24014, attached to remote volume '/opt/gluster-data/snake-puppet/bucket'. [2013-01-11 16:40:02.383213] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.383284] I [afr-common.c:3628:afr_notify] 1-puppet-bucket-replicate-0: Subvolume 'puppet-bucket-client-0' came back up; going online. [2013-01-11 16:40:02.384825] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-bucket-client-0: Server lk version = 1 [2013-01-11 16:40:02.384999] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-dist-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.385614] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-1: Connected to 10.136.200.16:24012, attached to remote volume '/opt/gluster-data/puppet/dist'. [2013-01-11 16:40:02.385646] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.385725] I [afr-common.c:3628:afr_notify] 1-puppet-dist-replicate-0: Subvolume 'puppet-dist-client-1' came back up; going online. [2013-01-11 16:40:02.386268] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-1: Server lk version = 1 [2013-01-11 16:40:02.386381] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-bucket-client-1: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.386710] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-ssl-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.386817] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-1: Connected to 10.136.200.16:24014, attached to remote volume '/opt/gluster-data/puppet/bucket'. [2013-01-11 16:40:02.386842] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-1: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.387051] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-0: Connected to 10.136.200.27:24012, attached to remote volume '/opt/gluster-data/snake-puppet/ssl'. [2013-01-11 16:40:02.387087] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.387222] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-bucket-client-1: Server lk version = 1 [2013-01-11 16:40:02.387345] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-0: Server lk version = 1 [2013-01-11 16:40:02.387427] I [client-handshake.c:1636:select_server_supported_programs] 1-puppet-dist-client-0: Using Program GlusterFS 3.3.1, Num (1298437), Version (330) [2013-01-11 16:40:02.388029] I [client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-0: Connected to 10.136.200.27:24010, attached to remote volume '/opt/gluster-data/snake-puppet/dist'. [2013-01-11 16:40:02.388058] I [client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-0: Server and Client lk-version numbers are not same, reopening the fds [2013-01-11 16:40:02.389682] I [client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-0: Server lk version = 1 ^C -- Yang Orange Key: 35745318S1 On Fri, Jan 11, 2013 at 11:00 AM, YANG ChengFu <youngseph at gmail.com> wrote: > Hello Fu Yong Tao, > > thanks for your suggest, after I did your steps, I got the following: > > gluster> volume sync new-host > please delete all the volumes before full sync > gluster> peer status > Number of Peers: 1 > > Hostname: 10.136.200.27 > Uuid: 184a81f4-ff0f-48d6-adb8-798b98957b1a > State: Accepted peer request (Connected) > > I still can not put the server in the truested pool ! > > -- > Yang > Orange Key: 35745318S1 > > > On Fri, Jan 11, 2013 at 5:22 AM, ??? <yongtaofu at gmail.com> wrote: > >> Reinstall gluster server or upgrade is a dangerous task before it it's >> better to backup /etc/glusterfs /var/lib/glusterd. >> >> /var/lib/glusterd/glusterd.info contains the uuid of current server >> and /var/lib/glusterd/peers contain it's peers >> make sure above two files are all correct >> >> If other servers status are fine then with only above configuration >> files you can start current host and gluster volumes files will >> automatically sync to current host. >> >> Always remember backup >> >> 2013/1/11, YANG ChengFu <youngseph at gmail.com>: >> > Hello, >> > >> > I did an upgrade glusterfs from 3.0.5 to 3.3.1, before I did it, I have >> > other two 3.3.1 hosts(new-host) ready and made a cluster. >> > >> > After I upgraded old hosts, I tried to add them to the cluster, I >> > got State: Peer Rejected (Connected), for sure it could be about same >> > volumes on the old hosts, but I have tried to stop glusterd, remove >> > everything from the old host, such /etc/glusterd, /etc/glusterfs >> > and /var/lib/glusterd/, and start glusterd, then I readded it to >> cluster, >> > the problem is still there. >> > >> > I also did 'volume sync', but I failed, because of the following error >> > message >> > >> > gluster> volume sync new-hosts >> > please delete all the volumes before full sync >> > >> > I can not do it, or I will lose all my data! >> > >> > The most funny thing I found, even if the peer status is rejected, but I >> > can mount the volume from the old host. >> > >> > Any ideas ! >> > >> > -- >> > Yang >> > Orange Key: 35745318S1 >> > >> >> >> -- >> ??? >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130111/d3e959ea/attachment-0001.html>