Hi - Do you have passwordless ssh login to slave machine? After setting passwordless login ,please try this - #gluster volume geo-replication athena root@$(hostname):/soft/venus start or #gluster volume geo-replication athena $(hostname):/soft/venus start wait for few seconds then verify the status. For minimum requirement ,checkout this http://www.gluster.com/community/documentation/index.php/Gluster_3.2:_Checking_Geo-replication_Minimum_Requirements HTH -- ---- Cheers, Lakshmipathi.G FOSS Programmer. ----- Original Message ----- From: "anthony garnier" <sokar6012 at hotmail.com> To: gluster-users at gluster.org Sent: Monday, May 16, 2011 5:06:22 PM Subject: G?o-rep fail Hi, I'm currently trying to use g?o-rep on the local data-node into a directory but it fails with status "faulty" Volume : Volume Name: athena Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: ylal3020:/users/exp1 Brick2: yval1010:/users/exp3 Brick3: ylal3030:/users/exp2 Brick4: yval1000:/users/exp4 Options Reconfigured: geo-replication.indexing: on diagnostics.count-fop-hits: on diagnostics.latency-measurement: on performance.cache-max-file-size: 256MB network.ping-timeout: 5 performance.cache-size: 512MB performance.cache-refresh-timeout: 60 nfs.port: 2049 I've done this cmd : # gluster volume geo-replication athena /soft/venus config # gluster volume geo-replication athena /soft/venus start # gluster volume geo-replication athena /soft/venus status MASTER SLAVE STATUS -------------------------------------------------------------------------------- athena /soft/venus faulty Here is the log file in Debug mod : [2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor: ------------------------------------------------------------ [2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: starting gsyncd worker [2011-05-16 13:28:55.326309] I [gsyncd:287:main_i] <top>: syncing: gluster://localhost:athena -> file:///soft/venus [2011-05-16 13:28:55.327905] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__() ... [2011-05-16 13:28:55.462613] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__ -> 1.0 [2011-05-16 13:28:55.462886] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.46 version() ... [2011-05-16 13:28:55.463330] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.46 version -> 1.0 [2011-05-16 13:28:55.480202] D [resource:381:connect] GLUSTER: auxiliary glusterfs mount in place [2011-05-16 13:28:55.682863] D [resource:393:connect] GLUSTER: auxiliary glusterfs mount prepared [2011-05-16 13:28:55.684926] D [monitor(monitor):57:monitor] Monitor: worker got connected in 0 sec, waiting 59 more to make sure it's fine [2011-05-16 13:28:55.685096] D [repce:131:push] RepceClient: call 10888:1115703616:1305545335.68 keep_alive(None,) ... [2011-05-16 13:28:55.685859] D [repce:141:__call__] RepceClient: call 10888:1115703616:1305545335.68 keep_alive -> 1 [2011-05-16 13:28:59.546574] D [master:167:volinfo_state_machine] <top>: (None, None) << (None, 28521f8f) -> (None, 28521f8f) [2011-05-16 13:28:59.546863] I [master:184:crawl] GMaster: new master is 28521f8f-49d3-4e2a-b984-f664f44f5289 [2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary master with volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ... [2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering . [2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545339.55 xtime('.', '28521f8f-49d3-4e2a-b984-f664f44f5289') ... [2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] <top>: FAIL: Traceback (most recent call last): File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py", line 152, in twrap tf(*aa) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 118, in listen rid, exc, res = recv(self.inf) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 42, in recv return pickle.load(inf) EOFError Does anyone already got those errors ? _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users