Ok,
I was able to get geo-replication configured by changing /usr/local/libexec/glusterfs/gverify.sh to use ssh to access the local machine, instead of accessing bash -c directly. I then found that the hook script was missing for geo-replication, so I copied that over manually. I now have what appears to be a "configured" geo-rep setup:
# gluster volume geo-replication shares gfs-a-bkp::bkpshares statusMASTER NODE MASTER VOL MASTER BRICK SLAVE STATUS CHECKPOINT STATUS CRAWL STATUS--------------------------------------------------------------------------------------------------------------------------------------------------------gfs-a-3 shares /mnt/a-3-shares-brick-1/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-3 shares /mnt/a-3-shares-brick-2/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-3 shares /mnt/a-3-shares-brick-3/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-3 shares /mnt/a-3-shares-brick-4/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-2 shares /mnt/a-2-shares-brick-1/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-2 shares /mnt/a-2-shares-brick-2/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-2 shares /mnt/a-2-shares-brick-3/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-2 shares /mnt/a-2-shares-brick-4/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-4 shares /mnt/a-4-shares-brick-1/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-4 shares /mnt/a-4-shares-brick-2/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-4 shares /mnt/a-4-shares-brick-3/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-4 shares /mnt/a-4-shares-brick-4/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-1 shares /mnt/a-1-shares-brick-1/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-1 shares /mnt/a-1-shares-brick-2/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-1 shares /mnt/a-1-shares-brick-3/brick gfs-a-bkp::bkpshares Not Started N/A N/Agfs-a-1 shares /mnt/a-1-shares-brick-4/brick gfs-a-bkp::bkpshares Not Started N/A N/A
So that's a step in the right direction (and I can upload a patch for gverify to a bugzilla). However, gverify *should* have worked with bash-c, and I was not able to figure out why it didn't work, other than it didn't seem able to find some programs. I'm thinking that maybe the PATH variable is wrong for Gluster, and that's why gverify didn't work out of the box.
When I attempt to start geo-rep now, I get the following in the geo-rep log:
[2014-12-07 10:52:40.893594] E [syncdutils(monitor):218:log_raise_exception] <top>: execution of "gluster" failed with ENOENT (No such file or directory)
[2014-12-07 10:52:40.893886] I [syncdutils(monitor):192:finalize] <top>: exiting.
Which seems to agree that maybe gluster isn't running with the same path variable that my console session is running with. Is this possible? I know I'm grasping :).
Any nudge in the right direction would be very much appreciated!
Cheers,
Dave
Dave
On Sat, Dec 6, 2014 at 10:06 AM, David Gibbons <david.c.gibbons@xxxxxxxxx> wrote:
Good Morning,I am having some trouble getting geo-replication started on a 3.5.3 volume.I have verified that password-less SSH is functional in both directions from the backup gluster server, and all nodes in the production gluster. I have verified that all nodes in production and backup cluster are running the same version of gluster, and that name resolution works in both directions.When I attempt to start geo-replication with this command:gluster volume geo-replication shares gfs-a-bkp::bkpshares create push-pemI end up with the following in the logs:[2014-12-06 15:02:50.284426] E [glusterd-geo-rep.c:1889:glusterd_verify_slave] 0-: Not a valid slave[2014-12-06 15:02:50.284495] E [glusterd-geo-rep.c:2106:glusterd_op_stage_gsync_create] 0-: gfs-a-bkp::bkpshares is not a valid slave volume. Error: Unable to fetch master volume details. Please check the master cluster and master volume.[2014-12-06 15:02:50.284509] E [glusterd-syncop.c:912:gd_stage_op_phase] 0-management: Staging of operation 'Volume Geo-replication Create' failed on localhost : Unable to fetch master volume details. Please check the master cluster and master volume.Would someone be so kind as to point me in the right direction?Cheers,Dave
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users