Re: Geo-replication setup problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So I ran the following on palace and madonna and it appears to be syncing properly now

root@palace:~# sed -i -e 's:/usr/libexec/glusterfs/gsyncd:/usr/lib/x86_64-linux-gnu/glusterfs/gsyncd:' .ssh/authorized_keys

root@james:~/georepsetup# gluster volume geo-replication status

MASTER NODE    MASTER VOL    MASTER BRICK                    SLAVE USER    SLAVE                   SLAVE NODE    STATUS     CRAWL STATUS    LAST_SYNCED
------------------------------------------------------------------------------------------------------------------------------------------------------
james          static        /data/gluster1/static/brick1    root          ssh://palace::static    palace        Active     Hybrid Crawl    N/A
cupid          static        /data/gluster1/static/brick2    root          ssh://palace::static    madonna       Passive    N/A             N/A
hilton         static        /data/gluster1/static/brick3    root          ssh://palace::static    palace        Passive    N/A             N/A

present        static        /data/gluster1/static/brick4    root          ssh://palace::static    madonna       Passive    N/A             N/A


I am still a little confused by the SLAVE column though - should "palace" in that column really be an A record such as "gluster-remote" that has 2 addresses (for both palace and madonna)?

On 7/10/2015 8:10 am, Wade Fitzpatrick wrote:





Thanks for the response, it looks like the problem is



[2015-10-06 19:08:49.547874] E
[resource(/data/gluster1/static/brick1):222:errlog] Popen: command
"ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S
/tmp/gsyncd-aux-ssh-RX3NRr/eb9c0906d6265193f403278fc1489309.sock
root@palace /nonexistent/gsyncd --session-owner
3f9f810d-a988-4914-a5ca-5bd7b251a273 -N --listen --timeout 120
gluster://localhost:static" returned with 127, saying:

[2015-10-06 19:08:49.547976] E
[resource(/data/gluster1/static/brick1):226:logerr] Popen: ssh>
This system is monitored and logged. Any unauthorized usage will be
prosecuted.

[2015-10-06 19:08:49.548062] E
[resource(/data/gluster1/static/brick1):226:logerr] Popen: ssh>
bash: /usr/libexec/glusterfs/gsyncd: No such file or directory




All servers are Ubuntu 15.04 running glusterfs 3.7.4 built on Sep 
1 2015 12:08:58 but /usr/libexec/glusterfs does not exist so is
this a packaging error?




On 6/10/2015 6:11 pm, Aravinda wrote:




Please share the tracebacks/errors from the logs(Master nodes)
/var/log/glusterfs/geo-replication/static/*.log

regards
Aravinda

On 10/06/2015 01:19 PM, Wade
Fitzpatrick wrote:




I am trying to set up geo-replication of a striped-replicate
volume. I used https://github.com/aravindavk/georepsetup
to configure the replication.



root@james:~# gluster volume info



Volume Name: static

Type: Striped-Replicate

Volume ID: 3f9f810d-a988-4914-a5ca-5bd7b251a273

Status: Started

Number of Bricks: 1 x 2 x 2 = 4

Transport-type: tcp

Bricks:

Brick1: james:/data/gluster1/static/brick1

Brick2: cupid:/data/gluster1/static/brick2

Brick3: hilton:/data/gluster1/static/brick3

Brick4: present:/data/gluster1/static/brick4

Options Reconfigured:

changelog.changelog: on

geo-replication.ignore-pid-check: on

geo-replication.indexing: on

performance.readdir-ahead: on



root@james:~# gluster volume geo-replication status



MASTER NODE    MASTER VOL    MASTER BRICK                   
SLAVE USER    SLAVE                   SLAVE NODE    STATUS    
CRAWL STATUS    LAST_SYNCED

------------------------------------------------------------------------------------------------------------------------------------------------------

james          static        /data/gluster1/static/brick1   
root          ssh://palace::static   

N/A           Created    N/A             N/A

cupid          static        /data/gluster1/static/brick2   
root          ssh://palace::static   

N/A           Created    N/A             N/A

hilton         static        /data/gluster1/static/brick3   
root          ssh://palace::static   

N/A           Created    N/A             N/A

present        static        /data/gluster1/static/brick4   
root          ssh://palace::static   

N/A           Created    N/A             N/A



So of the 4 bricks, data is striped over brick1 and brick3, also
brick1=brick2 is a mirror and brick3=brick4 is a mirror.
Therefore I have no need to geo-replicate bricks 2 and 4.



At the other site, palace and madonna form a stripe volume (no
replication):



root@palace:~# gluster volume info



Volume Name: static

Type: Stripe

Volume ID: 0e91c6f2-3499-4fc4-9630-9da8b7f57db5

Status: Started

Number of Bricks: 1 x 2 = 2

Transport-type: tcp

Bricks:

Brick1: palace:/data/gluster1/static/brick1

Brick2: madonna:/data/gluster1/static/brick2

Options Reconfigured:

performance.readdir-ahead: on



However, when I try to start geo-replication, it fails as below.



root@james:~# gluster volume geo-replication static ssh://palace::static start

Starting geo-replication session between static & ssh://palace::static has been
successful

root@james:~# gluster volume geo-replication status



MASTER NODE    MASTER VOL    MASTER BRICK                   
SLAVE USER    SLAVE                   SLAVE NODE    STATUS   
CRAWL STATUS    LAST_SYNCED

-----------------------------------------------------------------------------------------------------------------------------------------------------

james          static        /data/gluster1/static/brick1   
root          ssh://palace::static   

N/A           Faulty    N/A             N/A

cupid          static        /data/gluster1/static/brick2   
root          ssh://palace::static   

N/A           Faulty    N/A             N/A

hilton         static        /data/gluster1/static/brick3   
root          ssh://palace::static   

N/A           Faulty    N/A             N/A

present        static        /data/gluster1/static/brick4   
root          ssh://palace::static   

N/A           Faulty    N/A             N/A





What should I do to set this up properly so that
james:/data/gluster1/static/brick1 gets replicated to
palace:/data/gluster1/static/brick1 ; and
hilton:/data/gluster1/static/brick3 gets replicated to
madonna:/data/gluster1/static/brick2 ???







--







_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users














--

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux