Geo-replication error=12 (rsync)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I have configured geo-replication with a non-privileged user, I used
this documentation:
https://docs.gluster.org/en/latest/Administrator%20Guide/Geo%20Replication/

My setup:
(Master)
2-Node Gluster replicated volume with Debian 9 and gluster 5.1 (from
gluster.org)

(Slave)
2-Node Gluster replicated volume for geo-replication Debian 9 and
gluster 5.1 (from gluster.org)

rsync and ssh are from the distribution on all hosts

As soon as I start geo-replication I got the following message:
--------------
root@master-01:~# gluster v geo-replication gv0 geouser@slave-01::geo start
Starting geo-replication session between gv0 & geouser@slave-01::geo has
been successful
--------------

But then, if i check the status I see:
--------------
root@master-01:~# gluster v geo-replication gv0 geouser@slave-01::geo status

MASTER NODE    MASTER VOL    MASTER BRICK      SLAVE USER    SLAVE
              SLAVE NODE    STATUS     CRAWL STATUS       LAST_SYNCED

----------------------------------------------------------------------------------------------------------------------------------------------------
master-01      gv0           /gluster/brick    geouser
geouser@slave-01::geo    slave-02      Active     Changelog Crawl
2018-12-12 14:08:16
master-02      gv0           /gluster/brick    geouser
geouser@slave-01::geo    slave-02      Passive    N/A                N/A



root@master-01:~# gluster v geo-replication gv0 geouser@slave-01::geo status

MASTER NODE    MASTER VOL    MASTER BRICK      SLAVE USER    SLAVE
              SLAVE NODE    STATUS     CRAWL STATUS    LAST_SYNCED
-----------------------------------------------------------------------------------------------------------------------------------------
master-01      gv0           /gluster/brick    geouser
geouser@slave-01::geo    N/A           Faulty     N/A             N/A

master-02      gv0           /gluster/brick    geouser
geouser@slave-01::geo    slave-02      Passive    N/A             N/A
--------------
And so on and so on ....
The status always switches between Active --> Faulty --> Initialiting...
--> Activ .......

I took a look in the log-file on the master and found this message:
-----------------
[2018-12-12 13:28:52.187111] I
[gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker
Status Change status=Initializing...
[2018-12-12 13:28:52.187377] I [monitor(monitor):157:monitor] Monitor:
starting gsyncd worker   slave_node=slave-02     brick=/gluster/brick
[2018-12-12 13:28:52.271593] I [gsyncd(agent /gluster/brick):308:main]
<top>: Using session config file
path=/var/lib/glusterd/geo-replication/gv0_slave-01_geo/gsyncd.conf
[2018-12-12 13:28:52.275005] I [gsyncd(worker /gluster/brick):308:main]
<top>: Using session config file
path=/var/lib/glusterd/geo-replication/gv0_slave-01_geo/gsyncd.conf
[2018-12-12 13:28:52.281762] I [changelogagent(agent
/gluster/brick):72:__init__] ChangelogAgent: Agent listining...
[2018-12-12 13:28:52.282340] I [resource(worker
/gluster/brick):1401:connect_remote] SSH: Initializing SSH connection
between master and slave...
[2018-12-12 13:28:53.621534] I [resource(worker
/gluster/brick):1448:connect_remote] SSH: SSH connection between master
and slave established.  duration=1.3390
[2018-12-12 13:28:53.621803] I [resource(worker
/gluster/brick):1120:connect] GLUSTER: Mounting gluster volume locally...
[2018-12-12 13:28:54.637206] I [resource(worker
/gluster/brick):1143:connect] GLUSTER: Mounted gluster volume
duration=1.0152
[2018-12-12 13:28:54.637604] I [subcmds(worker
/gluster/brick):80:subcmd_worker] <top>: Worker spawn successful.
Acknowledging back to monitor
[2018-12-12 13:28:56.646581] I [master(worker
/gluster/brick):1603:register] _GMaster: Working dir
path=/var/lib/misc/gluster/gsyncd/gv0_slave-01_geo/gluster-brick
[2018-12-12 13:28:56.647150] I [resource(worker
/gluster/brick):1306:service_loop] GLUSTER: Register time
time=1544621336
[2018-12-12 13:28:56.663329] I [gsyncdstatus(worker
/gluster/brick):281:set_active] GeorepStatus: Worker Status Change
status=Active
[2018-12-12 13:28:56.714960] I [gsyncdstatus(worker
/gluster/brick):253:set_worker_crawl_status] GeorepStatus: Crawl Status
Change      status=History Crawl
[2018-12-12 13:28:56.715329] I [master(worker
/gluster/brick):1517:crawl] _GMaster: starting history crawl
turns=1 entry_stime=None        stime=(1544620096, 0)   etime=1544621336
[2018-12-12 13:28:57.717641] I [master(worker
/gluster/brick):1546:crawl] _GMaster: slave's time
stime=(1544620096, 0)
[2018-12-12 13:28:58.193830] I [master(worker
/gluster/brick):1954:syncjob] Syncer: Sync Time Taken     job=2
duration=0.0457 num_files=1     return_code=12
[2018-12-12 13:28:58.194150] E [syncdutils(worker
/gluster/brick):809:errlog] Popen: command returned error     cmd=rsync
-aR0 --inplace --files-from=- --super --stats --numeric-ids
--no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e
ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto
-S /tmp/gsyncd-aux-ssh-bte8zsl6/ecd4cc061ecc1734ea1c9d1a5332f983.sock
geouser@slave-01:/proc/6186/cwd   error=12
[2018-12-12 13:28:58.197854] I [repce(agent
/gluster/brick):97:service_loop] RepceServer: terminating on reaching EOF.
[2018-12-12 13:28:58.644061] I [monitor(monitor):278:monitor] Monitor:
worker died in startup phase     brick=/gluster/brick
[2018-12-12 13:28:58.713629] I
[gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker
Status Change status=Faulty
-----------------
This massage is repeating

Here the error-message I think is the problem:
-----------------
[2018-12-12 13:28:58.194150] E [syncdutils(worker
/gluster/brick):809:errlog] Popen: command returned error     cmd=rsync
-aR0 --inplace --files-from=- --super --stats --numeric-ids
--no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e
ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto
-S /tmp/gsyncd-aux-ssh-bte8zsl6/ecd4cc061ecc1734ea1c9d1a5332f983.sock
geouser@slave-01:/proc/6186/cwd   error=12
-----------------

On the slave:
-----------------
root@slave-01:~# tail -f /var/log/auth.log

Dec 12 14:42:08 slave-01 sshd[8869]: Did not receive identification
string from 192.168.80.81 port 60496
Dec 12 14:42:08 slave-01 sshd[8870]: Accepted publickey for geouser from
192.168.80.81 port 60498 ssh2: RSA
SHA256:tnCjr2rrPp+FaYLnxKSaCSxbQ0YJtEDtSWNi5m2jDQ8
Dec 12 14:42:08 slave-01 sshd[8870]: pam_unix(sshd:session): session
opened for user geouser by (uid=0)
Dec 12 14:42:09 slave-01 systemd-logind[437]: New session 174 of user
geouser.
Dec 12 14:42:14 slave-01 sshd[8876]: Received disconnect from
192.168.80.81 port 60498:11: disconnected by user
Dec 12 14:42:14 slave-01 sshd[8876]: Disconnected from 192.168.80.81
port 60498
Dec 12 14:42:14 slave-01 sshd[8870]: pam_unix(sshd:session): session
closed for user geouser
Dec 12 14:42:14 slave-01 systemd-logind[437]: Removed session 174.
-------------------

and

-----------------
root@slave-01:~# tail -f
/var/log/glusterfs/geo-replication-slaves/gv0_slave-01_geo/gsyncd.log

[2018-12-12 13:43:15.576101] W [gsyncd(slave
master-01/gluster/brick):304:main] <top>: Session config file not
exists, using the default config
path=/var/lib/glusterd/geo-replication/gv0_slave-01_geo/gsyncd.conf
[2018-12-12 13:43:15.596193] I [resource(slave
master-01/gluster/brick):1120:connect] GLUSTER: Mounting gluster volume
locally...
[2018-12-12 13:43:16.706221] I [resource(slave
master-01/gluster/brick):1143:connect] GLUSTER: Mounted gluster volume
 duration=1.1099
[2018-12-12 13:43:16.706489] I [resource(slave
master-01/gluster/brick):1170:service_loop] GLUSTER: slave listening
[2018-12-12 13:43:21.278398] I [repce(slave
master-01/gluster/brick):97:service_loop] RepceServer: terminating on
reaching EOF.
-------------------
This message is repeating

I took a look and the file
/var/lib/glusterd/geo-replication/gv0_slave-01_geo/gsyncd.conf is not
existing. How do I get the file?

I already restarted all 4 Nodes (master and Slave) but it's still the same.

any help would be fine

Stefan

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux