found a solution: it is a bug in rsync which fixed in version 3.1.1 when compression is on, and heavy changes on big files happens, it breaks. https://bugs.launchpad.net/ubuntu/+source/rsync/+bug/1384503 fix for ubuntu 14.04 LTS: install rsync from vivid repro http://packages.ubuntu.com/vivid/amd64/rsync/download chris. On 2015-11-23 11:25, Chris wrote:
hello! Ubuntu 14.04.3 LTS 3.19.0-30-generic #34~14.04.1-Ubuntu SMP glusterfs 3.6.6 built on Sep 23 2015 13:31:06 a simple geo-replication .. passwordless ssh setup work fine >obiwanvpn to soco hostA (obiwanvpn) gluster volume create rep transport tcp obiwanvpn:/home/gluster/geo gluster volume geo-replication rep soco::rep create push-pem force gluster volume geo-replication rep soco::rep start hostB (soco) gluster volume create rep transport tcp soco:/home/gluster/geo force gluster volume start rep work without error. KVM host on obiwanvpn: setup a glusterfs pool with virt-manager create vm, and start within the glusterfs pool geo-replication works fine due setup vm, but after a while, status switch to faulty. the only entry with an error: [2015-11-23 10:51:21.661911] E [resource(/home/gluster/geo):221:errlog] Popen: command "rsync -avR0 --inplace --files-from=- --super --stats --numeric-ids --no-implied-dirs . -e ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S /tmp/gsyncd-aux-ssh-2zxXdt/c2dd32adfb68384a8f45d8809a9a6539.sock --compress root@soco:/proc/22557/cwd" returned with 12, saying: [2015-11-23 10:51:21.662156] E [resource(/home/gluster/geo):225:logerr] Popen: rsync> rsync: [sender] write error: Broken pipe (32) [2015-11-23 10:51:21.662218] E [resource(/home/gluster/geo):225:logerr] Popen: rsync> rsync error: error in rsync protocol data stream (code 12) at io.c(837) [sender=3.1.0] [2015-11-23 10:51:21.662393] I [syncdutils(/home/gluster/geo):214:finalize] <top>: exiting. and thats it!. sync is gone forever! :( gluster volume info /status (hostA obiwanvpn) Volume Name: rep Type: Distribute Volume ID: 96b7f2ab-8c27-48f9-8071-627c1f5bc7fa Status: Started Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: obiwanvpn:/home/gluster/geo Options Reconfigured: nfs.register-with-portmap: off nfs.disable: on server.allow-insecure: on nfs.rpc-auth-allow: on changelog.changelog: on geo-replication.ignore-pid-check: on geo-replication.indexing: on Status of volume: rep Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick obiwanvpn:/home/gluster/geo 49153 Y 4303 Task Status of Volume rep ------------------------------------------------------------------------------ There are no active volume tasks gluster volume info / status (hostB soco) Volume Name: rep Type: Distribute Volume ID: 5d95e882-5f90-4372-8807-dfb9746ce517 Status: Started Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: soco:/data/gluster/geo Options Reconfigured: nfs.disable: on Status of volume: rep Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick soco:/data/gluster/geo 49154 Y 7474 Task Status of Volume rep ------------------------------------------------------------------------------ MASTER NODE MASTER VOL MASTER BRICK SLAVE STATUS CHECKPOINT STATUS CRAWL STATUS ----------------------------------------------------------------------------------------------------------- obiwan rep /home/gluster/geo soco::rep faulty N/A N/A any idea? _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users