found a solution:
it is a bug in rsync which fixed in version 3.1.1
when compression is on, and heavy changes on big files happens, it breaks.
https://bugs.launchpad.net/ubuntu/+source/rsync/+bug/1384503
fix for ubuntu 14.04 LTS:
install rsync from vivid repro
http://packages.ubuntu.com/vivid/amd64/rsync/download
chris.
On 2015-11-23 11:25, Chris wrote:> hello!
>
> Ubuntu 14.04.3 LTS
> 3.19.0-30-generic #34~14.04.1-Ubuntu SMP
> glusterfs 3.6.6 built on Sep 23 2015 13:31:06
>
>
> a simple geo-replication ..
>
> passwordless ssh setup work fine >obiwanvpn to soco
>
> hostA (obiwanvpn)
> gluster volume create rep transport tcp obiwanvpn:/home/gluster/geo
> gluster volume geo-replication rep soco::rep create push-pem force
> gluster volume geo-replication rep soco::rep start
>
> hostB (soco)
> gluster volume create rep transport tcp soco:/home/gluster/geo force
> gluster volume start rep
>
> work without error.
>
>
>
> KVM host on obiwanvpn:
> setup a glusterfs pool with virt-manager
> create vm, and start within the glusterfs pool
> geo-replication works fine due setup vm, but after a while, status
> switch to faulty.
>
> the only entry with an error:
>
> [2015-11-23 10:51:21.661911] E [resource(/home/gluster/geo):221:errlog]
> Popen: command "rsync -avR0 --inplace --files-from=- --super --stats
> --numeric-ids --no-implied-dirs . -e ssh -oPasswordAuthentication=no
> -oStrictHostKeyChecking=no -i
> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S
> /tmp/gsyncd-aux-ssh-2zxXdt/c2dd32adfb68384a8f45d8809a9a6539.sock
> --compress root at soco:/proc/22557/cwd" returned with 12, saying:
> [2015-11-23 10:51:21.662156] E [resource(/home/gluster/geo):225:logerr]
> Popen: rsync> rsync: [sender] write error: Broken pipe (32)
> [2015-11-23 10:51:21.662218] E [resource(/home/gluster/geo):225:logerr]
> Popen: rsync> rsync error: error in rsync protocol data stream (code 12)
> at io.c(837) [sender=3.1.0]
> [2015-11-23 10:51:21.662393] I
> [syncdutils(/home/gluster/geo):214:finalize] <top>: exiting.
>
> and thats it!. sync is gone forever! :(
>
>
> gluster volume info /status (hostA obiwanvpn)
>
> Volume Name: rep
> Type: Distribute
> Volume ID: 96b7f2ab-8c27-48f9-8071-627c1f5bc7fa
> Status: Started
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: obiwanvpn:/home/gluster/geo
> Options Reconfigured:
> nfs.register-with-portmap: off
> nfs.disable: on
> server.allow-insecure: on
> nfs.rpc-auth-allow: on
> changelog.changelog: on
> geo-replication.ignore-pid-check: on
> geo-replication.indexing: on
>
>
> Status of volume: rep
> Gluster process Port Online Pid
>
------------------------------------------------------------------------------
>
> Brick obiwanvpn:/home/gluster/geo 49153 Y 4303
>
> Task Status of Volume rep
>
------------------------------------------------------------------------------
>
> There are no active volume tasks
>
>
>
>
>
> gluster volume info / status (hostB soco)
>
> Volume Name: rep
> Type: Distribute
> Volume ID: 5d95e882-5f90-4372-8807-dfb9746ce517
> Status: Started
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: soco:/data/gluster/geo
> Options Reconfigured:
> nfs.disable: on
>
>
> Status of volume: rep
> Gluster process Port Online Pid
>
------------------------------------------------------------------------------
>
> Brick soco:/data/gluster/geo 49154 Y 7474
>
> Task Status of Volume rep
>
------------------------------------------------------------------------------
>
>
>
>
>
> MASTER NODE MASTER VOL MASTER BRICK SLAVE STATUS
> CHECKPOINT STATUS CRAWL STATUS
>
-----------------------------------------------------------------------------------------------------------
>
> obiwan rep /home/gluster/geo soco::rep faulty
> N/A N/A
>
>
>
>
> any idea?
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users