So I should do a compare out-of-band from Gluster and see what is actually in-sync vs out of sync? Is there any easy way just to start it over? I am assuming removing and re-adding geo-rep is the easiest way. Is that correct? Thanks, Dave On Mon, May 4, 2015 at 10:09 PM, Aravinda <avishwan@xxxxxxxxxx> wrote: > Status output has issue showing exact number of files in sync. Please check > the numbers on disk and let us know if difference exists between Master and > Secondary Volume. > > -- > regards > Aravinda > > On 05/05/2015 06:58 AM, David Gibbons wrote: >> >> I am having an issue with geo-replication. There were a number of >> complications when I upgraded to 3.5.3, but geo-replication was (I >> think) working at some point. The volume is accessed via samba using >> vfs_glusterfs. >> >> The main issue is that geo-replication has not been sending updated >> copies of old files to the replicated server. So in the scenario where >> file created -> time passes -> file is modified -> file is saved, the >> new version is not replicated. >> >> Is it possible that one brick is having a geo-rep issue and the others >> are not? Consider this output: >> >> MASTER NODE MASTER VOL MASTER BRICK >> SLAVE STATUS CHECKPOINT STATUS CRAWL >> STATUS FILES SYNCD FILES PENDING BYTES PENDING DELETES >> PENDING FILES SKIPPED >> >> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ >> gfs-a-1 shares >> /mnt/a-1-shares-brick-1/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2309456 0 0 >> 0 0 >> gfs-a-1 shares >> /mnt/a-1-shares-brick-2/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2315557 0 0 >> 0 0 >> gfs-a-1 shares >> /mnt/a-1-shares-brick-3/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2362884 0 0 >> 0 0 >> gfs-a-1 shares >> /mnt/a-1-shares-brick-4/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2407600 0 0 >> 0 0 >> gfs-a-2 shares >> /mnt/a-2-shares-brick-1/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2409430 0 0 >> 0 0 >> gfs-a-2 shares >> /mnt/a-2-shares-brick-2/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2308969 0 0 >> 0 0 >> gfs-a-2 shares >> /mnt/a-2-shares-brick-3/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2079576 8191 0 >> 0 0 >> gfs-a-2 shares >> /mnt/a-2-shares-brick-4/brick gfs-a-bkp::bkpshares Active >> N/A Hybrid Crawl 2340597 0 0 >> 0 0 >> gfs-a-3 shares >> /mnt/a-3-shares-brick-1/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-3 shares >> /mnt/a-3-shares-brick-2/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-3 shares >> /mnt/a-3-shares-brick-3/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-3 shares >> /mnt/a-3-shares-brick-4/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-4 shares >> /mnt/a-4-shares-brick-1/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-4 shares >> /mnt/a-4-shares-brick-2/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-4 shares >> /mnt/a-4-shares-brick-3/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> gfs-a-4 shares >> /mnt/a-4-shares-brick-4/brick gfs-a-bkp::bkpshares Passive >> N/A N/A 0 0 0 >> 0 0 >> >> This seems to show that there are 8191 files_pending on just one >> brick, and the others are up to date. I am suspicious of the 8191 >> number because it's looks like we're at a bucket-size boundary on the >> backend. I've tried stopping and re-starting the rep session. I've >> also tried changing the change_detector from xsync to changelog. >> Neither seems to have had an effect. >> >> It seems like geo-replication is quite wonky in 3.5.x. Is there light >> at the end of the tunnel, or should I find another solution to >> replicate? >> >> Cheers, >> Dave >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@xxxxxxxxxxx >> http://www.gluster.org/mailman/listinfo/gluster-users > > _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users