Re: Geo-Rep. 3.5.3, Missing Files, Incorrect "Files Pending"

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Status output has issue showing exact number of files in sync. Please check the numbers on disk and let us know if difference exists between Master and Secondary Volume.

--
regards
Aravinda

On 05/05/2015 06:58 AM, David Gibbons wrote:
I am having an issue with geo-replication. There were a number of
complications when I upgraded to 3.5.3, but geo-replication was (I
think) working at some point. The volume is accessed via samba using
vfs_glusterfs.

The main issue is that geo-replication has not been sending updated
copies of old files to the replicated server. So in the scenario where
file created -> time passes -> file is modified -> file is saved, the
new version is not replicated.

Is it possible that one brick is having a geo-rep issue and the others
are not? Consider this output:

MASTER NODE                     MASTER VOL    MASTER BRICK
         SLAVE                   STATUS     CHECKPOINT STATUS    CRAWL
STATUS    FILES SYNCD    FILES PENDING    BYTES PENDING    DELETES
PENDING    FILES SKIPPED
------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
gfs-a-1                         shares
/mnt/a-1-shares-brick-1/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2309456        0                0
                0                  0
gfs-a-1                         shares
/mnt/a-1-shares-brick-2/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2315557        0                0
                0                  0
gfs-a-1                         shares
/mnt/a-1-shares-brick-3/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2362884        0                0
                0                  0
gfs-a-1                         shares
/mnt/a-1-shares-brick-4/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2407600        0                0
                0                  0
gfs-a-2                         shares
/mnt/a-2-shares-brick-1/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2409430        0                0
                0                  0
gfs-a-2                         shares
/mnt/a-2-shares-brick-2/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2308969        0                0
                0                  0
gfs-a-2                         shares
/mnt/a-2-shares-brick-3/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2079576        8191             0
                0                  0
gfs-a-2                         shares
/mnt/a-2-shares-brick-4/brick    gfs-a-bkp::bkpshares    Active
N/A                  Hybrid Crawl    2340597        0                0
                0                  0
gfs-a-3                         shares
/mnt/a-3-shares-brick-1/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-3                         shares
/mnt/a-3-shares-brick-2/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-3                         shares
/mnt/a-3-shares-brick-3/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-3                         shares
/mnt/a-3-shares-brick-4/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-4                         shares
/mnt/a-4-shares-brick-1/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-4                         shares
/mnt/a-4-shares-brick-2/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-4                         shares
/mnt/a-4-shares-brick-3/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0
gfs-a-4                         shares
/mnt/a-4-shares-brick-4/brick    gfs-a-bkp::bkpshares    Passive
N/A                  N/A             0              0                0
                0                  0

This seems to show that there are 8191 files_pending on just one
brick, and the others are up to date. I am suspicious of the 8191
number because it's looks like we're at a bucket-size boundary on the
backend. I've tried stopping and re-starting the rep session. I've
also tried changing the change_detector from xsync to changelog.
Neither seems to have had an effect.

It seems like geo-replication is quite wonky in 3.5.x. Is there light
at the end of the tunnel, or should I find another solution to
replicate?

Cheers,
Dave
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux