Re: Geo replication stuck (rsync: link_stat "(unreachable)")

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Answers inline.

Thanks and Regards,
Kotresh H R

----- Original Message -----
> From: "mabi" <mabi@xxxxxxxxxxxxx>
> To: "Kotresh Hiremath Ravishankar" <khiremat@xxxxxxxxxx>
> Cc: "Gluster Users" <gluster-users@xxxxxxxxxxx>
> Sent: Thursday, April 13, 2017 8:51:29 PM
> Subject: Re:  Geo replication stuck (rsync: link_stat "(unreachable)")
> 
> Hi Kotresh,
> 
> Thanks for your feedback.
> 
> So do you mean I can simply login into the geo-replication slave node, mount
> the volume with fuse, and delete the problematic directory, and finally
> restart geo-replcation?
> 
   Trying to delete the problematic directory on slave might still result with
  the same ENOTEMPTY error. Try that out, if it does not work out, it needs to
  be deleted from backend bricks directly from all the nodes.

> I am planning to migrate to 3.8 as soon as I have a backup (geo-replication).
> Is this issue with DHT fixed in the latest 3.8.x release?
>
   Most of the issues are addressed.

> Regards,
> M.
> 
> -------- Original Message --------
> Subject: Re:  Geo replication stuck (rsync: link_stat
> "(unreachable)")
> Local Time: April 13, 2017 7:57 AM
> UTC Time: April 13, 2017 5:57 AM
> From: khiremat@xxxxxxxxxx
> To: mabi <mabi@xxxxxxxxxxxxx>
> Gluster Users <gluster-users@xxxxxxxxxxx>
> 
> Hi,
> 
> I think the directory Workhours_2017 is deleted on master and on
> slave it's failing to delete because there might be stale linkto files
> at the back end. These issues are fixed in DHT with latest versions.
> Upgrading to latest version would solve these issues.
> 
> To workaround the issue, you might need to cleanup the problematic
> directory on slave from the backend.
> 
> Thanks and Regards,
> Kotresh H R
> 
> ----- Original Message -----
> > From: "mabi" <mabi@xxxxxxxxxxxxx>
> > To: "Kotresh Hiremath Ravishankar" <khiremat@xxxxxxxxxx>
> > Cc: "Gluster Users" <gluster-users@xxxxxxxxxxx>
> > Sent: Thursday, April 13, 2017 12:28:50 AM
> > Subject: Re:  Geo replication stuck (rsync: link_stat
> > "(unreachable)")
> >
> > Hi Kotresh,
> >
> > Thanks for your hint, adding the "--ignore-missing-args" option to rsync
> > and
> > restarting geo-replication worked but it only managed to sync approximately
> > 1/3 of the data until it put the geo replication in status "Failed" this
> > time. Now I have a different type of error as you can see below from the
> > log
> > extract on my geo replication slave node:
> >
> > [2017-04-12 18:01:55.268923] I [MSGID: 109066]
> > [dht-rename.c:1574:dht_rename]
> > 0-myvol-private-geo-dht: renaming
> > /.gfid/1678ff37-f708-4197-bed0-3ecd87ae1314/Workhours_2017
> > empty.xls.ocTransferId2118183895.part
> > (hash=myvol-private-geo-client-0/cache=myvol-private-geo-client-0) =>
> > /.gfid/1678ff37-f708-4197-bed0-3ecd87ae1314/Workhours_2017 empty.xls
> > (hash=myvol-private-geo-client-0/cache=myvol-private-geo-client-0)
> > [2017-04-12 18:01:55.269842] W [fuse-bridge.c:1787:fuse_rename_cbk]
> > 0-glusterfs-fuse: 4786:
> > /.gfid/1678ff37-f708-4197-bed0-3ecd87ae1314/Workhours_2017
> > empty.xls.ocTransferId2118183895.part ->
> > /.gfid/1678ff37-f708-4197-bed0-3ecd87ae1314/Workhours_2017 empty.xls => -1
> > (Directory not empty)
> > [2017-04-12 18:01:55.314062] I [fuse-bridge.c:5016:fuse_thread_proc]
> > 0-fuse:
> > unmounting /tmp/gsyncd-aux-mount-PNSR8s
> > [2017-04-12 18:01:55.314311] W [glusterfsd.c:1251:cleanup_and_exit]
> > (-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x8064) [0x7f97d3129064]
> > -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x7f97d438a725]
> > -->/usr/sbin/glusterfs(cleanup_and_exit+0x57) [0x7f97d438a5a7] ) 0-:
> > received signum (15), shutting down
> > [2017-04-12 18:01:55.314335] I [fuse-bridge.c:5720:fini] 0-fuse: Unmounting
> > '/tmp/gsyncd-aux-mount-PNSR8s'.
> >
> > How can I fix now this issue and have geo-replication continue
> > synchronising
> > again?
> >
> > Best regards,
> > M.
> >
> > -------- Original Message --------
> > Subject: Re:  Geo replication stuck (rsync: link_stat
> > "(unreachable)")
> > Local Time: April 11, 2017 9:18 AM
> > UTC Time: April 11, 2017 7:18 AM
> > From: khiremat@xxxxxxxxxx
> > To: mabi <mabi@xxxxxxxxxxxxx>
> > Gluster Users <gluster-users@xxxxxxxxxxx>
> >
> > Hi,
> >
> > Then please use set the following rsync config and let us know if it helps.
> >
> > gluster vol geo-rep <mastervol> <slavehost>::<slavevol> config
> > rsync-options
> > "--ignore-missing-args"
> >
> > Thanks and Regards,
> > Kotresh H R
> >
> > ----- Original Message -----
> > > From: "mabi" <mabi@xxxxxxxxxxxxx>
> > > To: "Kotresh Hiremath Ravishankar" <khiremat@xxxxxxxxxx>
> > > Cc: "Gluster Users" <gluster-users@xxxxxxxxxxx>
> > > Sent: Tuesday, April 11, 2017 2:15:54 AM
> > > Subject: Re:  Geo replication stuck (rsync: link_stat
> > > "(unreachable)")
> > >
> > > Hi Kotresh,
> > >
> > > I am using the official Debian 8 (jessie) package which has rsync version
> > > 3.1.1.
> > >
> > > Regards,
> > > M.
> > >
> > > -------- Original Message --------
> > > Subject: Re:  Geo replication stuck (rsync: link_stat
> > > "(unreachable)")
> > > Local Time: April 10, 2017 6:33 AM
> > > UTC Time: April 10, 2017 4:33 AM
> > > From: khiremat@xxxxxxxxxx
> > > To: mabi <mabi@xxxxxxxxxxxxx>
> > > Gluster Users <gluster-users@xxxxxxxxxxx>
> > >
> > > Hi Mabi,
> > >
> > > What's the rsync version being used?
> > >
> > > Thanks and Regards,
> > > Kotresh H R
> > >
> > > ----- Original Message -----
> > > > From: "mabi" <mabi@xxxxxxxxxxxxx>
> > > > To: "Gluster Users" <gluster-users@xxxxxxxxxxx>
> > > > Sent: Saturday, April 8, 2017 4:20:25 PM
> > > > Subject:  Geo replication stuck (rsync: link_stat
> > > > "(unreachable)")
> > > >
> > > > Hello,
> > > >
> > > > I am using distributed geo replication with two of my GlusterFS 3.7.20
> > > > replicated volumes and just noticed that the geo replication for one
> > > > volume
> > > > is not working anymore. It is stuck since the 2017-02-23 22:39 and I
> > > > tried
> > > > to stop and restart geo replication but still it stays stuck at that
> > > > specific date and time under the DATA field of the geo replication
> > > > "status
> > > > detail" command I can see 3879 and that it has "Active" as STATUS but
> > > > still
> > > > nothing happens. I noticed that the rsync process is running but does
> > > > not
> > > > do
> > > > anything, then I did a strace on the PID of rsync and saw the
> > > > following:
> > > >
> > > > write(2, "rsync: link_stat \"(unreachable)/"..., 114
> > > >
> > > > It looks like rsync can't read or find a file and stays stuck on that.
> > > > In
> > > > the
> > > > geo-replication log files of GlusterFS master I can't find any error
> > > > messages just informational message. For example when I restart the geo
> > > > replication I see the following log entries:
> > > >
> > > > [2017-04-07 21:43:05.664541] I [monitor(monitor):443:distribute] <top>:
> > > > slave
> > > > bricks: [{'host': 'gfs1geo.domain', 'dir': '/data/private-geo/brick'}]
> > > > [2017-04-07 21:43:05.666435] I [monitor(monitor):468:distribute] <top>:
> > > > worker specs: [('/data/private/brick', 'ssh:// root@gfs1geo.domain
> > > > :gluster://localhost:private-geo', '1', False)]
> > > > [2017-04-07 21:43:05.823931] I [monitor(monitor):267:monitor] Monitor:
> > > > ------------------------------------------------------------
> > > > [2017-04-07 21:43:05.824204] I [monitor(monitor):268:monitor] Monitor:
> > > > starting gsyncd worker
> > > > [2017-04-07 21:43:05.930124] I [gsyncd(/data/private/brick):733:main_i]
> > > > <top>: syncing: gluster://localhost:private -> ssh://
> > > > root@gfs1geo.domain
> > > > :gluster://localhost:private-geo
> > > > [2017-04-07 21:43:05.931169] I [changelogagent(agent):73:__init__]
> > > > ChangelogAgent: Agent listining...
> > > > [2017-04-07 21:43:08.558648] I
> > > > [master(/data/private/brick):83:gmaster_builder] <top>: setting up
> > > > xsync
> > > > change detection mode
> > > > [2017-04-07 21:43:08.559071] I
> > > > [master(/data/private/brick):367:__init__]
> > > > _GMaster: using 'rsync' as the sync engine
> > > > [2017-04-07 21:43:08.560163] I
> > > > [master(/data/private/brick):83:gmaster_builder] <top>: setting up
> > > > changelog
> > > > change detection mode
> > > > [2017-04-07 21:43:08.560431] I
> > > > [master(/data/private/brick):367:__init__]
> > > > _GMaster: using 'rsync' as the sync engine
> > > > [2017-04-07 21:43:08.561105] I
> > > > [master(/data/private/brick):83:gmaster_builder] <top>: setting up
> > > > changeloghistory change detection mode
> > > > [2017-04-07 21:43:08.561391] I
> > > > [master(/data/private/brick):367:__init__]
> > > > _GMaster: using 'rsync' as the sync engine
> > > > [2017-04-07 21:43:11.354417] I
> > > > [master(/data/private/brick):1249:register]
> > > > _GMaster: xsync temp directory:
> > > > /var/lib/misc/glusterfsd/private/ssh%3A%2F%2Froot%40192.168.20.107%3Agluster%3A%2F%2F127.0.0.1%3Aprivate-geo/616931ac8f39da5dc5834f9d47fc7b1a/xsync
> > > > [2017-04-07 21:43:11.354751] I
> > > > [resource(/data/private/brick):1528:service_loop] GLUSTER: Register
> > > > time:
> > > > 1491601391
> > > > [2017-04-07 21:43:11.357630] I
> > > > [master(/data/private/brick):510:crawlwrap]
> > > > _GMaster: primary master with volume id
> > > > e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
> > > > ...
> > > > [2017-04-07 21:43:11.489355] I
> > > > [master(/data/private/brick):519:crawlwrap]
> > > > _GMaster: crawl interval: 1 seconds
> > > > [2017-04-07 21:43:11.516710] I [master(/data/private/brick):1163:crawl]
> > > > _GMaster: starting history crawl... turns: 1, stime: (1487885974, 0),
> > > > etime:
> > > > 1491601391
> > > > [2017-04-07 21:43:12.607836] I [master(/data/private/brick):1192:crawl]
> > > > _GMaster: slave's time: (1487885974, 0)
> > > >
> > > > Does anyone know how I can find out the root cause of this problem and
> > > > make
> > > > geo replication work again from the time point it got stuck?
> > > >
> > > > Many thanks in advance for your help.
> > > >
> > > > Best regards,
> > > > Mabi
> > > >
> > > >
> > > >
> > > >
> > > > _______________________________________________
> > > > Gluster-users mailing list
> > > > Gluster-users@xxxxxxxxxxx
> > > > http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux