Hi Kotresh,
I have been running 4.1.3 from the end of August.
Since then data has been synced to geo side with a couple of hundred GB per 24 hour, even with the errors I have reported in this thread.
Four days ago all data transfer to geo side stopped, and the logs repeats the same error over and over again (see below).
Both nodes toggle status Active/Faulty.
Thanks alot!
Best regards
Marcus
One master node, gsyncd.log:
[2018-09-10 10:53:38.409709] I [gsyncdstatus(monitor):244:set_worker_status] GeorepStatus: Worker Status Change status=Faulty
[2018-09-10 10:53:47.783914] I [gsyncd(config-get):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:47.852792] I [gsyncd(status):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:48.421061] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker brick=/urd-gds/gluster slave_node=urd-gds-geo-000
[2018-09-10 10:53:48.462655] I [gsyncd(agent /urd-gds/gluster):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:48.463366] I [changelogagent(agent /urd-gds/gluster):72:__init__] ChangelogAgent: Agent listining...
[2018-09-10 10:53:48.465905] I [gsyncd(worker /urd-gds/gluster):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:48.474558] I [resource(worker /urd-gds/gluster):1377:connect_remote] SSH: Initializing SSH connection between master and slave...
[2018-09-10 10:53:50.70219] I [resource(worker /urd-gds/gluster):1424:connect_remote] SSH: SSH connection between master and slave established. duration=1.5954
[2018-09-10 10:53:50.70777] I [resource(worker /urd-gds/gluster):1096:connect] GLUSTER: Mounting gluster volume locally...
[2018-09-10 10:53:51.170597] I [resource(worker /urd-gds/gluster):1119:connect] GLUSTER: Mounted gluster volume duration=1.0994
[2018-09-10 10:53:51.171158] I [subcmds(worker /urd-gds/gluster):70:subcmd_worker] <top>: Worker spawn successful. Acknowledging back to monitor
[2018-09-10 10:53:51.696057] I [gsyncd(config-get):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:51.764605] I [gsyncd(status):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 10:53:53.210553] I [master(worker /urd-gds/gluster):1593:register] _GMaster: Working dir path=/var/lib/misc/gluster/gsyncd/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/urd-gds-gluster
[2018-09-10 10:53:53.211148] I [resource(worker /urd-gds/gluster):1282:service_loop] GLUSTER: Register time time=1536576833
[2018-09-10 10:53:53.230945] I [gsyncdstatus(worker /urd-gds/gluster):277:set_active] GeorepStatus: Worker Status Change status=Active
[2018-09-10 10:53:53.233444] I [gsyncdstatus(worker /urd-gds/gluster):249:set_worker_crawl_status] GeorepStatus: Crawl Status Change status=History Crawl
[2018-09-10 10:53:53.233632] I [master(worker /urd-gds/gluster):1507:crawl] _GMaster: starting history crawl turns=1 stime=(1524272046, 0) entry_stime=(1524271940, 0) etime=1536576833
[2018-09-10 10:53:53.234951] I [master(worker /urd-gds/gluster):1536:crawl] _GMaster: slave's time stime=(1524272046, 0)
[2018-09-10 10:53:53.762105] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.0856 num_files=1 job=1 return_code=0
[2018-09-10 10:53:54.437858] I [master(worker /urd-gds/gluster):1374:process] _GMaster: Entry Time Taken MKD=0 MKN=0 LIN=0 SYM=0 REN=0 RMD=0 CRE=0 duration=0.0000 UNL=0
[2018-09-10 10:53:54.437973] I [master(worker /urd-gds/gluster):1384:process] _GMaster: Data/Metadata Time Taken SETA=0 SETX=0 meta_duration=0.0000 data_duration=1.1979 DATA="" XATT=0
[2018-09-10 10:53:54.438153] I [master(worker /urd-gds/gluster):1394:process] _GMaster: Batch Completed changelog_end=1524272047 entry_stime=(1524271940, 0) changelog_start=1524272047 stime=(152\
4272046, 0) duration=1.2029 num_changelogs=1 mode=history_changelog
[2018-09-10 10:53:54.482408] I [master(worker /urd-gds/gluster):1536:crawl] _GMaster: slave's time stime=(1524272046, 0)
[2018-09-10 10:53:54.583467] E [repce(worker /urd-gds/gluster):197:__call__] RepceClient: call failed call=1844:139973681694528:1536576834.54 method=entry_ops error=GsyncdError
[2018-09-10 10:53:54.583585] E [syncdutils(worker /urd-gds/gluster):300:log_raise_exception] <top>: execution of "gluster" failed with ENOENT (No such file or directory)
[2018-09-10 10:53:54.600353] I [repce(agent /urd-gds/gluster):80:service_loop] RepceServer: terminating on reaching EOF.
[2018-09-10 10:53:55.175978] I [monitor(monitor):279:monitor] Monitor: worker died in startup phase brick=/urd-gds/gluster
[2018-09-10 10:53:55.182988] I [gsyncdstatus(monitor):244:set_worker_status] GeorepStatus: Worker Status Change status=Faulty
[2018-09-10 10:53:56.24414] I [gsyncd(config-get):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
Other master node, gsyncd.log:
[2018-09-10 11:10:43.10458] I [gsyncdstatus(monitor):244:set_worker_status] GeorepStatus: Worker Status Change status=Faulty
[2018-09-10 11:10:53.28702] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker brick=/urd-gds/gluster slave_node=urd-gds-geo-000
[2018-09-10 11:10:53.69638] I [gsyncd(agent /urd-gds/gluster):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 11:10:53.70264] I [changelogagent(agent /urd-gds/gluster):72:__init__] ChangelogAgent: Agent listining...
[2018-09-10 11:10:53.71902] I [gsyncd(worker /urd-gds/gluster):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
[2018-09-10 11:10:53.80737] I [resource(worker /urd-gds/gluster):1377:connect_remote] SSH: Initializing SSH connection between master and slave...
[2018-09-10 11:10:54.621948] I [resource(worker /urd-gds/gluster):1424:connect_remote] SSH: SSH connection between master and slave established. duration=1.5410
[2018-09-10 11:10:54.622504] I [resource(worker /urd-gds/gluster):1096:connect] GLUSTER: Mounting gluster volume locally...
[2018-09-10 11:10:55.721349] I [resource(worker /urd-gds/gluster):1119:connect] GLUSTER: Mounted gluster volume duration=1.0984
[2018-09-10 11:10:55.721913] I [subcmds(worker /urd-gds/gluster):70:subcmd_worker] <top>: Worker spawn successful. Acknowledging back to monitor
[2018-09-10 11:10:58.543606] I [master(worker /urd-gds/gluster):1593:register] _GMaster: Working dir path=/var/lib/misc/gluster/gsyncd/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/urd-gds-gluster
[2018-09-10 11:10:58.545701] I [resource(worker /urd-gds/gluster):1282:service_loop] GLUSTER: Register time time=1536577858
[2018-09-10 11:10:58.564208] I [gsyncdstatus(worker /urd-gds/gluster):277:set_active] GeorepStatus: Worker Status Change status=Active
[2018-09-10 11:10:58.565689] I [gsyncdstatus(worker /urd-gds/gluster):249:set_worker_crawl_status] GeorepStatus: Crawl Status Change status=History Crawl
[2018-09-10 11:10:58.565876] I [master(worker /urd-gds/gluster):1507:crawl] _GMaster: starting history crawl turns=1 stime=(1527128725, 0) entry_stime=(1527128815, 0) etime=1536577858
[2018-09-10 11:10:59.593652] I [master(worker /urd-gds/gluster):1536:crawl] _GMaster: slave's time stime=(1527128725, 0)
[2018-09-10 11:11:01.755116] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.5233 num_files=103 job=1 return_code=0
[2018-09-10 11:11:02.897665] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.6648 num_files=116 job=2 return_code=0
[2018-09-10 11:11:03.98150] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.2003 num_files=59 job=2 return_code=23
[2018-09-10 11:11:03.219059] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.1207 num_files=16 job=2 return_code=0
[2018-09-10 11:11:03.841105] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.1212 num_files=32 job=2 return_code=23
[2018-09-10 11:11:04.951658] I [master(worker /urd-gds/gluster):1944:syncjob] Syncer: Sync Time Taken duration=0.2160 num_files=24 job=3 return_code=0
[2018-09-10 11:11:05.2938] E [repce(worker /urd-gds/gluster):197:__call__] RepceClient: call failed call=2935:140696531339072:1536577864.67 method=entry_ops error=GsyncdError
[2018-09-10 11:11:05.3125] E [syncdutils(worker /urd-gds/gluster):300:log_raise_exception] <top>: execution of "gluster" failed with ENOENT (No such file or directory)
[2018-09-10 11:11:05.17061] I [repce(agent /urd-gds/gluster):80:service_loop] RepceServer: terminating on reaching EOF.
[2018-09-10 11:11:05.733716] I [monitor(monitor):279:monitor] Monitor: worker died in startup phase brick=/urd-gds/gluster
[2018-09-10 11:11:05.768186] I [gsyncdstatus(monitor):244:set_worker_status] GeorepStatus: Worker Status Change status=Faulty
[2018-09-10 11:11:15.788830] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker brick=/urd-gds/gluster slave_node=urd-gds-geo-000
[2018-09-10 11:11:15.829871] I [gsyncd(agent /urd-gds/gluster):297:main] <top>: Using session config file path=/var/lib/glusterd/geo-replication/urd-gds-volume_urd-gds-geo-001_urd-gds-volume/gsyncd.conf
Från: Kotresh Hiremath Ravishankar <khiremat@xxxxxxxxxx>
Skickat: den 3 september 2018 07:58
Till: Marcus Pedersén
Kopia: gluster-users@xxxxxxxxxxx
Ämne: Re: Was: Upgrade to 4.1.2 geo-replication does not work Now: Upgraded to 4.1.3 geo node Faulty
Hi Marcus,
Geo-rep had few important fixes in 4.1.3. Is it possible to upgrade and check whether the issue is still seen?
Thanks,
Kotresh HR
---
När du skickar e-post till SLU så innebär detta att SLU behandlar dina personuppgifter. För att läsa mer om hur detta går till, klicka
här
E-mailing SLU will result in SLU processing your personal data. For more information on how this is done, click
here
|