Re: Geo-Replication issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Krishna,

The logs indicate that it has received gluster-poc-sj::gluster as the slave volume where as the actual slave volume is gluster-poc-sj::glusterdist

what's you irc nick in #gluster?

On Thu, Sep 6, 2018 at 1:45 PM, Krishna Verma <kverma@xxxxxxxxxxx> wrote:

Hi Kotresh,

 

[root@gluster-poc-noida repvol]# tailf /var/log/glusterfs/glusterd.log

[2018-09-06 07:57:03.443256] W [MSGID: 106028] [glusterd-geo-rep.c:2568:glusterd_get_statefile_name] 0-management: Config file (/var/lib/glusterd/geo-replication/glusterdist_gluster-poc-sj_gluster/gsyncd.conf) missing. Looking for template config file (/var/lib/glusterd/geo-replication/gsyncd_template.conf) [No such file or directory]

[2018-09-06 07:57:03.443339] I [MSGID: 106294] [glusterd-geo-rep.c:2577:glusterd_get_statefile_name] 0-management: Using default config template(/var/lib/glusterd/geo-replication/gsyncd_template.conf).

[2018-09-06 07:57:03.512014] E [MSGID: 106028] [glusterd-geo-rep.c:3577:glusterd_op_stage_gsync_set] 0-management: Geo-replication session between glusterdist and gluster-poc-sj::gluster does not exist.. statefile = /var/lib/glusterd/geo-replication/glusterdist_gluster-poc-sj_gluster/monitor.status [No such file or directory]

[2018-09-06 07:57:03.512049] E [MSGID: 106322] [glusterd-geo-rep.c:3778:glusterd_op_stage_gsync_set] 0-management: Geo-replication session between glusterdist and gluster-poc-sj::gluster does not exist.

[2018-09-06 07:57:03.512063] E [MSGID: 106301] [glusterd-syncop.c:1352:gd_stage_op_phase] 0-management: Staging of operation 'Volume Geo-replication' failed on localhost : Geo-replication session between glusterdist and gluster-poc-sj::gluster does not exist.

[2018-09-06 07:57:24.869113] E [MSGID: 106316] [glusterd-geo-rep.c:2761:glusterd_verify_slave] 0-management: Not a valid slave

[2018-09-06 07:57:24.869289] E [MSGID: 106316] [glusterd-geo-rep.c:3152:glusterd_op_stage_gsync_create] 0-management: gluster-poc-sj::gluster is not a valid slave volume. Error: Unable to mount and fetch slave volume details. Please check the log: /var/log/glusterfs/geo-replication/gverify-slavemnt.log

[2018-09-06 07:57:24.869313] E [MSGID: 106301] [glusterd-syncop.c:1352:gd_stage_op_phase] 0-management: Staging of operation 'Volume Geo-replication Create' failed on localhost : Unable to mount and fetch slave volume details. Please check the log: /var/log/glusterfs/geo-replication/gverify-slavemnt.log

[2018-09-06 07:56:38.421045] I [MSGID: 106308] [glusterd-geo-rep.c:4881:glusterd_get_gsync_status_mst_slv] 0-management: geo-replication status glusterdist gluster-poc-sj::gluster : session is not active

[2018-09-06 07:56:38.486229] I [MSGID: 106028] [glusterd-geo-rep.c:4903:glusterd_get_gsync_status_mst_slv] 0-management: /var/lib/glusterd/geo-replication/glusterdist_gluster-poc-sj_gluster/monitor.status statefile not present. [No such file or directory]

 

/Krishna

 

From: Kotresh Hiremath Ravishankar <khiremat@xxxxxxxxxx>
Sent: Thursday, September 6, 2018 1:20 PM
To: Krishna Verma <kverma@xxxxxxxxxxx>
Cc: Gluster Users <gluster-users@xxxxxxxxxxx>
Subject: Re: Geo-Replication issue

 

EXTERNAL MAIL

Hi Krishna,

glusterd log file would help here

Thanks,

Kotresh HR

 

On Thu, Sep 6, 2018 at 1:02 PM, Krishna Verma <kverma@xxxxxxxxxxx> wrote:

Hi All,

 

I am getting issue in geo-replication distributed gluster volume. In a session status it shows only peer node instead of 2. And I am also not able to delete/start/stop or anything on this session.

 

geo-replication distributed gluster volume “glusterdist” status

[root@gluster-poc-noida ~]# gluster volume status glusterdist

Status of volume: glusterdist

Gluster process                             TCP Port  RDMA Port  Online  Pid

------------------------------------------------------------------------------

Brick gluster-poc-noida:/data/gluster-dist/

distvol                                     49154     0          Y       23138

Brick noi-poc-gluster:/data/gluster-dist/di

stvol                                       49154     0          Y       14637

 

Task Status of Volume glusterdist

------------------------------------------------------------------------------

There are no active volume tasks

 

Geo-replication session status

[root@gluster-poc-noida ~]# gluster volume geo-replication glusterdist gluster-poc-sj::glusterdist status

 

MASTER NODE        MASTER VOL     MASTER BRICK                  SLAVE USER    SLAVE                          SLAVE NODE    STATUS     CRAWL STATUS    LAST_SYNCED

----------------------------------------------------------------------------------------------------------------------------------------------------------------

noi-poc-gluster    glusterdist    /data/gluster-dist/distvol    root          gluster-poc-sj::glusterdist    N/A           Stopped    N/A             N/A

[root@gluster-poc-noida ~]#

 

Can’t stop/start/delete the session:

[root@gluster-poc-noida ~]# gluster volume geo-replication glusterdist gluster-poc-sj::glusterdist stop

Staging failed on localhost. Please check the log file for more details.

geo-replication command failed

[root@gluster-poc-noida ~]# gluster volume geo-replication glusterdist gluster-poc-sj::glusterdist stop force

pid-file entry mising in config file and template config file.

geo-replication command failed

[root@gluster-poc-noida ~]# gluster volume geo-replication glusterdist gluster-poc-sj::glusterdist delete

Staging failed on localhost. Please check the log file for more details.

geo-replication command failed

[root@gluster-poc-noida ~]# gluster volume geo-replication glusterdist gluster-poc-sj::glusterdist start

Staging failed on localhost. Please check the log file for more details.

geo-replication command failed

[root@gluster-poc-noida ~]#

 

gsyncd.log errors

[2018-09-06 06:17:21.757195] I [monitor(monitor):269:monitor] Monitor: worker died before establishing connection       brick=/data/gluster-dist/distvol

[2018-09-06 06:17:32.312093] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker   brick=/data/gluster-dist/distvol        slave_node=gluster-poc-sj

[2018-09-06 06:17:32.441817] I [monitor(monitor):261:monitor] Monitor: Changelog Agent died, Aborting Worker    brick=/data/gluster-dist/distvol

[2018-09-06 06:17:32.442193] I [monitor(monitor):279:monitor] Monitor: worker died in startup phase     brick=/data/gluster-dist/distvol

[2018-09-06 06:17:43.1177] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker     brick=/data/gluster-dist/distvol        slave_node=gluster-poc-sj

[2018-09-06 06:17:43.137794] I [monitor(monitor):261:monitor] Monitor: Changelog Agent died, Aborting Worker    brick=/data/gluster-dist/distvol

[2018-09-06 06:17:43.138214] I [monitor(monitor):279:monitor] Monitor: worker died in startup phase     brick=/data/gluster-dist/distvol

[2018-09-06 06:17:53.144072] I [monitor(monitor):158:monitor] Monitor: starting gsyncd worker   brick=/data/gluster-dist/distvol        slave_node=gluster-poc-sj

[2018-09-06 06:17:53.276853] I [monitor(monitor):261:monitor] Monitor: Changelog Agent died, Aborting Worker    brick=/data/gluster-dist/distvol

[2018-09-06 06:17:53.277327] I [monitor(monitor):279:monitor] Monitor: worker died in startup phase     brick=/data/gluster-dist/distvol

 

Could anyone please help?

 

/Krishna

 


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




--

Thanks and Regards,

Kotresh H R




--
Thanks and Regards,
Kotresh H R
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux