cannot resolve split-brain via 'select one brick as source'

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi All,

I'm currently playing around with Gluster 8.3 and geo-replication.
I have created a 'replica 3' volume as source and a 'replica 2' as a destination.
In this case geo-replication is working quite fine , but during my tests I have managed to cause a split-brain in some of the files:

[root@glustere mnt]# gluster volume info secondary
 
Volume Name: secondary
Type: Distributed-Replicate
Volume ID: 1b5717ee-aa9b-4eff-9989-ad4f0388b86c
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: glusterd:/bricks/brick-d1/brick
Brick2: glustere:/bricks/brick-e2/brick
Brick3: glustere:/bricks/brick-e1/brick
Brick4: glusterd:/bricks/brick-d2/brick
Options Reconfigured:
cluster.quorum-count: 1
cluster.quorum-type: fixed
features.read-only: on
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
performance.quick-read: off
performance.client-io-threads: off
cluster.enable-shared-storage: enable

[root@glustere mnt]# gluster volume heal secondary info summary
Brick glusterd:/bricks/brick-d1/brick
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0


Brick glustere:/bricks/brick-e2/brick
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0


Brick glustere:/bricks/brick-e1/brick
Status: Connected
Total Number of entries: 2641
Number of entries in heal pending: 0
Number of entries in split-brain: 2641
Number of entries possibly healing: 0


Brick glusterd:/bricks/brick-d2/brick
Status: Connected
Total Number of entries: 2641
Number of entries in heal pending: 0
Number of entries in split-brain: 2641
Number of entries possibly healing: 0

As node "glustere" was last reboot , the source of truth should be 'glusterd:/bricks/brick-d2/brick'.

I have tried to tell gluster that , but it doesn't want to resolve the split-brain:

[root@glustere mnt]# gluster volume heal secondary split-brain source-brick glusterd:/bricks/brick-d2/brick | tail -n 5
Lookup failed on gfid:dbfb3794-4a1a-4540-b235-2fdce4d21d6a:Transport endpoint is not connected.
Lookup failed on gfid:d78fdf9c-1c00-4712-8229-cfc10b009ad3:Transport endpoint is not connected.
Status: Connected
Number of healed entries: 1

In the logs it's clearly that something is wrong:

[2020-12-27 20:14:07.100074] W [MSGID: 108027] [afr-common.c:2857:afr_attempt_readsubvol_set] 0-secondary-replicate-1: no read subvols for dbfb3794-4a1a-4540-b235-2fdce4d21d6a 
[2020-12-27 20:14:07.100106] I [MSGID: 109063] [dht-layout.c:641:dht_layout_normalize] 0-secondary-dht: Found anomalies [{path=dbfb3794-4a1a-4540-b235-2fdce4d21d6a}, {gfid=dbfb3794-4a1a-4540-b235-2fdce4d21d6a}, {holes=1}, {overlaps=0}] 
[2020-12-27 20:14:07.100903] W [MSGID: 108027] [afr-common.c:2857:afr_attempt_readsubvol_set] 0-secondary-replicate-1: no read subvols for d78fdf9c-1c00-4712-8229-cfc10b009ad3 
[2020-12-27 20:14:07.100935] I [MSGID: 109063] [dht-layout.c:641:dht_layout_normalize] 0-secondary-dht: Found anomalies [{path=d78fdf9c-1c00-4712-8229-cfc10b009ad3}, {gfid=d78fdf9c-1c00-4712-8229-cfc10b009ad3}, {holes=1}, {overlaps=0}]


Yet, if I specify the file in the previous command the heal is OK:

[root@glustere mnt]# gluster volume heal secondary split-brain source-brick glusterd:/bricks/brick-d2/brick gfid:dbfb3794-4a1a-4540-b235-2fdce4d21d6a
Healed gfid:dbfb3794-4a1a-4540-b235-2fdce4d21d6a.
[root@glustere mnt]# gluster volume heal secondary split-brain source-brick glusterd:/bricks/brick-d2/brick gfid:d78fdf9c-1c00-4712-8229-cfc10b009ad3
Healed gfid:d78fdf9c-1c00-4712-8229-cfc10b009ad3.

[2020-12-27 20:16:27.471113] I [MSGID: 108026] [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] 0-secondary-replicate-1: performing metadata selfheal on dbfb3794-4a1a-4540-b235-2fdce4d21d6a 
[2020-12-27 20:16:27.473157] I [MSGID: 108026] [afr-self-heal-common.c:1744:afr_log_selfheal] 0-secondary-replicate-1: Completed metadata selfheal on dbfb3794-4a1a-4540-b235-2fdce4d21d6a. sources=[1]  sinks=0  
[2020-12-27 20:16:38.303151] I [MSGID: 108026] [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] 0-secondary-replicate-1: performing metadata selfheal on d78fdf9c-1c00-4712-8229-cfc10b009ad3 
[2020-12-27 20:16:38.305144] I [MSGID: 108026] [afr-self-heal-common.c:1744:afr_log_selfheal] 0-secondary-replicate-1: Completed metadata selfheal on d78fdf9c-1c00-4712-8229-cfc10b009ad3. sources=[1]  sinks=0 



I thought that 'source-brick' solves both data and metadata split-brains. Am I wrong ?

Best Regards,
Strahil Nikolov
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux