Hi,
A client has a glusterfs cluster that's behaving weirdly after
some issues during upgrade.
They upgraded a glusterfs 2+1 cluster (replica with arbiter) from
3.10.9 to 3.12.4 on Centos and now have weird issues and some
files maybe being corrupted. They also switched from nfs ganesha
that crashed every couple of days to glusterfs subdirectory
mounting. Subdirectory mounting was the point of the upgrade.
On a client when I go switch to a user and go to the home folder,
I do a ls -lart and get a list of files, when I do it the second
time I get "ls: cannot open directory .: Permission denied" until
I logout and login again. Then it works mostly one time only,
somtimes it fails on the first time. Deploying something in this
folder works sometimes, other times I get write permission is
denied on /home/user.
I've enabled the bitrot daemon but it's still running:
Volume name : home
State of scrub: Active (In Progress)
Scrub impact: normal
Scrub frequency: biweekly
Bitrot error log location: /var/log/glusterfs/bitd.log
Scrubber error log location: /var/log/glusterfs/scrub.log
=========================================================
Node: localhost
Number of Scrubbed files: 152451
Number of Skipped files: 197
Last completed scrub time: Scrubber pending to complete.
Duration of last scrub (D:M:H:M:S): 0:0:0:0
Error count: 0
=========================================================
Node: gluster01
Number of Scrubbed files: 150198
Number of Skipped files: 190
Last completed scrub time: Scrubber pending to complete.
Duration of last scrub (D:M:H:M:S): 0:0:0:0
Error count: 0
=========================================================
Node: gluster02
Number of Scrubbed files: 0
Number of Skipped files: 153939
Last completed scrub time: Scrubber pending to complete.
Duration of last scrub (D:M:H:M:S): 0:0:0:0
Error count: 0
=========================================================
Gluster volume heal has one failed entries.
Ending time of crawl: Thu Jan 18 11:49:04 2018
Type of crawl: INDEX
No. of entries healed: 0
No. of entries in split-brain: 0
No. of heal failed entries: 0
Starting time of crawl: Thu Jan 18 11:59:04 2018
Ending time of crawl: Thu Jan 18 11:59:06 2018
Type of crawl: INDEX
No. of entries healed: 0
No. of entries in split-brain: 0
No. of heal failed entries: 1
Starting time of crawl: Thu Jan 18 12:09:06 2018
Ending time of crawl: Thu Jan 18 12:09:07 2018
Type of crawl: INDEX
No. of entries healed: 0
No. of entries in split-brain: 0
No. of heal failed entries: 0
but I'm unable to trace which file as "gluster volume heal
<volume> info heal-failed" doesn't exist anymore
gluster volume heal home info heal-failed
Usage:
volume heal <VOLNAME> [enable | disable | full |statistics
[heal-count [replica <HOSTNAME:BRICKNAME>]] |info
[split-brain] |split-brain {bigger-file <FILE> |
latest-mtime <FILE> |source-brick
<HOSTNAME:BRICKNAME> [<FILE>]} |granular-entry-heal
{enable | disable}]
The things I'm seeing in the logfiles on the client:
[2018-01-18 08:59:53.360864] W [MSGID: 114031]
[client-rpc-fops.c:2151:client3_3_seek_cbk] 0-home-client-0:
remote operation failed [No such device or address]
[2018-01-18 09:00:17.512636] W [MSGID: 114031]
[client-rpc-fops.c:2151:client3_3_seek_cbk] 0-home-client-0:
remote operation failed [No such device or address]
[2018-01-18 09:00:27.473702] W [MSGID: 114031]
[client-rpc-fops.c:2151:client3_3_seek_cbk] 0-home-client-0:
remote operation failed [No such device or address]
[2018-01-18 09:00:40.372756] W [MSGID: 114031]
[client-rpc-fops.c:2151:client3_3_seek_cbk] 0-home-client-0:
remote operation failed [No such device or address]
[2018-01-18 09:00:50.344597] W [MSGID: 114031]
[client-rpc-fops.c:2151:client3_3_seek_cbk] 0-home-client-0:
remote operation failed [No such device or address]
Thes one worry me, there are multiple but without any gfid:
[MSGID: 109063] [dht-layout.c:716:dht_layout_normalize]
5-home-dht: Found anomalies in (null) (gfid =
00000000-0000-0000-0000-000000000000). Holes=1 overlaps=0
Also it seems to disconnect and
[2018-01-18 08:38:41.210848] I [MSGID: 114057]
[client-handshake.c:1478:select_server_supported_programs]
0-home-client-1: Using Program GlusterFS 3.3, Num (1298437),
Version (330)
[2018-01-18 08:38:41.214548] I [MSGID: 114057]
[client-handshake.c:1478:select_server_supported_programs]
0-home-client-2: Using Program GlusterFS 3.3, Num (1298437),
Version (330)
[2018-01-18 08:38:41.255458] I [MSGID: 114046]
[client-handshake.c:1231:client_setvolume_cbk] 0-home-client-0:
Connected to home-client-0, attached to remote volume
'/data/home/brick1'.
[2018-01-18 08:38:41.255505] I [MSGID: 114047]
[client-handshake.c:1242:client_setvolume_cbk] 0-home-client-0:
Server and Client lk-version numbers are not same, reopening the
fds
[2018-01-18 08:38:41.255643] I [MSGID: 108005]
[afr-common.c:4929:__afr_handle_child_up_event] 0-home-replicate-0:
Subvolume 'home-client-0' came back up; going online.
[2018-01-18 08:38:41.255791] I [MSGID: 114035]
[client-handshake.c:202:client_set_lk_version_cbk]
0-home-client-0: Server lk version = 1
[2018-01-18 08:38:41.285774] I [MSGID: 114046]
[client-handshake.c:1231:client_setvolume_cbk] 0-home-client-1:
Connected to home-client-1, attached to remote volume
'/data/home/brick1'.
[2018-01-18 08:38:41.285815] I [MSGID: 114047]
[client-handshake.c:1242:client_setvolume_cbk] 0-home-client-1:
Server and Client lk-version numbers are not same, reopening the
fds
[2018-01-18 08:38:41.286672] I [MSGID: 114035]
[client-handshake.c:202:client_set_lk_version_cbk]
0-home-client-1: Server lk version = 1
[2018-01-18 08:38:41.291928] I [MSGID: 114046]
[client-handshake.c:1231:client_setvolume_cbk] 0-home-client-2:
Connected to home-client-2, attached to remote volume
'/data/home/brick1'.
[2018-01-18 08:38:41.291970] I [MSGID: 114047]
[client-handshake.c:1242:client_setvolume_cbk] 0-home-client-2:
Server and Client lk-version numbers are not same, reopening the
fds
[2018-01-18 08:38:41.292750] I [MSGID: 114035]
[client-handshake.c:202:client_set_lk_version_cbk]
0-home-client-2: Server lk version = 1
[2018-01-18 08:38:41.764822] I [MSGID: 108031]
[afr-common.c:2376:afr_local_discovery_cbk] 0-home-replicate-0:
selecting local read_child home-client-0
[2018-01-18 08:38:39.039054] I
[addr.c:55:compare_addr_and_update] 0-/data/home/brick1: allowed =
"*", received addr = "*.*.*.*"
[2018-01-18 08:38:39.040813] I [login.c:76:gf_auth] 0-auth/login:
allowed user names: 2d225def-3f34-472a-b8e4-c183acafc151
[2018-01-18 08:38:39.040853] I [MSGID: 115029]
[server-handshake.c:793:server_setvolume] 0-home-server: accepted
client from
gluster00.cluster.local-21570-2018/01/18-08:38:38:940331-home-client-0-0-0
(version: 3.12.4)
[2018-01-18 08:38:39.412234] I [MSGID: 115036]
[server.c:527:server_rpc_notify] 0-home-server: disconnecting
connection from
gluster00.cluster.local-21570-2018/01/18-08:38:38:940331-home-client-0-0-0
[2018-01-18 08:38:39.435180] I [MSGID: 101055]
[client_t.c:443:gf_client_unref] 0-home-server: Shutting down
connection
storage00.railscluster.local-21570-2018/01/18-08:38:38:940331-home-client-0-0-0
[2018-01-18 08:38:41.203430] I [addr.c:55:compare_addr_and_update]
0-/data/home/brick1: allowed = "*", received addr = "*.*.*.*"
[2018-01-18 08:38:41.203464] I [login.c:76:gf_auth] 0-auth/login:
allowed user names: 2d225def-3f34-472a-b8e4-c183acafc151
[2018-01-18 08:38:41.203502] I [MSGID: 115029]
[server-handshake.c:793:server_setvolume] 0-home-server: accepted
client from
gluster00.cluster.locall-21590-2018/01/18-08:38:41:174726-home-client-0-0-0
(version: 3.12.4)
[2018-01-18 08:38:41.787891] I [MSGID: 115036]
[server.c:527:server_rpc_notify] 0-home-server: disconnecting
connection from
gluster00.cluster.local-21590-2018/01/18-08:38:41:174726-home-client-0-0-0
[2018-01-18 08:38:41.790071] I [MSGID: 101055]
[client_t.c:443:gf_client_unref] 0-home-server: Shutting down
connection
gluster00.cluster.local.local-21590-2018/01/18-08:38:41:174726-home-client-0-0-0
[2018-01-18 08:38:56.298125] I
[glusterfsd-mgmt.c:52:mgmt_cbk_spec] 0-mgmt: Volume file changed
[2018-01-18 08:38:56.384120] I
[glusterfsd-mgmt.c:52:mgmt_cbk_spec] 0-mgmt: Volume file changed
[2018-01-18 08:38:56.394284] I
[glusterfsd-mgmt.c:1821:mgmt_getspec_cbk] 0-glusterfs: No change
in volfile,continuing
[2018-01-18 08:38:56.450621] I
[glusterfsd-mgmt.c:1821:mgmt_getspec_cbk] 0-glusterfs: No change
in volfile,continuing
[2018-01-18 08:39:17.606237] E [MSGID: 113107]
[posix.c:1150:posix_seek] 0-home-posix: seek failed on fd 639
length 1166 [No such device or address]
[2018-01-18 08:39:17.606310] E [MSGID: 115089]
[server-rpc-fops.c:2090:server_seek_cbk] 0-home-server: 1432655:
SEEK-2 (c20376ae-7db9-4340-950e-dba6aa95e848), client:
client1.cluster.local-5710-2018/01/17-22:02:07:122452-home-client-0-0-0,
error-xlator: home-posix [No such device or address]
[2018-01-18 08:39:17.610233] E [MSGID: 113107]
[posix.c:1150:posix_seek] 0-home-posix: seek failed on fd 639
length 1166 [No such device or address]
[2018-01-18 08:39:17.610285] E [MSGID: 115089]
[server-rpc-fops.c:2090:server_seek_cbk] 0-home-server: 1432656:
SEEK-2 (c20376ae-7db9-4340-950e-dba6aa95e848), client:
client1.cluster.local-5710-2018/01/17-22:02:07:122452-home-client-0-0-0,
error-xlator: home-posix [No such device or address]
[2018-01-18 08:37:53.592700] E [MSGID: 113107]
[posix.c:1150:posix_seek] 0-home-posix: seek failed on fd 1993
length 2279 [No such device or address]
[2018-01-18 08:39:53.583741] E [MSGID: 113107]
[posix.c:1150:posix_seek] 0-home-posix: seek failed on fd 1948
length 1758 [No such device or address]
[2018-01-18 08:39:53.583822] E [MSGID: 115089]
[server-rpc-fops.c:2090:server_seek_cbk] 0-home-server: 641734:
SEEK-2 (96dc4c82-ab85-4477-a6e6-1116466f68ba), client:
client1.cluster.local-5772-2018/01/17-22:02:07:201351-home-client-0-0-0,
error-xlator: home-posix [No such device or address]
Met vriendelijke groet, With kind regards,
Jorick Astrego
Netbulae Virtualization Experts
Tel: 053 20 30 270 | info@xxxxxxxxxxx | Staalsteden 4-3A | KvK 08198180 |
Fax: 053 20 30 271 | www.netbulae.eu | 7547 TA Enschede | BTW NL821234584B01 |