Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
**
This message contains confidential
information, which also may be privileged,
and is intended only for the person(s)
addressed above. Any unauthorized use,
distribution, copying or disclosure of
confidential and/or privileged information
is strictly prohibited. If you have received
this communication in error, please erase
all copies of the message and its
attachments and notify the sender
immediately via reply e-mail. **
From: Jonathan Heese
Sent: Tuesday, March 17, 2015 12:36
PM
To: 'Ravishankar N';
gluster-users@xxxxxxxxxxx
Subject: RE: I/O
error on replicated volume
Ravi,
The
last lines in the mount log before the massive
vomit of I/O errors are from 22 minutes prior,
and seem innocuous to me:
[2015-03-16
01:37:07.126340] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-0: failed to get the port
number for remote subvolume. Please run 'gluster
volume status' on server to see if brick process
is running.
[2015-03-16
01:37:07.126587] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a]))) 0-gluster_disk-client-0:
disconnect called (peer:10.10.10.1:24008)
[2015-03-16
01:37:07.126687] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-1: failed to get the port
number for remote subvolume. Please run 'gluster
volume status' on server to see if brick process
is running.
[2015-03-16
01:37:07.126737] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a]))) 0-gluster_disk-client-1:
disconnect called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.730165] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-0: changing port to 49152
(from 0)
[2015-03-16
01:37:10.730276] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a]))) 0-gluster_disk-client-0:
disconnect called (peer:10.10.10.1:24008)
[2015-03-16
01:37:10.739500] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-1: changing port to 49152
(from 0)
[2015-03-16
01:37:10.739560] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a]))) 0-gluster_disk-client-1:
disconnect called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.741883] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-0: Using Program GlusterFS
3.3, Num (1298437), Version (330)
[2015-03-16
01:37:10.744524] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-0: Connected to
10.10.10.1:49152, attached to remote volume
'/bricks/brick1'.
[2015-03-16
01:37:10.744537] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-0: Server and Client
lk-version numbers are not same, reopening the
fds
[2015-03-16
01:37:10.744566] I
[afr-common.c:4267:afr_notify]
0-gluster_disk-replicate-0: Subvolume
'gluster_disk-client-0' came back up; going
online.
[2015-03-16
01:37:10.744627] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-0: Server lk version = 1
[2015-03-16
01:37:10.753037] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-1: Using Program GlusterFS
3.3, Num (1298437), Version (330)
[2015-03-16
01:37:10.755657] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-1: Connected to
10.10.10.2:49152, attached to remote volume
'/bricks/brick1'.
[2015-03-16
01:37:10.755676] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-1: Server and Client
lk-version numbers are not same, reopening the
fds
[2015-03-16
01:37:10.761945] I
[fuse-bridge.c:5016:fuse_graph_setup] 0-fuse:
switched to graph 0
[2015-03-16
01:37:10.762144] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-1: Server lk version = 1
[2015-03-16
01:37:10.762279] I
[fuse-bridge.c:3953:fuse_init] 0-glusterfs-fuse:
FUSE inited with protocol versions: glusterfs
7.22 kernel 7.14
[2015-03-16
01:59:26.098670] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 292084: WRITE => -1
(Input/output error)
…
I’ve
seen no indication of split-brain on any files
at any point in this (ever since downdating from
3.6.2 to 3.5.3, which is when this particular
issue started):
[root@duke
gfapi-module-for-linux-target-driver-]# gluster
v heal gluster_disk info
Brick
duke.jonheese.local:/bricks/brick1/
Number
of entries: 0
Brick
duchess.jonheese.local:/bricks/brick1/
Number
of entries: 0
Thanks.
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
**
This message contains confidential
information, which also may be privileged,
and is intended only for the person(s)
addressed above. Any unauthorized use,
distribution, copying or disclosure of
confidential and/or privileged information
is strictly prohibited. If you have received
this communication in error, please erase
all copies of the message and its
attachments and notify the sender
immediately via reply e-mail. **
On 03/17/2015 02:14 AM,
Jonathan Heese wrote:
Hello,
So I resolved my previous issue with
split-brains and the lack of self-healing
by dropping my installed glusterfs*
packages from 3.6.2 to 3.5.3, but now I've
picked up a new issue, which actually
makes normal use of the volume practically
impossible.
A little background for those not already
paying close attention:
I have a 2 node 2 brick replicating volume
whose purpose in life is to hold iSCSI
target files, primarily for use to provide
datastores to a VMware ESXi cluster. The
plan is to put a handful of image files on
the Gluster volume, mount them locally on
both Gluster nodes, and run tgtd on both,
pointed to the image files on the mounted
gluster volume. Then the ESXi boxes will
use multipath (active/passive) iSCSI to
connect to the nodes, with automatic
failover in case of planned or unplanned
downtime of the Gluster nodes.
In my most recent round of testing with
3.5.3, I'm seeing a massive failure to
write data to the volume after about 5-10
minutes, so I've simplified the scenario a
bit (to minimize the variables) to: both
Gluster nodes up, only one node (duke)
mounted and running tgtd, and just regular
(single path) iSCSI from a single ESXi
server.
About 5-10 minutes into migration a VM
onto the test datastore, /var/log/messages
on duke gets blasted with a ton of
messages exactly like this:
Mar
15 22:24:06 duke tgtd: bs_rdwr_request(180)
io error 0x1781e00 2a -1 512 22971904,
Input/output error
And
/var/log/glusterfs/mnt-gluster_disk.log gets
blased with a ton of messages exactly like
this:
[2015-03-16
02:24:07.572279] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 635299: WRITE => -1
(Input/output error)
Are there any messages in the mount log from AFR
about split-brain just before the above line
appears?
Does `gluster v heal <VOLNAME> info` show
any files? Performing I/O on files that are in
split-brain fail with EIO.
-Ravi
And
the write operation from VMware's side fails
as soon as these messages start.
I
don't see any other errors (in the log files
I know of) indicating the root cause of
these i/o errors. I'm sure that this is not
enough information to tell what's going on,
but can anyone help me figure out what to
look at next to figure this out?
I've
also considered using Dan Lambright's
libgfapi gluster module for tgtd (or
something similar) to avoid going through
FUSE, but I'm not sure whether that would be
irrelevant to this problem, since I'm not
100% sure if it lies in FUSE or elsewhere.
Thanks!
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
**
This message contains confidential
information, which also may be
privileged, and is intended only for the
person(s) addressed above. Any
unauthorized use, distribution, copying
or disclosure of confidential and/or
privileged information is strictly
prohibited. If you have received this
communication in error, please erase all
copies of the message and its
attachments and notify the sender
immediately via reply e-mail. **
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users