Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This message
contains confidential information,
which also may be privileged, and
is intended only for the person(s)
addressed above. Any unauthorized
use, distribution, copying or
disclosure of confidential and/or
privileged information is strictly
prohibited. If you have received
this communication in error,
please erase all copies of the
message and its attachments and
notify the sender immediately via
reply e-mail. **
From: Jonathan
Heese
Sent: Tuesday, March 17,
2015 12:36 PM
To: 'Ravishankar N';
gluster-users@xxxxxxxxxxx
Subject: RE:
I/O error on
replicated volume
Ravi,
The last lines
in the mount log before the massive
vomit of I/O errors are from 22
minutes prior, and seem innocuous to
me:
[2015-03-16
01:37:07.126340] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-0: failed to get
the port number for remote subvolume.
Please run 'gluster volume status' on
server to see if brick process is
running.
[2015-03-16
01:37:07.126587] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0: disconnect
called (peer:10.10.10.1:24008)
[2015-03-16
01:37:07.126687] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-1: failed to get
the port number for remote subvolume.
Please run 'gluster volume status' on
server to see if brick process is
running.
[2015-03-16
01:37:07.126737] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1: disconnect
called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.730165] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-0: changing port
to 49152 (from 0)
[2015-03-16
01:37:10.730276] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0: disconnect
called (peer:10.10.10.1:24008)
[2015-03-16
01:37:10.739500] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-1: changing port
to 49152 (from 0)
[2015-03-16
01:37:10.739560] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1: disconnect
called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.741883] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-0: Using Program
GlusterFS 3.3, Num (1298437), Version
(330)
[2015-03-16
01:37:10.744524] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-0: Connected to
10.10.10.1:49152, attached to remote
volume '/bricks/brick1'.
[2015-03-16
01:37:10.744537] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-0: Server and
Client lk-version numbers are not
same, reopening the fds
[2015-03-16
01:37:10.744566] I
[afr-common.c:4267:afr_notify]
0-gluster_disk-replicate-0: Subvolume
'gluster_disk-client-0' came back up;
going online.
[2015-03-16
01:37:10.744627] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-0: Server lk
version = 1
[2015-03-16
01:37:10.753037] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-1: Using Program
GlusterFS 3.3, Num (1298437), Version
(330)
[2015-03-16
01:37:10.755657] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-1: Connected to
10.10.10.2:49152, attached to remote
volume '/bricks/brick1'.
[2015-03-16
01:37:10.755676] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-1: Server and
Client lk-version numbers are not
same, reopening the fds
[2015-03-16
01:37:10.761945] I
[fuse-bridge.c:5016:fuse_graph_setup]
0-fuse: switched to graph 0
[2015-03-16
01:37:10.762144] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-1: Server lk
version = 1
[2015-03-16
01:37:10.762279] I
[fuse-bridge.c:3953:fuse_init]
0-glusterfs-fuse: FUSE inited with
protocol versions: glusterfs 7.22
kernel 7.14
[2015-03-16
01:59:26.098670] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 292084: WRITE =>
-1 (Input/output error)
…
I’ve seen no
indication of split-brain on any files
at any point in this (ever since
downdating from 3.6.2 to 3.5.3, which
is when this particular issue
started):
[root@duke
gfapi-module-for-linux-target-driver-]#
gluster v heal gluster_disk info
Brick
duke.jonheese.local:/bricks/brick1/
Number of
entries: 0
Brick
duchess.jonheese.local:/bricks/brick1/
Number of
entries: 0
Thanks.
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This message
contains confidential information,
which also may be privileged, and
is intended only for the person(s)
addressed above. Any unauthorized
use, distribution, copying or
disclosure of confidential and/or
privileged information is strictly
prohibited. If you have received
this communication in error,
please erase all copies of the
message and its attachments and
notify the sender immediately via
reply e-mail. **
On 03/17/2015 02:14
AM, Jonathan Heese wrote:
Hello,
So I resolved my previous issue
with split-brains and the lack
of self-healing by dropping my
installed glusterfs* packages
from 3.6.2 to 3.5.3, but now
I've picked up a new issue,
which actually makes normal use
of the volume practically
impossible.
A little background for those
not already paying close
attention:
I have a 2 node 2 brick
replicating volume whose purpose
in life is to hold iSCSI target
files, primarily for use to
provide datastores to a VMware
ESXi cluster. The plan is to
put a handful of image files on
the Gluster volume, mount them
locally on both Gluster nodes,
and run tgtd on both, pointed to
the image files on the mounted
gluster volume. Then the ESXi
boxes will use multipath
(active/passive) iSCSI to
connect to the nodes, with
automatic failover in case of
planned or unplanned downtime of
the Gluster nodes.
In my most recent round of
testing with 3.5.3, I'm seeing a
massive failure to write data to
the volume after about 5-10
minutes, so I've simplified the
scenario a bit (to minimize the
variables) to: both Gluster
nodes up, only one node (duke)
mounted and running tgtd, and
just regular (single path) iSCSI
from a single ESXi server.
About 5-10 minutes into
migration a VM onto the test
datastore, /var/log/messages on
duke gets blasted with a ton of
messages exactly like this:
Mar 15
22:24:06 duke tgtd:
bs_rdwr_request(180) io error
0x1781e00 2a -1 512 22971904,
Input/output error
And
/var/log/glusterfs/mnt-gluster_disk.log
gets blased with a ton of messages
exactly like this:
[2015-03-16
02:24:07.572279] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 635299: WRITE
=> -1 (Input/output error)
Are there any messages in the mount
log from AFR about split-brain just
before the above line appears?
Does `gluster v heal <VOLNAME>
info` show any files? Performing I/O
on files that are in split-brain fail
with EIO.
-Ravi
And the
write operation from VMware's side
fails as soon as these messages
start.
I don't
see any other errors (in the log
files I know of) indicating the
root cause of these i/o errors.
I'm sure that this is not enough
information to tell what's going
on, but can anyone help me figure
out what to look at next to figure
this out?
I've also
considered using Dan Lambright's
libgfapi gluster module for tgtd
(or something similar) to avoid
going through FUSE, but I'm not
sure whether that would be
irrelevant to this problem, since
I'm not 100% sure if it lies in
FUSE or elsewhere.
Thanks!
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This message
contains confidential
information, which also may be
privileged, and is intended
only for the person(s)
addressed above. Any
unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged information
is strictly prohibited. If you
have received this
communication in error, please
erase all copies of the
message and its attachments
and notify the sender
immediately via reply e-mail.
**
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users