Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This message
contains confidential
information, which also may be
privileged, and is intended only
for the person(s) addressed
above. Any unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged information is
strictly prohibited. If you have
received this communication in
error, please erase all copies
of the message and its
attachments and notify the
sender immediately via reply
e-mail. **
From:
Jonathan Heese
Sent: Tuesday, March 17,
2015 12:36 PM
To: 'Ravishankar N'; gluster-users@xxxxxxxxxxx
Subject: RE:
I/O error on
replicated volume
Ravi,
The last lines
in the mount log before the massive
vomit of I/O errors are from 22
minutes prior, and seem innocuous to
me:
[2015-03-16
01:37:07.126340] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-0: failed to
get the port number for remote
subvolume. Please run 'gluster
volume status' on server to see if
brick process is running.
[2015-03-16
01:37:07.126587] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0: disconnect
called (peer:10.10.10.1:24008)
[2015-03-16
01:37:07.126687] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-1: failed to
get the port number for remote
subvolume. Please run 'gluster
volume status' on server to see if
brick process is running.
[2015-03-16
01:37:07.126737] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1: disconnect
called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.730165] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-0: changing
port to 49152 (from 0)
[2015-03-16
01:37:10.730276] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0: disconnect
called (peer:10.10.10.1:24008)
[2015-03-16
01:37:10.739500] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-1: changing
port to 49152 (from 0)
[2015-03-16
01:37:10.739560] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1: disconnect
called (peer:10.10.10.2:24008)
[2015-03-16
01:37:10.741883] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-0: Using
Program GlusterFS 3.3, Num
(1298437), Version (330)
[2015-03-16
01:37:10.744524] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-0: Connected
to 10.10.10.1:49152, attached to
remote volume '/bricks/brick1'.
[2015-03-16
01:37:10.744537] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-0: Server and
Client lk-version numbers are not
same, reopening the fds
[2015-03-16
01:37:10.744566] I
[afr-common.c:4267:afr_notify]
0-gluster_disk-replicate-0:
Subvolume 'gluster_disk-client-0'
came back up; going online.
[2015-03-16
01:37:10.744627] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-0: Server lk
version = 1
[2015-03-16
01:37:10.753037] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-1: Using
Program GlusterFS 3.3, Num
(1298437), Version (330)
[2015-03-16
01:37:10.755657] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-1: Connected
to 10.10.10.2:49152, attached to
remote volume '/bricks/brick1'.
[2015-03-16
01:37:10.755676] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-1: Server and
Client lk-version numbers are not
same, reopening the fds
[2015-03-16
01:37:10.761945] I
[fuse-bridge.c:5016:fuse_graph_setup]
0-fuse: switched to graph 0
[2015-03-16
01:37:10.762144] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-1: Server lk
version = 1
[2015-03-16
01:37:10.762279] I
[fuse-bridge.c:3953:fuse_init]
0-glusterfs-fuse: FUSE inited with
protocol versions: glusterfs 7.22
kernel 7.14
[2015-03-16
01:59:26.098670] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 292084: WRITE
=> -1 (Input/output error)
…
I’ve seen no
indication of split-brain on any
files at any point in this (ever
since downdating from 3.6.2 to
3.5.3, which is when this particular
issue started):
[root@duke
gfapi-module-for-linux-target-driver-]#
gluster v heal gluster_disk info
Brick
duke.jonheese.local:/bricks/brick1/
Number of
entries: 0
Brick
duchess.jonheese.local:/bricks/brick1/
Number of
entries: 0
Thanks.
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This message
contains confidential
information, which also may be
privileged, and is intended only
for the person(s) addressed
above. Any unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged information is
strictly prohibited. If you have
received this communication in
error, please erase all copies
of the message and its
attachments and notify the
sender immediately via reply
e-mail. **
On 03/17/2015
02:14 AM, Jonathan Heese wrote:
Hello,
So I resolved my previous
issue with split-brains and
the lack of self-healing by
dropping my installed
glusterfs* packages from 3.6.2
to 3.5.3, but now I've picked
up a new issue, which actually
makes normal use of the volume
practically impossible.
A little background for those
not already paying close
attention:
I have a 2 node 2 brick
replicating volume whose
purpose in life is to hold
iSCSI target files, primarily
for use to provide datastores
to a VMware ESXi cluster. The
plan is to put a handful of
image files on the Gluster
volume, mount them locally on
both Gluster nodes, and run
tgtd on both, pointed to the
image files on the mounted
gluster volume. Then the ESXi
boxes will use multipath
(active/passive) iSCSI to
connect to the nodes, with
automatic failover in case of
planned or unplanned downtime
of the Gluster nodes.
In my most recent round of
testing with 3.5.3, I'm seeing
a massive failure to write
data to the volume after about
5-10 minutes, so I've
simplified the scenario a bit
(to minimize the variables)
to: both Gluster nodes up,
only one node (duke) mounted
and running tgtd, and just
regular (single path) iSCSI
from a single ESXi server.
About 5-10 minutes into
migration a VM onto the test
datastore, /var/log/messages
on duke gets blasted with a
ton of messages exactly like
this:
Mar 15
22:24:06 duke tgtd:
bs_rdwr_request(180) io error
0x1781e00 2a -1 512 22971904,
Input/output error
And
/var/log/glusterfs/mnt-gluster_disk.log
gets blased with a ton of
messages exactly like this:
[2015-03-16
02:24:07.572279] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 635299: WRITE
=> -1 (Input/output error)
Are there any messages in the mount
log from AFR about split-brain just
before the above line appears?
Does `gluster v heal <VOLNAME>
info` show any files? Performing I/O
on files that are in split-brain
fail with EIO.
-Ravi
And the
write operation from VMware's
side fails as soon as these
messages start.
I don't
see any other errors (in the log
files I know of) indicating the
root cause of these i/o errors.
I'm sure that this is not enough
information to tell what's going
on, but can anyone help me
figure out what to look at next
to figure this out?
I've
also considered using Dan
Lambright's libgfapi gluster
module for tgtd (or something
similar) to avoid going through
FUSE, but I'm not sure whether
that would be irrelevant to this
problem, since I'm not 100% sure
if it lies in FUSE or elsewhere.
Thanks!
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This
message contains
confidential information,
which also may be
privileged, and is intended
only for the person(s)
addressed above. Any
unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged
information is strictly
prohibited. If you have
received this communication
in error, please erase all
copies of the message and
its attachments and notify
the sender immediately via
reply e-mail. **
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users