Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This
message contains
confidential information,
which also may be
privileged, and is
intended only for the
person(s) addressed above.
Any unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged
information is strictly
prohibited. If you have
received this
communication in error,
please erase all copies of
the message and its
attachments and notify the
sender immediately via
reply e-mail. **
From:
Jonathan Heese
Sent: Tuesday,
March 17, 2015 12:36 PM
To: 'Ravishankar
N';
gluster-users@xxxxxxxxxxx
Subject: RE:
I/O error
on replicated volume
Ravi,
The last
lines in the mount log before
the massive vomit of I/O
errors are from 22 minutes
prior, and seem innocuous to
me:
[2015-03-16
01:37:07.126340] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-0:
failed to get the port number
for remote subvolume. Please
run 'gluster volume status' on
server to see if brick process
is running.
[2015-03-16
01:37:07.126587] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0:
disconnect called
(peer:10.10.10.1:24008)
[2015-03-16
01:37:07.126687] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-1:
failed to get the port number
for remote subvolume. Please
run 'gluster volume status' on
server to see if brick process
is running.
[2015-03-16
01:37:07.126737] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1:
disconnect called
(peer:10.10.10.2:24008)
[2015-03-16
01:37:10.730165] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-0:
changing port to 49152 (from
0)
[2015-03-16
01:37:10.730276] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0:
disconnect called
(peer:10.10.10.1:24008)
[2015-03-16
01:37:10.739500] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-1:
changing port to 49152 (from
0)
[2015-03-16
01:37:10.739560] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1:
disconnect called
(peer:10.10.10.2:24008)
[2015-03-16
01:37:10.741883] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-0: Using
Program GlusterFS 3.3, Num
(1298437), Version (330)
[2015-03-16
01:37:10.744524] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-0:
Connected to 10.10.10.1:49152,
attached to remote volume
'/bricks/brick1'.
[2015-03-16
01:37:10.744537] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-0:
Server and Client lk-version
numbers are not same,
reopening the fds
[2015-03-16
01:37:10.744566] I
[afr-common.c:4267:afr_notify]
0-gluster_disk-replicate-0:
Subvolume
'gluster_disk-client-0' came
back up; going online.
[2015-03-16
01:37:10.744627] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-0:
Server lk version = 1
[2015-03-16
01:37:10.753037] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-1: Using
Program GlusterFS 3.3, Num
(1298437), Version (330)
[2015-03-16
01:37:10.755657] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-1:
Connected to 10.10.10.2:49152,
attached to remote volume
'/bricks/brick1'.
[2015-03-16
01:37:10.755676] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-1:
Server and Client lk-version
numbers are not same,
reopening the fds
[2015-03-16
01:37:10.761945] I
[fuse-bridge.c:5016:fuse_graph_setup]
0-fuse: switched to graph 0
[2015-03-16
01:37:10.762144] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-1:
Server lk version = 1
[2015-03-16
01:37:10.762279] I
[fuse-bridge.c:3953:fuse_init]
0-glusterfs-fuse: FUSE inited
with protocol versions:
glusterfs 7.22 kernel 7.14
[2015-03-16
01:59:26.098670] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 292084:
WRITE => -1 (Input/output
error)
…
I’ve
seen no indication of
split-brain on any files at
any point in this (ever since
downdating from 3.6.2 to
3.5.3, which is when this
particular issue started):
[root@duke
gfapi-module-for-linux-target-driver-]#
gluster v heal gluster_disk
info
Brick
duke.jonheese.local:/bricks/brick1/
Number
of entries: 0
Brick
duchess.jonheese.local:/bricks/brick1/
Number
of entries: 0
Thanks.
Jon Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This
message contains
confidential information,
which also may be
privileged, and is
intended only for the
person(s) addressed above.
Any unauthorized use,
distribution, copying or
disclosure of confidential
and/or privileged
information is strictly
prohibited. If you have
received this
communication in error,
please erase all copies of
the message and its
attachments and notify the
sender immediately via
reply e-mail. **
On
03/17/2015 02:14 AM, Jonathan
Heese wrote:
Hello,
So I resolved my
previous issue with
split-brains and the
lack of self-healing by
dropping my installed
glusterfs* packages from
3.6.2 to 3.5.3, but now
I've picked up a new
issue, which actually
makes normal use of the
volume practically
impossible.
A little background for
those not already paying
close attention:
I have a 2 node 2 brick
replicating volume whose
purpose in life is to
hold iSCSI target files,
primarily for use to
provide datastores to a
VMware ESXi cluster.
The plan is to put a
handful of image files
on the Gluster volume,
mount them locally on
both Gluster nodes, and
run tgtd on both,
pointed to the image
files on the mounted
gluster volume. Then the
ESXi boxes will use
multipath
(active/passive) iSCSI
to connect to the nodes,
with automatic failover
in case of planned or
unplanned downtime of
the Gluster nodes.
In my most recent round
of testing with 3.5.3,
I'm seeing a massive
failure to write data to
the volume after about
5-10 minutes, so I've
simplified the scenario
a bit (to minimize the
variables) to: both
Gluster nodes up, only
one node (duke) mounted
and running tgtd, and
just regular (single
path) iSCSI from a
single ESXi server.
About 5-10 minutes into
migration a VM onto the
test datastore,
/var/log/messages on
duke gets blasted with a
ton of messages exactly
like this:
Mar
15 22:24:06 duke tgtd:
bs_rdwr_request(180) io
error 0x1781e00 2a -1 512
22971904, Input/output
error
And
/var/log/glusterfs/mnt-gluster_disk.log
gets blased with a ton of
messages exactly like
this:
[2015-03-16
02:24:07.572279] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse: 635299:
WRITE => -1
(Input/output error)
Are there any messages in the
mount log from AFR about
split-brain just before the
above line appears?
Does `gluster v heal
<VOLNAME> info` show any
files? Performing I/O on files
that are in split-brain fail
with EIO.
-Ravi
And
the write operation from
VMware's side fails as
soon as these messages
start.
I
don't see any other errors
(in the log files I know
of) indicating the root
cause of these i/o
errors. I'm sure that
this is not enough
information to tell what's
going on, but can anyone
help me figure out what to
look at next to figure
this out?
I've
also considered using Dan
Lambright's libgfapi
gluster module for tgtd
(or something similar) to
avoid going through FUSE,
but I'm not sure whether
that would be irrelevant
to this problem, since I'm
not 100% sure if it lies
in FUSE or elsewhere.
Thanks!
Jon
Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
** This
message contains
confidential
information, which
also may be
privileged, and is
intended only for the
person(s) addressed
above. Any
unauthorized use,
distribution, copying
or disclosure of
confidential and/or
privileged information
is strictly
prohibited. If you
have received this
communication in
error, please erase
all copies of the
message and its
attachments and notify
the sender immediately
via reply e-mail. **
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users