Jon
Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
**
This message
contains
confidential
information,
which also may
be privileged,
and is intended
only for the
person(s)
addressed above.
Any unauthorized
use,
distribution,
copying or
disclosure of
confidential
and/or
privileged
information is
strictly
prohibited. If
you have
received this
communication in
error, please
erase all copies
of the message
and its
attachments and
notify the
sender
immediately via
reply e-mail. **
From:
Jonathan Heese
Sent:
Tuesday, March
17, 2015 12:36
PM
To:
'Ravishankar N';
gluster-users@xxxxxxxxxxx
Subject:
RE:
I/O error on
replicated
volume
Ravi,
The last lines in the mount log before the massive
vomit of I/O errors
are from 22 minutes
prior, and seem
innocuous to me:
[2015-03-16 01:37:07.126340] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-0:
failed to get the
port number for
remote subvolume.
Please run 'gluster
volume status' on
server to see if
brick process is
running.
[2015-03-16 01:37:07.126587] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0:
disconnect called
(peer:10.10.10.1:24008)
[2015-03-16 01:37:07.126687] E
[client-handshake.c:1760:client_query_portmap_cbk]
0-gluster_disk-client-1:
failed to get the
port number for
remote subvolume.
Please run 'gluster
volume status' on
server to see if
brick process is
running.
[2015-03-16 01:37:07.126737] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1:
disconnect called
(peer:10.10.10.2:24008)
[2015-03-16 01:37:10.730165] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-0:
changing port to
49152 (from 0)
[2015-03-16 01:37:10.730276] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-0:
disconnect called
(peer:10.10.10.1:24008)
[2015-03-16 01:37:10.739500] I
[rpc-clnt.c:1729:rpc_clnt_reconfig]
0-gluster_disk-client-1:
changing port to
49152 (from 0)
[2015-03-16 01:37:10.739560] W
[rdma.c:4273:gf_rdma_disconnect]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f)
[0x7fd9c557bccf]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)
[0x7fd9c557a995]
(-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea)
[0x7fd9c0d8fb9a])))
0-gluster_disk-client-1:
disconnect called
(peer:10.10.10.2:24008)
[2015-03-16 01:37:10.741883] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-0:
Using Program
GlusterFS 3.3, Num
(1298437), Version
(330)
[2015-03-16 01:37:10.744524] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-0:
Connected to
10.10.10.1:49152,
attached to remote
volume
'/bricks/brick1'.
[2015-03-16 01:37:10.744537] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-0:
Server and Client
lk-version numbers
are not same,
reopening the fds
[2015-03-16 01:37:10.744566] I
[afr-common.c:4267:afr_notify]
0-gluster_disk-replicate-0:
Subvolume
'gluster_disk-client-0'
came back up; going
online.
[2015-03-16 01:37:10.744627] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-0:
Server lk version =
1
[2015-03-16 01:37:10.753037] I
[client-handshake.c:1677:select_server_supported_programs]
0-gluster_disk-client-1:
Using Program
GlusterFS 3.3, Num
(1298437), Version
(330)
[2015-03-16 01:37:10.755657] I
[client-handshake.c:1462:client_setvolume_cbk]
0-gluster_disk-client-1:
Connected to
10.10.10.2:49152,
attached to remote
volume
'/bricks/brick1'.
[2015-03-16 01:37:10.755676] I
[client-handshake.c:1474:client_setvolume_cbk]
0-gluster_disk-client-1:
Server and Client
lk-version numbers
are not same,
reopening the fds
[2015-03-16 01:37:10.761945] I
[fuse-bridge.c:5016:fuse_graph_setup]
0-fuse: switched to
graph 0
[2015-03-16 01:37:10.762144] I
[client-handshake.c:450:client_set_lk_version_cbk]
0-gluster_disk-client-1:
Server lk version =
1
[2015-03-16 01:37:10.762279] I
[fuse-bridge.c:3953:fuse_init]
0-glusterfs-fuse:
FUSE inited with
protocol versions:
glusterfs 7.22
kernel 7.14
[2015-03-16 01:59:26.098670] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse:
292084: WRITE =>
-1 (Input/output
error)
…
I’ve seen no indication of split-brain on any
files at any point
in this (ever since
downdating from
3.6.2 to 3.5.3,
which is when this
particular issue
started):
[root@duke gfapi-module-for-linux-target-driver-]#
gluster v heal
gluster_disk info
Brick duke.jonheese.local:/bricks/brick1/
Number of entries: 0
Brick duchess.jonheese.local:/bricks/brick1/
Number of entries: 0
Thanks.
Jon
Heese
Systems
Engineer
INetU
Managed Hosting
P:
610.266.7441 x 261
F:
610.266.7434
www.inetu.net
**
This message
contains
confidential
information,
which also may
be privileged,
and is intended
only for the
person(s)
addressed above.
Any unauthorized
use,
distribution,
copying or
disclosure of
confidential
and/or
privileged
information is
strictly
prohibited. If
you have
received this
communication in
error, please
erase all copies
of the message
and its
attachments and
notify the
sender
immediately via
reply e-mail. **
On
03/17/2015 02:14 AM,
Jonathan Heese
wrote:
Hello,
So I resolved
my previous
issue with
split-brains
and the lack
of
self-healing
by dropping my
installed
glusterfs*
packages from
3.6.2 to
3.5.3, but now
I've picked up
a new issue,
which actually
makes normal
use of the
volume
practically
impossible.
A little
background for
those not
already paying
close
attention:
I have a 2
node 2 brick
replicating
volume whose
purpose in
life is to
hold iSCSI
target files,
primarily for
use to provide
datastores to
a VMware ESXi
cluster. The
plan is to put
a handful of
image files on
the Gluster
volume, mount
them locally
on both
Gluster nodes,
and run tgtd
on both,
pointed to the
image files on
the mounted
gluster
volume. Then
the ESXi boxes
will use
multipath
(active/passive)
iSCSI to
connect to the
nodes, with
automatic
failover in
case of
planned or
unplanned
downtime of
the Gluster
nodes.
In my most
recent round
of testing
with 3.5.3,
I'm seeing a
massive
failure to
write data to
the volume
after about
5-10 minutes,
so I've
simplified the
scenario a bit
(to minimize
the variables)
to: both
Gluster nodes
up, only one
node (duke)
mounted and
running tgtd,
and just
regular
(single path)
iSCSI from a
single ESXi
server.
About 5-10
minutes into
migration a VM
onto the test
datastore,
/var/log/messages
on duke gets
blasted with a
ton of
messages
exactly like
this:
Mar 15 22:24:06 duke tgtd: bs_rdwr_request(180)
io error
0x1781e00 2a -1
512 22971904,
Input/output
error
And /var/log/glusterfs/mnt-gluster_disk.log
gets blased with
a ton of
messages exactly
like this:
[2015-03-16 02:24:07.572279] W
[fuse-bridge.c:2242:fuse_writev_cbk]
0-glusterfs-fuse:
635299: WRITE
=> -1
(Input/output
error)
Are there any
messages in the
mount log from AFR
about split-brain
just before the
above line appears?
Does `gluster v heal
<VOLNAME>
info` show any
files? Performing
I/O on files that
are in split-brain
fail with EIO.
-Ravi
And the write operation from VMware's side
fails as soon as
these messages
start.
I don't see any other errors (in the log files
I know of)
indicating the
root cause of
these i/o
errors. I'm
sure that this
is not enough
information to
tell what's
going on, but
can anyone help
me figure out
what to look at
next to figure
this out?
I've also considered using Dan Lambright's
libgfapi gluster
module for tgtd
(or something
similar) to
avoid going
through FUSE,
but I'm not sure
whether that
would be
irrelevant to
this problem,
since I'm not
100% sure if it
lies in FUSE or
elsewhere.
Thanks!
Jon
Heese
Systems Engineer
INetU Managed Hosting
P:
610.266.7441 x
261
F:
610.266.7434
www.inetu.net
**
This message
contains
confidential
information,
which also may
be privileged,
and is
intended only
for the
person(s)
addressed
above. Any
unauthorized
use,
distribution,
copying or
disclosure of
confidential
and/or
privileged
information is
strictly
prohibited. If
you have
received this
communication
in error,
please erase
all copies of
the message
and its
attachments
and notify the
sender
immediately
via reply
e-mail. **
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users