No subject

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Is there any configuration option to prevent this problem? Is there some
enhancements done in gluster 3.3 that can prevent the "split brain" issue
happen again?

Thank you very much in advance for any link or information,
Samuel.

--0016e64f80a82b7d0c04acd34c3a
Content-Type: text/html; charset=ISO-8859-1
Content-Transfer-Encoding: quoted-printable

Hi folks,<br><br>We&#39;re researching the use of glusterfs (glusterfs 3.2.=
2 installed from sources) as storage solution for open nebula (open source =
cloud solution version 2.2 also installed from sources) and faced one weird=
 issue were some files get corrupted. We were performing some tests and dis=
connected the subnet were the glusterfd run (2 servers) from the subnets of=
 the clients (4 nodes).<br>
Apologies if the mail is too big but I&#39;m trying to attach as much infor=
mation as possible. I might have left some important stuff out so, if anyon=
e can shed some light, I would happily provide more information.<br><br>
Several disconnection messsages appeared in the logs:<br><br>client1:<br>[2=
011-09-12 16:20:39.861342] W [socket.c:1494:__socket_proto_state_machine] 0=
-glusterfs: reading from socket failed. Error (Connection reset by peer), p=
eer (A.B.C31:24007)<br>
[2011-09-12 16:22:48.101327] W [socket.c:204:__socket_rwv] 0-glusterfs: rea=
dv failed (Connection timed out)<br>[2011-09-12 16:22:48.101405] W [socket.=
c:1494:__socket_proto_state_machine] 0-glusterfs: reading from socket faile=
d. Error (Connection timed out), peer (A.B.C31:24007)<br>
[2011-09-12 16:23:09.938010] C [client-handshake.c:121:rpc_client_ping_time=
r_expired] 0-cloud-client-1: server A.B.C32:24010 has not responded in the =
last 42 seconds, disconnecting.<br>[2011-09-12 16:23:09.938196] C [client-h=
andshake.c:121:rpc_client_ping_timer_expired] 0-cloud-client-0: server A.B.=
C31:24010 has not responded in the last 42 seconds, disconnecting.<br>
<br>server1:<br>2011-09-12 16:19:26.658858] W [socket.c:204:__socket_rwv] 0=
-socket.management: readv failed (Connection timed out)<br>[2011-09-12 16:1=
9:26.688463] W [socket.c:1494:__socket_proto_state_machine] 0-socket.manage=
ment: reading from socket failed. Error (Connection timed out), peer (A.B.C=
21:1023)<br>
[2011-09-12 16:19:26.835551] W [socket.c:204:__socket_rwv] 0-management: re=
adv failed (Connection timed out)<br>[2011-09-12 16:19:26.835586] W [socket=
.c:1494:__socket_proto_state_machine] 0-management: reading from socket fai=
led. Error (Connection timed out), peer (A.B.C32:24007)<br>
[2011-09-12 16:19:27.255418] W [socket.c:204:__socket_rwv] 0-socket.managem=
ent: readv failed (Connection timed out)<br>[2011-09-12 16:19:27.255454] W =
[socket.c:1494:__socket_proto_state_machine] 0-socket.management: reading f=
rom socket failed. Error (Connection timed out), peer (A.B.C32:1023)<br>
<br>Then both gluster servers see each other:<br>server 1:<br>[2011-09-12 1=
6:19:40.34860] I [glusterd-handler.c:2797:glusterd_handle_incoming_friend_r=
eq] 0-glusterd: Received probe from uuid: 91a43282-2a05-4e7d-b4eb-3fbd3cef0=
055<br>
[2011-09-12 16:19:49.775930] I [glusterd-handshake.c:317:glusterd_set_clnt_=
mgmt_program] 0-: Using Program glusterd clnt mgmt, Num (1238433), Version =
(1)<br>[2011-09-12 16:19:49.776026] I [glusterd-handler.c:3567:glusterd_xfe=
r_friend_add_resp] 0-glusterd: Responded to A.B.C31 (0), ret: 0<br>
[2011-09-12 16:19:49.811642] I [glusterd-sm.c:492:glusterd_ac_send_friend_u=
pdate] 0-: Added uuid: 91a43282-2a05-4e7d-b4eb-3fbd3cef0055, host: store2<b=
r>[2011-09-12 16:19:49.811878] I [glusterd-rpc-ops.c:523:glusterd3_1_friend=
_add_cbk] 0-glusterd: Received ACC from uuid: 91a43282-2a05-4e7d-b4eb-3fbd3=
cef0055, host: store2, port: 0<br>
[2011-09-12 16:19:49.812122] I [glusterd-handler.c:2944:glusterd_handle_fri=
end_update] 0-glusterd: Received friend update from uuid: 91a43282-2a05-4e7=
d-b4eb-3fbd3cef0055<br>[2011-09-12 16:19:49.812156] I [glusterd-handler.c:2=
989:glusterd_handle_friend_update] 0-: Received uuid: ede0b9eb-c93c-40f2-a0=
65-0f04a541b784, hostname:A.B.C31<br>
[2011-09-12 16:19:49.812171] I [glusterd-handler.c:2992:glusterd_handle_fri=
end_update] 0-: Received my uuid as Friend<br>[2011-09-12 16:19:49.812234] =
I [glusterd-rpc-ops.c:705:glusterd3_1_friend_update_cbk] 0-glusterd: Receiv=
ed ACC from uuid:<br>
<br>server 2:<br>[2011-09-12 16:22:13.252695] I [glusterd-handshake.c:317:g=
lusterd_set_clnt_mgmt_program] 0-: Using Program glusterd clnt mgmt, Num (1=
238433), Version (1)<br>[2011-09-12 16:22:13.264336] I [glusterd-rpc-ops.c:=
523:glusterd3_1_friend_add_cbk] 0-glusterd: Received ACC from uuid: ede0b9e=
b-c93c-40f2-a065-0f04a541b784, host: A.B.C31, port: 0<br>
[2011-09-12 16:22:13.264635] I [glusterd-handler.c:2797:glusterd_handle_inc=
oming_friend_req] 0-glusterd: Received probe from uuid: ede0b9eb-c93c-40f2-=
a065-0f04a541b784<br>[2011-09-12 16:22:13.264708] I [glusterd-handler.c:356=
7:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to store2 (0), ret: =
0<br>
[2011-09-12 16:22:13.264855] I [glusterd-sm.c:492:glusterd_ac_send_friend_u=
pdate] 0-: Added uuid: ede0b9eb-c93c-40f2-a065-0f04a541b784, host: A.B.C31<=
br>[2011-09-12 16:22:13.265151] I [glusterd-handler.c:2944:glusterd_handle_=
friend_update] 0-glusterd: Received friend update from uuid: ede0b9eb-c93c-=
40f2-a065-0f04a541b784<br>
[2011-09-12 16:22:13.265184] I [glusterd-handler.c:2989:glusterd_handle_fri=
end_update] 0-: Received uuid: 91a43282-2a05-4e7d-b4eb-3fbd3cef0055, hostna=
me:store2<br>[2011-09-12 16:22:13.265197] I [glusterd-handler.c:2992:gluste=
rd_handle_friend_update] 0-: Received my uuid as Friend<br>
<br>Then on client 4 the &quot;problematic&quot; issue start to appear (Fil=
e descriptor in bad state)<br><br>client 4:<br>[2011-09-12 16:23:09.985706]=
 E [rpc-clnt.c:338:saved_frames_unwind] (--&gt;/usr/local/lib/libgfrpc.so.0=
(rpc_clnt_notify+0x6d) [0x7f20c5ba750d] (--&gt;/usr/local/lib/libgfrpc.so.0=
(rpc_clnt_connection_cleanup+0x7d) [0x7f20c5ba52dd] (--&gt;/usr/local/lib/l=
ibgfrpc.so.0(saved_frames_destroy+0xe) [0x7f20c5ba523e]))) 0-cloud-client-0=
: forced unwinding frame type(GlusterFS 3.1) op(LOOKUP(27)) called at 2011-=
09-12 16:22:22.945095<br>
[2011-09-12 16:23:09.985770] W [fuse-bridge.c:184:fuse_entry_cbk] 0-gluster=
fs-fuse: 5657287: LOOKUP() /one/.pam_environment =3D&gt; -1 (Transport endp=
oint is not connected)<br>[2011-09-12 16:23:09.985853] E [rpc-clnt.c:338:sa=
ved_frames_unwind] (--&gt;/usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0x6d=
) [0x7f20c5ba750d] (--&gt;/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_=
cleanup+0x7d) [0x7f20c5ba52dd] (--&gt;/usr/local/lib/libgfrpc.so.0(saved_fr=
ames_destroy+0xe) [0x7f20c5ba523e]))) 0-cloud-client-0: forced unwinding fr=
ame type(GlusterFS 3.1) op(FINODELK(30)) called at 2011-09-12 16:22:24.6901=
97<br>
[2011-09-12 16:23:09.986420] W [rpc-clnt.c:1411:rpc_clnt_submit] 0-cloud-cl=
ient-0: failed to submit rpc-request (XID: 0x7926489x Program: GlusterFS 3.=
1, ProgVers: 310, Proc: 30) to rpc-transport (cloud-client-0)<br>[2011-09-1=
2 16:23:09.986463] W [client3_1-fops.c:4868:client3_1_finodelk] 0-cloud-cli=
ent-1: (2760): failed to get fd ctx. EBADFD<br>
[2011-09-12 16:23:09.986480] W [client3_1-fops.c:4915:client3_1_finodelk] 0=
-cloud-client-1: failed to send the fop: File descriptor in bad state<br><b=
r><br>and later on:<br>[2011-09-12 16:23:09.993817] W [rpc-clnt.c:1411:rpc_=
clnt_submit] 0-cloud-client-0: failed to submit rpc-request (XID: 0x7926670=
x Program: GlusterFS 3.1, ProgVers: 310, Proc: 30) to rpc-transport (cloud-=
client-0)<br>
[2011-09-12 16:23:09.993837] W [rpc-clnt.c:1411:rpc_clnt_submit] 0-cloud-cl=
ient-0: failed to submit rpc-request (XID: 0x7926671x Program: GlusterFS 3.=
1, ProgVers: 310, Proc: 30) to rpc-transport (cloud-client-0)<br>[2011-09-1=
2 16:23:09.993970] E [rpc-clnt.c:338:saved_frames_unwind] (--&gt;/usr/local=
/lib/libgfrpc.so.0(rpc_clnt_notify+0x6d) [0x7f20c5ba750d] (--&gt;/usr/local=
/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7d) [0x7f20c5ba52dd] (--&g=
t;/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x7f20c5ba523e]))=
) 0-cloud-client-0: forced unwinding frame type(GlusterFS Handshake) op(PIN=
G(3)) called at 2011-09-12 16:22:27.931756<br>
[2011-09-12 16:23:09.993992] W [client-handshake.c:264:client_ping_cbk] 0-c=
loud-client-0: timer must have expired<br>[2011-09-12 16:23:09.994013] E [a=
fr-common.c:2634:afr_notify] 0-cloud-replicate-0: All subvolumes are down. =
Going offline until atleast one of them comes back up.<br>
[2011-09-12 16:23:10.577233] W [fuse-bridge.c:1825:fuse_writev_cbk] 0-glust=
erfs-fuse: 5657529: WRITE =3D&gt; -1 (Transport endpoint is not connected)<=
br>[2011-09-12 16:23:11.577562] W [afr-open.c:624:afr_openfd_flush] 0-cloud=
-replicate-0: fd not open on any subvolume 0x7f20c0f16174 (/one/var/30/imag=
es/disk.0)<br>
[2011-09-12 16:23:11.607857] W [dict.c:418:dict_unref] (--&gt;/usr/local/li=
b/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) [0x7f20c5ba7145] (--&gt;/usr/lo=
cal/lib/glusterfs/3.2.2/xlator/protocol/client.so(client3_1_fstat_cbk+0x31b=
) [0x7f20c292ee3b] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/rep=
licate.so(afr_sh_data_fstat_cbk+0x164) [0x7f20c26d8284]))) 0-dict: dict is =
NULL<br>
[2011-09-12 16:23:11.608416] W [afr-common.c:122:afr_set_split_brain] (--&g=
t;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_sh_data_fl=
ush_cbk+0x79) [0x7f20c26d7239] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator=
/cluster/replicate.so(afr_sh_data_done+0x4e) [0x7f20c26d696e] (--&gt;/usr/l=
ocal/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_self_heal_completi=
on_cbk+0x271) [0x7f20c26dd171]))) 0-cloud-replicate-0: invalid argument: in=
ode<br>
[2011-09-12 16:23:11.608446] W [afr-open.c:326:afr_openfd_sh_unwind] 0-clou=
d-replicate-0: fd not open on any subvolume 0x7f20c0f16174 (/one/var/30/ima=
ges/disk.0)<br>[2011-09-12 16:23:14.180613] W [fuse-bridge.c:1825:fuse_writ=
ev_cbk] 0-glusterfs-fuse: 5657804: WRITE =3D&gt; -1 (Transport endpoint is =
not connected)<br>
[2011-09-12 16:23:14.184627] W [afr-open.c:624:afr_openfd_flush] 0-cloud-re=
plicate-0: fd not open on any subvolume 0x7f20c0f160cc (/one/var/25/images/=
disk.0)<br>[2011-09-12 16:23:14.186908] W [dict.c:418:dict_unref] (--&gt;/u=
sr/local/lib/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) [0x7f20c5ba7145] (--=
&gt;/usr/local/lib/glusterfs/3.2.2/xlator/protocol/client.so(client3_1_fsta=
t_cbk+0x31b) [0x7f20c292ee3b] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/=
cluster/replicate.so(afr_sh_data_fstat_cbk+0x164) [0x7f20c26d8284]))) 0-dic=
t: dict is NULL<br>
[2011-09-12 16:23:14.187411] W [afr-common.c:122:afr_set_split_brain] (--&g=
t;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_sh_data_fl=
ush_cbk+0x79) [0x7f20c26d7239] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator=
/cluster/replicate.so(afr_sh_data_done+0x4e) [0x7f20c26d696e] (--&gt;/usr/l=
ocal/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_self_heal_completi=
on_cbk+0x271) [0x7f20c26dd171]))) 0-cloud-replicate-0: invalid argument: in=
ode<br>
[2011-09-12 16:23:14.187447] W [afr-open.c:326:afr_openfd_sh_unwind] 0-clou=
d-replicate-0: fd not open on any subvolume 0x7f20c0f160cc (/one/var/25/ima=
ges/disk.0)<br>[2011-09-12 16:23:21.10661] W [afr-open.c:624:afr_openfd_flu=
sh] 0-cloud-replicate-0: fd not open on any subvolume 0x7f20c0f160cc (/one/=
var/25/images/disk.0)<br>
[2011-09-12 16:23:21.544266] W [afr-open.c:624:afr_openfd_flush] 0-cloud-re=
plicate-0: fd not open on any subvolume 0x7f20c0f16174 (/one/var/30/images/=
disk.0)<br>[2011-09-12 16:53:31.252358] E [rpc-clnt.c:197:call_bail] 0-clou=
d-client-0: bailing out frame type(GlusterFS 3.1) op(FINODELK(30)) xid =3D =
0x7927047x sent =3D 2011-09-12 16:23:21.545077. timeout =3D 1800<br>
[2011-09-12 16:53:31.252463] E [rpc-clnt.c:197:call_bail] 0-cloud-client-0:=
 bailing out frame type(GlusterFS 3.1) op(FINODELK(30)) xid =3D 0x7927028x =
sent =3D 2011-09-12 16:23:21.11527. timeout =3D 1800<br>[2011-09-12 16:53:3=
1.255321] W [dict.c:418:dict_unref] (--&gt;/usr/local/lib/libgfrpc.so.0(rpc=
_clnt_handle_reply+0xa5) [0x7f20c5ba7145] (--&gt;/usr/local/lib/glusterfs/3=
.2.2/xlator/protocol/client.so(client3_1_fstat_cbk+0x31b) [0x7f20c292ee3b] =
(--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_sh_da=
ta_fstat_cbk+0x164) [0x7f20c26d8284]))) 0-dict: dict is NULL<br>
[2011-09-12 16:53:31.255463] W [dict.c:418:dict_unref] (--&gt;/usr/local/li=
b/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) [0x7f20c5ba7145] (--&gt;/usr/lo=
cal/lib/glusterfs/3.2.2/xlator/protocol/client.so(client3_1_fstat_cbk+0x31b=
) [0x7f20c292ee3b] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/rep=
licate.so(afr_sh_data_fstat_cbk+0x164) [0x7f20c26d8284]))) 0-dict: dict is =
NULL<br>
[2011-09-12 16:53:31.255875] W [afr-common.c:122:afr_set_split_brain] (--&g=
t;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_sh_data_fl=
ush_cbk+0x79) [0x7f20c26d7239] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator=
/cluster/replicate.so(afr_sh_data_done+0x4e) [0x7f20c26d696e] (--&gt;/usr/l=
ocal/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_self_heal_completi=
on_cbk+0x271) [0x7f20c26dd171]))) 0-cloud-replicate-0: invalid argument: in=
ode<br>
[2011-09-12 16:53:31.255909] W [afr-open.c:326:afr_openfd_sh_unwind] 0-clou=
d-replicate-0: fd not open on any subvolume 0x7f20c0f16174 (/one/var/30/ima=
ges/disk.0)<br>[2011-09-12 16:53:31.256041] W [afr-common.c:122:afr_set_spl=
it_brain] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so=
(afr_sh_data_flush_cbk+0x79) [0x7f20c26d7239] (--&gt;/usr/local/lib/gluster=
fs/3.2.2/xlator/cluster/replicate.so(afr_sh_data_done+0x4e) [0x7f20c26d696e=
] (--&gt;/usr/local/lib/glusterfs/3.2.2/xlator/cluster/replicate.so(afr_sel=
f_heal_completion_cbk+0x271) [0x7f20c26dd171]))) 0-cloud-replicate-0: inval=
id argument: inode<br>
[2011-09-12 16:53:31.256085] W [afr-open.c:326:afr_openfd_sh_unwind] 0-clou=
d-replicate-0: fd not open on any subvolume 0x7f20c0f160cc (/one/var/25/ima=
ges/disk.0)<br><br><br>And on the server side:<br>[2011-09-12 16:21:47.8355=
21] W [socket.c:1494:__socket_proto_state_machine] 0-tcp.cloud-server: read=
ing from socket failed. Error (Connection timed out), peer (A.B.C21:1021)<b=
r>
[2011-09-12 16:21:47.835563] I [server-helpers.c:485:do_fd_cleanup] 0-cloud=
-server: fd cleanup on /one/var/46/images/disk.0<br>[2011-09-12 16:22:36.48=
6483] I [server-helpers.c:343:do_lock_table_cleanup] 0-cloud-server: finode=
lk released on /one/var/30/images/disk.0<br>
[2011-09-12 16:22:36.486507] I [server-helpers.c:343:do_lock_table_cleanup]=
 0-cloud-server: finodelk released on /one/var/25/images/disk.0<br>[2011-09=
-12 16:22:36.497377] I [server-helpers.c:485:do_fd_cleanup] 0-cloud-server:=
 fd cleanup on /one/var/25/images/disk.0<br>
[2011-09-12 16:22:36.497413] I [server-helpers.c:485:do_fd_cleanup] 0-cloud=
-server: fd cleanup on /one/var/30/images/disk.0<br><br>What exactlly are t=
hese &quot;fd cleanup&quot; messages?<br><br>And this morning still side ef=
fects:<br>
[2011-09-13 13:18:59.205656] E [afr-self-heal-data.c:637:afr_sh_data_fix] 0=
-cloud-replicate-0: Unable to self-heal contents of &#39;/one/var/25/images=
/disk.0&#39; (possible split-brain). Please delete the file from all but th=
e preferred subvolume.<br>
[2011-09-13 13:18:59.206990] W [afr-open.c:168:afr_open] 0-cloud-replicate-=
0: failed to open as split brain seen, returning EIO<br>[2011-09-13 13:18:5=
9.207019] W [fuse-bridge.c:585:fuse_fd_cbk] 0-glusterfs-fuse: 62384: OPEN()=
 /one/var/25/images/disk.0 =3D&gt; -1 (Input/output error)<br>
<br><br>From the application layer, following error appears:<br>Sep 13 13:1=
8:58 node4 libvirtd: 13:18:58.417: 1111: error : qemudDomainSaveImageClose:=
5668 : cannot close file: Bad file descriptor<br>Sep 13 13:18:59 node4 libv=
irtd: 13:18:59.207: 1113: error : virDomainDiskDefForeachPath:9010 : unable=
 to open disk path /srv/cloud/one/var//25/images/disk.0: Input/output error=
<br>
Sep 13 13:18:59 node4 libvirtd: 13:18:59.213: 1113: error : qemudDomainSave=
ImageClose:5668 : cannot close file: Bad file descriptor<br><br><br>Is ther=
e any configuration option to prevent this problem? Is there some enhanceme=
nts done in gluster 3.3 that can prevent the &quot;split brain&quot; issue =
happen again?<br>
<br>Thank you very much in advance for any link or information,<br>Samuel.<=
br>

--0016e64f80a82b7d0c04acd34c3a--


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux