Thanks a bunch! So this would lead to that the mount point would "loose" it's connection ? Kindly //Marcus On Sat, Sep 13, 2008 at 7:49 PM, Amar S. Tumballi <amar@xxxxxxxxxxxxx>wrote: > this will always lead to EIO as it fails to meet the criteria for unify's > functioning. > > Unify wants a file to be present only on one of its subvolumes, and in this > case, you have done afr of (v1 v2), (v2 v3), (v3 v1), which means, if a file > is present on (v1 v2) pair, it will be seen by other two afrs too, (v2 in > second pair, and v1 in third pair), so unify sees file to be present on all > of its subvolume, and gets confused which file to open, and returns EIO. > > the fix is, you need to export two volumes (instead of currently present 1) > per server, and make pairs of (v1-1 v2-2), (v2-1 v3-2) (v3-1 v1-2), hope i > am clear > > Regards, > > > >> Client: >> volume v1 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.30 >> option remote-subvolume home >> end-volume >> >> volume v2 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.31 >> option remote-subvolume home >> end-volume >> >> volume v3 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.32 >> option remote-subvolume home >> end-volume >> >> volume afr-1 >> type cluster/afr >> subvolumes v1 v2 >> end-volume >> >> volume afr-2 >> type cluster/afr >> subvolumes v2 v3 >> end-volume >> >> volume afr-3 >> type cluster/afr >> subvolumes v3 v1 >> end-volume >> >> volume ns1 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.30 >> option remote-subvolume home-namespace >> end-volume >> >> volume ns2 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.31 >> option remote-subvolume home-namespace >> end-volume >> >> volume ns3 >> type protocol/client >> option transport-type tcp/client >> option remote-host 192.168.10.32 >> option remote-subvolume home-namespace >> end-volume >> >> volume namespace >> type cluster/afr >> subvolumes ns1 ns2 ns3 >> end-volume >> >> volume v >> type cluster/unify >> option scheduler rr >> option namespace namespace >> subvolumes afr-1 afr-2 afr-3 >> end-volume >> >> I really hope we have misconfigured something since that is the easiest >> fix :) >> >> Kindly >> >> //Marcus >> >> >> >> >> On Sat, Sep 13, 2008 at 12:50 AM, Amar S. Tumballi <amar@xxxxxxxxxxxxx>wrote: >> >>> Also which version of GlusterFS? >>> >>> ber@xxxxxxxxxxxxx> >>> >>>> may be configuration issue... lets start with config, what does you >>>> config look like on client and server? >>>> >>>> Marcus Herou wrote: >>>> >>>>> Lots of these on server >>>>> 2008-09-12 20:48:14 E [protocol.c:271:gf_block_unserialize_transport] >>>>> server: EOF from peer (*MailScanner has detected a possible fraud attempt >>>>> from "192.168.10.4:1007" claiming to be* *MailScanner warning: >>>>> numerical links are often malicious:* 192.168.10.4:1007 < >>>>> http://192.168.10.4:1007>) >>>>> ... >>>>> 2008-09-12 20:50:12 E [server-protocol.c:4153:server_closedir] server: >>>>> not getting enough data, returning EINVAL >>>>> ... >>>>> 2008-09-12 20:50:12 E [server-protocol.c:4148:server_closedir] server: >>>>> unresolved fd 6 >>>>> ... >>>>> 2008-09-12 20:51:47 E [protocol.c:271:gf_block_unserialize_transport] >>>>> server: EOF from peer (*MailScanner has detected a possible fraud attempt >>>>> from "192.168.10.10:1015" claiming to be* *MailScanner warning: >>>>> numerical links are often malicious:* 192.168.10.10:1015 < >>>>> http://192.168.10.10:1015>) >>>>> >>>>> ... >>>>> >>>>> And lots of these on client >>>>> >>>>> 2008-09-12 19:54:45 E [afr.c:2201:afr_open] home-namespace: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3954: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3956: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3958: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3987: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3989: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3991: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3993: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 C [client-protocol.c:212:call_bail] home3: bailing >>>>> transport >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3970: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3971: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3972: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 3974: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 4001: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 4002: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup] >>>>> home3: forced unwinding frame type(2) op(5) reply=@0x809abb0 >>>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: >>>>> no proper reply from server, returning ENOTCONN >>>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3: >>>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107 >>>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal >>>>> failed, returning EIO >>>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse: >>>>> 4004: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5) >>>>> 2008-09-12 19:55:01 E [unify.c:335:unify_lookup] home: returning ESTALE >>>>> for /rsyncer/.ssh/authorized_keys2: file count is 4 >>>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home: >>>>> /rsyncer/.ssh/authorized_keys2: found on home-namespace >>>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home: >>>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-2 >>>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home: >>>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-1 >>>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home: >>>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-3 >>>>> >>>>> >>>>> Both server and client are spitting out tons of these. Thought "E" was >>>>> Error level, seems like DEBUG ? >>>>> >>>>> Kindly >>>>> >>>>> //Marcus >>>>> >>>>> >>>>> >>>>> >>>>> On Fri, Sep 12, 2008 at 8:01 PM, Brian Taber <btaber@xxxxxxxxxxxxx<mailto: >>>>> btaber@xxxxxxxxxxxxx>> wrote: >>>>> >>>>> What do you see in your server and client logs for gluster? >>>>> >>>>> ------------------------- >>>>> Brian Taber >>>>> Owner/IT Specialist >>>>> Diverse Computer Group >>>>> Office: 774-206-5592 >>>>> Cell: 508-496-9221 >>>>> btaber@xxxxxxxxxxxxx <mailto:btaber@xxxxxxxxxxxxx> >>>>> >>>>> >>>>> >>>>> >>>>> Marcus Herou wrote: >>>>> > Hi. >>>>> > >>>>> > We have just recently installed a 3 node cluster with 16 SATA >>>>> disks each. >>>>> > >>>>> > We are using Hardy and the glusterfs-3.10 Ubuntu package on both >>>>> client(s) >>>>> > and server. >>>>> > >>>>> > We have only created one export (/home) yet since we want to >>>>> test it a while >>>>> > before putting it into a live high performance environment. >>>>> > >>>>> > The problem is currently that the client looses /home once a day >>>>> or so. This >>>>> > is really bad since this is a machine which all other connect to >>>>> with ssh >>>>> > keys thus making them unable to log in. >>>>> > >>>>> > Anyone seen something similar ? >>>>> > >>>>> > Kindly >>>>> > >>>>> > //Marcus >>>>> > _______________________________________________ >>>>> > Gluster-devel mailing list >>>>> > Gluster-devel@xxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxx> >>>>> > http://lists.nongnu.org/mailman/listinfo/gluster-devel >>>>> > >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Marcus Herou CTO and co-founder Tailsweep AB >>>>> +46702561312 >>>>> marcus.herou@xxxxxxxxxxxxx <mailto:marcus.herou@xxxxxxxxxxxxx> >>>>> http://www.tailsweep.com/ >>>>> http://blogg.tailsweep.com/ >>>>> >>>> _______________________________________________ >>>> Gluster-devel mailing list >>>> Gluster-devel@xxxxxxxxxx >>>> http://lists.nongnu.org/mailman/listinfo/gluster-devel >>>> >>> >>> >>> >>> -- >>> Amar Tumballi >>> Gluster/GlusterFS Hacker >>> [bulde on #gluster/irc.gnu.org] >>> http://www.zresearch.com - Commoditizing Super Storage! >>> >> >> >> >> -- >> Marcus Herou CTO and co-founder Tailsweep AB >> +46702561312 >> marcus.herou@xxxxxxxxxxxxx >> http://www.tailsweep.com/ >> http://blogg.tailsweep.com/ >> > > > > -- > Amar Tumballi > Gluster/GlusterFS Hacker > [bulde on #gluster/irc.gnu.org] > http://www.zresearch.com - Commoditizing Super Storage! > -- Marcus Herou CTO and co-founder Tailsweep AB +46702561312 marcus.herou@xxxxxxxxxxxxx http://www.tailsweep.com/ http://blogg.tailsweep.com/