Re: Problems with self-heal

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We are testing the same setup:

DATA SERVER1 : disk1,disk2,disk3,disk4 unify-ed with namespace
(disk1-ns,disk2-ns afr-ed)
DATA SERVER2 : disk1,disk2,disk3,disk4 unify-ed with namespace
(disk1-ns,disk2-ns afr-ed)
DATA CLIENT: afr with SERVER1 and SERVER2

We have found one case where a file is duplicated in one of the dataservers.
You can reproduce it following the next steps:

- CLIENT: create a new file called "file" on glusterfs
- SERVER2: delete file from the namespaces
- CLIENT: force self-heal with touch /path_to_glusterfs/file

After touch command ends "file" is duplicated on SERVER2.
I don't know if this is normal.

Sorry for my bad english :P


On Tue, Feb 26, 2008 at 2:08 PM, Raghavendra G <raghavendra.hg@xxxxxxxxx>
wrote:

> Hi,
> please do use glusterfs--mainline--2.5--patch-685, since I've fixed some
> compiler warnings in this patch.
>
> regards,
>
>
> On Tue, Feb 26, 2008 at 1:35 PM, Raghavendra G <raghavendra.hg@xxxxxxxxx>
> wrote:
>
> > Hi,
> >
> > Fix committed in glusterfs--mainline--patch-683
> >
> > regards,
> >
> >
> > On Tue, Feb 26, 2008 at 12:15 PM, E-Comm Factory <
> > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> >
> > >
> > > Happy to read it. Thank you for your time!
> > >
> > > Congratulations for this good job.
> > >
> > >
> > > On Tue, Feb 26, 2008 at 5:33 AM, Raghavendra G <
> > > raghavendra.hg@xxxxxxxxx> wrote:
> > >
> > > > Hi,
> > > >
> > > > Thanks for the access. We have found the issue and will soon commit
> > > > a fix to tla repository.
> > > >
> > > > regards,
> > > >
> > > >
> > > > On Thu, Feb 21, 2008 at 10:20 PM, E-Comm Factory <
> > > > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> > > >
> > > > > Hello Raghavendra,
> > > > >
> > > > > Now client box is down because we are in non working hours.
> > > > > You can check it from 9:00 to 14:00 and from 14:00 to 19:00 GMT+1.
> > > > >
> > > > > Thanks in advance,
> > > > >
> > > > >
> > > > > **
> > > > > On Thu, Feb 21, 2008 at 12:00 PM, Raghavendra G <
> > > > > raghavendra.hg@xxxxxxxxx> wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > I tried to reproduce the issue on my local system with the
> > > > > > configuration provided by you without any success. Self-heal is working fine
> > > > > > on my system. Is it possible to provide us access to your system, so that we
> > > > > > can look at the issue first hand?
> > > > > >
> > > > > > regards,
> > > > > >
> > > > > >
> > > > > > On Thu, Feb 21, 2008 at 1:34 PM, E-Comm Factory <
> > > > > > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> > > > > >
> > > > > > > This is my desired configuration:
> > > > > > >
> > > > > > > - Server Box 1: 4 HD: with 4 storage partition a 4 namespace
> > > > > > > partitions
> > > > > > >
> > > > > > >  AFR with the 4 ns partitions
> > > > > > >  UNIFY with 4 ds partitions and previous AFR volume
> > > > > > >
> > > > > > > - Server Box 2: 4 HD: with 4 storage partition a 4 namespace
> > > > > > > partitions
> > > > > > >
> > > > > > >  AFR with the 4 ns partitions
> > > > > > >  UNIFY with 4 ds partitions and previous AFR volume
> > > > > > >
> > > > > > > - Client Box:
> > > > > > >
> > > > > > >  AFR with the two UNIFY volumes
> > > > > > >
> > > > > > >
> > > > > > > Seems that i CAN'T use an afr volume under my two unify
> > > > > > > volumes if then the
> > > > > > > client use afr again with this 2 unified volumes. So i have
> > > > > > > deleted the
> > > > > > > namespace afr and now i use a single posix volume as ns.
> > > > > > >
> > > > > > > Is it the way it shold be done? or maybe i missed something.
> > > > > > >
> > > > > > > I will gratefully accept eny help.
> > > > > > >
> > > > > > > On Tue, Feb 19, 2008 at 7:12 PM, E-Comm Factory <
> > > > > > > sistemas@xxxxxxxxxxxxxxxxx>
> > > > > > > wrote:
> > > > > > >
> > > > > > > >
> > > > > > > > More info about this issue:
> > > > > > > >
> > > > > > > > 2008-02-19 19:06:09 D [inode.c:356:__active_inode]
> > > > > > > disk-fs11/inode:
> > > > > > > > activating inode(1064961), lru=6/1024
> > > > > > > > 2008-02-19 19:06:09 D [lock.c:128:mop_lock_impl] lock: Lock
> > > > > > > request to
> > > > > > > > /file.img queued
> > > > > > > > 2008-02-19 19:09:10 E [protocol.c
> > > > > > > :259:gf_block_unserialize_transport]
> > > > > > > > server: EOF from peer (192.168.1.103:1023)
> > > > > > > > 2008-02-19 19:09:10 C [tcp.c:87:tcp_disconnect] server:
> > > > > > > connection
> > > > > > > > disconnected
> > > > > > > > 2008-02-19 19:09:10 D [tcp-server.c:145:tcp_server_notify]
> > > > > > > server:
> > > > > > > > Registering socket (4) for new transport object of
> > > > > > > 192.168.1.103
> > > > > > > > 2008-02-19 19:09:10 D [ip.c:98:gf_auth] disk-fs11: allowed =
> > > > > > > "*", received
> > > > > > > > ip addr = "192.168.1.103"
> > > > > > > > 2008-02-19 19:09:10 D [server-protocol.c:5487:mop_setvolume]
> > > > > > > server:
> > > > > > > > accepted client from 192.168.1.103:1021
> > > > > > > > 2008-02-19 19:09:10 E [server-protocol.c:183:generic_reply]
> > > > > > > server:
> > > > > > > > transport_writev failed
> > > > > > > > 2008-02-19 19:09:10 D [server-protocol.c
> > > > > > > :6067:server_protocol_cleanup]
> > > > > > > > server: cleaned up transport state for client
> > > > > > > 192.168.1.103:1023
> > > > > > > > 2008-02-19 19:09:10 D [tcp-server.c:248:gf_transport_fini]
> > > > > > > server:
> > > > > > > > destroying transport object for 192.168.1.103:1023 (fd=4)
> > > > > > > > 2008-02-19 19:09:10 D [inode.c:386:__passive_inode]
> > > > > > > disk-fs11/inode:
> > > > > > > > passivating inode(1064961), lru=7/1024
> > > > > > > > 2008-02-19 19:09:10 D [inode.c:356:__active_inode]
> > > > > > > disk-fs11/inode:
> > > > > > > > activating inode(1064961), lru=6/1024
> > > > > > > > 2008-02-19 19:09:12 D [inode.c:386:__passive_inode]
> > > > > > > disk-fs11/inode:
> > > > > > > > passivating inode(1064961), lru=7/1024
> > > > > > > >
> > > > > > > > I don't know if it could help. Any help will be apreciated!
> > > > > > > >
> > > > > > > > Finally, thanks guys for this fantastic project. GlusterFS
> > > > > > > is amazing.
> > > > > > > >
> > > > > > > >
> > > > > > > > On Tue, Feb 19, 2008 at 5:26 PM, E-Comm Factory <
> > > > > > > > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> > > > > > > >
> > > > > > > > >
> > > > > > > > > I also tested glusterfs-mainline-2.5 PATCH 674 with the
> > > > > > > same results.
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > On Tue, Feb 19, 2008 at 5:23 PM, Toni Valverde <
> > > > > > > > > tvalverde@xxxxxxxxxxxxxxxxx> wrote:
> > > > > > > > >
> > > > > > > > > > I also tested glusterfs-mainline-2.5 PATCH 674 with the
> > > > > > > same
> > > > > > > > > > results.
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On Tue, Feb 19, 2008 at 4:05 PM, E-Comm Factory <
> > > > > > > > > > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> > > > > > > > > >
> > > > > > > > > > > thanks amar
> > > > > > > > > > >
> > > > > > > > > > > - glusterfs--mainline--2.5 PATCH 665
> > > > > > > > > > > - fuse-2.7.2glfs8
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > On Mon, Feb 18, 2008 at 8:48 PM, Amar S. Tumballi <
> > > > > > > > > > > amar@xxxxxxxxxxxxx> wrote:
> > > > > > > > > > >
> > > > > > > > > > > > Can you please let us know what version of fuse and
> > > > > > > glusterfs you
> > > > > > > > > > > > are running these tests from?
> > > > > > > > > > > >
> > > > > > > > > > > > -amar
> > > > > > > > > > > >
> > > > > > > > > > > > On Feb 18, 2008 11:03 PM, E-Comm Factory <
> > > > > > > > > > > > sistemas@xxxxxxxxxxxxxxxxx> wrote:
> > > > > > > > > > > >
> > > > > > > > > > > > > Hello,
> > > > > > > > > > > > >
> > > > > > > > > > > > > I have 2 boxes with 4 unified disks (so i have 2
> > > > > > > volumes). Then,
> > > > > > > > > > > > > in client
> > > > > > > > > > > > > side, i have set afr with this 2 virtual volumes.
> > > > > > > > > > > > >
> > > > > > > > > > > > > For testing purposes I've deleted one file on the
> > > > > > > second afr
> > > > > > > > > > > > > volume and then
> > > > > > > > > > > > > tried to self-heal the global afr but it crashes
> > > > > > > with this
> > > > > > > > > > > > > error:
> > > > > > > > > > > > >
> > > > > > > > > > > > > [afr.c:2754:afr_open] disk: self heal failed,
> > > > > > > returning EIOç
> > > > > > > > > > > > > [fuse-bridge.c:675:fuse_fd_cbk] glusterfs-fuse:
> > > > > > > 98:
> > > > > > > > > > > > > /fichero4.img => -1 (5)
> > > > > > > > > > > > >
> > > > > > > > > > > > > An strace to the pid of the glusterfs-server
> > > > > > > running on the
> > > > > > > > > > > > > first afr volume
> > > > > > > > > > > > > crashes too when selfhealing:
> > > > > > > > > > > > >
> > > > > > > > > > > > > epoll_wait(6, {{EPOLLIN, {u32=6304624,
> > > > > > > u64=6304624}}}, 2,
> > > > > > > > > > > > > 4294967295) = 1
> > > > > > > > > > > > > read(4, out of memory
> > > > > > > > > > > > > 0x7fff9a3b1d90, 113)            = 113
> > > > > > > > > > > > > read(4, Segmentation fault
> > > > > > > > > > > > >
> > > > > > > > > > > > > My server config file (same on both server boxes):
> > > > > > > > > > > > >
> > > > > > > > > > > > > # datastores
> > > > > > > > > > > > > volume disk1
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk1
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > > volume disk2
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk2
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > > volume disk3
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk3
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > > volume disk4
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk4
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > # namespaces
> > > > > > > > > > > > > volume disk1-ns
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk1-ns
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > > volume disk2-ns
> > > > > > > > > > > > >  type storage/posix
> > > > > > > > > > > > >  option directory /mnt/disk2-ns
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > > #volume disk3-ns
> > > > > > > > > > > > > #  type storage/posix
> > > > > > > > > > > > > #  option directory /mnt/disk3-ns
> > > > > > > > > > > > > #end-volume
> > > > > > > > > > > > > #volume disk4-ns
> > > > > > > > > > > > > #  type storage/posix
> > > > > > > > > > > > > #  option directory /mnt/disk4-ns
> > > > > > > > > > > > > #end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > # afr de namespaces
> > > > > > > > > > > > > volume disk-ns-afr
> > > > > > > > > > > > >  type cluster/afr
> > > > > > > > > > > > >  subvolumes disk1-ns disk2-ns
> > > > > > > > > > > > >  option scheduler random
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > # unify de datastores
> > > > > > > > > > > > > volume disk-unify
> > > > > > > > > > > > >  type cluster/unify
> > > > > > > > > > > > >  subvolumes disk1 disk2 disk3 disk4
> > > > > > > > > > > > >  option namespace disk-ns-afr
> > > > > > > > > > > > >  option scheduler rr
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > # performace para el disco
> > > > > > > > > > > > > volume disk-fs11
> > > > > > > > > > > > >  type performance/io-threads
> > > > > > > > > > > > >  option thread-count 8
> > > > > > > > > > > > >  option cache-size 64MB
> > > > > > > > > > > > >  subvolumes disk-unify
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > # permitimos acceso a cualquier cliente
> > > > > > > > > > > > > volume server
> > > > > > > > > > > > >  type protocol/server
> > > > > > > > > > > > >  option transport-type tcp/server
> > > > > > > > > > > > >  subvolumes disk-fs11
> > > > > > > > > > > > >  option auth.ip.disk-fs11.allow *
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > My client config file:
> > > > > > > > > > > > >
> > > > > > > > > > > > > volume disk-fs11
> > > > > > > > > > > > >  type protocol/client
> > > > > > > > > > > > >  option transport-type tcp/client
> > > > > > > > > > > > >  option remote-host 192.168.1.34
> > > > > > > > > > > > >  option remote-subvolume disk-fs11
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > volume disk-fs12
> > > > > > > > > > > > >  type protocol/client
> > > > > > > > > > > > >  option transport-type tcp/client
> > > > > > > > > > > > >  option remote-host 192.168.1.35
> > > > > > > > > > > > >  option remote-subvolume disk-fs12
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > volume disk
> > > > > > > > > > > > >  type cluster/afr
> > > > > > > > > > > > >  subvolumes disk-fs11 disk-fs12
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > volume trace
> > > > > > > > > > > > >  type debug/trace
> > > > > > > > > > > > >  subvolumes disk
> > > > > > > > > > > > > #  option includes
> > > > > > > open,close,create,readdir,opendir,closedir
> > > > > > > > > > > > > #  option excludes lookup,read,write
> > > > > > > > > > > > > end-volume
> > > > > > > > > > > > >
> > > > > > > > > > > > > Anyone could help me?
> > > > > > > > > > > > >
> > > > > > > > > > > > > Thanks in advance.
> > > > > > > > > > > > >
> > > > > > > > > > > > > --
> > > > > > > > > > > > > ecomm
> > > > > > > > > > > > > sistemas@xxxxxxxxxxxxxxxxx
> > > > > > > > > > > > >  _______________________________________________
> > > > > > > > > > > > > Gluster-devel mailing list
> > > > > > > > > > > > > Gluster-devel@xxxxxxxxxx
> > > > > > > > > > > > >
> > > > > > > http://lists.nongnu.org/mailman/listinfo/gluster-devel
> > > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > --
> > > > > > > > > > > > Amar Tumballi
> > > > > > > > > > > > Gluster/GlusterFS Hacker
> > > > > > > > > > > > [bulde on #gluster/irc.gnu.org]
> > > > > > > > > > > > http://www.zresearch.com - Commoditizing
> > > > > > > Supercomputing and
> > > > > > > > > > > > Superstorage!
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > --
> > > > > > > > > > > Toni Valverde
> > > > > > > > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > > > > > > > >
> > > > > > > > > > > Electronic Commerce Factory S.L.
> > > > > > > > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > > > > > > > 28008 - Madrid
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > --
> > > > > > > > > > Toni Valverde
> > > > > > > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > > > > > > >
> > > > > > > > > > Electronic Commerce Factory S.L.
> > > > > > > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > > > > > > 28008 - Madrid
> > > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > --
> > > > > > > > > Toni Valverde
> > > > > > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > > > > > >
> > > > > > > > > Electronic Commerce Factory S.L.
> > > > > > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > > > > > 28008 - Madrid
> > > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > > --
> > > > > > > > Toni Valverde
> > > > > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > > > > >
> > > > > > > > Electronic Commerce Factory S.L.
> > > > > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > > > > 28008 - Madrid
> > > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > --
> > > > > > > Toni Valverde
> > > > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > > > >
> > > > > > > Electronic Commerce Factory S.L.
> > > > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > > > 28008 - Madrid
> > > > > > > _______________________________________________
> > > > > > > Gluster-devel mailing list
> > > > > > > Gluster-devel@xxxxxxxxxx
> > > > > > > http://lists.nongnu.org/mailman/listinfo/gluster-devel
> > > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > --
> > > > > > Raghavendra G
> > > > > >
> > > > > > A centipede was happy quite, until a toad in fun,
> > > > > > Said, "Prey, which leg comes after which?",
> > > > > > This raised his doubts to such a pitch,
> > > > > > He fell flat into the ditch,
> > > > > > Not knowing how to run.
> > > > > > -Anonymous
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > Toni Valverde
> > > > > tvalverde@xxxxxxxxxxxxxxxxx
> > > > >
> > > > > Electronic Commerce Factory S.L.
> > > > > C/Martin de los Heros, 59bis - 1º nº 8
> > > > > 28008 - Madrid
> > > > >
> > > >
> > > >
> > > >
> > > > --
> > > > Raghavendra G
> > > >
> > > > A centipede was happy quite, until a toad in fun,
> > > > Said, "Prey, which leg comes after which?",
> > > > This raised his doubts to such a pitch,
> > > > He fell flat into the ditch,
> > > > Not knowing how to run.
> > > > -Anonymous
> > > >
> > >
> > >
> > >
> > > --
> > > Toni Valverde
> > > tvalverde@xxxxxxxxxxxxxxxxx
> > >
> > > Electronic Commerce Factory S.L.
> > > C/Martin de los Heros, 59bis - 1º nº 8
> > > 28008 - Madrid
> > >
> >
> >
> >
> > --
> > Raghavendra G
> >
> > A centipede was happy quite, until a toad in fun,
> > Said, "Prey, which leg comes after which?",
> > This raised his doubts to such a pitch,
> > He fell flat into the ditch,
> > Not knowing how to run.
> > -Anonymous
> >
>
>
>
> --
> Raghavendra G
>
> A centipede was happy quite, until a toad in fun,
> Said, "Prey, which leg comes after which?",
> This raised his doubts to such a pitch,
> He fell flat into the ditch,
> Not knowing how to run.
> -Anonymous
>



-- 
Toni Valverde
tvalverde@xxxxxxxxxxxxxxxxx

Electronic Commerce Factory S.L.
C/Martin de los Heros, 59bis - 1º nº 8
28008 - Madrid


[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux