Re: Files not available in all clients immediately

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Nope, thats the latest. But this should be fixed soon (at office time IST).
sorry for the inconvenience.

-amar

On Tue, Mar 18, 2008 at 5:07 PM, Claudio Cuqui <claudio@xxxxxxxxxxxxxxxx>
wrote:

> Hi Avati,
>
> I tried, but it don´t even allow me to start it:
>
> TLA Repo Revision: glusterfs--mainline--2.5--patch-709
> Time : 2008-03-18 20:52:32
> Signal Number : 11
>
> /C3Systems/gluster/bin/sbin/glusterfs -f
> /C3Systems/gluster/bin/etc/glusterfs/glusterfs-client.vol -l
> /C3Systems/gluster/bin-patch709/var/log/glusterfs/glusterfs.log -L
> WARNING /C3Systems/data/domains/webmail.pop.com.br/attachments
> volume fuse
>  type mount/fuse
>  option direct-io-mode 1
>  option entry-timeout 1
>  option attr-timeout 1
>  option mount-point /C3Systems/data/domains/webmail.pop.com.br/attachments
>  subvolumes iocache
> end-volume
>
> volume iocache
>  type performance/io-cache
>   option page-count 2
>  option page-size 256KB
>  subvolumes readahead
> end-volume
>
> volume readahead
>  type performance/read-ahead
>   option page-count 2
>  option page-size 1MB
>  subvolumes client
> end-volume
>
> volume client
>  type protocol/client
>   option remote-subvolume attachments
>  option remote-host 200.175.8.85
>   option transport-type tcp/client
> end-volume
>
> frame : type(1) op(34)
>
> /lib64/libc.so.6[0x3edca300c0]
> /lib64/libc.so.6(strcmp+0x0)[0x3edca75bd0]
>
> /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/mount/fuse.so[0x2aaaab302937]
>
> /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/mount/fuse.so[0x2aaaab302b42]
> /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/performance/io-
> cache.so(ioc_lookup_cbk+0x67)[0x2aaaab0f6557]
> /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0[0x2aaaaaab8344]
>
> /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/protocol/client.so(client_lookup_cbk+0x1b3)[0x2aaaaace93a3]
>
> /C3Systems/gluster/bin-patch709/lib/glusterfs/1.3.8/xlator/protocol/client.so(notify+0x8fc)[0x2aaaaace273c]
>
> /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0(sys_epoll_iteration+0xc0)[0x2aaaaaabdb90]
>
> /C3Systems/gluster/bin-patch709/lib/libglusterfs.so.0(poll_iteration+0x75)[0x2aaaaaabd095]
> [glusterfs](main+0x658)[0x4026b8]
> /lib64/libc.so.6(__libc_start_main+0xf4)[0x3edca1d8a4]
> [glusterfs][0x401b89]
> ---------
>
> Is there any other release that I should try ?
>
> Regards,
>
> Cuqui
>
> Anand Avati wrote:
> > Claudio,
> >  Can you try with glusterfs--mainline--2.5--patch-709 ? A similar
> > issue is addressed in that revision. We are interested to know if that
> > solves your issue as well.
> >
> > thanks,
> >
> > avati
> >
> > 2008/3/19, Claudio Cuqui <claudio@xxxxxxxxxxxxxxxx
> > <mailto:claudio@xxxxxxxxxxxxxxxx>>:
> >
> >     Hi there !
> >
> >     We are using gluster on an environment with multiple webservers
> >     and load
> >     balancer, where we have only one server and multiple clients (6).
> >     All servers are running Fedora Core 6 X86_64 with kernel
> >     2.6.22.14-72.fc6 (with exactly same packages installed in all
> server).
> >     The gluster version used is 1.3.8pre2 + 2.7.2glfs8 (both compiled
> >     locally). The underlying FS is reiserfs mounted with the follow
> >     options
> >     rw,noatime,nodiratime,notail. This filesystem has almost 4 thousand
> >     files from 2k - 10Mb in size. We are using gluster to export this
> >     filesystem for all other webservers. Below the config file used by
> >     gluster server:
> >
> >     ### Export volume "brick" with the contents of "/home/export"
> >     directory.
> >     volume attachments-nl
> >       type storage/posix                   # POSIX FS translator
> >       option directory
> >     /C3Systems/data/domains/webmail.pop.com.br/attachments
> >     end-volume
> >
> >     volume attachments
> >       type features/posix-locks
> >       subvolumes attachments-nl
> >       option mandatory on
> >     end-volume
> >
> >
> >     ### Add network serving capability to above brick.
> >     volume server
> >       type protocol/server
> >       option transport-type tcp/server     # For TCP/IP transport
> >       option client-volume-filename
> >     /C3Systems/gluster/bin/etc/glusterfs/glusterfs-client.vol
> >       subvolumes attachments-nl attachments
> >       option auth.ip.attachments-nl.allow * # Allow access to
> >     "attachments-nl" volume
> >       option auth.ip.attachments.allow * # Allow access to
> >     "attachments" volume
> >     end-volume
> >
> >     The problem happen when the LB sent the post (the uploaded file)
> >     to one
> >     webserver and than the next post goes to other webserver  that try
> to
> >     access the same file. When this happen, the other client got these
> >     messages:
> >
> >     PHP Warning:
> >
> fopen(/C3Systems/data/domains/c3systems.com.br/attachments/27gBgFQSIiOLDEo7AvxlpsFkqZw9jdnZ):
> >     failed to open stream: File Not Found.
> >     PHP Warning:
> >
> unlink(/C3Systems/data/domains/c3systems.com.br/attachments/5Dech7jNxjORZ2cZ9IAbR7kmgmgn2vTE):
> >     File Not Found.
> >
> >     LB is using RoundRobin to distribute the load between servers.
> >
> >     Below, you can find the gluster configuration file used by all
> >     clients:
> >
> >     ### file: client-volume.spec.sample
> >
> >     ##############################################
> >     ###  GlusterFS Client Volume Specification  ##
> >     ##############################################
> >
> >     #### CONFIG FILE RULES:
> >     ### "#" is comment character.
> >     ### - Config file is case sensitive
> >     ### - Options within a volume block can be in any order.
> >     ### - Spaces or tabs are used as delimitter within a line.
> >     ### - Each option should end within a line.
> >     ### - Missing or commented fields will assume default values.
> >     ### - Blank/commented lines are allowed.
> >     ### - Sub-volumes should already be defined above before referring.
> >
> >     ### Add client feature and attach to remote subvolume
> >     volume client
> >       type protocol/client
> >       option transport-type tcp/client     # for TCP/IP transport
> >     # option ib-verbs-work-request-send-size  1048576
> >     # option ib-verbs-work-request-send-count 16
> >     # option ib-verbs-work-request-recv-size  1048576
> >     # option ib-verbs-work-request-recv-count 16
> >     # option transport-type ib-sdp/client  # for Infiniband transport
> >     # option transport-type ib-verbs/client # for ib-verbs transport
> >       option remote-host 1.2.3.4 <http://1.2.3.4>      # IP address of
> >     the remote brick
> >     # option remote-port 6996              # default server port is 6996
> >
> >     # option transport-timeout 30          # seconds to wait for a reply
> >                                            # from server for each
> request
> >       option remote-subvolume attachments  # name of the remote volume
> >     end-volume
> >
> >     ### Add readahead feature
> >     volume readahead
> >       type performance/read-ahead
> >       option page-size 1MB     # unit in bytes
> >       option page-count 2       # cache per file  = (page-count x
> >     page-size)
> >       subvolumes client
> >     end-volume
> >
> >     ### Add IO-Cache feature
> >     volume iocache
> >       type performance/io-cache
> >       option page-size 256KB
> >       option page-count 2
> >       subvolumes readahead
> >     end-volume
> >
> >     ### Add writeback feature
> >     #volume writeback
> >     #  type performance/write-behind
> >     #  option aggregate-size 1MB
> >     #  option flush-behind off
> >     #  subvolumes iocache
> >     #end-volume
> >
> >     When I do the test manually, everything goes fine. What I think is
> >     happening is that gluster isn´t having enough time to sync all
> clients
> >     before clients trying to access the files (those servers are very
> busy
> >     ones.....they receive millions of requests per day).
> >
> >     Is this configuration appropriate for this situation ? a bug ? a
> >     feature
> >     ;-) ? Is there any option like the sync used in NFS that I can use
> in
> >     order guarantee that when the file is write down, all the clients
> >     already
> >     have it ?
> >
> >     TIA,
> >
> >     Claudio Cuqui
> >
> >
> >
> >
> >
> >     _______________________________________________
> >     Gluster-devel mailing list
> >     Gluster-devel@xxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxx>
> >     http://lists.nongnu.org/mailman/listinfo/gluster-devel
> >
> >
> >
> >
> > --
> > If I traveled to the end of the rainbow
> > As Dame Fortune did intend,
> > Murphy would be there to tell me
> > The pot's at the other end.
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel@xxxxxxxxxx
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>



-- 
Amar Tumballi
Gluster/GlusterFS Hacker
[bulde on #gluster/irc.gnu.org]
http://www.zresearch.com - Commoditizing Supercomputing and Superstorage!

[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux