Big I/O or gluster proccess problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



You should give a try to 3.0.4.

I used to run 3.0.2 before and I have the same behavior: when writting
in a directory, other processes couldn't access it untill the write is
finished..
Since I have upgraded to 3.0.4, I have not the issue any more.

cheers



On Thu, May 20, 2010 at 3:10 PM, Tejas N. Bhise <tejas at gluster.com> wrote:
> If I am not mistaken, you have a single server for glusterfs, and this is mirrored to a second ( non-glusterFS ) server using DRBD. If you have only a single server to export data from, why use Glusterfs ? Also, we don't officially support DRBD replication with glusterFS backend.
>
> Maybe you can consider GlusterFS replication across the two servers ?
>
>
> ----- Original Message -----
> From: "Ran" <smtp.test61 at gmail.com>
> To: "Tejas N. Bhise" <tejas at gluster.com>
> Sent: Thursday, May 20, 2010 6:30:21 PM
> Subject: Re: Big I/O or gluster proccess problem
>
>
> Tejas hi ,
> The 2 servers DRBD pair with HA so gluster actualy has 1 server that export 1 H.D ( 1 TB) this H.D is DRBD'd to the other server
> also this 1 TB hd is raid 1 with linux raid (i know its not optimal but rebust) in this setup if 1 servers go down the other continue , DRBD is more rebust then gluster replication specialy for VPS's etc..
> I didnt check iowait but the load of the server is about 5 while the CPU's are 10-50% only so that says it all(there are IO waits)
> I was thinking of breaking the raid 1 seens this H.D allready has full mirror with DRBD( to server 2) but im not sure it will resolve this problem seens with NFS its not the same , it slows thigs down but not to not functional .
>
> client vol file 192.168.0.9 is the HA IP of this pair iv tested also with plain config(no writebehind etc)
>
> <>
> # file: /etc/glusterfs/glusterfs.vol
> volume storage1-2
> type protocol/client
> option transport-type tcp
> option remote-host 192.168.0.9
> option remote-subvolume b1
> option ping-timeout 120
> option username ..........
> option password ..........
> end-volume
>
> volume cluster
> type cluster/distribute
> option lookup-unhashed yes
> subvolumes storage1-2
> end-volume
>
> #volume writebehind
> # type performance/write-behind
> # option cache-size 3MB
> # subvolumes cluster
> #end-volume
>
> #volume readahead
> # type performance/read-ahead
> # option page-count 4
> # subvolumes writebehind
> #end-volume
>
> volume iothreads
> type performance/io-threads
> option thread-count 4
> subvolumes cluster
> end-volume
>
>
> volume io-cache
> type performance/io-cache
> option cache-size 128MB
> option page-size 256KB #128KB is default option
> option force-revalidate-timeout 10 # default is 1
> subvolumes iothreads
> end-volume
>
>
> volume writebehind
> type performance/write-behind
> option aggregate-size 512KB # default is 0bytes
> option flush-behind on # default is 'off'
> subvolumes io-cache
> end-volume
>
> <>
>
> server vol file
>
> <>
> # file: /etc/glusterfs/glusterfs-server.vol
> volume posix
> type storage/posix
> option directory /data/gluster
> # option o-direct enable
> option background-unlink yes
> # option span-devices 8
> end-volume
>
> volume locks
> type features/locks
> subvolumes posix
> end-volume
>
> volume b1
> type performance/io-threads
> option thread-count 8
> subvolumes locks
> end-volume
>
> volume server
> type protocol/server
> option transport.socket.nodelay on
> option transport-type tcp
> # option auth.addr.b1.allow *
> option auth.login.b1.allow ..........
> option auth.login.gluster.password ................
> subvolumes b1
> end-volume
>
> <>
>
>
> 2010/5/20 Tejas N. Bhise < tejas at gluster.com >
>
>
> Ran,
>
> Can you please elaborate on - "2 servers in distrebute mod , each has 1 TB brick that replicate to each other using DRBD"
> Also, how many drives do you have and what does iowait look like when you write a big file ? Tell us more about the configs of your servers, share the volume files.
>
> Regards,
> Tejas.
>
>
>
>
> ----- Original Message -----
> From: "Ran" < smtp.test61 at gmail.com >
> To: Gluster-users at gluster.org
> Sent: Thursday, May 20, 2010 4:49:52 PM
> Subject: Big I/O or gluster proccess problem
>
> Hi all ,
> Our problem is simple but quite critical i posted few mounts ago regarding
> that issue and there was a good responses but not a fix
> What happen is that gluster stack when there is a big wrtite to it .
> For example time dd if=/dev/zero of=file bs=10240 count=1000000
> OR mv 20gig_file.img into gluster mount .
> When that happen the all storage freezes for the entire proccess , mails ,
> few vps's , simple dir etc..
> Our setup is quite simple at this point , 2 servers in distrebute mod , each
> has 1 TB brick that replicate to each other using DRBD .
> iv monitored closly everting on this big writes and noticed that its not mem
> , proc , network problem .
> Iv also checked the same setup with simple NFS and its not happening there .
> Anyone has any idea on how to fix this , if you cant write big files into
> gluster without making the storage unfunctional then you cant realy do
> anything .
>
> Please advise ,
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux