Big I/O or gluster proccess problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



If I am not mistaken, you have a single server for glusterfs, and this is mirrored to a second ( non-glusterFS ) server using DRBD. If you have only a single server to export data from, why use Glusterfs ? Also, we don't officially support DRBD replication with glusterFS backend. 

Maybe you can consider GlusterFS replication across the two servers ? 


----- Original Message ----- 
From: "Ran" <smtp.test61 at gmail.com> 
To: "Tejas N. Bhise" <tejas at gluster.com> 
Sent: Thursday, May 20, 2010 6:30:21 PM 
Subject: Re: Big I/O or gluster proccess problem 


Tejas hi , 
The 2 servers DRBD pair with HA so gluster actualy has 1 server that export 1 H.D ( 1 TB) this H.D is DRBD'd to the other server 
also this 1 TB hd is raid 1 with linux raid (i know its not optimal but rebust) in this setup if 1 servers go down the other continue , DRBD is more rebust then gluster replication specialy for VPS's etc.. 
I didnt check iowait but the load of the server is about 5 while the CPU's are 10-50% only so that says it all(there are IO waits) 
I was thinking of breaking the raid 1 seens this H.D allready has full mirror with DRBD( to server 2) but im not sure it will resolve this problem seens with NFS its not the same , it slows thigs down but not to not functional . 

client vol file 192.168.0.9 is the HA IP of this pair iv tested also with plain config(no writebehind etc) 

<> 
# file: /etc/glusterfs/glusterfs.vol 
volume storage1-2 
type protocol/client 
option transport-type tcp 
option remote-host 192.168.0.9 
option remote-subvolume b1 
option ping-timeout 120 
option username .......... 
option password .......... 
end-volume 

volume cluster 
type cluster/distribute 
option lookup-unhashed yes 
subvolumes storage1-2 
end-volume 

#volume writebehind 
# type performance/write-behind 
# option cache-size 3MB 
# subvolumes cluster 
#end-volume 

#volume readahead 
# type performance/read-ahead 
# option page-count 4 
# subvolumes writebehind 
#end-volume 

volume iothreads 
type performance/io-threads 
option thread-count 4 
subvolumes cluster 
end-volume 


volume io-cache 
type performance/io-cache 
option cache-size 128MB 
option page-size 256KB #128KB is default option 
option force-revalidate-timeout 10 # default is 1 
subvolumes iothreads 
end-volume 


volume writebehind 
type performance/write-behind 
option aggregate-size 512KB # default is 0bytes 
option flush-behind on # default is 'off' 
subvolumes io-cache 
end-volume 

<> 

server vol file 

<> 
# file: /etc/glusterfs/glusterfs-server.vol 
volume posix 
type storage/posix 
option directory /data/gluster 
# option o-direct enable 
option background-unlink yes 
# option span-devices 8 
end-volume 

volume locks 
type features/locks 
subvolumes posix 
end-volume 

volume b1 
type performance/io-threads 
option thread-count 8 
subvolumes locks 
end-volume 

volume server 
type protocol/server 
option transport.socket.nodelay on 
option transport-type tcp 
# option auth.addr.b1.allow * 
option auth.login.b1.allow .......... 
option auth.login.gluster.password ................ 
subvolumes b1 
end-volume 

<> 


2010/5/20 Tejas N. Bhise < tejas at gluster.com > 


Ran, 

Can you please elaborate on - "2 servers in distrebute mod , each has 1 TB brick that replicate to each other using DRBD" 
Also, how many drives do you have and what does iowait look like when you write a big file ? Tell us more about the configs of your servers, share the volume files. 

Regards, 
Tejas. 




----- Original Message ----- 
From: "Ran" < smtp.test61 at gmail.com > 
To: Gluster-users at gluster.org 
Sent: Thursday, May 20, 2010 4:49:52 PM 
Subject: Big I/O or gluster proccess problem 

Hi all , 
Our problem is simple but quite critical i posted few mounts ago regarding 
that issue and there was a good responses but not a fix 
What happen is that gluster stack when there is a big wrtite to it . 
For example time dd if=/dev/zero of=file bs=10240 count=1000000 
OR mv 20gig_file.img into gluster mount . 
When that happen the all storage freezes for the entire proccess , mails , 
few vps's , simple dir etc.. 
Our setup is quite simple at this point , 2 servers in distrebute mod , each 
has 1 TB brick that replicate to each other using DRBD . 
iv monitored closly everting on this big writes and noticed that its not mem 
, proc , network problem . 
Iv also checked the same setup with simple NFS and its not happening there . 
Anyone has any idea on how to fix this , if you cant write big files into 
gluster without making the storage unfunctional then you cant realy do 
anything . 

Please advise , 

_______________________________________________ 
Gluster-users mailing list 
Gluster-users at gluster.org 
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users 


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux