Re: Re :Problem in GlusterFS VS NFS Read Performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It's pretty straightforward actually... I ran many tests like this
myself with iozone and other tools.

Since GlusterFS is a FUSE-based filesystem, it does not (and unless
FUSE itself changes, cannot) make use of the normal Linux VFS cache
system like NFS can. That's why the iocache translator exists. Sadly,
the iocache translator is just incapable of the same kind of
performance the kernel can do. In a single server situation, unless
FUSE is somehow modified to have the same caching capability as a
native filesystem, NFS is always going to win in iozone. I haven't
followed FUSE development at all, but I'm of the impression this isn't
likely to happen any time soon as it's not a simple change.

GFS *does* use the standard VFS cache I believe, but it isn't really
comparable to GlusterFS. They don't offer the same feature set. If I
were looking to replace NFS but stick to a single-server solution, I
might look at GFS. The performance should be very good. What GlusterFS
is really good at though, is scaling to multiple content servers.
That's where it'll be better than either GFS or NFS.

My experience has been that if you approach GlusterFS from a
performance perspective only, you might be disappointed. If speed is
the only goal, you'd probably get more of it by intelligently breaking
up your data across multiple servers by hand, and using GFS or NFS.
The problem is managing all that manual tuning is usually a severe
pain. GlusterFS can effectively handle all that *for* you and deliver
performance almost as good (probably better, unless you do a very good
job balancing load between servers by hand).


As to how you can increase GlusterFS's performance, here's a few
things I found during my testing. Most (all?) of this is in the
glusterfs wiki:

1) Don't use the stock RHEL/CentOS FUSE packages/modules. Use one
patched with the patch on the glusterfs wiki... it's a very noticeable
difference. If you haven't done this yet, I'd do it and benchmark the
difference.

2) Over an gigabit ethernet network, I saw no apparent gain with
either readahead or writeback translators. I wouldn't bother with them
unless running over Infiniband or 10-gig ethernet.

3) Use iocache on the client, which you're doing. I also use iothreads
on the server *and* client.... my usage generally involves multiple
simultaneous requests from multiple frontends (web server frontends
connecting to content server backends over GlusterFS).

4) The server-side is mounting a normal ext3/xfs/whatever
filesystem... it's perfectly capable of using the normal VFS cache to
its fullest extent. If you have any trouble saturating the link
between server and client (due to insufficient server disk speed),
stock it up with RAM.... it will all be used as cache, and should help
make saturating that link easier.

5) If the link is already saturated and you need more performance,
either add glusterfs servers (Unify or AFR maybe), or add RAM in the
frontends and allocate a bigger iocache. If you've got the hardware,
you could go for channel bonding or even 10-gig ethernet.

Good luck,
Jake

On Wed, Dec 31, 2008 at 2:54 AM, mohan L <l.mohanphy@xxxxxxxxx> wrote:
> Dear All,
>
> we are currently using NFS to meet data sharing requirements.Now we are
> facing  some performance and scalability problem ,so this form does not meet
> the requirements of our network(performance).So we are finding the possible
> solutions to increase the performance and scalability .To give very strong
> solution to NFS issue I have analysed two File System one is GlusterFS and
> another one is Red Hat GFS.we conclude that GlusterFS will increase the
> performance and scalability ,It has all the features we are looking .For the
> testing purpose I am benchmarking NFS and GlusterFS to get better
> performance .My benchmark result shows that GlusterFS give better
> performance ,but i am getting some unacceptable read performance . I am not
> able to understand how exactly the read operation performs NFS and GlusterFS
> .even I don't know anything i am doing wrong.here i am showing the benchmark
> result to get better idea of my read performance issuee .i have attached the
> result of NFS and GlusterFS read  values .any one can please go thro this
> and give me some valuable guide .It will make my benchmarking very effective
> .
>
> This my server and client Hardware and software :
>
> HARDWARE CONFIG:
>
> Processor core speed  : Intel(R) Celeron(R) CPU 1.70GHz
>
> Number of cores  : Single Core (not dual-core)
>
> RAM size  : 384MB(128MB+256MB)
>
> RAM type  : DDR
>
> RAM Speed  : 266 MHz (3.8 ns)
>
> Swap  : 1027MB
>
> Storage controller  : ATA device
>
> Disk model/size  : SAMSUNG SV4012H /40 GB,2 MB Cache,
>
> Storage speed  : 52.4 MB/sec
>
> Spindle Speed  : 5400 rpm(Revolution per Minute)
>
> NIC Type  : VIA Rhine III chipset IRQ 18
>
> NIC Speed  : 100 Mbps/Full-Duplex Card
>
> SOFTWARE:
>
> Operation System : Fedora Core 9 GNU/Linux
>
> Linux version  : 2.6.9-42
>
> Local FS  : Ext3
>
> NFS version  : 1.1.2
>
> GlusterFS version: glusterfs 1.3.8 built on Feb 3 2008
>
> Iozone  : iozone-3-5.fc9.i386 (File System Benchmark Tool)
>
> ttcp  : ttcp-1.12-18.fc9.i386(RAW throughput measurement Tool)
>
> This is the server and client vol files i am using the benchmarking
>
> #GlusterFS Server Volume Specification
>
> volume brick
>   type storage/posix                   # POSIX FS translator
>   option directory /bench        # /bench dir contains 25,000 files with
> size 10 KB 15KB
> end-volume
>
> volume iot
>   type performance/io-threads
>   option thread-count 4
>   option cache-size 8MB
>   subvolumes brick
> end-volume
>
> volume server
>   type protocol/server
>   option transport-type tcp/server
>   subvolumes iot
>   option auth.ip.brick.allow * # Allow access to "brick" volume
> end-volume
>
>
>
> # GlusterFS Client Volume Specification
>
> volume client
>   type protocol/client
>   option transport-type tcp/client
>   option remote-host 192.xxx.x.xxx
>   option remote-subvolume brick
> end-volume
>
> volume readahead
>   type performance/read-ahead
>   option page-size 128KB     # 256KB is the default option
>   option page-count 4     # cache per file  = (page-count x page-size)  2 is
> default option
>   subvolumes client
> end-volume
>
> volume iocache
>   type performance/io-cache
>   #option page-size 128KB   ## default is 32MB
>   option cache-size 256MB  #128KB is default option
>   option page-count 4
>   subvolumes readahead
> end-volume
>
> volume writeback
>   type performance/write-behind
>   option aggregate-size 128KB
>   option flush-behind on
>   subvolumes iocache
> end-volume
>
>
> I am confusing this result .I don't have idea how to trace and get good
> comparable result is read performance .I think I am miss understanding the
> buffer cache concepts .
>
> From attached NFS read result , I understand that I have 348MB RAM  and I am
> doing the benchmark file size rage 128KB to 1GB .So up to file size 256MB I
> am getting  buffer cache performance and file size 512MB ,1GB I am getting
> with in  link speed .But in case of GlusterFS I not able to understand what
> is happening .
>
> Please any one can help me .
>
>
>
> Thanks for your time
> Mohan
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel@xxxxxxxxxx
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
>




[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux