Re: Performance question.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>
>
>       Caching won't help in the real appication I don't believe.
> Mostly it's read, crunch, write.  If I'm wrong here please let me
> know.  Although I don't believe it will hurt.  I'll give moving
> write-behind and io-cache to the client and see what happens.  Does it
> matter how they're stacked, i.e. the which comes first?


the order of them on the client side should not matter. (atleast with their
current featuresets)

avati


> You should also be loading io-cache on the client side with a decent
> > cache-size (like 256MB? depends on how much RAM you have to spare). this
> > will help re-read improve a lot.
> >
> > avati
> >
> > 2007/11/21, Anand Avati <avati@xxxxxxxxxxxxx>:
> >>
> >> Chris,
> >>  you shoud really be loading write-behind on the client side, that is
> wht
> >> improves write performance the most. do let us know the results with
> >> writebehind on the client side.
> >>
> >> avati
> >>
> >> 2007/11/21, Chris Johnson <johnson@xxxxxxxxxxxxxxxxxxx>:
> >>>
> >>>       Hi, again,
> >>>
> >>>       I asked about stack building philosophy.  Apparently there isn't
> >>> one.  So I tried a few things.  The configs are down the end here.
> >>>
> >>>       Two systems, CentOS5, both running fuse-devel-2.7.0-1 gluster
> >>> enhanced, glusterfs-1.3.5-2.  Both have gigabit ethernet, server runs
> >>> a SATABeast.  Currently I ge the following from from iozone.
> >>>
> >>> iozone -aN -r 32k -s 131072k -f /mnt/glusterfs/sdm1/junknstuff
> >>>
> >>>
> >>> random  random    bkwd  record  stride
> >>>                KB  reclen   write rewrite    read    reread    read
> >>> write    read rewrite    read   fwrite frewrite   fread  freread
> >>>            131072      32     589     587      345      343     818
> >>> 621     757     624     845      592      591     346      366
> >>>
> >>> Now, a similar test using NFS on a CentOS4.4 system running a 3ware
> >>> RAID card gives this
> >>>
> >>> iozone -aN -r 32k -s 131072k -f /space/sake/5/admin/junknstuff
> >>>
> >>>
> >>> random  random    bkwd  record  stride
> >>>                KB  reclen   write rewrite    read    reread    read
> >>> write    read rewrite    read   fwrite frewrite   fread  freread
> >>>            131072      32      27      26      292
> >>> 11      11      24     542       9     539       30       28     295
> >>> 11
> >>>
> >>> And you can see that the NFS system is faster.  Is this because of the
> >>> hardware 3ware RAID or is NFS really that much faster here?  Is there
> >>> a better way to stack this that would improve things?  And I tried
> with
> >>> and without striping.  No noticable difference in gluster performance.
> >>>
> >>>       Help appreciated.
> >>>
> >>> ============  server config
> >>>
> >>> volume brick1
> >>>    type storage/posix
> >>>    option directory /home/sdm1
> >>> end-volume
> >>>
> >>> volume brick2
> >>>    type storage/posix
> >>>    option directory /home/sdl1
> >>> end-volume
> >>>
> >>> volume brick3
> >>>    type storage/posix
> >>>    option directory /home/sdk1
> >>> end-volume
> >>>
> >>> volume brick4
> >>>    type storage/posix
> >>>    option directory /home/sdk1
> >>> end-volume
> >>>
> >>> volume ns-brick
> >>>    type storage/posix
> >>>    option directory /home/sdk1
> >>> end-volume
> >>>
> >>> volume stripe1
> >>>   type cluster/stripe
> >>>   subvolumes brick1 brick2
> >>> # option block-size *:10KB,
> >>> end-volume
> >>>
> >>> volume stripe2
> >>>   type cluster/stripe
> >>>   subvolumes brick3 brick4
> >>> # option block-size *:10KB,
> >>> end-volume
> >>>
> >>> volume unify0
> >>>   type cluster/unify
> >>>   subvolumes stripe1 stripe2
> >>>   option namespace ns-brick
> >>>   option scheduler rr
> >>> # option rr.limits.min-disk-free 5
> >>> end-volume
> >>>
> >>> volume iot
> >>>   type performance/io-threads
> >>>   subvolumes unify0
> >>>   option thread-count 8
> >>> end-volume
> >>>
> >>> volume writebehind
> >>>    type performance/write-behind
> >>>    option aggregate-size 131072 # in bytes
> >>>    subvolumes iot
> >>> end-volume
> >>>
> >>> volume readahead
> >>>    type performance/read-ahead
> >>> #  option page-size 65536 ### in bytes
> >>>    option page-size 128kb ### in bytes
> >>> #  option page-count 16 ### memory cache size is page-count x
> >>> page-size per file
> >>>    option page-count 2 ### memory cache size is page-count x page-size
> >>> per file
> >>>    subvolumes writebehind
> >>> end-volume
> >>>
> >>> volume server
> >>>    type protocol/server
> >>>    subvolumes readahead
> >>>    option transport-type tcp/server     # For TCP/IP transport
> >>> #  option client-volume-filename /etc/glusterfs/glusterfs-client.vol
> >>>    option auth.ip.readahead.allow *
> >>> end-volume
> >>>
> >>>
> >>> ============  client config
> >>>
> >>> volume client
> >>>    type protocol/client
> >>>    option transport-type tcp/client
> >>>    option remote-host xxx.xxx.xxx.xxx
> >>>    option remote-subvolume readahead
> >>> end-volume
> >>>
> >>>
> -------------------------------------------------------------------------------
> >>>
> >>> Chris Johnson               |Internet: johnson@xxxxxxxxxxxxxxxxxxx
> >>> Systems Administrator       |Web:
> http://www.nmr.mgh.harvard.edu/~johnson
> >>> <http://www.nmr.mgh.harvard.edu/%7Ejohnson>
> >>> NMR Center                  |Voice:    617.726.0949
> >>> Mass. General Hospital      |FAX:      617.726.7422
> >>> 149 (2301) 13th Street      |A compromise is a solution nobody is
> happy
> >>> with.
> >>> Charlestown, MA., 02129 USA |     Observation, Unknown
> >>>
> >>>
> -------------------------------------------------------------------------------
> >>>
> >>>
> >>> _______________________________________________
> >>> Gluster-devel mailing list
> >>> Gluster-devel@xxxxxxxxxx
> >>> http://lists.nongnu.org/mailman/listinfo/gluster-devel
> >>>
> >>
> >>
> >>
> >> --
> >> It always takes longer than you expect, even when you take into account
> >> Hofstadter's Law.
> >>
> >> -- Hofstadter's Law
> >
> >
> >
> >
> > --
> > It always takes longer than you expect, even when you take into account
> > Hofstadter's Law.
> >
> > -- Hofstadter's Law
> >
>
>
> -------------------------------------------------------------------------------
> Chris Johnson               |Internet: johnson@xxxxxxxxxxxxxxxxxxx
> Systems Administrator       |Web:
> http://www.nmr.mgh.harvard.edu/~johnson
> NMR Center                  |Voice:    617.726.0949
> Mass. General Hospital      |FAX:      617.726.7422
> 149 (2301) 13th Street      |For all sad words of tongue or pen, the
> saddest
> Charlestown, MA., 02129 USA |are these: "It might have been".  John G.
> Whittier
>
> -------------------------------------------------------------------------------
>



-- 
It always takes longer than you expect, even when you take into account
Hofstadter's Law.

-- Hofstadter's Law


[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux