Re: ceph/rbd benchmarks

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 24 Aug 2011, Marcus Sorensen wrote:
> I knew I read the acronym in a pdf somewhere (something from '08 I
> think), but I couldn't find it when I needed it, thanks.
> 
> Everything was running 3.1-rc1, I checked the kernel source before
> building and it already included the following patches, so I assumed I
> was good on the readahead thing.
> 
> https://patchwork.kernel.org/patch/1001462/
> https://patchwork.kernel.org/patch/1001432/

Yeah, those two help marginally, but the big fix is 

http://ceph.newdream.net/git/?p=ceph-client.git;a=commit;h=78e669966f994964581167c6e25c83d22ebb26c6

and you'd probably also want

http://ceph.newdream.net/git/?p=ceph-client.git;a=commitdiff;h=6468bfe33c8e674509e39e43fad6bc833398fee2

Those are in linux-next and will be sent upstream for 3.2-rc1.

Not sure if it's worth rerunning your tests just yet (I want to look at 
the MDS stuff still), but it should fix the sequential read performance.

sage


> 
> On Wed, Aug 24, 2011 at 10:46 AM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> > On Wed, 24 Aug 2011, Gregory Farnum wrote:
> >> On Wed, Aug 24, 2011 at 8:29 AM, Marcus Sorensen <shadowsor@xxxxxxxxx> wrote:
> >> > Just thought I'd share this basic testing I did, comparing cephfs 0.32
> >> > on 3.1-rc1 to nfs as well as rbd to iscsi. I'm sure you guys see a lot
> >> > of this. Any feedback would be appreciated.
> >> >
> >> > The data is here:
> >> >
> >> > http://learnitwithme.com/wp-content/uploads/2011/08/ceph-nfs-iscsi-benchmarks.ods
> >> >
> >> > and the writeup is here:
> >> >
> >> > http://learnitwithme.com/?p=303
> >>
> >> We see less of it than you'd think, actually. Thanks!
> >>
> >> To address a few things specifically
> >> Ceph is both the name of the project and of the POSIX-compliant
> >> filesystem. RADOS stands for Reliable Autonomous Distributed Object
> >> Store. Apparently we should publish this a bit more. :)
> >>
> >> Looks like most of the differences in your tests have to do with our
> >> relatively lousy read performance -- this is probably due to lousy
> >> readahead, which nobody's spent a lot of time optimizing as we focus
> >> on stability. Sage made some improvements a few weeks ago but I don't
> >> remember what version of stuff they ended up in. :) (Optimizing
> >> cross-server reads is hard!)
> >
> > The readahead improvements are in the 'master' branch of ceph-client.git,
> > and will go upstream for Linux 3.2-rc1 (I just missed the 3.1-rc1 cutoff).
> > In my tests I was limited by the wire speed with these patches.  I'm
> > guessing you were using 3.0 or earlier kernel?
> >
> > The file copy test was also surprising.  I think there is a regression
> > there somewhere, taking a look.
> >
> > sage
> >
> >
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 

[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux