Re: [Performance] Improvement on DB Performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 21 May 2014, Mike Dawson wrote:
> Haomai,
> 
> Thanks for finding this!
> 
> 
> Sage,
> 
> We have a client that runs an io intensive, closed-source software package
> that seems to issue overzealous flushes which may benefit from this patch (or
> the other methods you mention). If you were to spin a wip build based on
> Dumpling, I'll be a willing tester.

Pushed wip-librbd-flush-dumpling, should be built shortly.

sage

> 
> Thanks,
> Mike Dawson
> 
> On 5/21/2014 11:23 AM, Sage Weil wrote:
> > On Wed, 21 May 2014, Haomai Wang wrote:
> > > I pushed the commit to fix this
> > > problem(https://github.com/ceph/ceph/pull/1848).
> > > 
> > > With test program(Each sync request is issued with ten write request),
> > > a significant improvement is noticed.
> > > 
> > > aio_flush                          sum: 914750     avg: 1239   count:
> > > 738      max: 4714   min: 1011
> > > flush_set                          sum: 904200     avg: 1225   count:
> > > 738      max: 4698   min: 999
> > > flush                              sum: 641648     avg: 173    count:
> > > 3690     max: 1340   min: 128
> > > 
> > > Compared to last mail, it reduce each aio_flush request to 1239 ns
> > > instead of 24145 ns.
> > 
> > Good catch!  That's a great improvement.
> > 
> > The patch looks clearly correct.  We can probably do even better by
> > putting the Objects on a list when they get the first dirty buffer so that
> > we only cycle through the dirty ones.  Or, have a global list of dirty
> > buffers (instead of dirty objects -> dirty buffers).
> > 
> > sage
> > 
> > > 
> > > I hope it's the root cause for db on rbd performance.
> > > 
> > > On Wed, May 21, 2014 at 6:15 PM, Haomai Wang <haomaiwang@xxxxxxxxx> wrote:
> > > > Hi all,
> > > > 
> > > > I remember there exists discuss about DB(mysql) performance on rbd.
> > > > Recently I test mysql-bench with rbd and found awful performance. So I
> > > > dive into it and find that main cause is "flush" request from guest.
> > > > As we know, applications such as mysql, ceph has own journal for
> > > > durable and journal usually send sync&direct io. If fs barrier is on,
> > > > each sync io operation make kernel issue "sync"(barrier) request to
> > > > block device. Here, qemu will call "rbd_aio_flush" to apply.
> > > > 
> > > > Via systemtap, I found a amazing thing:
> > > > aio_flush                          sum: 4177085    avg: 24145  count:
> > > > 173      max: 28172  min: 22747
> > > > flush_set                          sum: 4172116    avg: 24116  count:
> > > > 173      max: 28034  min: 22733
> > > > flush                              sum: 3029910    avg: 4      count:
> > > > 670477   max: 1893   min: 3
> > > > 
> > > > This statistic info is gathered in 5s. Most of consuming time is on
> > > > "ObjectCacher::flush". What's more, with time increasing, the flush
> > > > count will be increasing.
> > > > 
> > > > After view source, I find the root cause is "ObjectCacher::flush_set",
> > > > it will iterator the "object_set" and look for dirty buffer. And
> > > > "object_set"  contains all objects ever opened.  For example:
> > > > 
> > > > 2014-05-21 18:01:37.959013 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5919 flushed: 5
> > > > 2014-05-21 18:01:37.999698 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5919 flushed: 5
> > > > 2014-05-21 18:01:38.038405 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5920 flushed: 5
> > > > 2014-05-21 18:01:38.080118 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5920 flushed: 5
> > > > 2014-05-21 18:01:38.119792 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5921 flushed: 5
> > > > 2014-05-21 18:01:38.162004 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5922 flushed: 5
> > > > 2014-05-21 18:01:38.202755 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5923 flushed: 5
> > > > 2014-05-21 18:01:38.243880 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5923 flushed: 5
> > > > 2014-05-21 18:01:38.284399 7f785c7c6700  0 objectcacher flush_set
> > > > total: 5923 flushed: 5
> > > > 
> > > > These logs record the iteration info, the loop will check 5920 objects
> > > > but only 5 objects are dirty.
> > > > 
> > > > So I think the solution is make "ObjectCacher::flush_set" only
> > > > iterator the objects which is dirty.
> > > > 
> > > > --
> > > > Best Regards,
> > > > 
> > > > Wheat
> > > 
> > > 
> > > 
> > > --
> > > Best Regards,
> > > 
> > > Wheat
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > 
> > > 
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux