Re: Slow performance during recovery operations

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>Recovery creates I/O performance drops in our VM too but it's manageable.
>What really hurts us are deep scrubs.
>Our current situation is Firefly 0.80.9 with a total of 24 identical OSDs
>evenly distributed on 4 servers with the following relevant configuration:
>
>    osd recovery max active      = 2
>    osd scrub load threshold      = 3
>    osd deep scrub interval       = 1209600 # 14 days
>    osd max backfills             = 4
>    osd disk thread ioprio class  = idle
>    osd disk thread ioprio priority = 7
>
>we managed to add several OSDs at once while deep scrubs were in practice
>disabled: we just increased deep scrub interval from 1 to 2 weeks which
>if I understand correctly had the effect of disabling them for 1 week
>(and indeed there were none while the backfilling
> went on for several hours).
>
>With these settings and no deep-scrubs the load increased a bit in the
>VMs doing non negligible I/Os but this was manageable. Even disk thread
>ioprio settings (which is what you want to get the ionice behaviour for
>deep scrubs) didn't seem to make much of a
> difference.

>From what I can tell, the 'osd disk thread' settings only applies to
scrubbing and 'snap trimming' operations.  I guess what I'm looking for is
a couple settings that may not exist yet:

  osd recovery thread ioprio class = idle
  osd recovery thread ioprio priority = 7

Or am I going down the wrong path here?


>Note : I don't believe Ceph will try to scatter the scrubs on the whole
>period you set with deep scrub interval, it seems its algorithm is much
>simpler than that and may lead to temporary salves of successive deep
>scrubs and it might generate some temporary
> I/O load which is hard to diagnose (by default scrubs and deep scrubs
>are logged by the OSD so you can correlate them with whatever you use to
>supervise your cluster).

It would definitely be nice to not have scrubs affect performance as much
either, so I'll probably add those to our config as well.


>I actually considered monitoring Ceph for backfills and using ceph set
>nodeep-scrub automatically when there are some and unset it when they
>disappear.

I'm pretty sure setting 'nodeep-scrub' doesn't cancel any current
deep-scrubs that are happening, but something like this would help prevent
the problem from getting worse.


Bryan


This E-mail and any of its attachments may contain Time Warner Cable proprietary information, which is privileged, confidential, or subject to copyright belonging to Time Warner Cable. This E-mail is intended solely for the use of the individual or entity to which it is addressed. If you are not the intended recipient of this E-mail, you are hereby notified that any dissemination, distribution, copying, or action taken in relation to the contents of and attachments to this E-mail is strictly prohibited and may be unlawful. If you have received this E-mail in error, please notify the sender immediately and permanently delete the original and any copy of this E-mail and any printout.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux