Re: Performance testing to tune osd recovery sleep

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Nana,

    Great testing.  One question,  do we have any idea why # of
degraded object keep increasing during recovery? and especially in
fio-rbd testing, there are multiple spike on each test.


Xiaoxi

2017-07-20 7:56 GMT+08:00 Neha Ojha <nojha@xxxxxxxxxx>:
> Hi all,
>
> The osd recovery sleep option has been re-implemented to make it
> asynchronous. This value determines the sleep time in seconds before
> the next recovery or backfill op.
>
> We have done rigorous testing on HDDs, SSDs and HDD+SSD setups, with
> both Filestore and Bluestore, in order to come up with better default
> values for this configuration option. Detailed performance results can
> be found here: https://drive.google.com/file/d/0B7I5sSnjMhmbN1ZOanF3T2JIZm8/view?usp=sharing
>
> Following are some of our conclusions:
>
> - We need separate default values of osd_recovery_sleep for HDDs, SSDs
> and hybrid(HDD+SSD) setups.
>
> - In setups with only HDDs, increasing the amount of sleep time, shows
> performance improvement. However, the total time taken by background
> recovery operation also increases. We found that recovery sleep value
> of 0.1 sec is optimal for these kind of setups.
>
> - In setups with only SSDs, with increase in sleep value, we do not
> see any drastic improvement in performance. Therefore, we have decided
> to keep the sleep value 0, and not pay any extra price in terms of
> increased recovery time.
>
> - In hybrid setups, where osd data is on HDDs and osd journal is on
> SSDs, increasing the sleep value more than 0 helps, but we would like
> to choose a default value lesser than 0.1 sec in order to not increase
> the recovery time too much. We haven't finalized this value yet.
> Introducing this configuration option would require some more work, in
> terms of determining whether the journal is on HDD or SSD.
>
> With https://github.com/ceph/ceph/pull/16328, we are introducing two
> new configuration options osd_recovery_sleep_hdd and
> osd_recovery_sleep_ssd.
>
> Please let me know if you any thoughts about it or have trouble
> accessing the link.
>
> Thanks,
> Neha
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux