Re: Question regarding Quincy mclock scheduler.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Philippe,

Your understanding is correct, 50% of IOPS are reserved for client
operations.
osd_mclock_max_capacity_iops_hdd defines the capacity per OSD.

There is a mClock queue for each OSD shard. The number of shards are
defined by osd_op_num_shards_hdd
<https://docs.ceph.com/en/latest/rados/configuration/osd-config-ref/#confval-osd_op_num_shards_hdd)>
which by default is set to 5.
So each queue has osd_mclock_max_capacity_iops_hdd/osd_op_num_shards_hdd
IOPS.
In your case, this would mean that the capacity of each mClock queue is
equal to 4578 IOPS (22889/5)
This would make osd_mclock_scheduler_client_res = 2289 (50%)

I hope that explains the numbers you are seeing.

We also have a few optimizations
<https://github.com/ceph/ceph/pull/48226/files> coming with regards to the
mClock profiles where we are increasing the reservation for client
operations in the high client profile. This change will also address the
inflated osd capacity numbers that are encountered in some cases with osd
bench.

Let me know if you have any other questions.

Regards,
Aishwarya

On Wed, Nov 9, 2022 at 1:46 PM philippe <philippe.vanhecke@xxxxxxxxx> wrote:

> Hi,
> We have a quincy 17.2.5 based cluster, and we have some question
> regarding the mclock iops scheduler.
> Looking into the documentation, the default profile is the HIGH_CLIENT_OPS
> that mean that 50% of IOPS for an OSD are reserved for clients operations.
>
> But looking into OSD configuration settings, it seems that this is not
> the case or probably there is something I don't understand.
>
> ceph config get osd.0 osd_mclock_profile
> high_client_ops
>
>
>
> ceph config show osd.0 | grep mclock
> osd_mclock_max_capacity_iops_hdd                 22889.222997       mon
> osd_mclock_scheduler_background_best_effort_lim  999999 default
> osd_mclock_scheduler_background_best_effort_res  1144
> default
> osd_mclock_scheduler_background_best_effort_wgt  2                 default
> osd_mclock_scheduler_background_recovery_lim     4578
> default
> osd_mclock_scheduler_background_recovery_res     1144
> default
> osd_mclock_scheduler_background_recovery_wgt     1                 default
> osd_mclock_scheduler_client_lim                  999999
> default
> osd_mclock_scheduler_client_res                  2289
> default
> osd_mclock_scheduler_client_wgt                  2                 default
>
> So i have osd_mclock_max_capacity_iops_hdd = 22889.222997 why
> osd_mclock_scheduler_client_res is not 11444 ?
>
> this value seem strange to me.
>
> Kr
> Philippe
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux