Re: snaptrim number of objects

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 23/08/2023 08:27, Sridhar Seshasayee wrote:

    This also leads me to agree with you there's 'something wrong' with
    the mclock scheduler. I was almost starting to suspect hardware issues
    or something like that, I was at my wit's end.

Could you update this thread with the exact quincy version by running:

$ ceph versions

and

$ ceph config show-with-defaults osd.N | grep osd_mclock

Please replace N with any valid OSD id.

I suspect that the quincy version you are running on doesn't
have the latest changes we made to the Reef upstream release.
Recent changes introduced significant improvements to the
mClock profiles and address slow recovery/backfill rates. The
improvements to the mClock profiles should also help throttle
snaptrim operations.

Snaptrim operation with mClock currently uses a static cost as
defined by osd_snap_trim_cost. There are improvements planned
around this soon. For e.g., the cost must be dynamic and reflect
the size of the object being trimmed.
-Sridhar


Here's the requested info, even though I'm going to stay on wpq for a while.


# ceph versions
{
    "mon": {
"ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)": 3
    },
    "mgr": {
"ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)": 3
    },
    "osd": {
"ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)": 117
    },
    "mds": {
"ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)": 3
    },
    "overall": {
"ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)": 126
    }
}

I'm using the docker image registry.<company>.local/quay-proxy/ceph/ceph@sha256:673b48521fd53e1b4bc7dda96335505c4d4b2e13d7bb92bf2e7782e2083094c9.


# ceph config show-with-defaults osd.0 | grep osd_mclock
osd_mclock_cost_per_byte_usec 0.000000



                      default
osd_mclock_cost_per_byte_usec_hdd 2.600000



                      default
osd_mclock_cost_per_byte_usec_ssd 0.011000



                      default
osd_mclock_cost_per_io_usec 0.000000



                      default
osd_mclock_cost_per_io_usec_hdd 11400.000000



                      default
osd_mclock_cost_per_io_usec_ssd 50.000000



                      default
osd_mclock_force_run_benchmark_on_init false



                      default
osd_mclock_iops_capacity_threshold_hdd 500.000000



                      default
osd_mclock_iops_capacity_threshold_ssd 80000.000000



                      default
osd_mclock_max_capacity_iops_hdd 250.000000



                      mon
osd_mclock_max_capacity_iops_ssd 21500.000000



                      default
osd_mclock_override_recovery_settings true



                      mon
osd_mclock_profile high_client_ops


                                  mon
osd_mclock_scheduler_anticipation_timeout 0.000000



                      default
osd_mclock_scheduler_background_best_effort_lim 999999



                      default
osd_mclock_scheduler_background_best_effort_res 1



                      default
osd_mclock_scheduler_background_best_effort_wgt 1



                      default
osd_mclock_scheduler_background_recovery_lim 999999



                      default
osd_mclock_scheduler_background_recovery_res 1



                      default
osd_mclock_scheduler_background_recovery_wgt 1



                      default
osd_mclock_scheduler_client_lim 999999



                      default
osd_mclock_scheduler_client_res 1



                      default
osd_mclock_scheduler_client_wgt 1



                      default
osd_mclock_skip_benchmark false



                      default


# ceph config show-with-defaults osd.0 | grep trim_cost
osd_snap_trim_cost 1048576


                                    default


Angelo.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux