Re: PG_SLOW_SNAP_TRIMMING and possible storage leakage on 16.2.5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dan,

  Thank you for replying. Since I posted I did some more digging. It
really seemed as if snaptrim simply wasn't being processed. The output
of "ceph health detail" showed that PG 3.9b had the longest queue. I
examined this PG and saw that it's primary was osd.8 so I manually
restarted that daemon. This seems to have kicked off snaptrim on some PGs:

----SNIP----
1513 pgs: 1 active+clean+scrubbing, 1 active+clean+scrubbing+snaptrim,
44 active+clean+snaptrim, 1 active+clean+scrubbing+deep+snaptrim_wait,
1406 active+clean, 2 active+clean+scrubbing+deep, 58
active+clean+snaptrim_wait; 114 TiB data, 344 TiB used, 93 TiB / 437 TiB
avail; 2.0 KiB/s rd, 64 KiB/s wr, 5 op/s
----SNIP----

I can see the "snaptrimq_len* value decreasing for that PG now. I will
look into the issue you posted as well as repeering the PGs. Does an osd
restart resulting in snaptrim proceeding seem consistent with the
behavior you saw?

I notice in the bug report you linked, that you are somehow monitoring
snaptrimq with grafana. Is this a global value that is readily avilable
for monitoring or are you calculating this somehow. If there is an easy
way to access it, I would greatly appreciate instructions.

Thank you,

-David

On 1/24/22 11:53 AM, Dan van der Ster wrote:
> Hi David,
>
> We observed the same here: https://tracker.ceph.com/issues/52026
> You can poke the trimming by repeering the PGs.
>
> Also, depending on your hardware, the defaults for osd_snap_trim_sleep
> might be far too conservative.
> We use osd_snap_trim_sleep = 0.1 on our mixed hdd block / ssd block.db OSDs.
>
> Cheers, Dan
>
> On Mon, Jan 24, 2022 at 4:54 PM David Prude <david@xxxxxxxxxxxxxxxx> wrote:
>> Hello,
>>
>>    We have a 5-node, 30 hdd (6 hdds/node) cluster running 16.2.5. We
>> utilize a snapshot scheme within cephfs that results in 24 hourly
>> snapshots, 7 daily snapshots, and 2 weekly snapshots. This has been
>> running without overt issues for several months. As of this weekend, we
>> started receiving a  PG_SLOW_SNAP_TRIMMING warning on a single PG. Over
>> the last 24 hours we are now seeing that this warning is associated with
>> 123 of our 1513 PGs. As recommended by the output of "ceph health
>> detail" we have tried tuning the following from their default values:
>>
>> osd_pg_max_concurrent_snap_trims=4 (default 2)
>> osd_snap_trim_sleep_hdd=3 (default 5)
>> osd_snap_trim_sleep=0.5 (default 0, it was suggested somewhere in a
>> search that 0 actually disables trim?)
>>
>> I am uncertain how to best measure if the above is having an effect on
>> the trimming process. I am unclear on how to clearly monitor the
>> progress of the snaptrim process or even of the total queue depth.
>> Interestingly, "ceph pg stat" does not show any PGs in the snaptrim state:
>>
>> ----SNIP----
>> 1513 pgs: 2 active+clean+scrubbing+deep, 1511 active+clean; 114 TiB
>> data, 344 TiB used, 93 TiB / 437 TiB avail; 6.2 KiB/s rd, 2.2 MiB/s wr,
>> 118 op/s
>>
>> ----SNIP----
>>
>> We have, for the time being, disabled our snapshots in the hopes that
>> the cluster will catch up with the trimming process. Two potential
>> things of note:
>>
>> 1. We are unaware of any particular action which would be associated
>> with this happening now (there were no unusual deletions of either live
>> data or snapshots).
>> 2. For the past month or two it has appeared as if there has been a
>> steady unchecked growth in storage utilization as if snapshots have not
>> been actually being trimmed.
>>
>> Any assistance in determining what exactly has prompted this behavior or
>> any guidance on how to evaluate the total snaptrim queue size to see if
>> we are making progress would be much appreciated.
>>
>> Thank you,
>>
>> -David
>>
>> --
>> David Prude
>> Systems Administrator
>> PGP Fingerprint: 1DAA 4418 7F7F B8AA F50C  6FDF C294 B58F A286 F847
>> Democracy Now!
>> www.democracynow.org
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx

-- 
David Prude
Systems Administrator
PGP Fingerprint: 1DAA 4418 7F7F B8AA F50C  6FDF C294 B58F A286 F847
Democracy Now!
www.democracynow.org


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux