Zitat von Cedric <yipikai7@xxxxxxxxx>:
> Yes the osd_scrub_invalid_stats is set to true.
>
> We are thinking about the use of "ceph pg_mark_unfound_lost revert"
> action, but we wonder if there is a risk of data loss.
>
> On Thu, Feb 22, 2024 at 11:50 AM Eugen Block <eblock@xxxxxx> wrote:
>>
>> I found a config to force scrub invalid PGs, what is your current
>> setting on that?
>>
>> ceph config get osd osd_scrub_invalid_stats
>> true
>>
>> The config reference states:
>>
>> > Forces extra scrub to fix stats marked as invalid.
>>
>> But the default seems to be true, so I'd expect it's true in your case
>> as well?
>>
>> Zitat von Cedric <yipikai7@xxxxxxxxx>:
>>
>> > Thanks Eugen for the suggestion, yes we have tried, also repeering
>> > concerned PGs, still the same issue.
>> >
>> > Looking at the code it seems the split-mode message is triggered when
>> > the PG as ""stats_invalid": true,", here is the result of a query:
>> >
>> > "stats_invalid": true,
>> > "dirty_stats_invalid": false,
>> > "omap_stats_invalid": false,
>> > "hitset_stats_invalid": false,
>> > "hitset_bytes_stats_invalid": false,
>> > "pin_stats_invalid": false,
>> > "manifest_stats_invalid": false,
>> >
>> > I also provide again cluster informations that was lost in previous
>> > missed reply all. Don't hesitate to ask more if needed I would be
>> > glade to provide them.
>> >
>> > Cédric
>> >
>> >
>> > On Thu, Feb 22, 2024 at 11:04 AM Eugen Block <eblock@xxxxxx> wrote:
>> >>
>> >> Hm, I wonder if setting (and unsetting after a while) noscrub and
>> >> nodeep-scrub has any effect. Have you tried that?
>> >>
>> >> Zitat von Cedric <yipikai7@xxxxxxxxx>:
>> >>
>> >> > Update: we have run fsck and re-shard on all bluestore volume, seems
>> >> > sharding were not applied.
>> >> >
>> >> > Unfortunately scrubs and deep-scrubs are still stuck on PGs of the
>> >> > pool that is suffering the issue, but other PGs scrubs well.
>> >> >
>> >> > The next step will be to remove the cache tier as suggested, but its
>> >> > not available yet as PGs needs to be scrubbed in order for the cache
>> >> > tier can be activated.
>> >> >
>> >> > As we are struggling to make this cluster works again, any help
>> >> > would be greatly appreciated.
>> >> >
>> >> > Cédric
>> >> >
>> >> >> On 20 Feb 2024, at 20:22, Cedric <yipikai7@xxxxxxxxx> wrote:
>> >> >>
>> >> >> Thanks Eugen, sorry about the missed reply to all.
>> >> >>
>> >> >> The reason we still have the cache tier is because we were not able
>> >> >> to flush all dirty entry to remove it (as per the procedure), so
>> >> >> the cluster as been migrated from HDD/SSD to NVME a while ago but
>> >> >> tiering remains, unfortunately.
>> >> >>
>> >> >> So actually we are trying to understand the root cause
>> >> >>
>> >> >> On Tue, Feb 20, 2024 at 1:43 PM Eugen Block <eblock@xxxxxx> wrote:
>> >> >>>
>> >> >>> Please don't drop the list from your response.
>> >> >>>
>> >> >>> The first question coming to mind is, why do you have a
cache-tier if
>> >> >>> all your pools are on nvme decices anyway? I don't see any
>> benefit here.
>> >> >>> Did you try the suggested workaround and disable the cache-tier?
>> >> >>>
>> >> >>> Zitat von Cedric <yipikai7@xxxxxxxxx>:
>> >> >>>
>> >> >>>> Thanks Eugen, see attached infos.
>> >> >>>>
>> >> >>>> Some more details:
>> >> >>>>
>> >> >>>> - commands that actually hangs: ceph balancer status ; rbd
>> -p vms ls ;
>> >> >>>> rados -p vms_cache cache-flush-evict-all
>> >> >>>> - all scrub running on vms_caches pgs are stall / start in a loop
>> >> >>>> without actually doing anything
>> >> >>>> - all io are 0 both from ceph status or iostat on nodes
>> >> >>>>
>> >> >>>> On Tue, Feb 20, 2024 at 10:00 AM Eugen Block
<eblock@xxxxxx> wrote:
>> >> >>>>>
>> >> >>>>> Hi,
>> >> >>>>>
>> >> >>>>> some more details would be helpful, for example what's the
>> pool size
>> >> >>>>> of the cache pool? Did you issue a PG split before or during the
>> >> >>>>> upgrade? This thread [1] deals with the same problem,
the described
>> >> >>>>> workaround was to set hit_set_count to 0 and disable the
>> cache layer
>> >> >>>>> until that is resolved. Afterwards you could enable the
cache layer
>> >> >>>>> again. But keep in mind that the code for cache tier is entirely
>> >> >>>>> removed in Reef (IIRC).
>> >> >>>>>
>> >> >>>>> Regards,
>> >> >>>>> Eugen
>> >> >>>>>
>> >> >>>>> [1]
>> >> >>>>>
>> >>
>>
https://ceph-users.ceph.narkive.com/zChyOq5D/ceph-strange-issue-after-adding-a-cache-osd
>> >> >>>>>
>> >> >>>>> Zitat von Cedric <yipikai7@xxxxxxxxx>:
>> >> >>>>>
>> >> >>>>>> Hello,
>> >> >>>>>>
>> >> >>>>>> Following an upgrade from Nautilus (14.2.22) to Pacific
>> (16.2.13), we
>> >> >>>>>> encounter an issue with a cache pool becoming completely stuck,
>> >> >>>>>> relevant messages below:
>> >> >>>>>>
>> >> >>>>>> pg xx.x has invalid (post-split) stats; must scrub before
>> tier agent
>> >> >>>>>> can activate
>> >> >>>>>>
>> >> >>>>>> In OSD logs, scrubs are starting in a loop without
>> succeeding for all
>> >> >>>>>> pg of this pool.
>> >> >>>>>>
>> >> >>>>>> What we already tried without luck so far:
>> >> >>>>>>
>> >> >>>>>> - shutdown / restart OSD
>> >> >>>>>> - rebalance pg between OSD
>> >> >>>>>> - raise the memory on OSD
>> >> >>>>>> - repeer PG
>> >> >>>>>>
>> >> >>>>>> Any idea what is causing this? any help will be greatly
>> appreciated
>> >> >>>>>>
>> >> >>>>>> Thanks
>> >> >>>>>>
>> >> >>>>>> Cédric
>> >> >>>>>> _______________________________________________
>> >> >>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>> >> >>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> _______________________________________________
>> >> >>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>> >> >>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >> >>>
>> >> >>>
>> >> >>>
>> >>
>> >>
>> >>
>>
>>
>>