Re: [RFC PATCH] madvise: make madvise_cold_or_pageout_pte_range() support large folio

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Jul 13, 2023 at 11:57 PM Yin, Fengwei <fengwei.yin@xxxxxxxxx> wrote:
>
>
> >> -               if (pageout_anon_only_filter && !folio_test_anon(folio))
> >> +               /* Do not interfere with other mappings of this folio */
> >> +               if (folio_mapcount(folio) != 1)
> >>                         continue;
> >>
> >> -               VM_BUG_ON_FOLIO(folio_test_large(folio), folio);
> >> -
> >> -               if (pte_young(ptent)) {
> >> -                       ptent = ptep_get_and_clear_full(mm, addr, pte,
> >> -                                                       tlb->fullmm);
> >> -                       ptent = pte_mkold(ptent);
> >> -                       set_pte_at(mm, addr, pte, ptent);
> >> -                       tlb_remove_tlb_entry(tlb, pte, addr);
> >> -               }
> >> -
> >> -               /*
> >> -                * We are deactivating a folio for accelerating reclaiming.
> >> -                * VM couldn't reclaim the folio unless we clear PG_young.
> >> -                * As a side effect, it makes confuse idle-page tracking
> >> -                * because they will miss recent referenced history.
> >> -                */
> >> -               folio_clear_referenced(folio);
> >> -               folio_test_clear_young(folio);
> >> -               if (folio_test_active(folio))
> >> -                       folio_set_workingset(folio);
> >> +pageout_cold_folio:
> >>                 if (pageout) {
> >>                         if (folio_isolate_lru(folio)) {
> >>                                 if (folio_test_unevictable(folio))
> >> @@ -529,8 +542,30 @@ static int madvise_cold_or_pageout_pte_range(pmd_t *pmd,
> >>                 arch_leave_lazy_mmu_mode();
> >>                 pte_unmap_unlock(start_pte, ptl);
> >>         }
> >> -       if (pageout)
> >> -               reclaim_pages(&folio_list);
> >> +
> >> +       if (pageout) {
> >> +               LIST_HEAD(reclaim_list);
> >> +
> >> +               while (!list_empty(&folio_list)) {
> >> +                       int refs;
> >> +                       unsigned long flags;
> >> +                       struct mem_cgroup *memcg = folio_memcg(folio);
> >> +
> >> +                       folio = lru_to_folio(&folio_list);
> >> +                       list_del(&folio->lru);
> >> +
> >> +                       refs = folio_referenced(folio, 0, memcg, &flags);
> >> +
> >> +                       if ((flags & VM_LOCKED) || (refs == -1)) {
> >> +                               folio_putback_lru(folio);
> >> +                               continue;
> >> +                       }
> >> +
> >> +                       folio_test_clear_referenced(folio);
> >> +                       list_add(&folio->lru, &reclaim_list);
> >> +               }
> >> +               reclaim_pages(&reclaim_list);
> >> +       }
> >
> > i overlooked the chunk above -- it's unnecessary: after we split the
> > large folio (and splice the base folios onto the same LRU list), we
> > continue at the position of the first base folio because of:
> >
> >   pte--;
> >   addr -= PAGE_SIZE;
> >   continue;
> >
> > And then we do pte_mkold(), which takes care of the A-bit.
> This patch moves the A-bit clear out of the folio isolation loop. So
> even the folio is split and loop restarts from the first base folio,
> the A-bit is not cleared. A-bit is only cleared in reclaim loop.
>
> There is one option for A-bit clearing:
>   - clear A-bit of base 4K page in isolation loop and leave large folio
>     A-bit clearing to reclaim loop.
>
> This patch didn't use it because don't want to introduce A-bit clearing
> in two places. But I am open about clearing base 4K page A-bit cleared in
> isolation loop. Thanks.

Sorry but why are we trying to do multiple things in one patch that I
assumed is supposed to simply fix madvise() for large anon folios? And
none of those things seems to have a clear rationale behind it.

The only patch that makes sense at the moment (or the first patch of a
series) is what I said before:

-  if (folio_mapcount(folio) != 1)
+  if (folio_estimated_sharers(folio) != 1)

And probably clarify (preferrably in the comments above) this is an
estimate because we think it's a better tradeoff if we do so (less
code/overhead from checking the mapcounts of the rest of folios within
the range).





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux