Re: [PATCH v2 2/2] mm: attempt to batch free swap entries for zap_pte_range()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Aug 7, 2024 at 9:29 PM David Hildenbrand <david@xxxxxxxxxx> wrote:
>
> >   mm/swapfile.c | 78 +++++++++++++++++++++++++++++++++++++++++++--------
> >   1 file changed, 67 insertions(+), 11 deletions(-)
> >
> > diff --git a/mm/swapfile.c b/mm/swapfile.c
> > index 35cb58373493..25c3f98fa8d5 100644
> > --- a/mm/swapfile.c
> > +++ b/mm/swapfile.c
> > @@ -156,6 +156,25 @@ static bool swap_is_has_cache(struct swap_info_struct *si,
> >       return true;
> >   }
> >
> > +static bool swap_is_last_map(struct swap_info_struct *si,
> > +                           unsigned long offset, int nr_pages,
> > +                           bool *has_cache)
>
> Please use double tabs for indenting parameters on 2nd line on
> new/changed code:
>
>                 unsigned long offset, int nr_pages, bool *has_cache)
>
> Results in less churn when renaming functions and we can frequently
> avoid some lines.

ack.

>
> > +{
> > +     unsigned char *map = si->swap_map + offset;
> > +     unsigned char *map_end = map + nr_pages;
> > +     bool cached = false;
> > +
> > +     do {
> > +             if ((*map & ~SWAP_HAS_CACHE) != 1)
> > +                     return false;
> > +             if (*map & SWAP_HAS_CACHE)
> > +                     cached = true;
> > +     } while (++map < map_end);
> > +
> > +     *has_cache = cached;
> > +     return true;
> > +}
> > +
> >   /*
> >    * returns number of pages in the folio that backs the swap entry. If positive,
> >    * the folio was reclaimed. If negative, the folio was not reclaimed. If 0, no
> > @@ -1469,6 +1488,53 @@ static unsigned char __swap_entry_free(struct swap_info_struct *si,
> >       return usage;
> >   }
> >
> > +static bool __swap_entries_free(struct swap_info_struct *si,
> > +                             swp_entry_t entry, int nr)
>
> Dito.

ack.

>
> > +{
> > +     unsigned long offset = swp_offset(entry);
> > +     unsigned int type = swp_type(entry);
> > +     struct swap_cluster_info *ci;
> > +     bool has_cache = false;
> > +     unsigned char count;
> > +     bool can_batch;
> > +     int i;
> > +
> > +     if (nr <= 1 || swap_count(data_race(si->swap_map[offset])) != 1)
> > +             goto fallback;
> > +     /* cross into another cluster */
> > +     if (nr > SWAPFILE_CLUSTER - offset % SWAPFILE_CLUSTER)
> > +             goto fallback;
> > +
> > +     ci = lock_cluster_or_swap_info(si, offset);
> > +     can_batch = swap_is_last_map(si, offset, nr, &has_cache);
> > +     if (can_batch) {
> > +             for (i = 0; i < nr; i++)
> > +                     WRITE_ONCE(si->swap_map[offset + i], SWAP_HAS_CACHE);
> > +     }
> > +     unlock_cluster_or_swap_info(si, ci);
> > +
> > +     if (!can_batch)
> > +             goto fallback;
>
> I'd avoid "can_batch" and just do:
>
> ci = lock_cluster_or_swap_info(si, offset);
> if (!swap_is_last_map(si, offset, nr, &has_cache)) {
>         unlock_cluster_or_swap_info(si, ci);
>         goto fallback;
> }
> for (i = 0; i < nr; i++)
>         WRITE_ONCE(si->swap_map[offset + i], SWAP_HAS_CACHE);
> unlock_cluster_or_swap_info(si, ci);

ack.

>
> > +     if (!has_cache) {
> > +             spin_lock(&si->lock);
>
> I'm no expert on that code, but we might drop the cluster lock the take
> the swap_info lock and then retake the cluster lock. I assume there are
> no races we are worrying about here, right?

I suppose so. Even the original single-entry code follows the same pattern:

static unsigned char __swap_entry_free(struct swap_info_struct *p,
       swp_entry_t entry)
{
         struct swap_cluster_info *ci;
         unsigned long offset = swp_offset(entry);
         unsigned char usage;

         ci = lock_cluster_or_swap_info(p, offset);
         usage = __swap_entry_free_locked(p, offset, 1);
         unlock_cluster_or_swap_info(p, ci);
         if (!usage)
                  free_swap_slot(entry);

         return usage;
}

I assume that once we mark them as SWAP_HAS_CACHE, no one else
will touch them except ourselves.

>
> > +             swap_entry_range_free(si, entry, nr);
> > +             spin_unlock(&si->lock);
> > +     }
> > +     return has_cache;
> > +
> > +fallback:
> > +     for (i = 0; i  < nr; i++) {
>
> One space too much before the "<".

ack.

>
> > +             if (data_race(si->swap_map[offset + i])) {
> > +                     count = __swap_entry_free(si, swp_entry(type, offset + i));
> > +                     if (count == SWAP_HAS_CACHE)
> > +                             has_cache = true;
> > +             } else {
> > +                     WARN_ON_ONCE(1);
> > +             }
> > +     }
> > +     return has_cache;
> > +}
> > +
>
> --
> Cheers,
>
> David / dhildenb
>

Thanks
Barry





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux