On Thu, Dec 5, 2024 at 3:40 AM Chris Li <chrisl@xxxxxxxxxx> wrote: > > On Mon, Dec 2, 2024 at 10:42 AM Kairui Song <ryncsn@xxxxxxxxx> wrote: > > > > From: Kairui Song <kasong@xxxxxxxxxxx> > > > > commit e9e58a4ec3b1 ("memcg: avoid use cmpxchg in swap cgroup maintainance") > > replaced the cmpxchg/xchg with a global irq spinlock because some archs > > doesn't support 2 bytes cmpxchg/xchg. Clearly this won't scale well. > > > > And as commented in swap_cgroup.c, this lock is not needed for map > > synchronization. > > > > Emulation of 2 bytes cmpxchg/xchg with atomic isn't hard, so implement > > it to get rid of this lock. > > > > Testing using 64G brd and build with build kernel with make -j96 in 1.5G > > memory cgroup using 4k folios showed below improvement (10 test run): > > > > Before this series: > > Sys time: 10730.08 (stdev 49.030728) > > Real time: 171.03 (stdev 0.850355) > > > > After this commit: > > Sys time: 9612.24 (stdev 66.310789), -10.42% > > Real time: 159.78 (stdev 0.577193), -6.57% > > > > With 64k folios and 2G memcg: > > Before this series: > > Sys time: 7626.77 (stdev 43.545517) > > Real time: 136.22 (stdev 1.265544) > > > > After this commit: > > Sys time: 6936.03 (stdev 39.996280), -9.06% > > Real time: 129.65 (stdev 0.880039), -4.82% > > > > Sequential swapout of 8G 4k zero folios (24 test run): > > Before this series: > > 5461409.12 us (stdev 183957.827084) > > > > After this commit: > > 5420447.26 us (stdev 196419.240317) > > > > Sequential swapin of 8G 4k zero folios (24 test run): > > Before this series: > > 19736958.916667 us (stdev 189027.246676) > > > > After this commit: > > 19662182.629630 us (stdev 172717.640614) > > > > Performance is better or at least not worse for all tests above. > > > > Signed-off-by: Kairui Song <kasong@xxxxxxxxxxx> > > --- > > mm/swap_cgroup.c | 56 +++++++++++++++++++++++++++++++++++------------- > > 1 file changed, 41 insertions(+), 15 deletions(-) > > > > diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c > > index a76afdc3666a..028f5e6be3f0 100644 > > --- a/mm/swap_cgroup.c > > +++ b/mm/swap_cgroup.c > > @@ -5,6 +5,15 @@ > > > > #include <linux/swapops.h> /* depends on mm.h include */ > > > > +#define ID_PER_UNIT (sizeof(atomic_t) / sizeof(unsigned short)) > > You might want to have some compile time assert that (sizeof(atomic_t) > % sizeof(unsigned short)) is zero. Could not hurt. > > > +struct swap_cgroup_unit { > > + union { > > + int raw; > > + atomic_t val; > > + unsigned short __id[ID_PER_UNIT]; > > + }; > > +}; > > I suggest just getting rid of this complicated struct/union and using > bit shift and mask to get the u16 out from the atomic_t. Good suggestion. > > > + > > static DEFINE_MUTEX(swap_cgroup_mutex); > > > > struct swap_cgroup { > > @@ -12,8 +21,10 @@ struct swap_cgroup { > > }; > > > > struct swap_cgroup_ctrl { > > - unsigned short *map; > > - spinlock_t lock; > > + union { > > + struct swap_cgroup_unit *units; > > + unsigned short *map; > > You really shouldn't access the map as an "unsigned short" array, > therefore, I suggest changing the array pointer to "atomic_t". > > > + }; > > }; > > > > static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES]; > > @@ -31,6 +42,24 @@ static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES]; > > * > > * TODO: we can push these buffers out to HIGHMEM. > > */ > > +static unsigned short __swap_cgroup_xchg(void *map, > > + pgoff_t offset, > > + unsigned int new_id) > > +{ > > + unsigned int old_id; > > + struct swap_cgroup_unit *units = map; > > + struct swap_cgroup_unit *unit = &units[offset / ID_PER_UNIT]; > > + struct swap_cgroup_unit new, old = { .raw = atomic_read(&unit->val) }; > > + > > + do { > > + new.raw = old.raw; > > + old_id = old.__id[offset % ID_PER_UNIT]; > > + new.__id[offset % ID_PER_UNIT] = new_id; > > + } while (!atomic_try_cmpxchg(&unit->val, &old.raw, new.raw)); > > I suggest just calculating the atomic_t offset (offset / > ID_PER_UNIT) and getting the address of the atomic_t. > Then use the mask and shift to construct the new atomic_t value. It is > likely to generate better code. > You don't want the compiler to generate memory load and store for > constructing the temporary new value. > I haven't checked the machine generated code, I suspect the compiler > is not smart enough to convert those into register shift here. Which > is what you really want. > > > + > > + return old_id; > > +} > > + > > /** > > * swap_cgroup_record - record mem_cgroup for a set of swap entries > > * @ent: the first swap entry to be recorded into > > @@ -44,22 +73,19 @@ unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, > > unsigned int nr_ents) > > { > > struct swap_cgroup_ctrl *ctrl; > > - unsigned short *map; > > - unsigned short old; > > - unsigned long flags; > > pgoff_t offset = swp_offset(ent); > > pgoff_t end = offset + nr_ents; > > + unsigned short old, iter; > > + unsigned short *map; > > Make it an atomic_t pointer here as well. > > > > > ctrl = &swap_cgroup_ctrl[swp_type(ent)]; > > map = ctrl->map; > > > > - spin_lock_irqsave(&ctrl->lock, flags); > > - old = map[offset]; > > + old = READ_ONCE(map[offset]); > > Ah, you shouldn't perform u16 reading directly. That will get into the > endian problem of how the u16 is arranged into atomic_t. You should do > atomic reading then shift the bits out so you don't have the endian > problem. It is a bad idea mixing atomic updates and reading the middle > of the atomic address location. Good suggestion, convert the whole map into atomic_t and access / xchg with bit shifts is also OK, mixing atomic with other types may lead to misuse indeed. > > Chris > > > do { > > - VM_BUG_ON(map[offset] != old); > > - map[offset] = id; > > + iter = __swap_cgroup_xchg(map, offset, id); > > + VM_BUG_ON(iter != old); > > } while (++offset != end); > > - spin_unlock_irqrestore(&ctrl->lock, flags); > > > > return old; > > } > > @@ -85,20 +111,20 @@ unsigned short lookup_swap_cgroup_id(swp_entry_t ent) > > > > int swap_cgroup_swapon(int type, unsigned long max_pages) > > { > > - void *map; > > + struct swap_cgroup_unit *units; > > struct swap_cgroup_ctrl *ctrl; > > > > if (mem_cgroup_disabled()) > > return 0; > > > > - map = vzalloc(max_pages * sizeof(unsigned short)); > > - if (!map) > > + units = vzalloc(DIV_ROUND_UP(max_pages, ID_PER_UNIT) * > > + sizeof(struct swap_cgroup_unit)); > > + if (!units) > > goto nomem; > > > > ctrl = &swap_cgroup_ctrl[type]; > > mutex_lock(&swap_cgroup_mutex); > > - ctrl->map = map; > > - spin_lock_init(&ctrl->lock); > > + ctrl->units = units; > > mutex_unlock(&swap_cgroup_mutex); > > > > return 0; > > -- > > 2.47.0 > > >