Re: [PATCH v1] mm/slub: enable debugging memory wasting of kmalloc

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Jul 14, 2022 at 10:11:32PM +0200, Vlastimil Babka wrote:
> On 7/13/22 09:36, Feng Tang wrote:
> > Hi Vlastimil,
> > 
> > On Mon, Jul 11, 2022 at 10:15:21AM +0200, Vlastimil Babka wrote:
> >> On 7/1/22 15:59, Feng Tang wrote:
> >> > kmalloc's API family is critical for mm, with one shortcoming that
> >> > its object size is fixed to be power of 2. When user requests memory
> >> > for '2^n + 1' bytes, actually 2^(n+1) bytes will be allocated, so
> >> > in worst case, there is around 50% memory space waste.
> >> > 
> >> > We've met a kernel boot OOM panic (v5.10), and from the dumped slab info:
> >> > 
> >> >     [   26.062145] kmalloc-2k            814056KB     814056KB
> >> > 
> >> > From debug we found there are huge number of 'struct iova_magazine',
> >> > whose size is 1032 bytes (1024 + 8), so each allocation will waste
> >> > 1016 bytes. Though the issue was solved by giving the right (bigger)
> >> > size of RAM, it is still nice to optimize the size (either use a
> >> > kmalloc friendly size or create a dedicated slab for it).
> > [...]
> >> 
> >> Hi and thanks.
> >> I would suggest some improvements to consider:
> >> 
> >> - don't use the struct track to store orig_size, although it's an obvious
> >> first choice. It's unused waste for the free_track, and also for any
> >> non-kmalloc caches. I'd carve out an extra int next to the struct tracks.
> >> Only for kmalloc caches (probably a new kmem cache flag set on creation will
> >> be needed to easily distinguish them).
> >> Besides the saved space, you can then set the field from ___slab_alloc()
> >> directly and not need to pass the orig_size also to alloc_debug_processing()
> >> etc.
> >  
> > Here is a draft patch fowlling your suggestion, please check if I missed
> > anything? (Quick test showed it achived similar effect as v1 patch). Thanks!
> 
> Thanks, overal it looks at first glance!

Thanks!

> > ---
> > diff --git a/include/linux/slab.h b/include/linux/slab.h
> > index 0fefdf528e0d..d3dacb0f013f 100644
> > --- a/include/linux/slab.h
> > +++ b/include/linux/slab.h
> > @@ -29,6 +29,8 @@
> >  #define SLAB_RED_ZONE		((slab_flags_t __force)0x00000400U)
> >  /* DEBUG: Poison objects */
> >  #define SLAB_POISON		((slab_flags_t __force)0x00000800U)
> > +/* Indicate a slab of kmalloc */
> 
> "Indicate a kmalloc cache" would be more precise.
 
Will use this in next version.

> > +#define SLAB_KMALLOC		((slab_flags_t __force)0x00001000U)
> >  /* Align objs on cache lines */
> >  #define SLAB_HWCACHE_ALIGN	((slab_flags_t __force)0x00002000U)
> >  /* Use GFP_DMA memory */
> > diff --git a/mm/slub.c b/mm/slub.c
> > index 26b00951aad1..3b0f80927817 100644
> 
> <snip>
> 
> > 
> >> - the knowledge of actual size could be used to improve poisoning checks as
> >> well, detect cases when there's buffer overrun over the orig_size but not
> >> cache's size. e.g. if you kmalloc(48) and overrun up to 64 we won't detect
> >> it now, but with orig_size stored we could?
> > 
> > The above patch doesn't touch this. As I have a question, for the
> > [orib_size, object_size) area, shall we fill it with POISON_XXX no matter
> > REDZONE flag is set or not?
> 
> Ah, looks like we use redzoning, not poisoning, for padding from
> s->object_size to word boundary. So it would be more consistent to use the
> redzone pattern (RED_ACTIVE) and check with the dynamic orig_size. Probably
> no change for RED_INACTIVE handling is needed though.

Thanks for clarifying, will go this way and do more test. Also I'd 
make it a separate patch, as it is logically different from the space
wastage.

Thanks,
Feng




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux