On Thu, Oct 22, 2020 at 3:20 PM Andrey Konovalov <andreyknvl@xxxxxxxxxx> wrote: > > Rename kasan_poison_kfree() into kasan_slab_free_mempool() as it better > reflects what this annotation does. > > No functional changes. > > Signed-off-by: Andrey Konovalov <andreyknvl@xxxxxxxxxx> > Link: https://linux-review.googlesource.com/id/I5026f87364e556b506ef1baee725144bb04b8810 Reviewed-by: Dmitry Vyukov <dvyukov@xxxxxxxxxx> > --- > include/linux/kasan.h | 16 ++++++++-------- > mm/kasan/common.c | 16 ++++++++-------- > mm/mempool.c | 2 +- > 3 files changed, 17 insertions(+), 17 deletions(-) > > diff --git a/include/linux/kasan.h b/include/linux/kasan.h > index 8654275aa62e..2ae92f295f76 100644 > --- a/include/linux/kasan.h > +++ b/include/linux/kasan.h > @@ -162,6 +162,13 @@ static inline bool kasan_slab_free(struct kmem_cache *s, void *object, unsigned > return false; > } > > +void __kasan_slab_free_mempool(void *ptr, unsigned long ip); > +static inline void kasan_slab_free_mempool(void *ptr, unsigned long ip) > +{ > + if (static_branch_likely(&kasan_enabled)) > + __kasan_slab_free_mempool(ptr, ip); > +} > + > void * __must_check __kasan_slab_alloc(struct kmem_cache *s, > void *object, gfp_t flags); > static inline void * __must_check kasan_slab_alloc(struct kmem_cache *s, > @@ -202,13 +209,6 @@ static inline void * __must_check kasan_krealloc(const void *object, > return (void *)object; > } > > -void __kasan_poison_kfree(void *ptr, unsigned long ip); > -static inline void kasan_poison_kfree(void *ptr, unsigned long ip) > -{ > - if (static_branch_likely(&kasan_enabled)) > - __kasan_poison_kfree(ptr, ip); > -} > - > void __kasan_kfree_large(void *ptr, unsigned long ip); > static inline void kasan_kfree_large(void *ptr, unsigned long ip) > { > @@ -244,6 +244,7 @@ static inline bool kasan_slab_free(struct kmem_cache *s, void *object, > { > return false; > } > +static inline void kasan_slab_free_mempool(void *ptr, unsigned long ip) {} > static inline void *kasan_slab_alloc(struct kmem_cache *s, void *object, > gfp_t flags) > { > @@ -264,7 +265,6 @@ static inline void *kasan_krealloc(const void *object, size_t new_size, > { > return (void *)object; > } > -static inline void kasan_poison_kfree(void *ptr, unsigned long ip) {} > static inline void kasan_kfree_large(void *ptr, unsigned long ip) {} > > #endif /* CONFIG_KASAN */ > diff --git a/mm/kasan/common.c b/mm/kasan/common.c > index b82dbae0c5d6..5622b0ec0907 100644 > --- a/mm/kasan/common.c > +++ b/mm/kasan/common.c > @@ -334,6 +334,14 @@ bool __kasan_slab_free(struct kmem_cache *cache, void *object, unsigned long ip) > return ____kasan_slab_free(cache, object, ip, true); > } > > +void __kasan_slab_free_mempool(void *ptr, unsigned long ip) > +{ > + struct page *page; > + > + page = virt_to_head_page(ptr); > + ____kasan_slab_free(page->slab_cache, ptr, ip, false); > +} > + > static void set_alloc_info(struct kmem_cache *cache, void *object, gfp_t flags) > { > kasan_set_track(&kasan_get_alloc_meta(cache, object)->alloc_track, flags); > @@ -436,14 +444,6 @@ void * __must_check __kasan_krealloc(const void *object, size_t size, gfp_t flag > flags, true); > } > > -void __kasan_poison_kfree(void *ptr, unsigned long ip) > -{ > - struct page *page; > - > - page = virt_to_head_page(ptr); > - ____kasan_slab_free(page->slab_cache, ptr, ip, false); > -} > - > void __kasan_kfree_large(void *ptr, unsigned long ip) > { > if (ptr != page_address(virt_to_head_page(ptr))) > diff --git a/mm/mempool.c b/mm/mempool.c > index 79bff63ecf27..0e8d877fbbc6 100644 > --- a/mm/mempool.c > +++ b/mm/mempool.c > @@ -106,7 +106,7 @@ static inline void poison_element(mempool_t *pool, void *element) > static __always_inline void kasan_poison_element(mempool_t *pool, void *element) > { > if (pool->alloc == mempool_alloc_slab || pool->alloc == mempool_kmalloc) > - kasan_poison_kfree(element, _RET_IP_); > + kasan_slab_free_mempool(element, _RET_IP_); > if (pool->alloc == mempool_alloc_pages) > kasan_free_pages(element, (unsigned long)pool->pool_data); > } > -- > 2.29.0.rc1.297.gfa9743e501-goog >