The patch titled Subject: slab: introduce __kmemcg_cache_deactivate() has been added to the -mm tree. Its filename is slab-introduce-__kmemcg_cache_deactivate.patch This patch should soon appear at http://ozlabs.org/~akpm/mmots/broken-out/slab-introduce-__kmemcg_cache_deactivate.patch and later at http://ozlabs.org/~akpm/mmotm/broken-out/slab-introduce-__kmemcg_cache_deactivate.patch Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/SubmitChecklist when testing your code *** The -mm tree is included into linux-next and is updated there every 3-4 working days ------------------------------------------------------ From: Tejun Heo <tj@xxxxxxxxxx> Subject: slab: introduce __kmemcg_cache_deactivate() __kmem_cache_shrink() is called with %true @deactivate only for memcg caches. Remove @deactivate from __kmem_cache_shrink() and introduce __kmemcg_cache_deactivate() instead. Each memcg-supporting allocator should implement it and it should deactivate and drain the cache. This is to allow memcg cache deactivation behavior to further deviate from simple shrinking without messing up __kmem_cache_shrink(). This is pure reorganization and doesn't introduce any observable behavior changes. v2: Dropped unnecessary ifdef in mm/slab.h as suggested by Vladimir. Link: http://lkml.kernel.org/r/20170117235411.9408-8-tj@xxxxxxxxxx Signed-off-by: Tejun Heo <tj@xxxxxxxxxx> Acked-by: Vladimir Davydov <vdavydov.dev@xxxxxxxxx> Cc: Christoph Lameter <cl@xxxxxxxxx> Cc: Pekka Enberg <penberg@xxxxxxxxxx> Cc: David Rientjes <rientjes@xxxxxxxxxx> Cc: Joonsoo Kim <iamjoonsoo.kim@xxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/slab.c | 11 +++++++++-- mm/slab.h | 3 ++- mm/slab_common.c | 4 ++-- mm/slob.c | 2 +- mm/slub.c | 39 ++++++++++++++++++++++----------------- 5 files changed, 36 insertions(+), 23 deletions(-) diff -puN mm/slab.c~slab-introduce-__kmemcg_cache_deactivate mm/slab.c --- a/mm/slab.c~slab-introduce-__kmemcg_cache_deactivate +++ a/mm/slab.c @@ -2315,7 +2315,7 @@ out: return nr_freed; } -int __kmem_cache_shrink(struct kmem_cache *cachep, bool deactivate) +int __kmem_cache_shrink(struct kmem_cache *cachep) { int ret = 0; int node; @@ -2333,9 +2333,16 @@ int __kmem_cache_shrink(struct kmem_cach return (ret ? 1 : 0); } +#ifdef CONFIG_MEMCG +void __kmemcg_cache_deactivate(struct kmem_cache *cachep) +{ + __kmem_cache_shrink(cachep); +} +#endif + int __kmem_cache_shutdown(struct kmem_cache *cachep) { - return __kmem_cache_shrink(cachep, false); + return __kmem_cache_shrink(cachep); } void __kmem_cache_release(struct kmem_cache *cachep) diff -puN mm/slab.h~slab-introduce-__kmemcg_cache_deactivate mm/slab.h --- a/mm/slab.h~slab-introduce-__kmemcg_cache_deactivate +++ a/mm/slab.h @@ -162,7 +162,8 @@ static inline unsigned long kmem_cache_f int __kmem_cache_shutdown(struct kmem_cache *); void __kmem_cache_release(struct kmem_cache *); -int __kmem_cache_shrink(struct kmem_cache *, bool); +int __kmem_cache_shrink(struct kmem_cache *); +void __kmemcg_cache_deactivate(struct kmem_cache *s); void slab_kmem_cache_release(struct kmem_cache *); struct seq_file; diff -puN mm/slab_common.c~slab-introduce-__kmemcg_cache_deactivate mm/slab_common.c --- a/mm/slab_common.c~slab-introduce-__kmemcg_cache_deactivate +++ a/mm/slab_common.c @@ -646,7 +646,7 @@ void memcg_deactivate_kmem_caches(struct if (!c) continue; - __kmem_cache_shrink(c, true); + __kmemcg_cache_deactivate(c); arr->entries[idx] = NULL; } mutex_unlock(&slab_mutex); @@ -794,7 +794,7 @@ int kmem_cache_shrink(struct kmem_cache get_online_cpus(); get_online_mems(); kasan_cache_shrink(cachep); - ret = __kmem_cache_shrink(cachep, false); + ret = __kmem_cache_shrink(cachep); put_online_mems(); put_online_cpus(); return ret; diff -puN mm/slob.c~slab-introduce-__kmemcg_cache_deactivate mm/slob.c --- a/mm/slob.c~slab-introduce-__kmemcg_cache_deactivate +++ a/mm/slob.c @@ -634,7 +634,7 @@ void __kmem_cache_release(struct kmem_ca { } -int __kmem_cache_shrink(struct kmem_cache *d, bool deactivate) +int __kmem_cache_shrink(struct kmem_cache *d) { return 0; } diff -puN mm/slub.c~slab-introduce-__kmemcg_cache_deactivate mm/slub.c --- a/mm/slub.c~slab-introduce-__kmemcg_cache_deactivate +++ a/mm/slub.c @@ -3890,7 +3890,7 @@ EXPORT_SYMBOL(kfree); * being allocated from last increasing the chance that the last objects * are freed in them. */ -int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate) +int __kmem_cache_shrink(struct kmem_cache *s) { int node; int i; @@ -3902,21 +3902,6 @@ int __kmem_cache_shrink(struct kmem_cach unsigned long flags; int ret = 0; - if (deactivate) { - /* - * Disable empty slabs caching. Used to avoid pinning offline - * memory cgroups by kmem pages that can be freed. - */ - s->cpu_partial = 0; - s->min_partial = 0; - - /* - * s->cpu_partial is checked locklessly (see put_cpu_partial), - * so we have to make sure the change is visible. - */ - synchronize_sched(); - } - flush_all(s); for_each_kmem_cache_node(s, node, n) { INIT_LIST_HEAD(&discard); @@ -3967,13 +3952,33 @@ int __kmem_cache_shrink(struct kmem_cach return ret; } +#ifdef CONFIG_MEMCG +void __kmemcg_cache_deactivate(struct kmem_cache *s) +{ + /* + * Disable empty slabs caching. Used to avoid pinning offline + * memory cgroups by kmem pages that can be freed. + */ + s->cpu_partial = 0; + s->min_partial = 0; + + /* + * s->cpu_partial is checked locklessly (see put_cpu_partial), so + * we have to make sure the change is visible. + */ + synchronize_sched(); + + __kmem_cache_shrink(s); +} +#endif + static int slab_mem_going_offline_callback(void *arg) { struct kmem_cache *s; mutex_lock(&slab_mutex); list_for_each_entry(s, &slab_caches, list) - __kmem_cache_shrink(s, false); + __kmem_cache_shrink(s); mutex_unlock(&slab_mutex); return 0; _ Patches currently in -mm which might be from tj@xxxxxxxxxx are revert-slub-move-synchronize_sched-out-of-slab_mutex-on-shrink.patch slub-separate-out-sysfs_slab_release-from-sysfs_slab_remove.patch slab-remove-synchronous-rcu_barrier-call-in-memcg-cache-release-path.patch slab-reorganize-memcg_cache_params.patch slab-link-memcg-kmem_caches-on-their-associated-memory-cgroup.patch slab-implement-slab_root_caches-list.patch slab-introduce-__kmemcg_cache_deactivate.patch slab-remove-synchronous-synchronize_sched-from-memcg-cache-deactivation-path.patch slab-remove-slub-sysfs-interface-files-early-for-empty-memcg-caches.patch slab-use-memcg_kmem_cache_wq-for-slab-destruction-operations.patch -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html