There are concerns about memory leaks from extensive use of memory cgroups as each memory cgroup creates its own set of kmem caches. There is a possiblity that the memcg kmem caches may remain even after the memory cgroup removal. Therefore, it will be useful to show how many memcg caches are present for each of the kmem caches. As slabinfo reporting code has to iterate through all the memcg caches to get the final numbers anyway, there is no additional cost in reporting the number of memcg caches available. The slabinfo version is bumped up to 2.2 as a new "<num_caches>" column is added at the end. Signed-off-by: Waiman Long <longman@xxxxxxxxxx> --- mm/slab_common.c | 24 ++++++++++++++++-------- 1 file changed, 16 insertions(+), 8 deletions(-) diff --git a/mm/slab_common.c b/mm/slab_common.c index 58251ba63e4a..c7aa47a99b2b 100644 --- a/mm/slab_common.c +++ b/mm/slab_common.c @@ -1308,13 +1308,13 @@ static void print_slabinfo_header(struct seq_file *m) * without _too_ many complaints. */ #ifdef CONFIG_DEBUG_SLAB - seq_puts(m, "slabinfo - version: 2.1 (statistics)\n"); + seq_puts(m, "slabinfo - version: 2.2 (statistics)\n"); #else - seq_puts(m, "slabinfo - version: 2.1\n"); + seq_puts(m, "slabinfo - version: 2.2\n"); #endif seq_puts(m, "# name <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>"); seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>"); - seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail>"); + seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail> <num_caches>"); #ifdef CONFIG_DEBUG_SLAB seq_puts(m, " : globalstat <listallocs> <maxobjs> <grown> <reaped> <error> <maxfreeable> <nodeallocs> <remotefrees> <alienoverflow>"); seq_puts(m, " : cpustat <allochit> <allocmiss> <freehit> <freemiss>"); @@ -1338,14 +1338,18 @@ void slab_stop(struct seq_file *m, void *p) mutex_unlock(&slab_mutex); } -static void +/* + * Return number of memcg caches. + */ +static unsigned int memcg_accumulate_slabinfo(struct kmem_cache *s, struct slabinfo *info) { struct kmem_cache *c; struct slabinfo sinfo; + unsigned int cnt = 0; if (!is_root_cache(s)) - return; + return 0; for_each_memcg_cache(c, s) { memset(&sinfo, 0, sizeof(sinfo)); @@ -1356,17 +1360,20 @@ memcg_accumulate_slabinfo(struct kmem_cache *s, struct slabinfo *info) info->shared_avail += sinfo.shared_avail; info->active_objs += sinfo.active_objs; info->num_objs += sinfo.num_objs; + cnt++; } + return cnt; } static void cache_show(struct kmem_cache *s, struct seq_file *m) { struct slabinfo sinfo; + unsigned int nr_memcg_caches; memset(&sinfo, 0, sizeof(sinfo)); get_slabinfo(s, &sinfo); - memcg_accumulate_slabinfo(s, &sinfo); + nr_memcg_caches = memcg_accumulate_slabinfo(s, &sinfo); seq_printf(m, "%-17s %6lu %6lu %6u %4u %4d", cache_name(s), sinfo.active_objs, sinfo.num_objs, s->size, @@ -1374,8 +1381,9 @@ static void cache_show(struct kmem_cache *s, struct seq_file *m) seq_printf(m, " : tunables %4u %4u %4u", sinfo.limit, sinfo.batchcount, sinfo.shared); - seq_printf(m, " : slabdata %6lu %6lu %6lu", - sinfo.active_slabs, sinfo.num_slabs, sinfo.shared_avail); + seq_printf(m, " : slabdata %6lu %6lu %6lu %3u", + sinfo.active_slabs, sinfo.num_slabs, sinfo.shared_avail, + nr_memcg_caches); slabinfo_show_stats(m, s); seq_putc(m, '\n'); } -- 2.18.1