Hi, This is the 13th iteration of Glauber Costa's patch-set implementing slab shrinking on memcg pressure. The main idea is to make the list_lru structure used by most FS shrinkers per-memcg. When adding or removing an element from a list_lru, we use the page information to figure out which memcg it belongs to and relay it to the appropriate list. This allows scanning kmem objects accounted to different memcgs independently. Please note that in contrast to previous versions this patch-set implements slab shrinking only when we hit the user memory limit so that kmem allocations will still fail if we are below the user memory limit, but close to the kmem limit. This is, because the implementation of kmem-only reclaim was rather incomplete - we had to fail GFP_NOFS allocations since everything we could reclaim was only FS data. I will try to improve this and send in a separate patch-set, but currently it is only worthwhile setting the kmem limit to be greater than the user mem limit just to enable per-memcg slab accounting and reclaim. The patch-set is based on top of Linux-3.13-rc3 and organized as follows: - patches 1-9 prepare vmscan, memcontrol, list_lru to kmemcg reclaim; - patch 10 implements the kmemcg reclaim core; - patch 11 makes the list_lru struct per-memcg and patch 12 marks all list_lru-based shrinkers as memcg-aware; - patches 13-16 slightly improve memcontrol behavior regarding mem reclaim. Changes in v13: - fix NUMA-unaware shrinkers not being called when node 0 is not set in the nodemask; - rework list_lru API to require a shrink_control - make list_lru automatically handle memcgs w/o introducing a separate struct; - simplify walk over all memcg LRUs of a list_lru; - cleanup shrink_slab() - remove kmem-only reclaim as explained above; Previous iterations of this patch-set can be found here: - https://lkml.org/lkml/2013/12/2/141 - https://lkml.org/lkml/2013/11/25/214 Thanks. Glauber Costa (7): memcg: make cache index determination more robust memcg: consolidate callers of memcg_cache_id memcg: move initialization to memcg creation vmscan: take at least one pass with shrinkers vmpressure: in-kernel notifications memcg: reap dead memcgs upon global memory pressure memcg: flush memcg items upon memcg destruction Vladimir Davydov (9): memcg: move memcg_caches_array_size() function vmscan: move call to shrink_slab() to shrink_zones() vmscan: remove shrink_control arg from do_try_to_free_pages() vmscan: call NUMA-unaware shrinkers irrespective of nodemask mm: list_lru: require shrink_control in count, walk functions fs: consolidate {nr,free}_cached_objects args in shrink_control vmscan: shrink slab on memcg pressure mm: list_lru: add per-memcg lists fs: mark list_lru based shrinkers memcg aware fs/dcache.c | 17 +- fs/gfs2/quota.c | 10 +- fs/inode.c | 8 +- fs/internal.h | 9 +- fs/super.c | 24 ++- fs/xfs/xfs_buf.c | 16 +- fs/xfs/xfs_qm.c | 8 +- fs/xfs/xfs_super.c | 6 +- include/linux/fs.h | 6 +- include/linux/list_lru.h | 83 ++++++---- include/linux/memcontrol.h | 35 ++++ include/linux/shrinker.h | 10 +- include/linux/vmpressure.h | 5 + mm/list_lru.c | 263 +++++++++++++++++++++++++++--- mm/memcontrol.c | 384 +++++++++++++++++++++++++++++++++++++++----- mm/vmpressure.c | 53 +++++- mm/vmscan.c | 172 +++++++++++++------- 17 files changed, 888 insertions(+), 221 deletions(-) -- 1.7.10.4 -- To unsubscribe from this list: send the line "unsubscribe cgroups" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html