Hi, SLAB and SLUB use hardwall cpuset check on fallback alloc, while the page allocator uses softwall check for all kernel allocations. This may result in falling into the page allocator even if there are free objects on other nodes. SLAB algorithm is especially affected: the number of objects allocated in vain is unlimited, so that they theoretically can eat up a whole NUMA node. For more details see comments to patches 3, 4. When I last sent a fix (https://lkml.org/lkml/2014/8/10/100), David found the whole cpuset API being cumbersome and proposed to simplify it before getting to fixing its users. So this patch set addresses both David's complain (patches 1, 2) and the SL[AU]B issues (patches 3, 4). Reviews are appreciated. Thanks, Vladimir Davydov (4): cpuset: convert callback_mutex to a spinlock cpuset: simplify cpuset_node_allowed API slab: fix cpuset check in fallback_alloc slub: fix cpuset check in get_any_partial include/linux/cpuset.h | 37 +++-------- kernel/cpuset.c | 162 +++++++++++++++++------------------------------- mm/hugetlb.c | 2 +- mm/oom_kill.c | 2 +- mm/page_alloc.c | 6 +- mm/slab.c | 2 +- mm/slub.c | 2 +- mm/vmscan.c | 5 +- 8 files changed, 74 insertions(+), 144 deletions(-) -- 1.7.10.4 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>