Hi, This is a follow up to my previous series (http://lkml.kernel.org/g/<1453770913-32287-1-git-send-email-labbott@xxxxxxxxxxxxxxxxx>) This series takes the suggestion of Christoph Lameter and only focuses on optimizing the slow path where the debug processing runs. The two main optimizations in this series are letting the consistency checks be skipped and relaxing the cmpxchg restrictions when we are not doing consistency checks. With hackbench -g 20 -l 1000 averaged over 100 runs: Before slub_debug=P mean 15.607 variance .086 stdev .294 After slub_debug=P mean 10.836 variance .155 stdev .394 This still isn't as fast as what is in grsecurity unfortunately so there's still work to be done. Profiling ___slab_alloc shows that 25-50% of time is spent in deactivate_slab. I haven't looked too closely to see if this is something that can be optimized. My plan for now is to focus on getting all of this merged (if appropriate) before digging in to another task. As always feedback is appreciated. Laura Abbott (4): slub: Drop lock at the end of free_debug_processing slub: Fix/clean free_debug_processing return paths sl[aob]: Convert SLAB_DEBUG_FREE to SLAB_CONSISTENCY_CHECKS slub: Relax CMPXCHG consistency restrictions Documentation/vm/slub.txt | 4 +- include/linux/slab.h | 2 +- mm/slab.h | 5 +- mm/slub.c | 126 ++++++++++++++++++++++++++++------------------ tools/vm/slabinfo.c | 2 +- 5 files changed, 83 insertions(+), 56 deletions(-) -- 2.5.0 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>