This is an update of the MCS lock patch series posted in November. Proper passing of the mcs lock is now done with smp_load_acquire() in mcs_spin_lock() and smp_store_release() in mcs_spin_unlock. Note that this is not sufficient to form a full memory barrier across cpus on many architectures (except x86) for the mcs_unlock and mcs_lock pair. For code that needs a full memory barrier, smp_mb__after_unlock_lock() should be used after mcs_lock. I will appreciate Paul and other experts review this portion of the code. Will also added hooks to allow for architecture specific implementation and optimization of the of the contended paths of lock and unlock of mcs_spin_lock and mcs_spin_unlock functions. The original mcs lock code has potential leaks between critical sections, which was not a problem when MCS was embedded within the mutex but needs to be corrected when allowing the MCS lock to be used by itself for other locking purposes. The MCS lock code was previously embedded in the mutex.c and is now sepearted. This allows for easier reuse of MCS lock in other places like rwsem and qrwlock. We also did some micro optimizations and barrier cleanup. Tim v7: 1. Update architecture specific hooks with concise architecture specific arch_mcs_spin_lock_contended and arch_mcs_spin_lock_uncontended functions. v6: 1. Fix a bug of improper xchg_acquire and extra space in barrier fixing patch. 2. Added extra hooks to allow for architecture specific version of mcs_spin_lock and mcs_spin_unlock to be used. v5: 1. Rework barrier correction patch. We now use smp_load_acquire() in mcs_spin_lock() and smp_store_release() in mcs_spin_unlock() to allow for architecture dependent barriers to be automatically used. This is clean and will provide the right barriers for all architecture. v4: 1. Move patch series to the latest tip after v3.12 v3: 1. modified memory barriers to support non x86 architectures that have weak memory ordering. v2: 1. change export mcs_spin_lock as a GPL export symbol 2. corrected mcs_spin_lock to references Jason Low (1): MCS Lock: optimizations and extra comments Tim Chen (1): MCS Lock: Restructure the MCS lock defines and locking code into its own file Waiman Long (2): MCS Lock: Move mcs_lock/unlock function into its own file MCS Lock: Barrier corrections Will Deacon (2): MCS Lock: allow architectures to hook in to contended paths MCS Lock: add Kconfig entries to allow arch-specific hooks arch/Kconfig | 3 ++ include/linux/mcs_spinlock.h | 33 ++++++++++++++++ include/linux/mutex.h | 5 ++- kernel/locking/Makefile | 6 +-- kernel/locking/mcs_spinlock.c | 89 +++++++++++++++++++++++++++++++++++++++++++ kernel/locking/mutex.c | 60 ++++------------------------- 6 files changed, 138 insertions(+), 58 deletions(-) create mode 100644 include/linux/mcs_spinlock.h create mode 100644 kernel/locking/mcs_spinlock.c -- 1.7.11.7 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>