The logic is currently broken so that PREEMPT_RT disables preemptible spinlock waiters, which is counter intuitive. Signed-off-by: Gregory Haskins <ghaskins@xxxxxxxxxx> --- kernel/spinlock.c | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/kernel/spinlock.c b/kernel/spinlock.c index c9bcf1b..b0e7f02 100644 --- a/kernel/spinlock.c +++ b/kernel/spinlock.c @@ -117,7 +117,7 @@ EXPORT_SYMBOL(__write_trylock_irqsave); * not re-enabled during lock-acquire (which the preempt-spin-ops do): */ #if !defined(CONFIG_PREEMPT) || !defined(CONFIG_SMP) || \ - defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_PREEMPT_RT) + defined(CONFIG_DEBUG_LOCK_ALLOC) void __lockfunc __read_lock(raw_rwlock_t *lock) { - To unsubscribe from this list: send the line "unsubscribe linux-rt-users" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html