On 07/11/2014 04:25 AM, Peter Zijlstra wrote: > On Thu, Jul 10, 2014 at 03:02:29PM -0400, Sasha Levin wrote: >> What if we move lockdep's acquisition point to after it actually got the >> lock? > > NAK, you want to do deadlock detection _before_ you're stuck in a > deadlock. I didn't suggest to do it in the general case, but just for debugging the issue we have here. >> We'd miss deadlocks, but we don't care about them right now. Anyways, doesn't >> lockdep have anything built in to allow us to separate between locks which >> we attempt to acquire and locks that are actually acquired? >> >> (cc PeterZ) >> >> We can treat locks that are in the process of being acquired the same as >> acquired locks to avoid races, but when we print something out it would >> be nice to have annotation of the read state of the lock. > > I'm missing the problem here I think. The problem here is that lockdep reports tasks waiting on lock as ones that already have the lock. So we have a list of about 500 different tasks looking like this: [ 367.805809] 2 locks held by trinity-c214/9083: [ 367.805811] #0: (sb_writers#9){.+.+.+}, at: do_fallocate (fs/open.c:298) [ 367.805824] #1: (&sb->s_type->i_mutex_key#16){+.+.+.}, at: shmem_fallocate (mm/shmem.c:1738) While they haven't actually acquired i_mutex, but are merely blocking on it: [ 367.644150] trinity-c214 D 0000000000000002 13528 9083 8490 0x00000000 [ 367.644171] ffff880018757ce8 0000000000000002 ffffffff91a01d70 0000000000000001 [ 367.644178] ffff880018757fd8 00000000001d7740 00000000001d7740 00000000001d7740 [ 367.644188] ffff880006428000 ffff880018758000 ffff880018757cd8 ffff880031fdc210 [ 367.644213] Call Trace: [ 367.644218] schedule (kernel/sched/core.c:2832) [ 367.644229] schedule_preempt_disabled (kernel/sched/core.c:2859) [ 367.644237] mutex_lock_nested (kernel/locking/mutex.c:535 kernel/locking/mutex.c:587) [ 367.644240] ? shmem_fallocate (mm/shmem.c:1738) [ 367.644248] ? get_parent_ip (kernel/sched/core.c:2546) [ 367.644255] ? shmem_fallocate (mm/shmem.c:1738) [ 367.644264] shmem_fallocate (mm/shmem.c:1738) [ 367.644268] ? SyS_madvise (mm/madvise.c:334 mm/madvise.c:384 mm/madvise.c:534 mm/madvise.c:465) [ 367.644280] ? put_lock_stats.isra.12 (./arch/x86/include/asm/preempt.h:98 kernel/locking/lockdep.c:254) [ 367.644291] ? SyS_madvise (mm/madvise.c:334 mm/madvise.c:384 mm/madvise.c:534 mm/madvise.c:465) [ 367.644298] do_fallocate (include/linux/fs.h:1281 fs/open.c:299) [ 367.644303] SyS_madvise (mm/madvise.c:335 mm/madvise.c:384 mm/madvise.c:534 mm/madvise.c:465) [ 367.644309] ? context_tracking_user_exit (./arch/x86/include/asm/paravirt.h:809 (discriminator 2) kernel/context_tracking.c:184 (discriminator 2)) [ 367.644315] ? trace_hardirqs_on (kernel/locking/lockdep.c:2607) [ 367.644321] tracesys (arch/x86/kernel/entry_64.S:543) There's no easy way to see whether a given task is actually holding a lock or is just blocking on it without going through all those tasks one by one and looking at their trace. I agree with you that "The call trace is very clear on it that its not", but when you have 500 call traces you really want something better than going through it one call trace at a time. Thanks, Sasha -- To unsubscribe from this list: send the line "unsubscribe stable" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html