Hi all, When offline the whole memory of a movable numa node on kernel stable-3.10-y, the following possible deadlock warning occurs. [ 2457.467359] [ 2457.485175] ================================= [ 2457.537325] [ INFO: inconsistent lock state ] [ 2457.589476] 3.10.39+ #4 Not tainted [ 2457.631218] --------------------------------- [ 2457.683370] inconsistent {RECLAIM_FS-ON-W} -> {IN-RECLAIM_FS-R} usage. [ 2457.761540] kswapd2/1151 [HC0[0]:SC0[0]:HE1:SE1] takes: [ 2457.824102] (&sig->group_rwsem){+++++?}, at: [<ffffffff81071864>] exit_signals+0x24/0x130 [ 2457.923538] {RECLAIM_FS-ON-W} state was registered at: [ 2457.985055] [<ffffffff810bfc99>] mark_held_locks+0xb9/0x140 [ 2458.053976] [<ffffffff810c1e3a>] lockdep_trace_alloc+0x7a/0xe0 [ 2458.126015] [<ffffffff81194f47>] kmem_cache_alloc_trace+0x37/0x240 [ 2458.202214] [<ffffffff812c6e89>] flex_array_alloc+0x99/0x1a0 [ 2458.272175] [<ffffffff810da563>] cgroup_attach_task+0x63/0x430 [ 2458.344214] [<ffffffff810dcca0>] attach_task_by_pid+0x210/0x280 [ 2458.417294] [<ffffffff810dcd26>] cgroup_procs_write+0x16/0x20 [ 2458.488287] [<ffffffff810d8410>] cgroup_file_write+0x120/0x2c0 [ 2458.560320] [<ffffffff811b21a0>] vfs_write+0xc0/0x1f0 [ 2458.622994] [<ffffffff811b2bac>] SyS_write+0x4c/0xa0 [ 2458.684618] [<ffffffff815ec3c0>] tracesys+0xdd/0xe2 [ 2458.745214] irq event stamp: 49 [ 2458.782794] hardirqs last enabled at (49): [<ffffffff815e2b56>] _raw_spin_unlock_irqrestore+0x36/0x70 [ 2458.894388] hardirqs last disabled at (48): [<ffffffff815e337b>] _raw_spin_lock_irqsave+0x2b/0xa0 [ 2459.000771] softirqs last enabled at (0): [<ffffffff81059247>] copy_process.part.24+0x627/0x15f0 [ 2459.107161] softirqs last disabled at (0): [< (null)>] (null) [ 2459.195852] [ 2459.195852] other info that might help us debug this: [ 2459.274024] Possible unsafe locking scenario: [ 2459.274024] [ 2459.344911] CPU0 [ 2459.374161] ---- [ 2459.403408] lock(&sig->group_rwsem); [ 2459.448490] <Interrupt> [ 2459.479825] lock(&sig->group_rwsem); [ 2459.526979] [ 2459.526979] *** DEADLOCK *** [ 2459.526979] [ 2459.597866] no locks held by kswapd2/1151. [ 2459.646896] [ 2459.646896] stack backtrace: [ 2459.699049] CPU: 30 PID: 1151 Comm: kswapd2 Not tainted 3.10.39+ #4 [ 2459.774098] Hardware name: FUJITSU PRIMEQUEST2800E/SB, BIOS PRIMEQUEST 2000 Series BIOS Version 01.48 05/07/2014 [ 2459.895983] ffffffff82284bf0 ffff88085856bbf8 ffffffff815dbcf6 ffff88085856bc48 [ 2459.985003] ffffffff815d67c6 0000000000000000 ffff880800000001 ffff880800000001 [ 2460.074024] 000000000000000a ffff88085edc9600 ffffffff810be0e0 0000000000000009 [ 2460.163087] Call Trace: [ 2460.192345] [<ffffffff815dbcf6>] dump_stack+0x19/0x1b [ 2460.253874] [<ffffffff815d67c6>] print_usage_bug+0x1f7/0x208 [ 2460.322679] [<ffffffff810be0e0>] ? check_usage_backwards+0x160/0x160 [ 2460.399807] [<ffffffff810bfb5d>] mark_lock+0x21d/0x2a0 [ 2460.462369] [<ffffffff810c076a>] __lock_acquire+0x52a/0xb60 [ 2460.530136] [<ffffffff8101acd3>] ? native_sched_clock+0x13/0x80 [ 2460.602065] [<ffffffff8101ad49>] ? sched_clock+0x9/0x10 [ 2460.665668] [<ffffffff81096f05>] ? sched_clock_cpu+0xb5/0x100 [ 2460.735516] [<ffffffff810c1592>] lock_acquire+0xa2/0x140 [ 2460.800156] [<ffffffff81071864>] ? exit_signals+0x24/0x130 [ 2460.866885] [<ffffffff81158ca0>] ? balance_pgdat+0x5e0/0x5e0 [ 2460.935691] [<ffffffff815e01e1>] down_read+0x51/0xa0 [ 2460.996166] [<ffffffff81071864>] ? exit_signals+0x24/0x130 [ 2461.062888] [<ffffffff81071864>] exit_signals+0x24/0x130 [ 2461.127536] [<ffffffff81060d55>] do_exit+0xb5/0xa50 [ 2461.186976] [<ffffffff810841e0>] ? wake_up_bit+0x30/0x30 [ 2461.251629] [<ffffffff81158ca0>] ? balance_pgdat+0x5e0/0x5e0 [ 2461.320433] [<ffffffff8108303b>] kthread+0xdb/0x100 [ 2461.379870] [<ffffffff815e12eb>] ? wait_for_completion+0x3b/0x110 [ 2461.453879] [<ffffffff81082f60>] ? kthread_create_on_node+0x140/0x140 [ 2461.532049] [<ffffffff815ec0ec>] ret_from_fork+0x7c/0xb0 [ 2461.596689] [<ffffffff81082f60>] ? kthread_create_on_node+0x140/0x140 And when reference to the related code(kernel-3.10.y), it seems that cgroup_attach_task(thread-2, attach kswapd) trigger kswapd(reclaim memory?) when trying to alloc memory(flex_array_alloc) under the protection of sig->group_rwsem, but meanwhile the kswapd(thread-1) is in the exit routine (because it was marked SHOULD STOP when offline pages completed), which needs to acquire sig->group_rwsem in exit_signals(), so the deadlock occurs. thread-1 | thread-2 | __offline_pages(): | system_call_fastpath() |-> kswapd_stop(node); | |-> ...... |-> kthread_stop(kswapd) | |-> cgroup_file_write() |-> set_bit(KTHREAD_SHOULD_STOP, &kthread->flags); | |-> ...... |-> wake_up_process(k) | |-> attach_task_by_pid() | | |-> threadgroup_lock(tsk) |<----------| | // Here, got the lock. |-> kswapd() | |-> ... |-> if (kthread_should_stop()) | |-> cgroup_attach_task() return; | |-> flex_array_alloc() | | |-> kzalloc() |<----------| | |-> wait for kswapd to reclaim memory |-> kthread() | |-> do_exit(ret) | |-> exit_signals() | |-> threadgroup_change_begin(tsk) | |-> down_read(&tsk->signal->group_rwsem) | // Here, acquire the lock. If my analysis is correct, the latest kernel may have the same issue, though the flex_array was replaced by list, but we still need to alloc memory(e.g. in find_css_set()), so the race may still occur. Any comments about this? If I missed something, please correct me.:) Regards, Gu -- To unsubscribe from this list: send the line "unsubscribe stable" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html