Re: Please add to stable: module: don't unlink the module until we've removed all exposure.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 06/04/2013 09:53 AM, Ben Greear wrote:
On 06/04/2013 07:07 AM, Joe Lawrence wrote:
On Tue, 04 Jun 2013 15:26:28 +0930
Rusty Russell <rusty@xxxxxxxxxxxxxxx> wrote:

Do you have a backtrace of the 3.9.4 crash?  You can add "CFLAGS_module.o
= -O0" to get a clearer backtrace if you want...

Hi Rusty,

See my 3.9 stack traces below, which may or may not be what Ben had
been seeing.  If you like, I can try a similar loop as the one you were
testing in the other email.

My stack traces are similar.  I had better luck reproducing the problem
once I enabled lots of debugging (slub memory poisoning, lockdep,
object debugging, etc).

I'm using Fedora 17 on 2-core core-i7 (4 CPU threads total) for most of this
testing.  We reproduced on dual-core Atom system as well
(32-bit Fedora 14 and Fedora 17).  Relatively standard hardware as far
as I know.

I'll run the insmod/rmmod stress test on my patched systems
and see if I can reproduce with the patch in the title applied.

Rusty:  I'm also seeing lockups related to migration on stock 3.9.4+
(with and without the 'don't unlink the module...' patch.  Much harder
to reproduce.  But, that code appears to be mostly called during
module load/unload, so it's possible it is related.  The first
traces are from a system with local patches, applied, but a later
post by me has traces from clean upstream kernel.

Further debugging showed that this could be a race, because it seems
that all migration/ threads think they are done with their state machine,
but the atomic thread counter sits at 1, so no progress is ever made.

http://www.mail-archive.com/linux-kernel@xxxxxxxxxxxxxxx/msg443471.html

I reproduced the migration deadlock after a while (loading and unloading
the macvlan module with this command:

for i in `seq 10000`; do modprobe macvlan; rmmod macvlan; done

I did not see the kobj crash, but this kernel was running your patch
that makes the problem go away for me, for whatever reason.

I have some printk debugging in (see bottom of email) and was using a serial console, so things
were probably running a bit slower than on most systems.  Here is trace
from my kernel with local patches and not so much debugging enabled
(this is NOT a clean upstream kernel, though I reproduced the same thing
with a clean upstream 3.9.4 kernel plus your module unlink patch yesterday).

__stop_machine, num-threads: 4, fn: __try_stop_module  data: ffff8801c6ae7f28
cpu: 0 loops: 1 jiffies: 4299011449  timeout: 4299011448 curstate: 0  smdata->state: 1  thread_ack: 4
cpu: 1 loops: 1 jiffies: 4299011449  timeout: 4299011448 curstate: 0  smdata->state: 1  thread_ack: 4
cpu: 2 loops: 1 jiffies: 4299011449  timeout: 4299011448 curstate: 0  smdata->state: 1  thread_ack: 3
cpu: 3 loops: 1 jiffies: 4299011449  timeout: 4299011448 curstate: 0  smdata->state: 1  thread_ack: 2
__stop_machine, num-threads: 4, fn: __unlink_module  data: ffffffffa0aeeab0
cpu: 0 loops: 1 jiffies: 4299011501  timeout: 4299011500 curstate: 0  smdata->state: 1  thread_ack: 4
cpu: 1 loops: 1 jiffies: 4299011501  timeout: 4299011500 curstate: 0  smdata->state: 1  thread_ack: 4
cpu: 3 loops: 1 jiffies: 4299011501  timeout: 4299011500 curstate: 0  smdata->state: 1  thread_ack: 3
ath: wiphy0: Failed to stop TX DMA, queues=0x005!
cpu: 2 loops: 1 jiffies: 4299011501  timeout: 4299011500 curstate: 0  smdata->state: 1  thread_ack: 2
BUG: soft lockup - CPU#3 stuck for 23s! [migration/3:29]
Modules linked in: nf_nat_ipv4 nf_nat veth 8021q garp stp mrp llc wanlink(O) pktgen lockd sunrpc coretemp hwmon mperf intel_po]
CPU 3
Pid: 29, comm: migration/3 Tainted: G         C O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff8109d69c>]  [<ffffffff8109d69c>] tasklet_action+0x46/0xcc
RSP: 0000:ffff88022bd83ed8  EFLAGS: 00000282
RAX: ffff88022bd8e080 RBX: ffff8802222a4000 RCX: ffffffff81a90f06
RDX: ffff88021f48afa8 RSI: 0000000000000000 RDI: ffffffff81a050b0
RBP: ffff88022bd83ee8 R08: 0000000000000000 R09: 0000000000000000
R10: 00000000000005f2 R11: 00000000fd010018 R12: ffff88022bd83e48
R13: ffffffff815d145d R14: ffff88022bd83ee8 R15: ffff880222282000
FS:  0000000000000000(0000) GS:ffff88022bd80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fa8a06d5000 CR3: 00000002007ea000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/3 (pid: 29, threadinfo ffff880222282000, task ffff880222279770)
Stack:
 ffffffff81a050b0 ffff880222282000 ffff88022bd83f78 ffffffff8109db1f
 ffff88022bd83f08 ffff880222282010 ffff880222283fd8 04208040810b79ef
 00000001003db5d8 000000032bd8e150 ffff880222282000 0000000000000030
Call Trace:
 <IRQ>
 [<ffffffff8109db1f>] __do_softirq+0x107/0x23c
 [<ffffffff8109dce6>] irq_exit+0x4b/0xa8
 [<ffffffff815d257f>] smp_apic_timer_interrupt+0x8b/0x99
 [<ffffffff815d145d>] apic_timer_interrupt+0x6d/0x80
 <EOI>
 [<ffffffff810f9984>] ? stop_machine_cpu_stop+0xc7/0x145
 [<ffffffff810f9974>] ? stop_machine_cpu_stop+0xb7/0x145
 [<ffffffff81016445>] ? sched_clock+0x9/0xd
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 48 8b 14 25 80 e0 00 00 65 48 c7 04 25 80 e0 00 00 00 00 00 00 65 48 03 04 25 d8 da 00 00 65 48 89 04 25 88 e0 00 00 fb
------------[ cut here ]------------
WARNING: at /home/greearb/git/linux-3.9.dev.y/kernel/watchdog.c:245 watchdog_overflow_callback+0x9b/0xa6()
Hardware name: To be filled by O.E.M.
Watchdog detected hard LOCKUP on cpu 1
Modules linked in: nf_nat_ipv4 nf_nat veth 8021q garp stp mrp llc wanlink(O) pktgen lockd sunrpc coretemp hwmon mperf intel_po]
Pid: 17, comm: migration/1 Tainted: G         C O 3.9.4+ #60
Call Trace:
 <NMI>  [<ffffffff810963e1>] warn_slowpath_common+0x85/0x9f
 [<ffffffff8109649e>] warn_slowpath_fmt+0x46/0x48
 [<ffffffff810c4dfb>] ? sched_clock_cpu+0x44/0xce
 [<ffffffff81103b70>] watchdog_overflow_callback+0x9b/0xa6
 [<ffffffff8113354f>] __perf_event_overflow+0x137/0x1cb
 [<ffffffff8101db3f>] ? x86_perf_event_set_period+0x107/0x113
 [<ffffffff81133a9e>] perf_event_overflow+0x14/0x16
 [<ffffffff810230dc>] intel_pmu_handle_irq+0x2b0/0x32d
 [<ffffffff815cbc51>] perf_event_nmi_handler+0x19/0x1b
 [<ffffffff815cb4ca>] nmi_handle+0x55/0x7e
 [<ffffffff815cb59b>] do_nmi+0xa8/0x2db
 [<ffffffff815cac31>] end_repeat_nmi+0x1e/0x2e
 [<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
 [<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
 [<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
 <<EOE>>  [<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
---[ end trace dcd772d6fdf499cf ]---

...


SysRq : Show backtrace of all active CPUs
sending NMI to all CPUs:
NMI backtrace for cpu 2
CPU 2
Pid: 23, comm: migration/2 Tainted: G        WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff810f9987>]  [<ffffffff810f9987>] stop_machine_cpu_stop+0xca/0x145
RSP: 0018:ffff880222219cf8  EFLAGS: 00000012
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000097
RDX: 0000000000000002 RSI: 0000000000000006 RDI: 0000000000000246
RBP: ffff880222219d68 R08: 00000001003dc935 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000002 R15: 0000000000000002
FS:  0000000000000000(0000) GS:ffff88022bd00000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00000000027e2178 CR3: 000000021955d000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/2 (pid: 23, threadinfo ffff880222218000, task ffff880222210000)
Stack:
 ffff880200000001 ffff880200000002 ffff880222219d28 ffffffff810c6b1c
 ffff880222219d88 00ffffff8100f8a4 000000051058262e 0000000000000292
 ffff880222219d58 ffff88022bd0e400 ffff8801c6ae7d08 ffff880222218000
Call Trace:
 [<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00 00 f3 90
NMI backtrace for cpu 0
CPU 0
Pid: 8, comm: migration/0 Tainted: G        WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E..
RIP: 0010:[<ffffffff810f9984>]  [<ffffffff810f9984>] stop_machine_cpu_stop+0xc7/0x145
RSP: 0000:ffff880222145cf8  EFLAGS: 00000006
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000004
RDX: 0000000000000002 RSI: ffff88022bc0de68 RDI: 0000000000000246
RBP: ffff880222145d68 R08: 00000001003dc95f R09: 0000000000000001
R10: ffff880222145bf8 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000002
FS:  0000000000000000(0000) GS:ffff88022bc00000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00000000006fb9a8 CR3: 000000021af43000 CR4: 00000000000007f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/0 (pid: 8, threadinfo ffff880222144000, task ffff88022213aee0)
Stack:
 ffff880200000001 ffff880200000004 ffff880222145d28 ffffffff810c6b1c
 ffff880222145d88 01ffffff8100f8a4 000000051023ce2c 0000000000000292
 ffff880222145d98 ffff88022bc0e400 ffff8801c6ae7d08 ffff880222144000
Call Trace:
 [<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 44 89 f6 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00
NMI backtrace for cpu 1
CPU 1
Pid: 17, comm: migration/1 Tainted: G        WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff810f9984>]  [<ffffffff810f9984>] stop_machine_cpu_stop+0xc7/0x145
RSP: 0000:ffff88022217dcf8  EFLAGS: 00000012
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000094
RDX: 0000000000000002 RSI: 0000000000000006 RDI: 0000000000000246
RBP: ffff88022217dd68 R08: 00000001003dc935 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000001 R15: 0000000000000002
FS:  0000000000000000(0000) GS:ffff88022bc80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fbe801fd000 CR3: 00000001be9bb000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/1 (pid: 17, threadinfo ffff88022217c000, task ffff88022216ddc0)
Stack:
 ffff880200000001 ffff880200000004 ffff88022217dd28 ffffffff810c6b1c
 ffff88022217dd88 00ffffff8100f8a4 00000004c3cbfc0c 0000000000000292
 ffff88022217dd58 ffff88022bc8e400 ffff8801c6ae7d08 ffff88022217c000
Call Trace:
 [<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 44 89 f6 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00
NMI backtrace for cpu 3
CPU 3
Pid: 29, comm: migration/3 Tainted: G        WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff812f6975>]  [<ffffffff812f6975>] delay_tsc+0x83/0xee
RSP: 0000:ffff88022bd83b60  EFLAGS: 00000046
RAX: 00000b04f0fdf718 RBX: ffff880222282000 RCX: ffff880222282010
RDX: 0000000000000000 RSI: 0000000000000002 RDI: 0000000000289672
RBP: ffff88022bd83bb0 R08: 0000000000000040 R09: 0000000000000001
R10: ffff88022bd83ad0 R11: 0000000000000000 R12: 00000000f0fdf6e8
R13: 0000000000000003 R14: ffff880222282000 R15: 0000000000000001
FS:  0000000000000000(0000) GS:ffff88022bd80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fa8a06d5000 CR3: 00000002007ea000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/3 (pid: 29, threadinfo ffff880222282000, task ffff880222279770)
Stack:
 000000000000000f 0000000000000002 ffff880222282010 0028967200000082
 ffff88022bd83b90 0000000000000001 000000000000006c 0000000000000007
 0000000000000086 0000000000000001 ffff88022bd83bc0 ffffffff812f68c5
Call Trace:
 <IRQ>
 [<ffffffff812f68c5>] __const_udelay+0x28/0x2a
 [<ffffffff8102ff25>] arch_trigger_all_cpu_backtrace+0x66/0x7d
 [<ffffffff813ae154>] sysrq_handle_showallcpus+0xe/0x10
 [<ffffffff813ae46b>] __handle_sysrq+0xbf/0x15b
 [<ffffffff813ae80e>] handle_sysrq+0x2c/0x2e
 [<ffffffff813c25a2>] serial8250_rx_chars+0x13c/0x1b9
 [<ffffffff813c2691>] serial8250_handle_irq+0x72/0xa8
 [<ffffffff813c2752>] serial8250_default_handle_irq+0x23/0x28
 [<ffffffff813c148c>] serial8250_interrupt+0x4d/0xc6
 [<ffffffff811046b8>] handle_irq_event_percpu+0x7a/0x1e5
 [<ffffffff81104864>] handle_irq_event+0x41/0x61
 [<ffffffff81107028>] handle_edge_irq+0xa6/0xcb
 [<ffffffff81011d9f>] handle_irq+0x24/0x2d
 [<ffffffff815d248d>] do_IRQ+0x4d/0xb4
 [<ffffffff815ca5ad>] common_interrupt+0x6d/0x6d
 [<ffffffff810a47ee>] ? run_timer_softirq+0x24/0x1df
 [<ffffffff8109dabd>] ? __do_softirq+0xa5/0x23c
 [<ffffffff8109db8a>] ? __do_softirq+0x172/0x23c
 [<ffffffff8109dce6>] irq_exit+0x4b/0xa8
 [<ffffffff815d257f>] smp_apic_timer_interrupt+0x8b/0x99
 [<ffffffff815d145d>] apic_timer_interrupt+0x6d/0x80
 <EOI>
 [<ffffffff810f9984>] ? stop_machine_cpu_stop+0xc7/0x145
 [<ffffffff810f9974>] ? stop_machine_cpu_stop+0xb7/0x145
 [<ffffffff81016445>] ? sched_clock+0x9/0xd
 [<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
 [<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
 [<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
 [<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
 [<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
 [<ffffffff810b4a09>] kthread+0xb5/0xbd
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
 [<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
 [<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: fa d1 ff 66 90 89 c2 44 29 e2 3b 55 cc 73 4a ff 4b 1c 48 8b 4d c0 48 8b 11 80 e2 08 74 0b 89 45 b8 e8 08 28 2d 00 8b 45


commit 5b783a651bb9941923cd8dbff6d0ce80d2617f97
Author: Ben Greear <greearb@xxxxxxxxxxxxxxx>
Date:   Mon Jun 3 13:31:09 2013 -0700

    debugging:  Instrument the 'migration/[0-x]' process a bit.

    Hoping to figure out the deadlock we see occassionally due
    to migration processes all spinning in busy loop.

    Signed-off-by: Ben Greear <greearb@xxxxxxxxxxxxxxx>

diff --git a/kernel/stop_machine.c b/kernel/stop_machine.c
index c09f295..f7a1de5 100644
--- a/kernel/stop_machine.c
+++ b/kernel/stop_machine.c
@@ -408,6 +408,9 @@ static int stop_machine_cpu_stop(void *data)
        int cpu = smp_processor_id(), err = 0;
        unsigned long flags;
        bool is_active;
+       unsigned long loops = 0;
+       unsigned long start_at = jiffies;
+       unsigned long timeout = start_at - 1;

        /*
         * When called from stop_machine_from_inactive_cpu(), irq might
@@ -422,6 +425,13 @@ static int stop_machine_cpu_stop(void *data)

        /* Simple state machine */
        do {
+               loops++;
+               if (time_after(jiffies, timeout)) {
+                       printk("cpu: %i loops: %lu jiffies: %lu  timeout: %lu curstate: %i  smdata->state:
+                              cpu, loops, jiffies, timeout, curstate, smdata->state,
+                              atomic_read(&smdata->thread_ack));
+                       timeout = jiffies + 5000;
+               }
                /* Chill out and ensure we re-read stopmachine_state. */
                cpu_relax();
                if (smdata->state != curstate) {
@@ -473,6 +483,14 @@ int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus)

        /* Set the initial state and stop all online cpus. */
        set_state(&smdata, STOPMACHINE_PREPARE);
+       {
+               char ksym_buf[KSYM_NAME_LEN];
+               printk("__stop_machine, num-threads: %i, fn: %s  data: %p\n",
+                      smdata.num_threads,
+                      kallsyms_lookup((unsigned long)fn, NULL, NULL, NULL,
+                                      ksym_buf),
+                      data);
+       }
        return stop_cpus(cpu_online_mask, stop_machine_cpu_stop, &smdata);
 }

@@ -517,10 +535,16 @@ int stop_machine_from_inactive_cpu(int (*fn)(void *), void *data,
                                            .active_cpus = cpus };
        struct cpu_stop_done done;
        int ret;
+       char ksym_buf[KSYM_NAME_LEN];

        /* Local CPU must be inactive and CPU hotplug in progress. */
        BUG_ON(cpu_active(raw_smp_processor_id()));
        smdata.num_threads = num_active_cpus() + 1;     /* +1 for local */
+       printk("stop-machine-from-inactive-cpu, num-threads: %i  fn: %s(%p)\n",
+              smdata.num_threads,
+              kallsyms_lookup((unsigned long)fn, NULL, NULL, NULL,
+                              ksym_buf),
+              data);

        /* No proper task established and can't sleep - busy wait for lock. */
        while (!mutex_trylock(&stop_cpus_mutex))


Thanks,
Ben


--
Ben Greear <greearb@xxxxxxxxxxxxxxx>
Candela Technologies Inc  http://www.candelatech.com

--
To unsubscribe from this list: send the line "unsubscribe stable" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux Kernel]     [Kernel Development Newbies]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite Hiking]     [Linux Kernel]     [Linux SCSI]