[ thanks for the analysis Johannes ... involving also Oliver here; really seems like a false positive to me ] On Wed, 1 Apr 2009, Johannes Berg wrote: > On Tue, 2009-03-31 at 10:40 +0200, Jiri Kosina wrote: > > > > > Could you please send me your config? > > > Sure, attached. I haven't yet tried to reproduce on .29 though, which > > > this config is for (but I haven't changed it since, only taken it > > > forward). > > I've now gotten it again on 2.6.29-wl-20327-g8f2487d-dirty. > > I've analysed a bit more. > > Let's start from the bottom: > > -> #0 (&dev->mutex){--..}: > [<ffffffff80276e87>] check_prev_add+0x57/0x770 > [<ffffffff80277b96>] validate_chain+0x5f6/0x6b0 > [<ffffffff8027808f>] __lock_acquire+0x43f/0xa10 > [<ffffffff802786f1>] lock_acquire+0x91/0xc0 > [<ffffffff805cb56c>] mutex_lock_nested+0xfc/0x390 > [<ffffffff804d19e1>] input_disconnect_device+0x31/0xf0 > [<ffffffff804d1aba>] input_unregister_device+0x1a/0x110 > [<ffffffffa00df329>] bcm5974_disconnect+0x29/0x90 [bcm5974] > [<ffffffffa001c9ed>] usb_unbind_interface+0x6d/0x180 [usbcore] > [<ffffffff80498661>] __device_release_driver+0x81/0xc0 > [<ffffffff804987c0>] device_release_driver+0x30/0x50 > [<ffffffffa001ce48>] usb_driver_release_interface+0xc8/0xf0 [usbcore] > [<ffffffffa001cf99>] usb_forced_unbind_intf+0x39/0x90 [usbcore] > [<ffffffffa00121a5>] usb_reset_device+0xd5/0x220 [usbcore] > [<ffffffffa00957fa>] hid_reset+0x18a/0x280 [usbhid] > [<ffffffff8025d28d>] run_workqueue+0x10d/0x250 > > > Here we have hid_reset being called off schedule_work. It eventually > calls into bcm5974 which will, from its usb_driver disconnect call, call > input_unregister_device(), which acquires &dev->mutex. > > > -> #1 (polldev_mutex){--..}: > [<ffffffff802771e7>] check_prev_add+0x3b7/0x770 > [<ffffffff80277b96>] validate_chain+0x5f6/0x6b0 > [<ffffffff8027808f>] __lock_acquire+0x43f/0xa10 > [<ffffffff802786f1>] lock_acquire+0x91/0xc0 > [<ffffffff805cb12c>] mutex_lock_interruptible_nested+0xec/0x430 > [<ffffffffa0058151>] input_open_polled_device+0x21/0xd0 [input_polldev] > [<ffffffff804d1528>] input_open_device+0x98/0xc0 > [<ffffffffa009f468>] evdev_open+0x1c8/0x1f0 [evdev] > [<ffffffff804d099f>] input_open_file+0x10f/0x200 > [<ffffffff802e6e47>] chrdev_open+0x147/0x220 > [<ffffffff802e18eb>] __dentry_open+0x11b/0x350 > [<ffffffff802e1c37>] nameidata_to_filp+0x57/0x70 > [<ffffffff802f0e6e>] do_filp_open+0x1fe/0x970 > [<ffffffff802e16d0>] do_sys_open+0x80/0x110 > [<ffffffff802e17a0>] sys_open+0x20/0x30 > > This is another code path -- evdev triggered here. Any input polldev > will acquire polldev_mutex within its struct input_dev->open() callback, > and thus create a dependency of &dev->mutex on polldev_mutex because > input_open_device() is called with &dev->mutex held. > > -> #2 (cpu_add_remove_lock){--..}: > [<ffffffff802771e7>] check_prev_add+0x3b7/0x770 > [<ffffffff80277b96>] validate_chain+0x5f6/0x6b0 > [<ffffffff8027808f>] __lock_acquire+0x43f/0xa10 > [<ffffffff802786f1>] lock_acquire+0x91/0xc0 > [<ffffffff805cb56c>] mutex_lock_nested+0xfc/0x390 > [<ffffffff8024a447>] cpu_maps_update_begin+0x17/0x20 > [<ffffffff8025d7c8>] destroy_workqueue+0x38/0xb0 > [<ffffffffa0058115>] input_close_polled_device+0x45/0x60 [input_polldev] > [<ffffffff804d145c>] input_close_device+0x5c/0x90 > [<ffffffffa009f1e9>] evdev_release+0xa9/0xd0 [evdev] > [<ffffffff802e4e25>] __fput+0xd5/0x1e0 > [<ffffffff802e4f55>] fput+0x25/0x30 > [<ffffffff802e14f8>] filp_close+0x58/0x90 > [<ffffffff802e15ee>] sys_close+0xbe/0x120 > [<ffffffff8020bf0b>] system_call_fastpath+0x16/0x1b > [<ffffffffffffffff>] 0xffffffffffffffff > > This is cute. > So input-polldev uses its own workqueue, and it's singlethread. But > destroy_workqueue must stop CPU hotplug anyway, calls > cpu_map_update_begin() which locks cpu_add_remove_lock. > > -> #3 (events){--..}: > [<ffffffff802771e7>] check_prev_add+0x3b7/0x770 > [<ffffffff80277b96>] validate_chain+0x5f6/0x6b0 > [<ffffffff8027808f>] __lock_acquire+0x43f/0xa10 > [<ffffffff802786f1>] lock_acquire+0x91/0xc0 > [<ffffffff8025d722>] cleanup_workqueue_thread+0x42/0x90 > [<ffffffff805bca8d>] workqueue_cpu_callback+0x9d/0x132 > [<ffffffff805d1205>] notifier_call_chain+0x65/0xa0 > [<ffffffff80267626>] raw_notifier_call_chain+0x16/0x20 > [<ffffffff805ba91b>] _cpu_down+0x1db/0x350 > [<ffffffff8024a5b5>] disable_nonboot_cpus+0xe5/0x170 > [<ffffffff80287ef5>] hibernation_snapshot+0x135/0x170 > [<ffffffff8028b8a5>] snapshot_ioctl+0x425/0x620 > [<ffffffff802f27a6>] vfs_ioctl+0x36/0xb0 > [<ffffffff802f2b89>] do_vfs_ioctl+0x89/0x350 > [<ffffffff802f2e9f>] sys_ioctl+0x4f/0x80 > [<ffffffff8020bf0b>] system_call_fastpath+0x16/0x1b > [<ffffffffffffffff>] 0xffffffffffffffff > > Here we have hibernation, which needs to call disable_nonboot_cpus. This > takes down all CPUs, and causes the workqueue code, now running off the > workqueue_cpu_callback, to call cleanup_workqueue_thread(), which > "acquires" the workqueue. I suspect this will also happen if you go into > sysfs and disable a CPU manually, which may help you reproduce this. > > disable_nonboot_cpus calls cpu_map_update_begin() to avoid other things > interfering, and thus creates the dependency of the workqueue on that. > > > -> #4 (&usbhid->reset_work){--..}: > [<ffffffff802771e7>] check_prev_add+0x3b7/0x770 > [<ffffffff80277b96>] validate_chain+0x5f6/0x6b0 > [<ffffffff8027808f>] __lock_acquire+0x43f/0xa10 > [<ffffffff802786f1>] lock_acquire+0x91/0xc0 > [<ffffffff8025d287>] run_workqueue+0x107/0x250 > [<ffffffff8025d47f>] worker_thread+0xaf/0x130 > [<ffffffff80261dae>] kthread+0x4e/0x90 > [<ffffffff8020cfba>] child_rip+0xa/0x20 > [<ffffffffffffffff>] 0xffffffffffffffff > > Now, of course, usbhid->reset_work runs off the schedule_work workqueue, > which was stopped during hibernation, so it depends on that workqueue. > > > Finally, we're back at the top, with input_disconnect_device() acquiring > &dev->mutex. > > > > Now, how can a deadlock happen? > > I think it cannot -- unless you have a polled USB device. The two > "&dev->mutex" instances here are from difference devices, but lockdep > cannot tell them apart, and if you have a polled USB device then the > same can happen. > > Assume you had a polled USB driver using input_unregister_polled_device, > and thus input_unregister_device, in its usb_driver disconnect call. In > that case you could potentially trigger the deadlock when you manage to > get that usb device reset very very close before calling > disable_nonboot_cpus, so close that the usb reset_work is still > scheduled or something like that... > > > I don't really see a good way to solve it -- but I hope the analysis > helps some -- also adding lots of people to CC. > > johannes > -- To unsubscribe from this list: send the line "unsubscribe linux-input" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html