On Tue, Mar 5, 2024 at 8:20 AM Kai-Heng Feng <kai.heng.feng@xxxxxxxxxxxxx> wrote: > > On Tue, Mar 5, 2024 at 2:10 AM Rafael J. Wysocki <rafael@xxxxxxxxxx> wrote: > > > > On Mon, Mar 4, 2024 at 6:00 PM Rafael J. Wysocki <rafael@xxxxxxxxxx> wrote: > > > > > > On Mon, Mar 4, 2024 at 5:41 PM Rafael J. Wysocki <rafael@xxxxxxxxxx> wrote: > > > > > > > > On Mon, Mar 4, 2024 at 4:51 PM Bjorn Helgaas <helgaas@xxxxxxxxxx> wrote: > > > > > > > > > > On Mon, Mar 04, 2024 at 03:38:38PM +0100, Rafael J. Wysocki wrote: > > > > > > On Thu, Feb 29, 2024 at 7:23 AM Kai-Heng Feng > > > > > > <kai.heng.feng@xxxxxxxxxxxxx> wrote: > > > > > > > > > > > > > > When inserting an SD7.0 card to Realtek card reader, the card reader > > > > > > > unplugs itself and morph into a NVMe device. The slot Link down on hot > > > > > > > unplugged can cause the following error: > > > > > > > > > > > > > > pcieport 0000:00:1c.0: pciehp: Slot(8): Link Down > > > > > > > BUG: unable to handle page fault for address: ffffb24d403e5010 > > > > > > > PGD 100000067 P4D 100000067 PUD 1001fe067 PMD 100d97067 PTE 0 > > > > > > > Oops: 0000 [#1] PREEMPT SMP PTI > > > > > > > CPU: 3 PID: 534 Comm: kworker/3:10 Not tainted 6.4.0 #6 > > > > > > > Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./H370M Pro4, BIOS P3.40 10/25/2018 > > > > > > > Workqueue: pm pm_runtime_work > > > > > > > RIP: 0010:ioread32+0x2e/0x70 > > > > > > ... > > > > > > > Call Trace: > > > > > > > <TASK> > > > > > > > ? show_regs+0x68/0x70 > > > > > > > ? __die_body+0x20/0x70 > > > > > > > ? __die+0x2b/0x40 > > > > > > > ? page_fault_oops+0x160/0x480 > > > > > > > ? search_bpf_extables+0x63/0x90 > > > > > > > ? ioread32+0x2e/0x70 > > > > > > > ? search_exception_tables+0x5f/0x70 > > > > > > > ? kernelmode_fixup_or_oops+0xa2/0x120 > > > > > > > ? __bad_area_nosemaphore+0x179/0x230 > > > > > > > ? bad_area_nosemaphore+0x16/0x20 > > > > > > > ? do_kern_addr_fault+0x8b/0xa0 > > > > > > > ? exc_page_fault+0xe5/0x180 > > > > > > > ? asm_exc_page_fault+0x27/0x30 > > > > > > > ? ioread32+0x2e/0x70 > > > > > > > ? rtsx_pci_write_register+0x5b/0x90 [rtsx_pci] > > > > > > > rtsx_set_l1off_sub+0x1c/0x30 [rtsx_pci] > > > > > > > rts5261_set_l1off_cfg_sub_d0+0x36/0x40 [rtsx_pci] > > > > > > > rtsx_pci_runtime_idle+0xc7/0x160 [rtsx_pci] > > > > > > > ? __pfx_pci_pm_runtime_idle+0x10/0x10 > > > > > > > pci_pm_runtime_idle+0x34/0x70 > > > > > > > rpm_idle+0xc4/0x2b0 > > > > > > > pm_runtime_work+0x93/0xc0 > > > > > > > process_one_work+0x21a/0x430 > > > > > > > worker_thread+0x4a/0x3c0 > > > > > > ... > > > > > > > > > > > > This happens because scheduled pm_runtime_idle() is not cancelled. > > > > > > > > > > > > But rpm_resume() changes dev->power.request to RPM_REQ_NONE and if > > > > > > pm_runtime_work() sees this, it will not run rpm_idle(). > > > > > > > > > > > > However, rpm_resume() doesn't deactivate the autosuspend timer if it > > > > > > is running (see the comment in rpm_resume() regarding this), so it may > > > > > > queue up a runtime PM work later. > > > > > > > > > > > > If this is not desirable, you need to stop the autosuspend timer > > > > > > explicitly in addition to calling pm_runtime_get_sync(). > > > > > > > > > > I don't quite follow all this. I think the race is between > > > > > rtsx_pci_remove() (not resume) and rtsx_pci_runtime_idle(). > > > > > > > > I think so too and the latter is not expected to run. > > > > > > > > > rtsx_pci_remove() > > > > > { > > > > > pm_runtime_get_sync() > > > > > pm_runtime_forbid() > > > > > ... > > > > > > > > > > If this is an rtsx bug, what exactly should be added to > > > > > rtsx_pci_remove()? > > > > > > > > > > Is there ever a case where we want any runtime PM work to happen > > > > > during or after a driver .remove()? If not, maybe the driver core > > > > > should prevent that, which I think is basically what this patch does. > > > > > > > > No, it is not, because it doesn't actually prevent the race from > > > > occurring, it just narrows the window quite a bit. > > > > > > > > It would be better to call pm_runtime_dont_use_autosuspend() instead > > > > of pm_runtime_barrier(). > > > > > > > > > If this is an rtsx driver bug, I'm concerned there may be many other > > > > > drivers with a similar issue. rtsx exercises this path more than most > > > > > because the device switches between card reader and NVMe SSD using > > > > > hotplug add/remove based on whether an SD card is inserted (see [1]). > > > > > > > > This is a valid concern, so it is mostly a matter of where to disable > > > > autosuspend. > > > > > > > > It may be the driver core in principle, but note that it calls > > > > ->remove() after invoking pm_runtime_put_sync(), so why would it > > > > disable autosuspend when it allows runtime PM to race with device > > > > removal in general? > > > > > > > > Another way might be to add a pm_runtime_dont_use_autosuspend() call > > > > at the beginning of pci_device_remove(). > > > > > > > > Or just remove the optimization in question from rpm_resume() which is > > > > quite confusing and causes people to make assumptions that lead to > > > > incorrect behavior in this particular case. > > > > > > Well, scratch this. > > > > > > If rpm_idle() is already running at the time rpm_resume() is called, > > > the latter may return right away without waiting, which is incorrect. > > > > > > rpm_resume() needs to wait for the "idle" callback to complete, so > > > this (again, modulo GMail-induced whitespace mangling) should help: > > > > > > --- > > > drivers/base/power/runtime.c | 6 ++++-- > > > 1 file changed, 4 insertions(+), 2 deletions(-) > > > > > > Index: linux-pm/drivers/base/power/runtime.c > > > =================================================================== > > > --- linux-pm.orig/drivers/base/power/runtime.c > > > +++ linux-pm/drivers/base/power/runtime.c > > > @@ -798,7 +798,8 @@ static int rpm_resume(struct device *dev > > > } > > > > > > if (dev->power.runtime_status == RPM_RESUMING || > > > - dev->power.runtime_status == RPM_SUSPENDING) { > > > + dev->power.runtime_status == RPM_SUSPENDING || > > > + dev->power.idle_notification) { > > > DEFINE_WAIT(wait); > > > > > > if (rpmflags & (RPM_ASYNC | RPM_NOWAIT)) { > > > @@ -826,7 +827,8 @@ static int rpm_resume(struct device *dev > > > prepare_to_wait(&dev->power.wait_queue, &wait, > > > TASK_UNINTERRUPTIBLE); > > > if (dev->power.runtime_status != RPM_RESUMING && > > > - dev->power.runtime_status != RPM_SUSPENDING) > > > + dev->power.runtime_status != RPM_SUSPENDING && > > > + !dev->power.idle_notification) > > > break; > > > > > > spin_unlock_irq(&dev->power.lock); > > > > Well, not really. > > > > The problem is that rtsx_pci_runtime_idle() is not expected to be > > running after pm_runtime_get_sync(), but the latter doesn't really > > guarantee that. It only guarantees that the suspend/resume callbacks > > will not be running after it returns. > > > > As I said above, if the ->runtime_idle() callback is already running > > when pm_runtime_get_sync() runs, the latter will notice that the > > status is RPM_ACTIVE and will return right away without waiting for > > the former to complete. In fact, it cannot wait for it to complete, > > because it may be called from a ->runtime_idle() callback itself (it > > arguably does not make much sense to do that, but it is not strictly > > forbidden). > > > > So whoever is providing a ->runtime_idle() callback, they need to > > protect it from running in parallel with whatever code runs after > > pm_runtime_get_sync(). Note that ->runtime_idle() will not start > > after pm_runtime_get_sync(), but it may continue running then if it > > has started earlier already. > > > > Calling pm_runtime_barrier() after pm_runtime_get_sync() (not before > > it) should suffice, but once the runtime PM usage counter is dropped, > > rpm_idle() may run again, so this is only effective until the usage > > counter is greater than 1. This means that > > __device_release_driver(() is not the right place to call it, because > > the usage counter is dropped before calling device_remove() in that > > case. > > > > The PCI bus type can prevent the race between driver-provided > > ->runtime_idle() and ->remove() from occurring by adding a > > pm_runtime_probe() call in the following way: > > Thanks for the detailed explanation. Does this mean only PCI bus needs > this fix because other subsystems are not affected, or this needs to > be implemented for each different bus types? I don't think that it needs to be implemented in this form for any other bus types. There are not many cases in which non-trivial .runtime_idle() callbacks are used and the majority of them are PCI drivers anyway AFAICS. Also the drivers using .runtime_idle() can prevent it from racing with other code paths by themselves, this way or another (for example, with the help of locking), and it is generally better to address the races in question on a per-driver basis IMV. The PCI bus type is kind of exceptional, because it forces devices to resume and prevents them from suspending before removing drivers, which is necessary for the fix to be really effective.