On Tue, Feb 15, 2022 at 06:09:15PM +0100, Rafael J. Wysocki wrote: > On Tue, Feb 15, 2022 at 4:09 PM Bjorn Helgaas <helgaas@xxxxxxxxxx> wrote: > > On Mon, Feb 14, 2022 at 08:23:05AM +0800, Kai-Heng Feng wrote: > > > On Thu, Feb 10, 2022 at 5:36 AM Bjorn Helgaas <helgaas@xxxxxxxxxx> wrote: > > > > On Tue, Dec 07, 2021 at 02:15:04PM +0100, Rafael J. Wysocki wrote: > > > > > On Tue, Dec 7, 2021 at 12:12 AM Keith Busch <kbusch@xxxxxxxxxx> wrote: > > > > > > On Fri, Dec 03, 2021 at 11:15:41AM +0800, Kai-Heng Feng wrote: > > > > > > > When Samsung PCIe Gen4 NVMe is connected to Intel ADL VMD, the > > > > > > > combination causes AER message flood and drags the system performance > > > > > > > down. > > > > > > > > > > > > > > The issue doesn't happen when VMD mode is disabled in BIOS, since AER > > > > > > > isn't enabled by acpi_pci_root_create() . When VMD mode is enabled, AER > > > > > > > is enabled regardless of _OSC: > > > > > > > [ 0.410076] acpi PNP0A08:00: _OSC: platform does not support [AER] > > > > > > > ... > > > > > > > [ 1.486704] pcieport 10000:e0:06.0: AER: enabled with IRQ 146 > > > > > > > > > > > > > > Since VMD is an aperture to regular PCIe root ports, honor ACPI _OSC to > > > > > > > disable PCIe features accordingly to resolve the issue. > > > > > > > > > > > > At least for some versions of this hardare, I recall ACPI is unaware of > > > > > > any devices in the VMD domain; the platform can not see past the VMD > > > > > > endpoint, so I throught the driver was supposed to always let the VMD > > > > > > domain use OS native support regardless of the parent's ACPI _OSC. > > > > > > > > > > This is orthogonal to whether or not ACPI is aware of the VMD domain > > > > > or the devices in it. > > > > > > > > > > If the platform firmware does not allow the OS to control specific > > > > > PCIe features at the physical host bridge level, that extends to the > > > > > VMD "bus", because it is just a way to expose a hidden part of the > > > > > PCIe hierarchy. > > > > > > > > I don't understand what's going on here. Do we understand the AER > > > > message flood? Are we just papering over it by disabling AER? > > > > > > To be more precise, AER is disabled by the platform vendor in BIOS to > > > paper over the issue. > > > The only viable solution for us is to follow their settings. We may > > > never know what really happens underneath. > > > > > > Disabling ASPM/AER/PME etc is a normal practice for ODMs unfortunately. > > > > OK. So this patch actually has nothing in particular to do with AER. > > It's about making _OSC apply to *all* devices below a host bridge, > > even those below a VMD. > > Right. > > > This is slightly ambiguous because while "_OSC applies to the entire > > hierarchy originated by a PCI Host Bridge" (PCI Firmware spec r3.3, > > sec 4.5.1), vmd.c creates a logical view where devices below the VMD > > are in a separate hierarchy with a separate domain. > > But from the HW perspective they still are in the same hierarchy below > the original host bridge. I suppose in some sense it's the same hierarchy because the electrical connection all goes through a root port in the original host bridge, but it's a little muddy because according to [1], a VMD spawns a new hierarchy that can use an entirely new [bus 00-ff] space, and the hierarchy below VMD uses a new config access mechanism independent of ECAM or whatever the original host bridge uses. > > The interpretation that _OSC applies to devices below VMD should work, > > as long as it is possible for platform firmware to manage services > > (AER, pciehp, etc) for things below VMD without getting in the way of > > vmd.c. > > vmd.c actually exposes things hidden by the firmware and the point of > the patch is to still let the firmware control them if it wants/needs > to IIUC. My mental picture is that without vmd.c, Linux would enumerate the VMD RCiEP itself, but none of the devices below the VMD would be visible. With vmd.c, devices below the VMD RCiEP are visible. Maybe this picture is incorrect or too simple? Apparently there's a firmware toggle, but I don't know exactly what it does. Maybe if the toggle is set to disable VMD, the VMD device looks like a regular Root Port and the devices below are enumerated normally even without any vmd.c? > > But I think one implication of this is that we cannot support > > hot-added VMDs. For example, firmware that wants to manage AER will > > use _OSC to retain AER control. But if the firmware doesn't know how > > VMDs work, it will not be able to handle AER for devices below the > > VMD. > > Well, the firmware needs to know how stuff works to hide it in the > first place ... [1] does also say that VMD is a Root Complex *Integrated* Endpoint, which could not be hotplugged. But I don't see anything in the code that actually enforces or requires that, so I don't know what to make of it. If it's possible to hot-add a VMD device, firmware wouldn't be involved in configuring it (assuming pciehp hotplug). I assume the new VMD would look like an Endpoint, and if vmd.c is present, maybe it could construct a new hierarchy below that Endpoint? In that case, we have to assume firmware doesn't know how to operate VMD, so even if firmware manages AER in general, it wouldn't be able to do it for things below the VMD. > > > > If an error occurs below a VMD, who notices and reports it? If we > > > > disable native AER below VMD because of _OSC, as this patch does, I > > > > guess we're assuming the platform will handle AER events below VMD. > > > > Is that really true? Does the platform know how to find AER log > > > > registers of devices below VMD? > > > > > > > > > The platform firmware does that through ACPI _OSC under the host > > > > > bridge device (not under the VMD device) which it is very well aware > > > > > of. [1] https://git.kernel.org/linus/185a383ada2e