On 02/09/2016 08:31 AM, Arnd Bergmann wrote:
On Tuesday 09 February 2016 10:26:28 Bjorn Helgaas wrote:
On Tue, Feb 09, 2016 at 10:25:33AM +0100, Arnd Bergmann wrote:
On Monday 08 February 2016 17:24:30 Bjorn Helgaas wrote:
I assume your system conforms to expectations like these; I'm just
pointing them out because you mentioned buses with multiple devices on
them, which is definitely something one doesn't expect in PCIe.
The topology we have is currently working with the kernel's core PCI
code. I don't really want to get into discussing what the
definition of PCIe is. We have multiple devices (more than 32) on a
single bus, and they have PCI Express and ARI Capabilities. Is that
PCIe? I don't know.
I don't need to know the details of your topology. As long as it
conforms to the PCIe spec, it should be fine. If it *doesn't* conform
to the spec, but things currently seem to work, that's less fine,
because a future Linux change is liable to break something for you.
I was a little concerned about your statement that "there are multiple
devices residing on each bus, so from that point of view it cannot be
PCIe." That made it sound like you're doing something outside the
spec. If you're just using regular multi-function devices or ARI,
then I don't see any issue (or any reason to say it can't be PCIe).
It doesn't conform to the PCIe port spec, because there are no external
ports but just integrated devices in the host bridge.
Is there a spec section you have in mind? Based on sec 1.3.1, I don't
think there's a requirement to have PCI Express Ports (is that what
you mean by "external ports"?)
No, I was just assuming that ports are specified in their own document,
which would not be followed here if there are none. There is nothing in
here that leads me to believe that the hardware is actually noncompliant
with any relevant standard.
Root Complex Integrated Endpoints (sec 1.3.2.3) are clearly supported
and they would not be behind a Root Port. If you're using those, I
hope they're correctly identified via the PCIe capability Device/Port
Type (sec 7.8.2) because we rely on that type to figure out whether
the link-related registers are implemented.
The spec does include rules related to peer-to-peer transactions, MPS,
ASPM, error reporting, etc., and Linux relies on those, so I think it
would be important to get those right.
David can probably explain more if the registers are compliant with
those parts of the spec.
It is somewhat moot, but in the interest of keeping this thread alive:
None of the "on-chip" devices behind these root complexes implements
ASPM, AER, etc. The capability structures for all the features you
mention are not present.
All that is there beyond standard PCI capabilities are:
- PCI Express capability, to indicate presence of PCI Express
Extended Capabilities.
- ARI capability so we can fit more than 16 devices on a bus.
- SRIOV capability on devices that are virtualizable.
- That's it!
The reality is that they are not really PCI/PCIe devices at all. All
the device registers are at fixed addresses and are connected to
proprietary internal buses with various weird properties. Witness the
need for Enhanced Allocation capabilities to describe the fixed addressing.
The PCI config space is a veneer laid on top of it all to aid in device
discovery and interrupt routing.
So is it PCI or PCIe? It is not really important to say. All we want,
is to be able to get the pci-host-generic root complex driver to bind to
our ECAM/ECAM-like configuration space accessors.
David Daney
Arnd
--
To unsubscribe from this list: send the line "unsubscribe linux-pci" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html