On Mon, Feb 20, 2017 at 12:49:28PM +0100, Lukas Wunner wrote: > On Sat, Feb 18, 2017 at 10:27:45PM -0600, Bjorn Helgaas wrote: > > On Sat, Feb 18, 2017 at 10:12:24AM +0100, Lukas Wunner wrote: > > > On Fri, Feb 17, 2017 at 09:29:03AM -0600, Bjorn Helgaas wrote: > > > > On Sun, Feb 12, 2017 at 05:07:45PM +0100, Lukas Wunner wrote: > > > > > Detect on device probe whether a PCI device is part of a Thunderbolt > > > > > daisy chain (i.e. it is either part of a Thunderbolt controller or part > > > > > of the hierarchy below a Thunderbolt controller). > > > > > > > > My problem with this is that "is_thunderbolt" is not well-defined. > > > > The PCI specs lay out the common vocabulary and framework for this > > > > area. To keep the code maintainable, we need to connect things back > > > > to the specs somehow. > > > > > > > > For example, it might make sense to have a bit that means "this device > > > > can generate PME from D3hot", because PME and D3hot are part of that > > > > common understanding of how PCI works, and we can use that information > > > > to design the software. > > > > > > > > An "is_thunderbolt" bit doesn't have any of that context. It doesn't > > > > say anything about how the device behaves, so I can't evaluate the > > > > code that uses it. > > > > > > No, this *does* connect back to the spec, the Thunderbolt spec to be > > > specific, except that spec is not available publicly. (I assume it's > > > available under NDA.) FWIW the PCI SIG doesn't make its specs freely > > > available either, only to members or for $$$. As Andreas Noever has > > > pointed out before, there is plenty of precedent for including > > > (reverse-engineered) code in the kernel for stuff that isn't public. > > > > I'm not objecting to the fact that the Thunderbolt spec is not public. > > What I want to know is specifically how the Thunderbolt bridge behaves > > differently than a plain old PCIe bridge. I don't even want to know > > *all* the differences. You're proposing to make the PCI core work a > > slightly differently based on this bit, and in order to maintain that > > in the future, we need to know the details of *why* we need to do > > things differently. > > Okay, I think I'm starting to understand what you're driving at. > > Thunderbolt tunnels PCIe transparently, so in principle there should be > no difference behaviour-wise. > > As far as the PCI core is concerned, I'm only using is_thunderbolt to > whitelist Thunderbolt ports for runtime PM in patch [2/8]. (This just > concerns like a dozen chips whose behaviour is well understood, hence > enabling runtime PM should be safe.) If there's no known PM-related behavior difference between Thunderbolt and other PCIe ports, I'm hesitant to enable runtime PM only for Thunderbolt, especially since the proposal also enables runtime PM for non-Thunderbolt devices that happen to be downstream from a Thunderbolt port. If we think Thunderbolt ports and non-Thunderbolt ports work the same with respect to PM, we should operate them the same way to simplify the code and improve test coverage. In other words, if there are problems with new functionality, I would prefer not to just exclude the hardware platforms with the problems, especially when it's the huge generic class of "all non-Thunderbolt PCIe ports". I'd rather figure out the problems so we can enable the functionality everywhere (possibly with quirks for known defective platforms). > The other two upcoming use cases merely concern whether a device is on > a Thunderbolt daisy chain (vs. soldered to the mainboard), and to detect > presence of a Thunderbolt controller in the machine (their PCI devices > use PCI_CLASS_SYSTEM_OTHER and PCI_CLASS_BRIDGE_PCI, which is not > sufficiently unique to identify Thunderbolt controllers). If I understand correctly, this is another case where there's no actual functional difference because of Thunderbolt, but apple-gmux would use the fact that a GPU is connected via Thunderbolt to infer something about which GPUs vga_switcheroo can switch between. Since the PCI core doesn't care at all about this, could apple-gmux figure this out on its own by looking for the VSEC capability itself? It seems like doing it closer to where it's used would make things more understandable. > I decided to put the is_thunderbolt flag in struct pci_dev because any > PCI device might end up on a Thunderbolt daisy chain. The purpose of > the bit is merely to cache that status, it does not signify that the > device suffers from some particular PCI quirk. Assuming we need it, having it in struct pci_dev is fine. There's no point in looking up the VSEC capability more than once. > > Maybe D3hot means something different, maybe PME works differently, > > maybe hotplug interrupts are signaled differently, I dunno. If you > > want us to treat these devices differently, we have to know *why* so > > we can tell whether future changes in other areas also need to handle > > them differently. > > This all works fine. Once a Thunderbolt tunnel has been set up, the > hotplug port on the PCIe switch integrated into the Thunderbolt > controller signals "Card present" and "Link up" interrupts. On surprise > removal of an attached device, the Presence Detect and Link State bits > are cleared and an interrupt is signaled for each of them. > > There's a quirk wherein Thunderbolt controllers claim to support Command > Complete interrupts but they never send them, I don't think that's > intentional but rather a hardware bug. The kernel deals gracefully with > that though, no special treatment necessary. Many Intel ports have a similar issue (erratum CF118, see 3461a068661c ("PCI: pciehp: Wait for hotplug command completion lazily")) where they generate Command Completion interrupts for some commands but not all. I think pciehp should work even without those interrupts, though we emit timeout warnings (which maybe could be toned down or omitted). > > > > If runtime code needs to know whether any upstream bridge is > > > > Thunderbolt, it can always search up the hierarchy. I think that > > > > would improve readability because the software model would map more > > > > closely to the hardware situation. > > > > > > That would be more expensive than just checking a bit that is searched > > > only once and then cached. We have is_hotplug_bridge for just eight > > > places where we check presence of hotplug capability, why can't we have > > > is_thunderbolt? Beats me. > > > > Searching up the tree is more expensive, but it looks like we only > > check it in enumeration-type situations, so I doubt this is a > > performance issue. > > In patch [3/8] of v5 of this series, I used the is_thunderbolt bit in > pci_dev_check_d3cold(), which is not only executed during ->probe and > ->remove but also whenver the D3cold status of a device is changed via > sysfs. > > However I dropped that patch and the remaining use cases are indeed > limited to ->probe paths, so I no longer feel strongly about avoiding > walking up the hierarchy. > > The set_pcie_thunderbolt() function in this commit essentially does > two things: (a) Detect presence of the Thunderbolt VSEC on a device > and (b) walking up the hierarchy to detect whether that VSEC is present > on a parent. > > Do you want me to set the is_thunderbolt bit only on devices belonging > to a Thunderbolt controller and use a separate function to walk up the > hierarchy? Let's figure out what information we need and then figure out where to store it. If Thunderbolt devices have a functional difference we need to know about, struct pci_dev seems like a good place to store that. Bjorn