On 12/9/20 11:55 PM, Hinko Kocevar wrote:
On 12/9/20 10:32 PM, Bjorn Helgaas wrote:
On Wed, Dec 09, 2020 at 09:50:17PM +0100, Hinko Kocevar wrote:
On 12/9/20 6:40 PM, Bjorn Helgaas wrote:
Hmm. Yeah, I don't see anything in the path that would restore the
01:00.0 config space, which seems like a problem. Probably its
secondary/subordinate bus registers and window registers get reset to
0 ("lspci -vv" would show that) so config/mem/io space below it is
accessible.
I guess the save/restore of the switch registers would be in the core
PCI(e)
code; and not a switch vendor/part specific.
I still have interest in getting this to work. With my limited
knowledge of
the PCI code base, is there anything I could do to help? Is there a
particular code base area where you expect this should be managed;
pci.c or
pcie/aer.c or maybe hotplug/ source? I can try and tinker with it..
Here is the 'lspci -vv' of the 01:00.0 device at the time after the
reset:
01:00.0 PCI bridge: PLX Technology, Inc. Device 8725 (rev ca)
(prog-if 00
[Normal decode])
Control: I/O- Mem- BusMaster- SpecCycle- MemWINV- VGASnoop-
ParErr-
Stepping- SERR- FastB2B- DisINTx-
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
Interrupt: pin A routed to IRQ 128
Region 0: Memory at df200000 (32-bit, non-prefetchable)
[virtual]
[size=256K]
Bus: primary=00, secondary=00, subordinate=00, sec-latency=0
I/O behind bridge: 0000f000-00000fff [disabled]
Memory behind bridge: fff00000-000fffff [disabled]
Prefetchable memory behind bridge:
00000000fff00000-00000000000fffff
[disabled]
Yep, this explains why nothing below the bridge is accessible.
Secondary/subordinate=0 means no config transactions will be
forwarded, and the windows are all disabled, so no MMIO transactions
will be forwarded either.
Looks like many fields have been zeroed out. Should a bare
restoration of the previous registers values be enough or does some
other magic (i.e. calling PCI functions/performing initialization
tasks/etc) needs to happen to get that port back to sanity?
Good question. Documentation/PCI/pci-error-recovery.rst has some
details about the design, including this:
It is important for the platform to restore the PCI config space to
the "fresh poweron" state, rather than the "last state". After a
slot reset, the device driver will almost always use its standard
device initialization routines, and an unusual config space setup
may result in hung devices, kernel panics, or silent data
corruption.
I'm not sure exactly how to interpret that. It seems like "restoring
to 'fresh poweron' state" is basically just doing a reset, which gives
results like your 01:00.0 device above. IIUC, we call a driver's
.slot_reset() callback after a reset, and it's supposed to be able to
initialize the device from power-on state.
01:00.0 is a Switch Upstream Port, so portdrv should claim it, and it
looks like pcie_portdrv_slot_reset() should be restoring some state.
But maybe dev->state_saved is false so pci_restore_state() bails out?
Maybe add some printks in that path to see where it's broken?
Adding a bunch of printk()'s to portdrv_pci.c led to (partial) success!
So, the pcie_portdrv_error_detected() returns PCI_ERS_RESULT_CAN_RECOVER
and therefore the pcie_portdrv_slot_reset() is not called.
But the pcie_portdrv_err_resume() is called! Adding these two lines to
pcie_portdrv_err_resume(), before the call to device_for_each_child():
pci_restore_state(dev);
pci_save_state(dev);
gets the secondary buses and the endpoint restored to the following state:
00:01.0 PCI bridge: Intel Corporation Xeon E3-1200 v5/E3-1500 v5/6th Gen
Core Processor PCIe Controller (x16) (rev 05) (prog-if 00 [Normal decode])
Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx+
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
Latency: 0, Cache Line Size: 64 bytes
01:00.0 PCI bridge: PLX Technology, Inc. Device 8725 (rev ca) (prog-if
00 [Normal decode])
Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx+
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
Latency: 0, Cache Line Size: 64 bytes
02:01.0 PCI bridge: PLX Technology, Inc. Device 8725 (rev ca) (prog-if
00 [Normal decode])
Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx+
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
Latency: 0, Cache Line Size: 64 bytes
03:00.0 PCI bridge: PLX Technology, Inc. PEX 8748 48-Lane, 12-Port PCI
Express Gen 3 (8 GT/s) Switch, 27 x 27mm FCBGA (rev ca) (prog-if 00
[Normal decode])
Physical Slot: 1
Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx+
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
04:11.0 PCI bridge: PLX Technology, Inc. PEX 8748 48-Lane, 12-Port PCI
Express Gen 3 (8 GT/s) Switch, 27 x 27mm FCBGA (rev ca) (prog-if 00
[Normal decode])
Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx+
Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
<TAbort- <MAbort- >SERR- <PERR- INTx-
Latency: 0, Cache Line Size: 64 bytes
0e:00.0 Signal processing controller: Xilinx Corporation Device 7011
Subsystem: Device 1a3e:132c
Physical Slot: 2
Control: I/O- Mem- BusMaster- SpecCycle- MemWINV- VGASnoop- ParErr-
Stepping- SERR- FastB2B- DisINTx-
Very nice.
I have not verified if the buses are actually functional, yet, as I
still see the endpoint not happy; I think its driver is just not doing
any restoration currently. Knowing what it takes for switch ports to
recover makes me think I can apply similar save/restore logic to my
endpoint driver(s), too.
I've added the same PCI save/restore state calls to my drivers' error
resume callback, as well as save state to the probe. The results from
injecting the error are as expected; the recovery is successful.
Further, I'm running torture tests as we speak, with 5 AMCs as
endpoints, supported by two different kernel drivers.
As of this moment, 7 cases of:
[ 3745.957711] pcieport 0000:00:01.0: AER: Uncorrected (Fatal) error
received: 0000:00:01.0
[ 3746.149747] pcieport 0000:00:01.0: PCIe Bus Error:
severity=Uncorrected (Fatal), type=Transaction Layer, (Requester ID)
[ 3746.149749] pcieport 0000:00:01.0: device [8086:1901] error
status/mask=00004000/00000000
[ 3746.149751] pcieport 0000:00:01.0: [14] CmpltTO (First)
were observed (not provoked by injection) and successfully recovered
from. Before these changes PCI bus would essentially remain in
non-operational state while now the bus recovers and next test iteration
is successful without user intervention.
For the record, no other types of errors are observed on my side so far
during the tests.
I would like to see the log header in the error reports above, as it is
not visible for some reason, and that is what I'll look into next. At a
glance, looks like the 'info->tlp_header_valid' check in the
aer_print_error() is not passing, resulting in no call to
__print_tlp_header().
Thanks!
//hinko
Cheers,
//hinko
Bjorn