On 2/15/23 18:04, Rick Wertenbroek wrote: > On Wed, Feb 15, 2023 at 12:56 AM Damien Le Moal > <damien.lemoal@xxxxxxxxxxxxxxxxxx> wrote: >> >> I checked the TRM and indeed these registers are listed as unused. >> However, with this patch, nothing work for me using a Pine rockpro64 >> board. Keeping this patch, your series (modulo some other fixes, more >> emails coming) is making things work ! > > Hello, Thank you for testing the driver and commenting, I'll incorporate your > suggestions in the next version of this series. > > This patch alone does not make the driver work. Without the fixes to the > address windows and translation found in [PATCH v2 6/9] ("PCI: rockchip: > Fix window mapping and address translation for endpoint") transfers will not > work. However, as you said, with the patch series, the driver works. > Good to see that you have the driver working on the rockpro64 which is a > very similar but different board than the one I used (FriendlyElec NanoPC-T4). > >> So I think the bug is with the TRM, not the code. THinking logically about >> htis, it makes sense: this is programming the address translation unit to >> translate mmio & dma between host PCI address and local CPU space address. >> If we never set the PU address, how can that unit possibly ever translate >> anything ? > > No, the bug is not in the TRM: > The RK3399 PCIe endpoint core has the physical address space of 64MB > @ 0xF800'0000 to access the PCIe address space (TRM 17.5.4). > This space is split into 33 windows, one of 32MBytes and 32 of 1MByte. > Read-write accesses by the CPU to that region will be translated. Each > window has a mapping that is configured through the ATR Configuration > Register Address Map (TRM 17.6.8) and the registers addr0 and addr1 > will dictate the translation between the window (a physical CPU addr) > into a PCI space address (with this the unit can translate). The other > registers are for the PCIe header descriptor. > The translation process is documented in TRM 17.5.5.1.1 > The core will translate all read-write accesses to the windows that fall > in the 64MB space @ 0xF800'0000 and generate the PCIe addresses > and headers according to the values in the registers in the ATR > Configuration Register Address Map (@ 0xFDC0'0000). > > Translation does indeed take place and works > but requires the changes in [PATCH v2 6/9] ("PCI: rockchip: > Fix window mapping and address translation for endpoint") > because it was broken from the start... > > The two writes that were removed are to unused (read-only) registers. > The writes don't do anything, manually writing and reading back these > addresses will always lead to 0 (they are read-only). So they are removed. OK. Tested again and it is working-ish... ./pcitest.sh ## Loading pci_endpoint_test ## BAR tests BAR0: OKAY BAR1: OKAY BAR2: OKAY BAR3: OKAY BAR4: OKAY BAR5: OKAY ## Legacy interrupt tests SET IRQ TYPE TO LEGACY: OKAY LEGACY IRQ: OKAY ## MSI interrupt tests SET IRQ TYPE TO MSI: OKAY MSI1: OKAY MSI2: OKAY MSI3: OKAY MSI4: OKAY MSI5: OKAY MSI6: OKAY MSI7: OKAY MSI8: OKAY MSI9: OKAY MSI10: OKAY MSI11: OKAY MSI12: OKAY MSI13: OKAY MSI14: OKAY MSI15: OKAY MSI16: OKAY ## Read Tests (DMA) READ ( 1 bytes): OKAY READ ( 1024 bytes): OKAY READ ( 1025 bytes): OKAY READ ( 4096 bytes): OKAY READ ( 131072 bytes): OKAY READ (1024000 bytes): OKAY READ (1024001 bytes): OKAY READ (1048576 bytes): OKAY ## Write Tests (DMA) WRITE ( 1 bytes): OKAY WRITE ( 1024 bytes): OKAY WRITE ( 1025 bytes): OKAY WRITE ( 4096 bytes): OKAY WRITE ( 131072 bytes): OKAY WRITE (1024000 bytes): OKAY Then stops here doing the 1024001 B case. The host waits for a completion that does not come. On the EP side, I see: [ 94.307215] pci_epf_test pci_epf_test.0: READ src addr 0xffd00000, 1024001 B [ 94.307960] pci_epc fd000000.pcie-ep: Map region 1 phys addr 0xfa100000 to pci addr 0xffd00000, 1024001 B [ 94.308924] rockchip-pcie-ep fd000000.pcie-ep: Set atu: region 1, cpu addr 0xfa100000, pci addr 0xffd00000, 1024001 B [ 132.309645] dma-pl330 ff6e0000.dma-controller: Reset Channel-2 CS-20000e FTC-40000 ^^^^^^^^^^^^^^^ The DMA engine does not like something at all. Back to where I was when I tried your series initially, which is why I tried removing patch 1 to see what happens... [ 132.370479] pci_epf_test pci_epf_test.0: READ => Size: 1024001 B, DMA: YES, Time: 38.059623935 s, Rate: 26 KB/s [ 132.372152] pci_epc fd000000.pcie-ep: Unmap region 1 [ 132.372780] pci_epf_test pci_epf_test.0: RAISE MSI IRQ 1 [ 132.373312] rockchip-pcie-ep fd000000.pcie-ep: Send MSI IRQ 1 [ 132.373844] rockchip-pcie-ep fd000000.pcie-ep: MSI disabled [ 132.374388] pci_epf_test pci_epf_test.0: Raise IRQ failed -22 And it looks like the PCI core crashed or something because MSI does not work anymore as well (note that this is wheat I see with my nvme epf driver too, but I do not have that DMA channel reset message...) If I run the tests without DMA (mmio only), everything seems fine: ## Read Tests (No DMA) READ ( 1 bytes): OKAY READ ( 1024 bytes): OKAY READ ( 1025 bytes): OKAY READ (1024000 bytes): OKAY READ (1024001 bytes): OKAY ## Write Tests (No DMA) WRITE ( 1 bytes): OKAY WRITE ( 1024 bytes): OKAY WRITE ( 1025 bytes): OKAY WRITE (1024000 bytes): OKAY WRITE (1024001 bytes): OKAY ## Copy Tests (No DMA) COPY ( 1 bytes): OKAY COPY ( 1024 bytes): OKAY COPY ( 1025 bytes): OKAY COPY (1024000 bytes): OKAY COPY (1024001 bytes): OKAY So it looks like translation is working with your patch, but that the driver is still missing something for DMA to work correctly... Will keep digging. Note that this is all tested with the patch series fixing pci-epf-test and pci_endpoint_test drivers that I posted earlier today. As mentioned, my host is an AMD Ryzen PC. -- Damien Le Moal Western Digital Research