Hello & Happy New Year, Bjorn
On 2024/12/19 10:34, Chen Wang wrote:
hello ~
On 2024/12/11 1:33, Bjorn Helgaas wrote:
On Mon, Dec 09, 2024 at 03:19:38PM +0800, Chen Wang wrote:
Add binding for Sophgo SG2042 PCIe host controller.
+ sophgo,pcie-port:
[......]
+ The Cadence IP has two modes of operation, selected by a
strap pin.
+
+ In the single-link mode, the Cadence PCIe core instance
associated
+ with Link0 is connected to all the lanes and the Cadence PCIe
core
+ instance associated with Link1 is inactive.
+
+ In the dual-link mode, the Cadence PCIe core instance associated
+ with Link0 is connected to the lower half of the lanes and the
+ Cadence PCIe core instance associated with Link1 is connected to
+ the upper half of the lanes.
I assume this means there are two separate Root Ports, one for Link0
and a second for Link1?
+ SG2042 contains 2 Cadence IPs and configures the Cores as below:
+
+ +-- Core(Link0) <---> pcie_rc0
+-----------------+
+ | | |
+ Cadence IP 1 --+ |
cdns_pcie0_ctrl |
+ | | |
+ +-- Core(Link1) <---> disabled
+-----------------+
+
+ +-- Core(Link0) <---> pcie_rc1
+-----------------+
+ | | |
+ Cadence IP 2 --+ |
cdns_pcie1_ctrl |
+ | | |
+ +-- Core(Link1) <---> pcie_rc2
+-----------------+
+
+ pcie_rcX is pcie node ("sophgo,sg2042-pcie-host") defined in
DTS.
+ cdns_pcie0_ctrl is syscon node ("sophgo,sg2042-pcie-ctrl")
defined in DTS
+
+ cdns_pcieX_ctrl contains some registers shared by pcie_rcX,
even two
+ RC(Link)s may share different bits of the same register. For
example,
+ cdns_pcie1_ctrl contains registers shared by link0 & link1
for Cadence IP 2.
An RC doesn't have a Link. A Root Port does.
+ "sophgo,pcie-port" is defined to flag which core(link) the rc
maps to, with
+ this we can know what registers(bits) we should use.
+
+ sophgo,syscon-pcie-ctrl:
+ $ref: /schemas/types.yaml#/definitions/phandle
+ description:
+ Phandle to the PCIe System Controller DT node. It's required to
+ access some MSI operation registers shared by PCIe RCs.
I think this probably means "shared by PCIe Root Ports", not RCs.
It's unlikely that this hardware has multiple Root Complexes.
hi, Bjorn,
I just double confirmed with sophgo engineers, they told me that the
actual PCIe design is that there is only one root port under a host
bridge. I am sorry that my original description and diagram may not
make this clear, so please allow me to introduce this historical
background in detail again. Please read it patiently :):
The IP provided by Cadence contains two independent cores (called
"links" according to the terminology of their manual, the first one is
called link0 and the second one is called link1). Each core
corresponds to a host bridge, and each host bridge has only one root
port, and their configuration registers are completely independent.
That is to say,one cadence IP encapsulates two independent host
bridges. SG2042 integrates two Cadence IPs, so there can actually be
up to four host bridges.
Taking a Cadence IP as an example, the two host bridges can be
connected to different lanes through configuration, which has been
described in the original message. At present, the configuration of
SG2042 is to let core0 (link0) in the first ip occupy all lanes in the
ip, and let core0 (link0) and core1 (link1) in the second ip each use
half of the lanes in the ip. So in the end we only use 3 cores, that's
why 3 host bridge nodes are configured in dts.
Because the configurations of these links are independent, the story
ends here, but unfortunately, sophgo engineers defined some new
register files to add support for their msi controller inside pcie.
The problem is they did not separate these register files according to
link0 and link1. These new register files are "cdns_pcie0_ctrl" /
"cdns_pcie1_ctrl" in the original picture and dts, where the register
of "cdns_pcie0_ctrl" is shared by link0 and link1 of the first ip, and
"cdns_pcie1_ctrl" is shared by link0 and link1 of the second ip.
According to my new description, "cdns_pcieX_ctrl" is not shared by
root ports, they are shared by host bridge/rc.
Because the register design of "cdns_pcieX_ctrl" is not strictly
segmented according to link0 and link1, in pcie host bridge driver
coding we must know whether the host bridge corresponds to link0 or
link1 in the ip, so the "sophgo,link-id" attribute is introduced.
Now I think it is not appropriate to change it to "sophgo,pcie-port".
The reason is that as mentioned above, there is only one root port
under each host bridge in the cadence ip. Link0 and link1 are actually
used to distinguish the two host bridges in one ip.
So I suggest to keep the original "sophgo,link-id" and with the prefix
because the introduction of this attribute is indeed caused by the
private design of sophgo.
Any other good idea please feel free let me know.
Thansk,
Chen
+required:
+ - compatible
+ - reg
+ - reg-names
+ - vendor-id
+ - device-id
+ - sophgo,syscon-pcie-ctrl
+ - sophgo,pcie-port
It looks like vendor-id and device-id apply to PCI devices, i.e.,
things that will show up in lspci, I assume Root Ports in this case.
Can we make this explicit in the DT, e.g., something like this?
pcie@62000000 {
compatible = "sophgo,sg2042-pcie-host";
port0: pci@0,0 {
vendor-id = <0x1f1c>;
device-id = <0x2042>;
};
As I mentioned above, there is actually only one root port under a
host bridge, so I think it is unnecessary to introduce the port subnode.
In addition, I found that it is also allowed to directly add the
vendor-id and device-id properties directly under the host bridge, see
https://github.com/devicetree-org/dt-schema/blob/main/dtschema/schemas/pci/pci-host-bridge.yaml
And refer to the dts for those products using cadence ip:
arch/arm64/boot/dts/ti/k3-j721e-main.dtsi
In this way, when executing lspci, the vendor id and device id will
appear in the line corresponding to the pci brdge device.
[......]
I see that you haven't replied since my last reply, so I will continue
to modify the code as I described. If you have any comments, please let
me know, thank you.
Regards,
Chen
_______________________________________________
linux-riscv mailing list
linux-riscv@xxxxxxxxxxxxxxxxxxx
http://lists.infradead.org/mailman/listinfo/linux-riscv