On Fri, Aug 02, 2019 at 05:36:43PM +0530, Vidya Sagar wrote: > On 7/30/2019 9:19 PM, Lorenzo Pieralisi wrote: > > On Tue, Jul 23, 2019 at 08:14:08PM +0530, Vidya Sagar wrote: > > > On 7/16/2019 4:52 PM, Lorenzo Pieralisi wrote: > > > > On Sat, Jul 13, 2019 at 12:34:34PM +0530, Vidya Sagar wrote: > > > > > > > > [...] > > > > > > > > > > > > > +static int tegra_pcie_bpmp_set_ctrl_state(struct tegra_pcie_dw *pcie, > > > > > > > > > + bool enable) > > > > > > > > > +{ > > > > > > > > > + struct mrq_uphy_response resp; > > > > > > > > > + struct tegra_bpmp_message msg; > > > > > > > > > + struct mrq_uphy_request req; > > > > > > > > > + int err; > > > > > > > > > + > > > > > > > > > + if (pcie->cid == 5) > > > > > > > > > + return 0; > > > > > > > > > > > > > > > > What's wrong with cid == 5 ? Explain please. > > > > > > > Controller with ID=5 doesn't need any programming to enable it which is > > > > > > > done here through calling firmware API. > > > > > > > > > > > > > > > > > > > > > > > > + memset(&req, 0, sizeof(req)); > > > > > > > > > + memset(&resp, 0, sizeof(resp)); > > > > > > > > > + > > > > > > > > > + req.cmd = CMD_UPHY_PCIE_CONTROLLER_STATE; > > > > > > > > > + req.controller_state.pcie_controller = pcie->cid; > > > > > > > > > + req.controller_state.enable = enable; > > > > > > > > > + > > > > > > > > > + memset(&msg, 0, sizeof(msg)); > > > > > > > > > + msg.mrq = MRQ_UPHY; > > > > > > > > > + msg.tx.data = &req; > > > > > > > > > + msg.tx.size = sizeof(req); > > > > > > > > > + msg.rx.data = &resp; > > > > > > > > > + msg.rx.size = sizeof(resp); > > > > > > > > > + > > > > > > > > > + if (irqs_disabled()) > > > > > > > > > > > > > > > > Can you explain to me what this check is meant to achieve please ? > > > > > > > Firmware interface provides different APIs to be called when there are > > > > > > > no interrupts enabled in the system (noirq context) and otherwise > > > > > > > hence checking that situation here and calling appropriate API. > > > > > > > > > > > > That's what I am questioning. Being called from {suspend/resume}_noirq() > > > > > > callbacks (if that's the code path this check caters for) does not mean > > > > > > irqs_disabled() == true. > > > > > Agree. > > > > > Actually, I got a hint of having this check from the following. > > > > > Both tegra_bpmp_transfer_atomic() and tegra_bpmp_transfer() are indirectly > > > > > called by APIs registered with .master_xfer() and .master_xfer_atomic() hooks of > > > > > struct i2c_algorithm and the decision to call which one of these is made using the > > > > > following check in i2c-core.h file. > > > > > static inline bool i2c_in_atomic_xfer_mode(void) > > > > > { > > > > > return system_state > SYSTEM_RUNNING && irqs_disabled(); > > > > > } > > > > > I think I should use this condition as is IIUC. > > > > > Please let me know if there are any concerns with this. > > > > > > > > It is not a concern, it is just that I don't understand how this code > > > > can be called with IRQs disabled, if you can give me an execution path I > > > > am happy to leave the check there. On top of that, when called from > > > > suspend NOIRQ context, it is likely to use the blocking API (because > > > > IRQs aren't disabled at CPU level) behind which there is most certainly > > > > an IRQ required to wake the thread up and if the IRQ in question was > > > > disabled in the suspend NOIRQ phase this code is likely to deadlock. > > > > > > > > I want to make sure we can justify adding this check, I do not > > > > want to add it because we think it can be needed when it may not > > > > be needed at all (and it gets copy and pasted over and over again > > > > in other drivers). > > > I had a discussion internally about this and the prescribed usage of these APIs > > > seem to be that > > > use tegra_bpmp_transfer() in .probe() and other paths where interrupts are > > > enabled as this API needs interrupts to be enabled for its working. > > > Use tegra_bpmp_transfer_atomic() surrounded by local_irq_save()/local_irq_restore() > > > in other paths where interrupt servicing is disabled. > > > > Why tegra_bpmp_transfer_atomic() needs IRQs to be disabled ? And why > > is it needed in this piece of code where IRQs are _never_ disabled > > at CPU level ? > > > > IRQs are enabled when you call a suspend_noirq() callback, so the > > blocking API can be used as long as the IRQ descriptor backing > > the IRQ that will wake-up the blocked call is marked as > > IRQF_NO_SUSPEND. > > > > The problem is not IRQs enabled/disabled at CPU level, the problem is > > the IRQ descriptor of the IRQ required to handle the blocking BPMP call, > > mark it as IRQF_NO_SUSPEND and remove the tegra_bpmp_transfer_atomic() > > call from this code (or please give me a concrete example pinpointing > > why it is needed). > Ideally, using tegra_bpmp_transfer() alone in all paths (.probe() as > well as .resume_noirq()) should have worked as the corresponding IRQ > is already flagged as IRQF_NO_SUSPEND, but, because of the way BPMP-FW > driver in kernel making its interface available through > .resume_early(), tegra_bpmp_transfer() wasn't working as expected and > I pushed a patch (CC'ing you) at > http://patchwork.ozlabs.org/patch/1140973/ to make it .resume_noirq() > from .resume_early(). With that in place, we can just use > tegra_bpmp_trasnfer(). I'll push a new patch with this change once my > BPMP-FW driver patch is approved. Does this leave you with a resume_noirq() callbacks ordering issue to sort out ? a.k.a How will you guarantee that the BPMP will resume before the host bridge ? Thanks, Lorenzo > Thanks, > Vidya Sagar > > > > Thanks, > > Lorenzo > > > > > I'll go ahead and make next patch series with this if this looks fine to you. > > > > > > > > > > > > > Actually, if tegra_bpmp_transfer() requires IRQs to be enabled you may > > > > > > even end up in a situation where that blocking call does not wake up > > > > > > because the IRQ in question was disabled in the NOIRQ suspend/resume > > > > > > phase. > > > > > > > > > > > > [...] > > > > > > > > > > > > > > > +static int tegra_pcie_dw_probe(struct platform_device *pdev) > > > > > > > > > +{ > > > > > > > > > + const struct tegra_pcie_soc *data; > > > > > > > > > + struct device *dev = &pdev->dev; > > > > > > > > > + struct resource *atu_dma_res; > > > > > > > > > + struct tegra_pcie_dw *pcie; > > > > > > > > > + struct resource *dbi_res; > > > > > > > > > + struct pcie_port *pp; > > > > > > > > > + struct dw_pcie *pci; > > > > > > > > > + struct phy **phys; > > > > > > > > > + char *name; > > > > > > > > > + int ret; > > > > > > > > > + u32 i; > > > > > > > > > + > > > > > > > > > + pcie = devm_kzalloc(dev, sizeof(*pcie), GFP_KERNEL); > > > > > > > > > + if (!pcie) > > > > > > > > > + return -ENOMEM; > > > > > > > > > + > > > > > > > > > + pci = &pcie->pci; > > > > > > > > > + pci->dev = &pdev->dev; > > > > > > > > > + pci->ops = &tegra_dw_pcie_ops; > > > > > > > > > + pp = &pci->pp; > > > > > > > > > + pcie->dev = &pdev->dev; > > > > > > > > > + > > > > > > > > > + data = (struct tegra_pcie_soc *)of_device_get_match_data(dev); > > > > > > > > > + if (!data) > > > > > > > > > + return -EINVAL; > > > > > > > > > + pcie->mode = (enum dw_pcie_device_mode)data->mode; > > > > > > > > > + > > > > > > > > > + ret = tegra_pcie_dw_parse_dt(pcie); > > > > > > > > > + if (ret < 0) { > > > > > > > > > + dev_err(dev, "Failed to parse device tree: %d\n", ret); > > > > > > > > > + return ret; > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pcie->pex_ctl_supply = devm_regulator_get(dev, "vddio-pex-ctl"); > > > > > > > > > + if (IS_ERR(pcie->pex_ctl_supply)) { > > > > > > > > > + dev_err(dev, "Failed to get regulator: %ld\n", > > > > > > > > > + PTR_ERR(pcie->pex_ctl_supply)); > > > > > > > > > + return PTR_ERR(pcie->pex_ctl_supply); > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pcie->core_clk = devm_clk_get(dev, "core"); > > > > > > > > > + if (IS_ERR(pcie->core_clk)) { > > > > > > > > > + dev_err(dev, "Failed to get core clock: %ld\n", > > > > > > > > > + PTR_ERR(pcie->core_clk)); > > > > > > > > > + return PTR_ERR(pcie->core_clk); > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pcie->appl_res = platform_get_resource_byname(pdev, IORESOURCE_MEM, > > > > > > > > > + "appl"); > > > > > > > > > + if (!pcie->appl_res) { > > > > > > > > > + dev_err(dev, "Failed to find \"appl\" region\n"); > > > > > > > > > + return PTR_ERR(pcie->appl_res); > > > > > > > > > + } > > > > > > > > > + pcie->appl_base = devm_ioremap_resource(dev, pcie->appl_res); > > > > > > > > > + if (IS_ERR(pcie->appl_base)) > > > > > > > > > + return PTR_ERR(pcie->appl_base); > > > > > > > > > + > > > > > > > > > + pcie->core_apb_rst = devm_reset_control_get(dev, "apb"); > > > > > > > > > + if (IS_ERR(pcie->core_apb_rst)) { > > > > > > > > > + dev_err(dev, "Failed to get APB reset: %ld\n", > > > > > > > > > + PTR_ERR(pcie->core_apb_rst)); > > > > > > > > > + return PTR_ERR(pcie->core_apb_rst); > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + phys = devm_kcalloc(dev, pcie->phy_count, sizeof(*phys), GFP_KERNEL); > > > > > > > > > + if (!phys) > > > > > > > > > + return PTR_ERR(phys); > > > > > > > > > + > > > > > > > > > + for (i = 0; i < pcie->phy_count; i++) { > > > > > > > > > + name = kasprintf(GFP_KERNEL, "p2u-%u", i); > > > > > > > > > + if (!name) { > > > > > > > > > + dev_err(dev, "Failed to create P2U string\n"); > > > > > > > > > + return -ENOMEM; > > > > > > > > > + } > > > > > > > > > + phys[i] = devm_phy_get(dev, name); > > > > > > > > > + kfree(name); > > > > > > > > > + if (IS_ERR(phys[i])) { > > > > > > > > > + ret = PTR_ERR(phys[i]); > > > > > > > > > + dev_err(dev, "Failed to get PHY: %d\n", ret); > > > > > > > > > + return ret; > > > > > > > > > + } > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pcie->phys = phys; > > > > > > > > > + > > > > > > > > > + dbi_res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dbi"); > > > > > > > > > + if (!dbi_res) { > > > > > > > > > + dev_err(dev, "Failed to find \"dbi\" region\n"); > > > > > > > > > + return PTR_ERR(dbi_res); > > > > > > > > > + } > > > > > > > > > + pcie->dbi_res = dbi_res; > > > > > > > > > + > > > > > > > > > + pci->dbi_base = devm_ioremap_resource(dev, dbi_res); > > > > > > > > > + if (IS_ERR(pci->dbi_base)) > > > > > > > > > + return PTR_ERR(pci->dbi_base); > > > > > > > > > + > > > > > > > > > + /* Tegra HW locates DBI2 at a fixed offset from DBI */ > > > > > > > > > + pci->dbi_base2 = pci->dbi_base + 0x1000; > > > > > > > > > + > > > > > > > > > + atu_dma_res = platform_get_resource_byname(pdev, IORESOURCE_MEM, > > > > > > > > > + "atu_dma"); > > > > > > > > > + if (!atu_dma_res) { > > > > > > > > > + dev_err(dev, "Failed to find \"atu_dma\" region\n"); > > > > > > > > > + return PTR_ERR(atu_dma_res); > > > > > > > > > + } > > > > > > > > > + pcie->atu_dma_res = atu_dma_res; > > > > > > > > > + pci->atu_base = devm_ioremap_resource(dev, atu_dma_res); > > > > > > > > > + if (IS_ERR(pci->atu_base)) > > > > > > > > > + return PTR_ERR(pci->atu_base); > > > > > > > > > + > > > > > > > > > + pcie->core_rst = devm_reset_control_get(dev, "core"); > > > > > > > > > + if (IS_ERR(pcie->core_rst)) { > > > > > > > > > + dev_err(dev, "Failed to get core reset: %ld\n", > > > > > > > > > + PTR_ERR(pcie->core_rst)); > > > > > > > > > + return PTR_ERR(pcie->core_rst); > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pp->irq = platform_get_irq_byname(pdev, "intr"); > > > > > > > > > + if (!pp->irq) { > > > > > > > > > + dev_err(dev, "Failed to get \"intr\" interrupt\n"); > > > > > > > > > + return -ENODEV; > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + ret = devm_request_irq(dev, pp->irq, tegra_pcie_irq_handler, > > > > > > > > > + IRQF_SHARED, "tegra-pcie-intr", pcie); > > > > > > > > > + if (ret) { > > > > > > > > > + dev_err(dev, "Failed to request IRQ %d: %d\n", pp->irq, ret); > > > > > > > > > + return ret; > > > > > > > > > + } > > > > > > > > > + > > > > > > > > > + pcie->bpmp = tegra_bpmp_get(dev); > > > > > > > > > + if (IS_ERR(pcie->bpmp)) > > > > > > > > > + return PTR_ERR(pcie->bpmp); > > > > > > > > > + > > > > > > > > > + platform_set_drvdata(pdev, pcie); > > > > > > > > > + > > > > > > > > > + if (pcie->mode == DW_PCIE_RC_TYPE) { > > > > > > > > > + ret = tegra_pcie_config_rp(pcie); > > > > > > > > > + if (ret && ret != -ENOMEDIUM) > > > > > > > > > + goto fail; > > > > > > > > > + else > > > > > > > > > + return 0; > > > > > > > > > > > > > > > > So if the link is not up we still go ahead and make probe > > > > > > > > succeed. What for ? > > > > > > > We may need root port to be available to support hot-plugging of > > > > > > > endpoint devices, so, we don't fail the probe. > > > > > > > > > > > > We need it or we don't. If you do support hotplugging of endpoint > > > > > > devices point me at the code, otherwise link up failure means > > > > > > failure to probe. > > > > > Currently hotplugging of endpoint is not supported, but it is one of > > > > > the use cases that we may add support for in future. > > > > > > > > You should elaborate on this, I do not understand what you mean, > > > > either the root port(s) supports hotplug or it does not. > > > > > > > > > But, why should we fail probe if link up doesn't happen? As such, > > > > > nothing went wrong in terms of root port initialization right? I > > > > > checked other DWC based implementations and following are not failing > > > > > the probe pci-dra7xx.c, pcie-armada8k.c, pcie-artpec6.c, pcie-histb.c, > > > > > pcie-kirin.c, pcie-spear13xx.c, pci-exynos.c, pci-imx6.c, > > > > > pci-keystone.c, pci-layerscape.c > > > > > > > > > > Although following do fail the probe if link is not up. pcie-qcom.c, > > > > > pcie-uniphier.c, pci-meson.c > > > > > > > > > > So, to me, it looks more like a choice we can make whether to fail the > > > > > probe or not and in this case we are choosing not to fail. > > > > > > > > I disagree. I had an offline chat with Bjorn and whether link-up should > > > > fail the probe or not depends on whether the root port(s) is hotplug > > > > capable or not and this in turn relies on the root port "Slot > > > > implemented" bit in the PCI Express capabilities register. > > > > > > > > It is a choice but it should be based on evidence. > > > > > > > > Lorenzo > > > With Bjorn's latest comment on top of this, I think we are good not to fail > > > the probe here. > > > > > > - Vidya Sagar > > > > > > > >