Re: [PATCH v9 3/4] drivers/perf: add DesignWare PCIe PMU driver

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]


+ Will, Jonathan, Bjorn and Yicong for probe and hotplug handing.

On 2023/10/24 02:51, Robin Murphy wrote:
> On 2023-10-23 13:32, Will Deacon wrote:
> [...]
>>> +
>>> +    /*
>>> +     * The Group#1 event measures the amount of data processed in 16-byte
>>> +     * units. Simplify the end-user interface by multiplying the counter
>>> +     * at the point of read.
>>> +     */
>>> +    if (event_id >= 0x20 && event_id <= 0x23)
>>> +        return (((u64)hi << 32) | lo) << 4;
>>> +    else
>>> +        return (((u64)hi << 32) | lo);
>> nit, but I think it would be clearer to do:
>>     ret = ((u64)hi << 32) | lo;
>>     /* ... */
>>     if (event_id >= 0x20 && event_id <= 0x23)
>>         ret <<= 4;
> Nit: "ret *= 16;" since the comment says it's multiplying a value, not moving a bitfield. The compiler already knows the most efficient way to implement constant multiplication.

Cool, will use multiplication directly.

>>     return ret;
> [...]
>>> +static int __init dwc_pcie_pmu_init(void)
>>> +{
>>> +    int ret;
>>> +
>>> +    ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN,
>>> +                      "perf/dwc_pcie_pmu:online",
>>> +                      dwc_pcie_pmu_online_cpu,
>>> +                      dwc_pcie_pmu_offline_cpu);
>>> +    if (ret < 0)
>>> +        return ret;
>>> +
>>> +    dwc_pcie_pmu_hp_state = ret;
>>> +
>>> +    ret = platform_driver_register(&dwc_pcie_pmu_driver);
>>> +    if (ret)
>>> +        goto platform_driver_register_err;
>>> +
>>> +    dwc_pcie_pmu_dev = platform_device_register_simple(
>>> +                "dwc_pcie_pmu", PLATFORM_DEVID_NONE, NULL, 0);
>>> +    if (IS_ERR(dwc_pcie_pmu_dev)) {
>>> +        ret = PTR_ERR(dwc_pcie_pmu_dev);
>>> +        goto platform_device_register_error;
>>> +    }
>> I'm a bit confused as to why you're having to create a platform device
>> for a PCI device -- is this because the main designware driver has already
>> bound to it? A comment here explaining why you need to do this would be
>> very helpful. In particular, is there any dependency on another driver
>> to make sure that e.g. config space accesses work properly? If so, we
>> probably need to enforce module load ordering or something like that.
> AFAICS the platform device/driver serve no purpose other than being a hilariously roundabout way to run the for_each_pci_dev() loop in dwc_pcie_pmu_probe() upon module init, and to save explicitly freeing the PMU name/data. Furthermore the devres action for dwc_pcie_pmu_remove_cpuhp_instance() is apparently going for even more style points at module exit by not even relying on the corresponding .remove callback of the tenuous platform driver to undo what its .probe did, but (ab)using the device's devres list to avoid having to keep track of an explicit list of PMU instances at all.

You are right.

> Frankly I think it would be a lot more straightforward to just maintain that explicit list of PMU instances, do the PMU creation directly in dwc_pcie_pmu_init(), then unregister and free them in dwc_pcie_pmu_exit(). Not every driver has to contain a literal struct device_driver.

Agreed, it might be more straightforward. But personally speaking, I prefer
current implementation.

    - standard driver creation / probe flow is more normal
    - it avoid maintaining list of PMU instances
    - IMHO, both of them are temporary solution, if PCI core addes a
      standard mechanism to discover and enbale PCIe VSEC/DVSEC capability,
      the driver will use the standard way.

> It also smells a bit odd that it handles PCI hot-remove but not hot-add - if the underlying device really is hotpluggable, wouldn't we also want to handle new ones turning up after module load? Conversely if it isn't, why pretend to handle it being removed? Even if it's not to do with physical hotplug of the PMU but with the user unloading the PCI controller driver itself (since there's no module/driver-level dependency enforced) and thus tearing down the whole PCI bus, then the same point still applies - if that *can* happen, then what if the user then re-loads it again, or indeed if this module loads first to begin with; wouldn't we want to be able to (re-)discover the PMUs rather than leave the whole PMU driver degraded to a useless state?

I see you point, there are three casees:
1. hot-remove PCI root port firstly and then load the PMU module, the PMU
of the removed PCI device will not be registered.

I think it is the expected behavior.

2. load the PMU module firstly and then hot-remove PCI root port, the PMU
of the removed PCI device will be unregistered.

it is what the dwc_pcie_pmu_unregister_nb() does upon BUS_NOTIFY_DEL_DEVICE

3. load the PMU module firstly, hot-remove PCI root port, and then hot-plug
the PCI root port the PMU of the hot-pluged device will not load again by
current design upon BUS_NOTIFY_DEL_DEVICE.

I guess it is the really problem. It can be workaround be reload the PMU
module. It has been a bit complex around how the following interact:

    - Driver loading/unloading
    - CPU hotplug events
    - PCI device add/del events

We can also add action for BUS_NOTIFY_ADD_DEVICE to address the problem,

    - scan all PCI device
     - check RAS_DES cap
     - check cached PMU node
     - registers its PMU

But I prefer leave as it is, just as x86 does in uncore_bus_notify().
Certainly, if the community deems it necessary to implement
BUS_NOTIFY_ADD_DEVICE action, I also would like to extend it.

[Index of Archives]     [DMA Engine]     [Linux Coverity]     [Linux USB]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [Greybus]

  Powered by Linux