From: Souradeep Chakrabarti <schakrabarti@xxxxxxxxxxxxxxxxxxx> Sent: Tuesday, November 21, 2023 5:55 AM > > Existing MANA design assigns IRQ to every CPUs, including sibling hyper-threads "assigns IRQs to every CPU" > in a core. This causes multiple IRQs to work on same CPU and may reduce the network "This may cause multiple IRQs to be active simultaneously in the same core and may reduce the network" > performance with RSS. > > Improve the performance by adhering the configuration for RSS, which assigns > IRQ on HT cores. This sentence still doesn't make any sense to me. > > Signed-off-by: Souradeep Chakrabarti <schakrabarti@xxxxxxxxxxxxxxxxxxx> > --- > V1 -> V2: > * Simplified the code by removing filter_mask_list and using avail_cpus. > * Addressed infinite loop issue when there are numa nodes with no CPUs. > * Addressed uses of local numa node instead of 0 to start. > * Removed uses of BUG_ON. > * Placed cpus_read_lock in parent function to avoid num_online_cpus > to get changed before function finishes the affinity assignment. > --- > .../net/ethernet/microsoft/mana/gdma_main.c | 134 ++++++++++++++++- > - > 1 file changed, 123 insertions(+), 11 deletions(-) > > diff --git a/drivers/net/ethernet/microsoft/mana/gdma_main.c > b/drivers/net/ethernet/microsoft/mana/gdma_main.c > index 6367de0c2c2e..8177502ffbd9 100644 > --- a/drivers/net/ethernet/microsoft/mana/gdma_main.c > +++ b/drivers/net/ethernet/microsoft/mana/gdma_main.c > @@ -1243,15 +1243,120 @@ void mana_gd_free_res_map(struct > gdma_resource *r) > r->size = 0; > } > > +static int irq_setup(int *irqs, int nvec, int start_numa_node) > +{ > + unsigned int *core_id_list; > + cpumask_var_t filter_mask, avail_cpus; > + int i, core_count = 0, cpu_count = 0, err = 0, node_count = 0; > + unsigned int cpu_first, cpu, irq_start, cores = 0, numa_node = start_numa_node; > + > + if(!alloc_cpumask_var(&filter_mask, GFP_KERNEL) > + || !alloc_cpumask_var(&avail_cpus, GFP_KERNEL)) { I think it's the case that you don't really need both filter_mask and avail_cpus. filter_mask is used to count the number of cores and set up core_id_list. But it isn't used anymore when the code starts working with avail_cpus. So a single allocated cpumask_var_t variable could serve both purposes. > + err = -ENOMEM; > + goto free_irq; This error path will check if core_id_list is NULL to decide if the core_id_list memory needs to be freed. But core_id_list is uninitialized at this point. > + } > + cpumask_copy(filter_mask, cpu_online_mask); > + cpumask_copy(avail_cpus, cpu_online_mask); > + /* count the number of cores > + */ > + for_each_cpu(cpu, filter_mask) { > + cpumask_andnot(filter_mask, filter_mask, topology_sibling_cpumask(cpu)); > + cores++; > + } > + core_id_list = kcalloc(cores, sizeof(unsigned int), GFP_KERNEL); Need to check for memory allocation failure. > + cpumask_copy(filter_mask, cpu_online_mask); > + /* initialize core_id_list array */ > + for_each_cpu(cpu, filter_mask) { > + core_id_list[core_count] = cpu; > + cpumask_andnot(filter_mask, filter_mask, topology_sibling_cpumask(cpu)); > + core_count++; > + } > + > + /* if number of cpus are equal to max_queues per port, then > + * one extra interrupt for the hardware channel communication. > + */ The "then" part of the above comment is missing some wording. I think what you are saying is that in this case, irq[0] is in the IRQ for the hardware communication channel and is treated specially by assigning it to the first online CPU. That IRQ then does not participate in the IRQ assignment algorithm that is implemented by the remaining code in this function. > + if (nvec - 1 == num_online_cpus()) { > + irq_start = 1; > + cpu_first = cpumask_first(cpu_online_mask); > + irq_set_affinity_and_hint(irqs[0], cpumask_of(cpu_first)); > + } else { > + irq_start = 0; > + } > + > + /* reset the core_count and num_node to 0. > + */ This comment seems out-of-date since num_node is gone. > + core_count = 0; > + > + /* for each interrupt find the cpu of a particular > + * sibling set and if it belongs to the specific numa > + * then assign irq to it and clear the cpu bit from > + * the corresponding avail_cpus. > + * Increase the cpu_count for that node. > + * Once all cpus for a numa node is assigned, then > + * move to different numa node and continue the same. > + */ > + for (i = irq_start; i < nvec; ) { > + > + /* check if the numa node has cpu or not > + * to avoid infinite loop. > + */ > + if (cpumask_empty(cpumask_of_node(numa_node))) { > + numa_node++; This doesn't work correctly. Just incrementing numa_node could produce a value that needs to wrap-around to zero or has wrapped back to the initial numa node. Also, the next numa node selected could *also* have zero CPUs and the code below would still get stuck in an infinite loop. This also seems like the wrong place to make this check as this check is executed every time through the loop, including when only moving to the next core. You really want to make this check in two places: 1) the initial NUMA node that is passed in as an argument, and 2) whenever the NUMA node is updated below. A suggestion: create a helper function "get_next_numa_node()". This function would do the following: 1) Wrap-around back to NUMA node 0 if appropriate 2) Then check for having visited all NUMA nodes -- i.e., having wrapped back to the initial NUMA node 3) Check for no CPUs in the selected NUMA node. If that's the case, increment the numa node, then retry starting at Step #1. This helper function would be called before starting the main "for" loop and again when all CPUs in a node are used. I haven't coded the above suggestion, so you'll have to see if it really works out. But I think getting all of the numa node selection code in one place would help make sure it is right. > + if (++node_count == num_online_nodes()) { > + err = -EAGAIN; > + goto free_irq; I don't understand what the above code is doing. What is the situation where you could "run out" of NUMA nodes and need to return an error? There always must be at least one NUMA node with CPUs. > + } > + } > + cpu_first = cpumask_first_and(avail_cpus, > + topology_sibling_cpumask(core_id_list[core_count])); > + if (cpu_first < nr_cpu_ids && cpu_to_node(cpu_first) == numa_node) { > + irq_set_affinity_and_hint(irqs[i], cpumask_of(cpu_first)); > + cpumask_clear_cpu(cpu_first, avail_cpus); This looks good. Getting rid of filter_mask_list worked out well. :-) > + cpu_count = cpu_count + 1; > + i = i + 1; Nit: Stylistically, "C" usually writes the above as just: cpu_count++; i++; > + > + /* checking if all the cpus are used from the > + * particular node. > + */ > + if (cpu_count == nr_cpus_node(numa_node)) { > + numa_node = numa_node + 1; Same here: just numa_node++ > + if (numa_node == num_online_nodes()) > + numa_node = 0; > + > + /* wrap around once numa nodes > + * are traversed. > + */ > + if (numa_node == start_numa_node) { > + node_count = 0; > + cpumask_copy(avail_cpus, cpu_online_mask); > + } > + cpu_count = 0; > + core_count = 0; > + continue; > + } > + } > + if (++core_count == cores) > + core_count = 0; > + } > +free_irq: > + free_cpumask_var(filter_mask); > + free_cpumask_var(avail_cpus); > + if (core_id_list) > + kfree(core_id_list); > + return err; > +} > + > static int mana_gd_setup_irqs(struct pci_dev *pdev) > { > - unsigned int max_queues_per_port = num_online_cpus(); > + unsigned int max_queues_per_port; > struct gdma_context *gc = pci_get_drvdata(pdev); > struct gdma_irq_context *gic; > - unsigned int max_irqs, cpu; > - int nvec, irq; > + unsigned int max_irqs; > + int nvec, *irqs, irq; > int err, i = 0, j; > > + cpus_read_lock(); > + max_queues_per_port = num_online_cpus(); > if (max_queues_per_port > MANA_MAX_NUM_QUEUES) > max_queues_per_port = MANA_MAX_NUM_QUEUES; > > @@ -1261,6 +1366,11 @@ static int mana_gd_setup_irqs(struct pci_dev *pdev) > nvec = pci_alloc_irq_vectors(pdev, 2, max_irqs, PCI_IRQ_MSIX); > if (nvec < 0) > return nvec; > + irqs = kmalloc_array(nvec, sizeof(int), GFP_KERNEL); > + if (!irqs) { > + err = -ENOMEM; > + goto free_irq_vector; > + } > > gc->irq_contexts = kcalloc(nvec, sizeof(struct gdma_irq_context), > GFP_KERNEL); > @@ -1281,27 +1391,27 @@ static int mana_gd_setup_irqs(struct pci_dev *pdev) > snprintf(gic->name, MANA_IRQ_NAME_SZ, "mana_q%d@pci:%s", > i - 1, pci_name(pdev)); > > - irq = pci_irq_vector(pdev, i); > - if (irq < 0) { > - err = irq; > + irqs[i] = pci_irq_vector(pdev, i); > + if (irqs[i] < 0) { > + err = irqs[i]; > goto free_irq; > } > > - err = request_irq(irq, mana_gd_intr, 0, gic->name, gic); > + err = request_irq(irqs[i], mana_gd_intr, 0, gic->name, gic); > if (err) > goto free_irq; > - > - cpu = cpumask_local_spread(i, gc->numa_node); > - irq_set_affinity_and_hint(irq, cpumask_of(cpu)); > } > > + err = irq_setup(irqs, nvec, gc->numa_node); > + if (err) > + goto free_irq; > err = mana_gd_alloc_res_map(nvec, &gc->msix_resource); > if (err) > goto free_irq; > > gc->max_num_msix = nvec; > gc->num_msix_usable = nvec; > - > + cpus_read_unlock(); > return 0; > > free_irq: > @@ -1314,8 +1424,10 @@ static int mana_gd_setup_irqs(struct pci_dev *pdev) > } > > kfree(gc->irq_contexts); > + kfree(irqs); > gc->irq_contexts = NULL; > free_irq_vector: > + cpus_read_unlock(); > pci_free_irq_vectors(pdev); > return err; > } > -- > 2.34.1