Hi, Bibo, On Mon, Jan 22, 2024 at 6:03 PM Bibo Mao <maobibo@xxxxxxxxxxx> wrote: > > This patch refines ipi handling on LoongArch platform, there are > three changes with this patch. > 1. Add generic get_percpu_irq api, replace some percpu irq function > such as get_ipi_irq/get_pmc_irq/get_timer_irq with get_percpu_irq. > > 2. Change parameter action definition with function > loongson_send_ipi_single and loongson_send_ipi_mask. Code encoding is used > here rather than bitmap encoding for ipi action, ipi hw sender uses action > code, and ipi receiver will get action bitmap encoding, the ipi hw will > convert it into bitmap in ipi message buffer. > > 3. Add smp_ops on LoongArch platform so that pv ipi can be used later. > > Signed-off-by: Bibo Mao <maobibo@xxxxxxxxxxx> > --- > arch/loongarch/include/asm/hardirq.h | 4 ++ > arch/loongarch/include/asm/irq.h | 10 ++++- > arch/loongarch/include/asm/smp.h | 31 +++++++-------- > arch/loongarch/kernel/irq.c | 22 +---------- > arch/loongarch/kernel/perf_event.c | 14 +------ > arch/loongarch/kernel/smp.c | 58 +++++++++++++++++++--------- > arch/loongarch/kernel/time.c | 12 +----- > 7 files changed, 71 insertions(+), 80 deletions(-) > > diff --git a/arch/loongarch/include/asm/hardirq.h b/arch/loongarch/include/asm/hardirq.h > index 0ef3b18f8980..9f0038e19c7f 100644 > --- a/arch/loongarch/include/asm/hardirq.h > +++ b/arch/loongarch/include/asm/hardirq.h > @@ -12,6 +12,10 @@ > extern void ack_bad_irq(unsigned int irq); > #define ack_bad_irq ack_bad_irq > > +enum ipi_msg_type { > + IPI_RESCHEDULE, > + IPI_CALL_FUNCTION, > +}; > #define NR_IPI 2 > > typedef struct { > diff --git a/arch/loongarch/include/asm/irq.h b/arch/loongarch/include/asm/irq.h > index 218b4da0ea90..00101b6d601e 100644 > --- a/arch/loongarch/include/asm/irq.h > +++ b/arch/loongarch/include/asm/irq.h > @@ -117,8 +117,16 @@ extern struct fwnode_handle *liointc_handle; > extern struct fwnode_handle *pch_lpc_handle; > extern struct fwnode_handle *pch_pic_handle[MAX_IO_PICS]; > > -extern irqreturn_t loongson_ipi_interrupt(int irq, void *dev); > +static inline int get_percpu_irq(int vector) > +{ > + struct irq_domain *d; > + > + d = irq_find_matching_fwnode(cpuintc_handle, DOMAIN_BUS_ANY); > + if (d) > + return irq_create_mapping(d, vector); > > + return -EINVAL; > +} > #include <asm-generic/irq.h> > > #endif /* _ASM_IRQ_H */ > diff --git a/arch/loongarch/include/asm/smp.h b/arch/loongarch/include/asm/smp.h > index f81e5f01d619..330f1cb3741c 100644 > --- a/arch/loongarch/include/asm/smp.h > +++ b/arch/loongarch/include/asm/smp.h > @@ -12,6 +12,13 @@ > #include <linux/threads.h> > #include <linux/cpumask.h> > > +struct smp_ops { > + void (*call_func_ipi)(const struct cpumask *mask, unsigned int action); > + void (*call_func_single_ipi)(int cpu, unsigned int action); To keep consistency, it is better to use call_func_ipi_single and call_func_ipi_mask. > + void (*ipi_init)(void); > +}; > + > +extern struct smp_ops smp_ops; > extern int smp_num_siblings; > extern int num_processors; > extern int disabled_cpus; > @@ -24,8 +31,6 @@ void loongson_prepare_cpus(unsigned int max_cpus); > void loongson_boot_secondary(int cpu, struct task_struct *idle); > void loongson_init_secondary(void); > void loongson_smp_finish(void); > -void loongson_send_ipi_single(int cpu, unsigned int action); > -void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action); > #ifdef CONFIG_HOTPLUG_CPU > int loongson_cpu_disable(void); > void loongson_cpu_die(unsigned int cpu); > @@ -59,9 +64,12 @@ extern int __cpu_logical_map[NR_CPUS]; > > #define cpu_physical_id(cpu) cpu_logical_map(cpu) > > -#define SMP_BOOT_CPU 0x1 > -#define SMP_RESCHEDULE 0x2 > -#define SMP_CALL_FUNCTION 0x4 > +#define ACTTION_BOOT_CPU 0 > +#define ACTTION_RESCHEDULE 1 > +#define ACTTION_CALL_FUNCTION 2 > +#define SMP_BOOT_CPU BIT(ACTTION_BOOT_CPU) > +#define SMP_RESCHEDULE BIT(ACTTION_RESCHEDULE) > +#define SMP_CALL_FUNCTION BIT(ACTTION_CALL_FUNCTION) > > struct secondary_data { > unsigned long stack; > @@ -71,7 +79,8 @@ extern struct secondary_data cpuboot_data; > > extern asmlinkage void smpboot_entry(void); > extern asmlinkage void start_secondary(void); > - > +extern void arch_send_call_function_single_ipi(int cpu); > +extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); Similarly, to keep consistency, it is better to use arch_send_function_ipi_single and arch_send_function_ipi_mask. Huacai > extern void calculate_cpu_foreign_map(void); > > /* > @@ -79,16 +88,6 @@ extern void calculate_cpu_foreign_map(void); > */ > extern void show_ipi_list(struct seq_file *p, int prec); > > -static inline void arch_send_call_function_single_ipi(int cpu) > -{ > - loongson_send_ipi_single(cpu, SMP_CALL_FUNCTION); > -} > - > -static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) > -{ > - loongson_send_ipi_mask(mask, SMP_CALL_FUNCTION); > -} > - > #ifdef CONFIG_HOTPLUG_CPU > static inline int __cpu_disable(void) > { > diff --git a/arch/loongarch/kernel/irq.c b/arch/loongarch/kernel/irq.c > index 883e5066ae44..1b58f7c3eed9 100644 > --- a/arch/loongarch/kernel/irq.c > +++ b/arch/loongarch/kernel/irq.c > @@ -87,23 +87,9 @@ static void __init init_vec_parent_group(void) > acpi_table_parse(ACPI_SIG_MCFG, early_pci_mcfg_parse); > } > > -static int __init get_ipi_irq(void) > -{ > - struct irq_domain *d = irq_find_matching_fwnode(cpuintc_handle, DOMAIN_BUS_ANY); > - > - if (d) > - return irq_create_mapping(d, INT_IPI); > - > - return -EINVAL; > -} > - > void __init init_IRQ(void) > { > int i; > -#ifdef CONFIG_SMP > - int r, ipi_irq; > - static int ipi_dummy_dev; > -#endif > unsigned int order = get_order(IRQ_STACK_SIZE); > struct page *page; > > @@ -113,13 +99,7 @@ void __init init_IRQ(void) > init_vec_parent_group(); > irqchip_init(); > #ifdef CONFIG_SMP > - ipi_irq = get_ipi_irq(); > - if (ipi_irq < 0) > - panic("IPI IRQ mapping failed\n"); > - irq_set_percpu_devid(ipi_irq); > - r = request_percpu_irq(ipi_irq, loongson_ipi_interrupt, "IPI", &ipi_dummy_dev); > - if (r < 0) > - panic("IPI IRQ request failed\n"); > + smp_ops.ipi_init(); > #endif > > for (i = 0; i < NR_IRQS; i++) > diff --git a/arch/loongarch/kernel/perf_event.c b/arch/loongarch/kernel/perf_event.c > index 0491bf453cd4..3265c8f33223 100644 > --- a/arch/loongarch/kernel/perf_event.c > +++ b/arch/loongarch/kernel/perf_event.c > @@ -456,16 +456,6 @@ static void loongarch_pmu_disable(struct pmu *pmu) > static DEFINE_MUTEX(pmu_reserve_mutex); > static atomic_t active_events = ATOMIC_INIT(0); > > -static int get_pmc_irq(void) > -{ > - struct irq_domain *d = irq_find_matching_fwnode(cpuintc_handle, DOMAIN_BUS_ANY); > - > - if (d) > - return irq_create_mapping(d, INT_PCOV); > - > - return -EINVAL; > -} > - > static void reset_counters(void *arg); > static int __hw_perf_event_init(struct perf_event *event); > > @@ -473,7 +463,7 @@ static void hw_perf_event_destroy(struct perf_event *event) > { > if (atomic_dec_and_mutex_lock(&active_events, &pmu_reserve_mutex)) { > on_each_cpu(reset_counters, NULL, 1); > - free_irq(get_pmc_irq(), &loongarch_pmu); > + free_irq(get_percpu_irq(INT_PCOV), &loongarch_pmu); > mutex_unlock(&pmu_reserve_mutex); > } > } > @@ -562,7 +552,7 @@ static int loongarch_pmu_event_init(struct perf_event *event) > if (event->cpu >= 0 && !cpu_online(event->cpu)) > return -ENODEV; > > - irq = get_pmc_irq(); > + irq = get_percpu_irq(INT_PCOV); > flags = IRQF_PERCPU | IRQF_NOBALANCING | IRQF_NO_THREAD | IRQF_NO_SUSPEND | IRQF_SHARED; > if (!atomic_inc_not_zero(&active_events)) { > mutex_lock(&pmu_reserve_mutex); > diff --git a/arch/loongarch/kernel/smp.c b/arch/loongarch/kernel/smp.c > index a16e3dbe9f09..46735ba49815 100644 > --- a/arch/loongarch/kernel/smp.c > +++ b/arch/loongarch/kernel/smp.c > @@ -66,11 +66,6 @@ static cpumask_t cpu_core_setup_map; > struct secondary_data cpuboot_data; > static DEFINE_PER_CPU(int, cpu_state); > > -enum ipi_msg_type { > - IPI_RESCHEDULE, > - IPI_CALL_FUNCTION, > -}; > - > static const char *ipi_types[NR_IPI] __tracepoint_string = { > [IPI_RESCHEDULE] = "Rescheduling interrupts", > [IPI_CALL_FUNCTION] = "Function call interrupts", > @@ -123,24 +118,19 @@ static u32 ipi_read_clear(int cpu) > > static void ipi_write_action(int cpu, u32 action) > { > - unsigned int irq = 0; > - > - while ((irq = ffs(action))) { > - uint32_t val = IOCSR_IPI_SEND_BLOCKING; > + uint32_t val; > > - val |= (irq - 1); > - val |= (cpu << IOCSR_IPI_SEND_CPU_SHIFT); > - iocsr_write32(val, LOONGARCH_IOCSR_IPI_SEND); > - action &= ~BIT(irq - 1); > - } > + val = IOCSR_IPI_SEND_BLOCKING | action; > + val |= (cpu << IOCSR_IPI_SEND_CPU_SHIFT); > + iocsr_write32(val, LOONGARCH_IOCSR_IPI_SEND); > } > > -void loongson_send_ipi_single(int cpu, unsigned int action) > +static void loongson_send_ipi_single(int cpu, unsigned int action) > { > ipi_write_action(cpu_logical_map(cpu), (u32)action); > } > > -void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action) > +static void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action) > { > unsigned int i; > > @@ -148,6 +138,16 @@ void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action) > ipi_write_action(cpu_logical_map(i), (u32)action); > } > > +void arch_send_call_function_single_ipi(int cpu) > +{ > + smp_ops.call_func_single_ipi(cpu, ACTTION_CALL_FUNCTION); > +} > + > +void arch_send_call_function_ipi_mask(const struct cpumask *mask) > +{ > + smp_ops.call_func_ipi(mask, ACTTION_CALL_FUNCTION); > +} > + > /* > * This function sends a 'reschedule' IPI to another CPU. > * it goes straight through and wastes no time serializing > @@ -155,11 +155,11 @@ void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action) > */ > void arch_smp_send_reschedule(int cpu) > { > - loongson_send_ipi_single(cpu, SMP_RESCHEDULE); > + smp_ops.call_func_single_ipi(cpu, ACTTION_RESCHEDULE); > } > EXPORT_SYMBOL_GPL(arch_smp_send_reschedule); > > -irqreturn_t loongson_ipi_interrupt(int irq, void *dev) > +static irqreturn_t loongson_ipi_interrupt(int irq, void *dev) > { > unsigned int action; > unsigned int cpu = smp_processor_id(); > @@ -179,6 +179,26 @@ irqreturn_t loongson_ipi_interrupt(int irq, void *dev) > return IRQ_HANDLED; > } > > +static void loongson_ipi_init(void) > +{ > + int r, ipi_irq; > + > + ipi_irq = get_percpu_irq(INT_IPI); > + if (ipi_irq < 0) > + panic("IPI IRQ mapping failed\n"); > + > + irq_set_percpu_devid(ipi_irq); > + r = request_percpu_irq(ipi_irq, loongson_ipi_interrupt, "IPI", &irq_stat); > + if (r < 0) > + panic("IPI IRQ request failed\n"); > +} > + > +struct smp_ops smp_ops = { > + .call_func_single_ipi = loongson_send_ipi_single, > + .call_func_ipi = loongson_send_ipi_mask, > + .ipi_init = loongson_ipi_init, > +}; > + > static void __init fdt_smp_setup(void) > { > #ifdef CONFIG_OF > @@ -256,7 +276,7 @@ void loongson_boot_secondary(int cpu, struct task_struct *idle) > > csr_mail_send(entry, cpu_logical_map(cpu), 0); > > - loongson_send_ipi_single(cpu, SMP_BOOT_CPU); > + loongson_send_ipi_single(cpu, ACTTION_BOOT_CPU); > } > > /* > diff --git a/arch/loongarch/kernel/time.c b/arch/loongarch/kernel/time.c > index e7015f7b70e3..fd5354f9be7c 100644 > --- a/arch/loongarch/kernel/time.c > +++ b/arch/loongarch/kernel/time.c > @@ -123,16 +123,6 @@ void sync_counter(void) > csr_write64(init_offset, LOONGARCH_CSR_CNTC); > } > > -static int get_timer_irq(void) > -{ > - struct irq_domain *d = irq_find_matching_fwnode(cpuintc_handle, DOMAIN_BUS_ANY); > - > - if (d) > - return irq_create_mapping(d, INT_TI); > - > - return -EINVAL; > -} > - > int constant_clockevent_init(void) > { > unsigned int cpu = smp_processor_id(); > @@ -142,7 +132,7 @@ int constant_clockevent_init(void) > static int irq = 0, timer_irq_installed = 0; > > if (!timer_irq_installed) { > - irq = get_timer_irq(); > + irq = get_percpu_irq(INT_TI); > if (irq < 0) > pr_err("Failed to map irq %d (timer)\n", irq); > } > -- > 2.39.3 >