On Mon, Jul 8, 2024 at 10:17 AM Tianyang Zhang <zhangtianyang@xxxxxxxxxxx> wrote: > > > 在 2024/7/6 下午12:50, Huacai Chen 写道: > > On Sat, Jul 6, 2024 at 10:46 AM Tianyang Zhang > > <zhangtianyang@xxxxxxxxxxx> wrote: > >> > >> 在 2024/7/5 下午2:58, Huacai Chen 写道: > >>> Hi, Tianyang, > >>> > >>> You can squash two patches into one, as Song does: > >>> https://lore.kernel.org/lkml/20240624084410.1026648-1-gaosong@xxxxxxxxxxx/ > >> Ok. I have resend the patch with two-in-one mode > > I ask several questions but you only take this one. > > > > Huacai > > Hi ,Huacai > > Sorry for my careless first. > > > > >>> On Fri, Jul 5, 2024 at 9:53 AM Tianyang Zhang <zhangtianyang@xxxxxxxxxxx> wrote: > >>>> Introduce the advanced extended interrupt controllers. This feature will > >>>> allow each core to have 256 independent interrupt vectors and MSI > >>>> interrupts can be independently routed to any vector on any CPU. > >>>> > >>>> Co-developed-by: Jianmin Lv <lvjianmin@xxxxxxxxxxx> > >>>> Signed-off-by: Jianmin Lv <lvjianmin@xxxxxxxxxxx> > >>>> Co-developed-by: Liupu Wang <wangliupu@xxxxxxxxxxx> > >>>> Signed-off-by: Liupu Wang <wangliupu@xxxxxxxxxxx> > >>>> Co-developed-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx> > >>>> Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx> > >>>> Signed-off-by: Tianyang Zhang <zhangtianyang@xxxxxxxxxxx> > >>>> --- > >>>> arch/loongarch/Kconfig | 1 + > >>>> arch/loongarch/include/asm/cpu-features.h | 1 + > >>>> arch/loongarch/include/asm/cpu.h | 2 + > >>>> arch/loongarch/include/asm/hw_irq.h | 10 + > >>>> arch/loongarch/include/asm/irq.h | 12 +- > >>>> arch/loongarch/include/asm/loongarch.h | 20 +- > >>>> arch/loongarch/include/asm/smp.h | 2 + > >>>> arch/loongarch/kernel/cpu-probe.c | 3 +- > >>>> arch/loongarch/kernel/smp.c | 5 + > >>>> drivers/irqchip/Makefile | 2 +- > >>>> drivers/irqchip/irq-loongarch-avec.c | 440 ++++++++++++++++++++++ > >>>> drivers/irqchip/irq-loongarch-cpu.c | 4 +- > >>>> drivers/irqchip/irq-loongson-eiointc.c | 3 + > >>>> drivers/irqchip/irq-loongson-pch-msi.c | 43 ++- > >>>> 14 files changed, 538 insertions(+), 10 deletions(-) > >>>> create mode 100644 drivers/irqchip/irq-loongarch-avec.c > >>>> > >>>> Changes log: > >>>> V0->V1: > >>>> 1.Modified some formats and declarations > >>>> 2.Removed kmalloc/kfree when adding affinity related data to pending_list, > >>>> and used moving tag to replace the original behavior > >>>> 3.Adjusted the process that enables AVEC interrupts, now it is at the end of all processes > >>>> 4.Removed CPUHP related callbacks, now irq_matrix_online/irq_matrix_offline is completed in start_secondary/loongson_cpu_disable > >>>> 5.Adjusted compatibility issues for CONFIG_ACPI > >>>> 6.About question: > >>>> > irr = csr_read64(LOONGARCH_CSR_IRR0 + vector / 64); > >>>> > should be good enough, no? > >>>> csr_read64 was built-in as __csrrd_d, it doesn't seem to support variables as parameters > >>>> >>>> > >>>> drivers/irqchip/irq-loongarch-avec.c: In function ‘complete_irq_moving’: > >>>> ./arch/loongarch/include/asm/loongarch.h:164:25: error: invalid argument to built-in function > >>>> 164 | #define csr_read64(reg) __csrrd_d(reg) > >>>> | ^~~~~~~~~ > >>>> drivers/irqchip/irq-loongarch-avec.c:170:23: note: in expansion of macro ‘csr_read64’ > >>>> 170 | irr = csr_read64(LOONGARCH_CSR_IRR_BASE + vector / VECTORS_PER_REG); > >>>> | ^~~~~~~~~~ > >>>> >>>> > >>>> So we have temporarily retained the previous implementation. > >>>> > >>>> V1->V2: > >>>> Fixed up coding style. Made on/offline functions void > >>>> Added compatibility when CONFIG_SMP is turned off > >>>> > >>>> diff --git a/arch/loongarch/Kconfig b/arch/loongarch/Kconfig > >>>> index e38139c576ee..a66e49b5a68c 100644 > >>>> --- a/arch/loongarch/Kconfig > >>>> +++ b/arch/loongarch/Kconfig > >>>> @@ -83,6 +83,7 @@ config LOONGARCH > >>>> select GENERIC_ENTRY > >>>> select GENERIC_GETTIMEOFDAY > >>>> select GENERIC_IOREMAP if !ARCH_IOREMAP > >>>> + select GENERIC_IRQ_MATRIX_ALLOCATOR > >>>> select GENERIC_IRQ_MULTI_HANDLER > >>>> select GENERIC_IRQ_PROBE > >>>> select GENERIC_IRQ_SHOW > >>>> diff --git a/arch/loongarch/include/asm/cpu-features.h b/arch/loongarch/include/asm/cpu-features.h > >>>> index 2eafe6a6aca8..16a716f88a5c 100644 > >>>> --- a/arch/loongarch/include/asm/cpu-features.h > >>>> +++ b/arch/loongarch/include/asm/cpu-features.h > >>>> @@ -65,5 +65,6 @@ > >>>> #define cpu_has_guestid cpu_opt(LOONGARCH_CPU_GUESTID) > >>>> #define cpu_has_hypervisor cpu_opt(LOONGARCH_CPU_HYPERVISOR) > >>>> #define cpu_has_ptw cpu_opt(LOONGARCH_CPU_PTW) > >>>> +#define cpu_has_avecint cpu_opt(LOONGARCH_CPU_AVECINT) > >>>> > >>>> #endif /* __ASM_CPU_FEATURES_H */ > >>>> diff --git a/arch/loongarch/include/asm/cpu.h b/arch/loongarch/include/asm/cpu.h > >>>> index 48b9f7168bcc..843f9c4ec980 100644 > >>>> --- a/arch/loongarch/include/asm/cpu.h > >>>> +++ b/arch/loongarch/include/asm/cpu.h > >>>> @@ -99,6 +99,7 @@ enum cpu_type_enum { > >>>> #define CPU_FEATURE_GUESTID 24 /* CPU has GuestID feature */ > >>>> #define CPU_FEATURE_HYPERVISOR 25 /* CPU has hypervisor (running in VM) */ > >>>> #define CPU_FEATURE_PTW 26 /* CPU has hardware page table walker */ > >>>> +#define CPU_FEATURE_AVECINT 27 /* CPU has avec interrupt */ > >>>> > >>>> #define LOONGARCH_CPU_CPUCFG BIT_ULL(CPU_FEATURE_CPUCFG) > >>>> #define LOONGARCH_CPU_LAM BIT_ULL(CPU_FEATURE_LAM) > >>>> @@ -127,5 +128,6 @@ enum cpu_type_enum { > >>>> #define LOONGARCH_CPU_GUESTID BIT_ULL(CPU_FEATURE_GUESTID) > >>>> #define LOONGARCH_CPU_HYPERVISOR BIT_ULL(CPU_FEATURE_HYPERVISOR) > >>>> #define LOONGARCH_CPU_PTW BIT_ULL(CPU_FEATURE_PTW) > >>>> +#define LOONGARCH_CPU_AVECINT BIT_ULL(CPU_FEATURE_AVECINT) > >>>> > >>>> #endif /* _ASM_CPU_H */ > >>>> diff --git a/arch/loongarch/include/asm/hw_irq.h b/arch/loongarch/include/asm/hw_irq.h > >>>> index af4f4e8fbd85..772692e765c0 100644 > >>>> --- a/arch/loongarch/include/asm/hw_irq.h > >>>> +++ b/arch/loongarch/include/asm/hw_irq.h > >>>> @@ -9,6 +9,16 @@ > >>>> > >>>> extern atomic_t irq_err_count; > >>>> > >>>> +/* > >>>> + * 256 vectors Map: > >>>> + * > >>>> + * 0 - 15: mapping legacy IPs, e.g. IP0-12. > >>>> + * 16 - 255: mapping a vector for external IRQ. > >>> In handle_cpu_irq() we handle CPUINTC irqs from low to high, which > >>> means when we handle AVECINTC (IP13), IP0-12 have been handled > >>> already, so should we really mapping IP012 to AVECINTC? > In fact, retaining 0-15 vector is for compatibility with future > extensions, and currently it has not been done in practice. > > The AVEC interrupt controller can completely replace the current core-ip > 0-12 interrupt. When we implement the > > function described above, core-ip 0-12 will no longer have its original > function > > Of course, this is a long-term plan > >>> > >>>> + * > >>>> + */ > >>>> +#define NR_VECTORS 256 > >>>> +#define IRQ_MATRIX_BITS NR_VECTORS > >>>> +#define NR_LEGACY_VECTORS 16 > >>> Should we define them in irq.h and update NR_IRQS there?] > Yes, you are right, this may indeed cause potential problems > >>> > >>>> /* > >>>> * interrupt-retrigger: NOP for now. This may not be appropriate for all > >>>> * machines, we'll see ... > >>>> diff --git a/arch/loongarch/include/asm/irq.h b/arch/loongarch/include/asm/irq.h > >>>> index 480418bc5071..cf3b635a9b86 100644 > >>>> --- a/arch/loongarch/include/asm/irq.h > >>>> +++ b/arch/loongarch/include/asm/irq.h > >>>> @@ -65,7 +65,7 @@ extern struct acpi_vector_group msi_group[MAX_IO_PICS]; > >>>> #define LOONGSON_LPC_LAST_IRQ (LOONGSON_LPC_IRQ_BASE + 15) > >>>> > >>>> #define LOONGSON_CPU_IRQ_BASE 16 > >>>> -#define LOONGSON_CPU_LAST_IRQ (LOONGSON_CPU_IRQ_BASE + 14) > >>>> +#define LOONGSON_CPU_LAST_IRQ (LOONGSON_CPU_IRQ_BASE + 15) > >>>> > >>>> #define LOONGSON_PCH_IRQ_BASE 64 > >>>> #define LOONGSON_PCH_ACPI_IRQ (LOONGSON_PCH_IRQ_BASE + 47) > >>>> @@ -101,6 +101,16 @@ int pch_msi_acpi_init(struct irq_domain *parent, > >>>> struct acpi_madt_msi_pic *acpi_pchmsi); > >>>> int pch_pic_acpi_init(struct irq_domain *parent, > >>>> struct acpi_madt_bio_pic *acpi_pchpic); > >>>> + > >>>> +#ifdef CONFIG_ACPI > >>> No #ifdef needed. > ok , I adjust here next patch > >>> > >>>> +int __init pch_msi_acpi_init_v2(struct irq_domain *parent, > >>>> + struct acpi_madt_msi_pic *pch_msi_entry); > >>> No line break needed. > ok > >>> > >>>> +int __init loongarch_avec_acpi_init(struct irq_domain *parent); > >>>> +void complete_irq_moving(void); > >>>> +void loongarch_avec_offline_cpu(unsigned int cpu); > >>>> +void loongarch_avec_online_cpu(unsigned int cpu); > >>>> +#endif > >>>> + > >>>> int find_pch_pic(u32 gsi); > >>>> struct fwnode_handle *get_pch_msi_handle(int pci_segment); > >>>> > >>>> diff --git a/arch/loongarch/include/asm/loongarch.h b/arch/loongarch/include/asm/loongarch.h > >>>> index eb09adda54b7..16a910359977 100644 > >>>> --- a/arch/loongarch/include/asm/loongarch.h > >>>> +++ b/arch/loongarch/include/asm/loongarch.h > >>>> @@ -72,7 +72,6 @@ > >>>> #define CPUCFG1_RPLV BIT(23) > >>>> #define CPUCFG1_HUGEPG BIT(24) > >>>> #define CPUCFG1_CRC32 BIT(25) > >>>> -#define CPUCFG1_MSGINT BIT(26) > >>>> > >>>> #define LOONGARCH_CPUCFG2 0x2 > >>>> #define CPUCFG2_FP BIT(0) > >>>> @@ -252,8 +251,8 @@ > >>>> #define CSR_ESTAT_EXC_WIDTH 6 > >>>> #define CSR_ESTAT_EXC (_ULCAST_(0x3f) << CSR_ESTAT_EXC_SHIFT) > >>>> #define CSR_ESTAT_IS_SHIFT 0 > >>>> -#define CSR_ESTAT_IS_WIDTH 14 > >>>> -#define CSR_ESTAT_IS (_ULCAST_(0x3fff) << CSR_ESTAT_IS_SHIFT) > >>>> +#define CSR_ESTAT_IS_WIDTH 15 > >>>> +#define CSR_ESTAT_IS (_ULCAST_(0x7fff) << CSR_ESTAT_IS_SHIFT) > >>>> > >>>> #define LOONGARCH_CSR_ERA 0x6 /* ERA */ > >>>> > >>>> @@ -999,10 +998,18 @@ > >>>> #define CSR_FWPC_SKIP_SHIFT 16 > >>>> #define CSR_FWPC_SKIP (_ULCAST_(1) << CSR_FWPC_SKIP_SHIFT) > >>>> > >>>> +#define LOONGARCH_CSR_IRR0 0xa0 > >>>> +#define LOONGARCH_CSR_IRR1 0xa1 > >>>> +#define LOONGARCH_CSR_IRR2 0xa2 > >>>> +#define LOONGARCH_CSR_IRR3 0xa3 > >>>> +#define LOONGARCH_CSR_IRR_BASE LOONGARCH_CSR_IRR0 > >>>> + > >>>> +#define LOONGARCH_CSR_ILR 0xa4 > >>> What is ILR stand for? > Interrupt Load Register, ILR According to the user manual, IRR here should be ISR, and ILR here should be IRR. > >>>> + > >>>> /* > >>>> * CSR_ECFG IM > >>>> */ > >>>> -#define ECFG0_IM 0x00001fff > >>>> +#define ECFG0_IM 0x00005fff > >>>> #define ECFGB_SIP0 0 > >>>> #define ECFGF_SIP0 (_ULCAST_(1) << ECFGB_SIP0) > >>>> #define ECFGB_SIP1 1 > >>>> @@ -1045,6 +1052,7 @@ > >>>> #define IOCSRF_EIODECODE BIT_ULL(9) > >>>> #define IOCSRF_FLATMODE BIT_ULL(10) > >>>> #define IOCSRF_VM BIT_ULL(11) > >>>> +#define IOCSRF_AVEC BIT_ULL(15) > >>>> > >>>> #define LOONGARCH_IOCSR_VENDOR 0x10 > >>>> > >>>> @@ -1055,6 +1063,7 @@ > >>>> #define LOONGARCH_IOCSR_MISC_FUNC 0x420 > >>>> #define IOCSR_MISC_FUNC_TIMER_RESET BIT_ULL(21) > >>>> #define IOCSR_MISC_FUNC_EXT_IOI_EN BIT_ULL(48) > >>>> +#define IOCSR_MISC_FUNC_AVEC_EN BIT_ULL(51) > >>>> > >>>> #define LOONGARCH_IOCSR_CPUTEMP 0x428 > >>>> > >>>> @@ -1375,9 +1384,10 @@ __BUILD_CSR_OP(tlbidx) > >>>> #define INT_TI 11 /* Timer */ > >>>> #define INT_IPI 12 > >>>> #define INT_NMI 13 > >>>> +#define INT_AVEC 14 > >>>> > >>>> /* ExcCodes corresponding to interrupts */ > >>>> -#define EXCCODE_INT_NUM (INT_NMI + 1) > >>>> +#define EXCCODE_INT_NUM (INT_AVEC + 1) > >>>> #define EXCCODE_INT_START 64 > >>>> #define EXCCODE_INT_END (EXCCODE_INT_START + EXCCODE_INT_NUM - 1) > >>>> > >>>> diff --git a/arch/loongarch/include/asm/smp.h b/arch/loongarch/include/asm/smp.h > >>>> index 278700cfee88..2399004596a3 100644 > >>>> --- a/arch/loongarch/include/asm/smp.h > >>>> +++ b/arch/loongarch/include/asm/smp.h > >>>> @@ -69,9 +69,11 @@ extern int __cpu_logical_map[NR_CPUS]; > >>>> #define ACTION_BOOT_CPU 0 > >>>> #define ACTION_RESCHEDULE 1 > >>>> #define ACTION_CALL_FUNCTION 2 > >>>> +#define ACTION_CLEAR_VECT 3 > >>>> #define SMP_BOOT_CPU BIT(ACTION_BOOT_CPU) > >>>> #define SMP_RESCHEDULE BIT(ACTION_RESCHEDULE) > >>>> #define SMP_CALL_FUNCTION BIT(ACTION_CALL_FUNCTION) > >>>> +#define SMP_CLEAR_VECT BIT(ACTION_CLEAR_VECT) > >>>> > >>>> struct secondary_data { > >>>> unsigned long stack; > >>>> diff --git a/arch/loongarch/kernel/cpu-probe.c b/arch/loongarch/kernel/cpu-probe.c > >>>> index 55320813ee08..3b2e72e8f9bd 100644 > >>>> --- a/arch/loongarch/kernel/cpu-probe.c > >>>> +++ b/arch/loongarch/kernel/cpu-probe.c > >>>> @@ -106,7 +106,6 @@ static void cpu_probe_common(struct cpuinfo_loongarch *c) > >>>> elf_hwcap |= HWCAP_LOONGARCH_CRC32; > >>>> } > >>>> > >>>> - > >>>> config = read_cpucfg(LOONGARCH_CPUCFG2); > >>>> if (config & CPUCFG2_LAM) { > >>>> c->options |= LOONGARCH_CPU_LAM; > >>>> @@ -176,6 +175,8 @@ static void cpu_probe_common(struct cpuinfo_loongarch *c) > >>>> c->options |= LOONGARCH_CPU_EIODECODE; > >>>> if (config & IOCSRF_VM) > >>>> c->options |= LOONGARCH_CPU_HYPERVISOR; > >>>> + if (config & IOCSRF_AVEC) > >>>> + c->options |= LOONGARCH_CPU_AVECINT; > >>>> > >>>> config = csr_read32(LOONGARCH_CSR_ASID); > >>>> config = (config & CSR_ASID_BIT) >> CSR_ASID_BIT_SHIFT; > >>>> diff --git a/arch/loongarch/kernel/smp.c b/arch/loongarch/kernel/smp.c > >>>> index 0dfe2388ef41..6dfedef306f3 100644 > >>>> --- a/arch/loongarch/kernel/smp.c > >>>> +++ b/arch/loongarch/kernel/smp.c > >>>> @@ -234,6 +234,9 @@ static irqreturn_t loongson_ipi_interrupt(int irq, void *dev) > >>>> per_cpu(irq_stat, cpu).ipi_irqs[IPI_CALL_FUNCTION]++; > >>>> } > >>>> > >>>> + if (action & SMP_CLEAR_VECT) > >>>> + complete_irq_moving(); > >>> SMP_CLEAR_VECT trigger complete_irq_moving(), but in > >>> complete_irq_moving() it calls mp_ops.send_ipi_single(cpu, > >>> ACTION_CLEAR_VECT), which will trigger loongson_ipi_interrupt() again. > >>> I'm a bit confused here. Could you please explain the design goal of > >>> SMP_CLEAR_VECT here? > SMP_CLEAR_VECT is used to release the affinity of interrupts on old CPUs > > Due to the possibility of certain interrupts occurring during interrupt > affinity modification, > > which can result in the original CPU receiving this interrupt vector, it > is necessary to > > determine whether the interrupt has been processed through the original CPU. > > When the original CPU discovers that an interrupt that needs to clear > affinity has not > > been processed locally in the SMP_CLEAR_VECT process, it will suspend > SMP_CLEAR_VECT > > again and end the process, allowing the avec interrupt that has already > been routed > > to the current CPU to be processed. > > Then, it will enter the SMP_CLEAR_VECT process again for judgment When setting affinity from CPU A to CPU B, trigger SMP_CLEAR_VECT on A, A calls complete_irq_moving() to see if the irq is processed. If not, it triggers SMP_CLEAR_VECT again and again, until B has processed this irq. Am I right (correct me if not)? Huacai > > >>> > >>>> + > >>>> return IRQ_HANDLED; > >>>> } > >>>> > >>>> @@ -388,6 +391,7 @@ int loongson_cpu_disable(void) > >>>> irq_migrate_all_off_this_cpu(); > >>>> clear_csr_ecfg(ECFG0_IM); > >>>> local_irq_restore(flags); > >>>> + loongarch_avec_offline_cpu(cpu); > >>>> local_flush_tlb_all(); > >>>> > >>>> return 0; > >>>> @@ -566,6 +570,7 @@ asmlinkage void start_secondary(void) > >>>> * early is dangerous. > >>>> */ > >>>> WARN_ON_ONCE(!irqs_disabled()); > >>>> + loongarch_avec_online_cpu(cpu); > >>>> loongson_smp_finish(); > >>>> > >>>> cpu_startup_entry(CPUHP_AP_ONLINE_IDLE); > >>>> diff --git a/drivers/irqchip/Makefile b/drivers/irqchip/Makefile > >>>> index 2df72b7b165b..de091a9f7c88 100644 > >>>> --- a/drivers/irqchip/Makefile > >>>> +++ b/drivers/irqchip/Makefile > >>>> @@ -110,7 +110,7 @@ obj-$(CONFIG_LS1X_IRQ) += irq-ls1x.o > >>>> obj-$(CONFIG_TI_SCI_INTR_IRQCHIP) += irq-ti-sci-intr.o > >>>> obj-$(CONFIG_TI_SCI_INTA_IRQCHIP) += irq-ti-sci-inta.o > >>>> obj-$(CONFIG_TI_PRUSS_INTC) += irq-pruss-intc.o > >>>> -obj-$(CONFIG_IRQ_LOONGARCH_CPU) += irq-loongarch-cpu.o > >>>> +obj-$(CONFIG_IRQ_LOONGARCH_CPU) += irq-loongarch-cpu.o irq-loongarch-avec.o > >>>> obj-$(CONFIG_LOONGSON_LIOINTC) += irq-loongson-liointc.o > >>>> obj-$(CONFIG_LOONGSON_EIOINTC) += irq-loongson-eiointc.o > >>>> obj-$(CONFIG_LOONGSON_HTPIC) += irq-loongson-htpic.o > >>>> diff --git a/drivers/irqchip/irq-loongarch-avec.c b/drivers/irqchip/irq-loongarch-avec.c > >>>> new file mode 100644 > >>>> index 000000000000..81dbccbea5a2 > >>>> --- /dev/null > >>>> +++ b/drivers/irqchip/irq-loongarch-avec.c > >>>> @@ -0,0 +1,440 @@ > >>>> +// SPDX-License-Identifier: GPL-2.0 > >>>> +/* > >>>> + * Copyright (C) 2020-2024 Loongson Technologies, Inc. > >>>> + */ > >>>> + > >>>> +#include <linux/cpuhotplug.h> > >>>> +#include <linux/init.h> > >>>> +#include <linux/interrupt.h> > >>>> +#include <linux/irq.h> > >>>> +#include <linux/irqchip.h> > >>>> +#include <linux/irqchip/chained_irq.h> > >>>> +#include <linux/irqdomain.h> > >>>> +#include <linux/kernel.h> > >>>> +#include <linux/msi.h> > >>>> +#include <linux/radix-tree.h> > >>>> +#include <linux/spinlock.h> > >>>> + > >>>> +#include <asm/loongarch.h> > >>>> +#include <asm/setup.h> > >>>> + > >>>> +#define VECTORS_PER_REG 64 > >>>> +#define ILR_INVALID_MASK 0x80000000UL > >>>> +#define ILR_VECTOR_MASK 0xffUL > >>>> +#define AVEC_MSG_OFFSET 0x100000 > >>>> + > >>>> +static phys_addr_t msi_base_v2; > >>>> +static DEFINE_PER_CPU(struct irq_desc * [NR_VECTORS], irq_map); > >>>> + > >>>> +#ifdef CONFIG_SMP > >>>> +struct pending_list { > >>>> + struct list_head head; > >>>> +}; > >>>> + > >>>> +static DEFINE_PER_CPU(struct pending_list, pending_list); > >>>> +#endif > >>>> + > >>>> +struct loongarch_avec_chip { > >>>> + struct fwnode_handle *fwnode; > >>>> + struct irq_domain *domain; > >>>> + struct irq_matrix *vector_matrix; > >>>> + raw_spinlock_t lock; > >>>> +}; > >>>> + > >>>> +static struct loongarch_avec_chip loongarch_avec; > >>>> + > >>>> +struct loongarch_avec_data { > >>>> + struct list_head entry; > >>>> + unsigned int cpu; > >>>> + unsigned int vec; > >>>> + unsigned int prev_cpu; > >>>> + unsigned int prev_vec; > >>>> + unsigned int moving : 1, > >>>> + managed : 1; > >>>> +}; > >>>> + > >>>> +static struct cpumask intersect_mask; > >>>> + > >>>> +static int assign_irq_vector(struct irq_data *irqd, const struct cpumask *dest, > >>>> + unsigned int *cpu) > >>>> +{ > >>>> + return irq_matrix_alloc(loongarch_avec.vector_matrix, dest, false, cpu); > >>>> +} > >>>> + > >>>> +static inline void loongarch_avec_ack_irq(struct irq_data *d) > >>>> +{ > >>>> +} > >>>> + > >>>> +static inline void loongarch_avec_unmask_irq(struct irq_data *d) > >>>> +{ > >>>> +} > >>>> + > >>>> +static inline void loongarch_avec_mask_irq(struct irq_data *d) > >>>> +{ > >>>> +} > >>>> + > >>>> +#ifdef CONFIG_SMP > >>>> +static inline void pending_list_init(int cpu) > >>>> +{ > >>>> + struct pending_list *plist = per_cpu_ptr(&pending_list, cpu); > >>>> + > >>>> + INIT_LIST_HEAD(&plist->head); > >>>> +} > >>>> + > >>>> +static void loongarch_avec_sync(struct loongarch_avec_data *adata) > >>>> +{ > >>>> + struct pending_list *plist; > >>>> + > >>>> + if (cpu_online(adata->prev_cpu)) { > >>>> + plist = per_cpu_ptr(&pending_list, adata->prev_cpu); > >>>> + list_add_tail(&adata->entry, &plist->head); > >>>> + adata->moving = true; > >>>> + mp_ops.send_ipi_single(adata->prev_cpu, ACTION_CLEAR_VECT); > >>>> + } > >>>> +} > >>>> + > >>>> +static int loongarch_avec_set_affinity(struct irq_data *data, const struct cpumask *dest, > >>>> + bool force) > >>>> +{ > >>>> + struct loongarch_avec_data *adata; > >>>> + unsigned int cpu, vector; > >>>> + unsigned long flags; > >>>> + int ret; > >>>> + > >>>> + raw_spin_lock_irqsave(&loongarch_avec.lock, flags); > >>>> + adata = irq_data_get_irq_chip_data(data); > >>>> + > >>>> + if (adata->vec && cpu_online(adata->cpu) && cpumask_test_cpu(adata->cpu, dest)) { > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + return 0; > >>>> + } > >>>> + if (adata->moving) > >>>> + return -EBUSY; > >>>> + > >>>> + cpumask_and(&intersect_mask, dest, cpu_online_mask); > >>>> + > >>>> + ret = assign_irq_vector(data, &intersect_mask, &cpu); > >>>> + if (ret < 0) { > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + return ret; > >>>> + } > >>>> + vector = ret; > >>>> + adata->cpu = cpu; > >>>> + adata->vec = vector; > >>>> + per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(data); > >>>> + loongarch_avec_sync(adata); > >>>> + > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + irq_data_update_effective_affinity(data, cpumask_of(cpu)); > >>>> + > >>>> + return IRQ_SET_MASK_OK; > >>>> +} > >>>> + > >>>> +void complete_irq_moving(void) > >>>> +{ > >>>> + struct pending_list *plist = this_cpu_ptr(&pending_list); > >>>> + struct loongarch_avec_data *adata, *tmp; > >>>> + int cpu, vector, bias; > >>>> + u64 irr; > >>>> + > >>>> + raw_spin_lock(&loongarch_avec.lock); > >>>> + > >>>> + list_for_each_entry_safe(adata, tmp, &plist->head, entry) { > >>>> + cpu = adata->prev_cpu; > >>>> + vector = adata->prev_vec; > >>>> + bias = vector / VECTORS_PER_REG; > >>>> + switch (bias) { > >>>> + case 0: > >>>> + irr = csr_read64(LOONGARCH_CSR_IRR0); > >>>> + case 1: > >>>> + irr = csr_read64(LOONGARCH_CSR_IRR1); > >>>> + case 2: > >>>> + irr = csr_read64(LOONGARCH_CSR_IRR2); > >>>> + case 3: > >>>> + irr = csr_read64(LOONGARCH_CSR_IRR3); > >>>> + } > >>>> + > >>>> + if (irr & (1UL << (vector % VECTORS_PER_REG))) { > >>>> + mp_ops.send_ipi_single(cpu, ACTION_CLEAR_VECT); > >>>> + continue; > >>>> + } > >>>> + list_del(&adata->entry); > >>>> + irq_matrix_free(loongarch_avec.vector_matrix, cpu, vector, adata->managed); > >>>> + this_cpu_write(irq_map[vector], NULL); > >>>> + adata->prev_cpu = adata->cpu; > >>>> + adata->prev_vec = adata->vec; > >>>> + adata->moving = 0; > >>>> + } > >>>> + raw_spin_unlock(&loongarch_avec.lock); > >>>> +} > >>>> + > >>>> +void loongarch_avec_offline_cpu(unsigned int cpu) > >>>> +{ > >>>> + struct pending_list *plist = per_cpu_ptr(&pending_list, cpu); > >>>> + unsigned long flags; > >>>> + > >>>> + if (!loongarch_avec.vector_matrix) > >>>> + return; > >>>> + > >>>> + raw_spin_lock_irqsave(&loongarch_avec.lock, flags); > >>>> + if (list_empty(&plist->head)) > >>>> + irq_matrix_offline(loongarch_avec.vector_matrix); > >>>> + else > >>>> + pr_warn("cpu %d advanced extioi is busy\n", cpu); > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> +} > >>>> + > >>>> +void loongarch_avec_online_cpu(unsigned int cpu) > >>>> +{ > >>>> + unsigned long flags; > >>>> + > >>>> + if (!loongarch_avec.vector_matrix) > >>>> + return; > >>>> + > >>>> + raw_spin_lock_irqsave(&loongarch_avec.lock, flags); > >>>> + > >>>> + irq_matrix_online(loongarch_avec.vector_matrix); > >>>> + > >>>> + pending_list_init(cpu); > >>>> + > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> +} > >>>> + > >>>> +#else > >>>> +#define loongarch_avec_set_affinity NULL > >>> Empty function is prefered here. > > ok ,I got it > > Thanks > > Tianyang > > >>> > >>> > >>> Huacai > >>> > >>>> +#endif > >>>> + > >>>> +static void loongarch_avec_compose_msg(struct irq_data *d, > >>>> + struct msi_msg *msg) > >>>> +{ > >>>> + struct loongarch_avec_data *avec_data; > >>>> + > >>>> + avec_data = irq_data_get_irq_chip_data(d); > >>>> + > >>>> + msg->address_hi = 0xfd; > >>>> + msg->address_lo = ((avec_data->vec & 0xff) << 4) | > >>>> + ((cpu_logical_map(avec_data->cpu & 0xffff)) << 12); > >>>> + msg->data = 0x0; > >>>> + > >>>> +} > >>>> + > >>>> +static struct irq_chip loongarch_avec_controller = { > >>>> + .name = "AVECINTC", > >>>> + .irq_ack = loongarch_avec_ack_irq, > >>>> + .irq_mask = loongarch_avec_mask_irq, > >>>> + .irq_unmask = loongarch_avec_unmask_irq, > >>>> + .irq_set_affinity = loongarch_avec_set_affinity, > >>>> + .irq_compose_msi_msg = loongarch_avec_compose_msg, > >>>> +}; > >>>> + > >>>> +static void loongarch_avec_dispatch(struct irq_desc *desc) > >>>> +{ > >>>> + struct irq_chip *chip = irq_desc_get_chip(desc); > >>>> + unsigned long vector; > >>>> + struct irq_desc *d; > >>>> + > >>>> + chained_irq_enter(chip, desc); > >>>> + vector = csr_read64(LOONGARCH_CSR_ILR); > >>>> + if (vector & ILR_INVALID_MASK) > >>>> + return; > >>>> + > >>>> + vector &= ILR_VECTOR_MASK; > >>>> + > >>>> + d = this_cpu_read(irq_map[vector]); > >>>> + if (d) { > >>>> + generic_handle_irq_desc(d); > >>>> + } else { > >>>> + pr_warn("IRQ ERROR:Unexpected irq occur on cpu %d[vector %ld]\n", > >>>> + smp_processor_id(), vector); > >>>> + } > >>>> + > >>>> + chained_irq_exit(chip, desc); > >>>> +} > >>>> + > >>>> +static int loongarch_avec_alloc(struct irq_domain *domain, unsigned int virq, > >>>> + unsigned int nr_irqs, void *arg) > >>>> +{ > >>>> + struct loongarch_avec_data *adata; > >>>> + struct irq_data *irqd; > >>>> + unsigned int cpu, vector, i, ret; > >>>> + unsigned long flags; > >>>> + > >>>> + raw_spin_lock_irqsave(&loongarch_avec.lock, flags); > >>>> + for (i = 0; i < nr_irqs; i++) { > >>>> + irqd = irq_domain_get_irq_data(domain, virq + i); > >>>> + adata = kzalloc(sizeof(*adata), GFP_KERNEL); > >>>> + if (!adata) { > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + return -ENOMEM; > >>>> + } > >>>> + ret = assign_irq_vector(irqd, cpu_online_mask, &cpu); > >>>> + if (ret < 0) { > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + return ret; > >>>> + } > >>>> + vector = ret; > >>>> + adata->prev_cpu = adata->cpu = cpu; > >>>> + adata->prev_vec = adata->vec = vector; > >>>> + adata->managed = irqd_affinity_is_managed(irqd); > >>>> + irq_domain_set_info(domain, virq + i, virq + i, &loongarch_avec_controller, > >>>> + adata, handle_edge_irq, NULL, NULL); > >>>> + adata->moving = 0; > >>>> + irqd_set_single_target(irqd); > >>>> + irqd_set_affinity_on_activate(irqd); > >>>> + > >>>> + per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(irqd); > >>>> + } > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> + > >>>> + return 0; > >>>> +} > >>>> + > >>>> +static void clear_free_vector(struct irq_data *irqd) > >>>> +{ > >>>> + struct loongarch_avec_data *adata = irq_data_get_irq_chip_data(irqd); > >>>> + bool managed = irqd_affinity_is_managed(irqd); > >>>> + > >>>> + per_cpu(irq_map, adata->cpu)[adata->vec] = NULL; > >>>> + irq_matrix_free(loongarch_avec.vector_matrix, adata->cpu, adata->vec, managed); > >>>> + adata->cpu = 0; > >>>> + adata->vec = 0; > >>>> +#ifdef CONFIG_SMP > >>>> + if (!adata->moving) > >>>> + return; > >>>> + > >>>> + per_cpu(irq_map, adata->prev_cpu)[adata->prev_vec] = 0; > >>>> + irq_matrix_free(loongarch_avec.vector_matrix, adata->prev_cpu, > >>>> + adata->prev_vec, adata->managed); > >>>> + adata->prev_vec = 0; > >>>> + adata->prev_cpu = 0; > >>>> + adata->moving = 0; > >>>> + list_del_init(&adata->entry); > >>>> +#endif > >>>> +} > >>>> + > >>>> +static void loongarch_avec_free(struct irq_domain *domain, unsigned int virq, > >>>> + unsigned int nr_irqs) > >>>> +{ > >>>> + struct irq_data *d; > >>>> + unsigned long flags; > >>>> + unsigned int i; > >>>> + > >>>> + raw_spin_lock_irqsave(&loongarch_avec.lock, flags); > >>>> + for (i = 0; i < nr_irqs; i++) { > >>>> + d = irq_domain_get_irq_data(domain, virq + i); > >>>> + if (d) { > >>>> + clear_free_vector(d); > >>>> + irq_domain_reset_irq_data(d); > >>>> + > >>>> + } > >>>> + } > >>>> + > >>>> + raw_spin_unlock_irqrestore(&loongarch_avec.lock, flags); > >>>> +} > >>>> + > >>>> +static const struct irq_domain_ops loongarch_avec_domain_ops = { > >>>> + .alloc = loongarch_avec_alloc, > >>>> + .free = loongarch_avec_free, > >>>> +}; > >>>> + > >>>> +static int __init irq_matrix_init(void) > >>>> +{ > >>>> + int i; > >>>> + > >>>> + loongarch_avec.vector_matrix = irq_alloc_matrix(NR_VECTORS, 0, NR_VECTORS - 1); > >>>> + if (!loongarch_avec.vector_matrix) > >>>> + return -ENOMEM; > >>>> + for (i = 0; i < NR_LEGACY_VECTORS; i++) > >>>> + irq_matrix_assign_system(loongarch_avec.vector_matrix, i, false); > >>>> + > >>>> + irq_matrix_online(loongarch_avec.vector_matrix); > >>>> + > >>>> + return 0; > >>>> +} > >>>> + > >>>> +static int __init loongarch_avec_init(struct irq_domain *parent) > >>>> +{ > >>>> + int ret = 0, parent_irq; > >>>> + unsigned long tmp; > >>>> + > >>>> + raw_spin_lock_init(&loongarch_avec.lock); > >>>> + > >>>> + loongarch_avec.fwnode = irq_domain_alloc_named_fwnode("CORE_AVEC"); > >>>> + if (!loongarch_avec.fwnode) { > >>>> + pr_err("Unable to allocate domain handle\n"); > >>>> + ret = -ENOMEM; > >>>> + goto out; > >>>> + } > >>>> + > >>>> + loongarch_avec.domain = irq_domain_create_tree(loongarch_avec.fwnode, > >>>> + &loongarch_avec_domain_ops, NULL); > >>>> + if (!loongarch_avec.domain) { > >>>> + pr_err("core-vec: cannot create IRQ domain\n"); > >>>> + ret = -ENOMEM; > >>>> + goto out_free_handle; > >>>> + } > >>>> + > >>>> + parent_irq = irq_create_mapping(parent, INT_AVEC); > >>>> + if (!parent_irq) { > >>>> + pr_err("Failed to mapping hwirq\n"); > >>>> + ret = -EINVAL; > >>>> + goto out_remove_domain; > >>>> + } > >>>> + irq_set_chained_handler_and_data(parent_irq, loongarch_avec_dispatch, NULL); > >>>> + > >>>> + ret = irq_matrix_init(); > >>>> + if (ret) { > >>>> + pr_err("Failed to init irq matrix\n"); > >>>> + goto out_free_matrix; > >>>> + } > >>>> +#ifdef CONFIG_SMP > >>>> + pending_list_init(0); > >>>> +#endif > >>>> + tmp = iocsr_read64(LOONGARCH_IOCSR_MISC_FUNC); > >>>> + tmp |= IOCSR_MISC_FUNC_AVEC_EN; > >>>> + iocsr_write64(tmp, LOONGARCH_IOCSR_MISC_FUNC); > >>>> + > >>>> + return ret; > >>>> + > >>>> +out_free_matrix: > >>>> + kfree(loongarch_avec.vector_matrix); > >>>> +out_remove_domain: > >>>> + irq_domain_remove(loongarch_avec.domain); > >>>> +out_free_handle: > >>>> + irq_domain_free_fwnode(loongarch_avec.fwnode); > >>>> +out: > >>>> + return ret; > >>>> +} > >>>> + > >>>> +static int __init pch_msi_parse_madt(union acpi_subtable_headers *header, > >>>> + const unsigned long end) > >>>> +{ > >>>> + struct acpi_madt_msi_pic *pchmsi_entry = (struct acpi_madt_msi_pic *)header; > >>>> + > >>>> + msi_base_v2 = pchmsi_entry->msg_address - AVEC_MSG_OFFSET; > >>>> + return pch_msi_acpi_init_v2(loongarch_avec.domain, pchmsi_entry); > >>>> +} > >>>> + > >>>> +static inline int __init acpi_cascade_irqdomain_init(void) > >>>> +{ > >>>> + return acpi_table_parse_madt(ACPI_MADT_TYPE_MSI_PIC, pch_msi_parse_madt, 1); > >>>> +} > >>>> + > >>>> +int __init loongarch_avec_acpi_init(struct irq_domain *parent) > >>>> +{ > >>>> + int ret = 0; > >>>> + > >>>> + ret = loongarch_avec_init(parent); > >>>> + if (ret) { > >>>> + pr_err("Failed to init irq domain\n"); > >>>> + return ret; > >>>> + } > >>>> + > >>>> + ret = acpi_cascade_irqdomain_init(); > >>>> + if (ret) { > >>>> + pr_err("Failed to cascade IRQ domain\n"); > >>>> + return ret; > >>>> + } > >>>> + > >>>> + return ret; > >>>> +} > >>>> diff --git a/drivers/irqchip/irq-loongarch-cpu.c b/drivers/irqchip/irq-loongarch-cpu.c > >>>> index 9d8f2c406043..1ecac59925c6 100644 > >>>> --- a/drivers/irqchip/irq-loongarch-cpu.c > >>>> +++ b/drivers/irqchip/irq-loongarch-cpu.c > >>>> @@ -138,7 +138,9 @@ static int __init acpi_cascade_irqdomain_init(void) > >>>> if (r < 0) > >>>> return r; > >>>> > >>>> - return 0; > >>>> + if (cpu_has_avecint) > >>>> + r = loongarch_avec_acpi_init(irq_domain); > >>>> + return r; > >>>> } > >>>> > >>>> static int __init cpuintc_acpi_init(union acpi_subtable_headers *header, > >>>> diff --git a/drivers/irqchip/irq-loongson-eiointc.c b/drivers/irqchip/irq-loongson-eiointc.c > >>>> index c7ddebf312ad..1f9a30488137 100644 > >>>> --- a/drivers/irqchip/irq-loongson-eiointc.c > >>>> +++ b/drivers/irqchip/irq-loongson-eiointc.c > >>>> @@ -359,6 +359,9 @@ static int __init acpi_cascade_irqdomain_init(void) > >>>> if (r < 0) > >>>> return r; > >>>> > >>>> + if (cpu_has_avecint) > >>>> + return 0; > >>>> + > >>>> r = acpi_table_parse_madt(ACPI_MADT_TYPE_MSI_PIC, pch_msi_parse_madt, 1); > >>>> if (r < 0) > >>>> return r; > >>>> diff --git a/drivers/irqchip/irq-loongson-pch-msi.c b/drivers/irqchip/irq-loongson-pch-msi.c > >>>> index dd4d699170f4..1926857f9a41 100644 > >>>> --- a/drivers/irqchip/irq-loongson-pch-msi.c > >>>> +++ b/drivers/irqchip/irq-loongson-pch-msi.c > >>>> @@ -16,7 +16,6 @@ > >>>> #include <linux/slab.h> > >>>> > >>>> static int nr_pics; > >>>> - > >>>> struct pch_msi_data { > >>>> struct mutex msi_map_lock; > >>>> phys_addr_t doorbell; > >>>> @@ -100,6 +99,17 @@ static struct irq_chip middle_irq_chip = { > >>>> .irq_compose_msi_msg = pch_msi_compose_msi_msg, > >>>> }; > >>>> > >>>> +static struct irq_chip pch_msi_irq_chip_v2 = { > >>>> + .name = "MSI", > >>>> + .irq_ack = irq_chip_ack_parent, > >>>> +}; > >>>> + > >>>> +static struct msi_domain_info pch_msi_domain_info_v2 = { > >>>> + .flags = MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS | > >>>> + MSI_FLAG_MULTI_PCI_MSI | MSI_FLAG_PCI_MSIX, > >>>> + .chip = &pch_msi_irq_chip_v2, > >>>> +}; > >>>> + > >>>> static int pch_msi_parent_domain_alloc(struct irq_domain *domain, > >>>> unsigned int virq, int hwirq) > >>>> { > >>>> @@ -268,6 +278,9 @@ struct fwnode_handle *get_pch_msi_handle(int pci_segment) > >>>> { > >>>> int i; > >>>> > >>>> + if (cpu_has_avecint) > >>>> + return pch_msi_handle[0]; > >>>> + > >>>> for (i = 0; i < MAX_IO_PICS; i++) { > >>>> if (msi_group[i].pci_segment == pci_segment) > >>>> return pch_msi_handle[i]; > >>>> @@ -289,4 +302,32 @@ int __init pch_msi_acpi_init(struct irq_domain *parent, > >>>> > >>>> return ret; > >>>> } > >>>> + > >>>> +int __init pch_msi_acpi_init_v2(struct irq_domain *parent, > >>>> + struct acpi_madt_msi_pic *msi_entry) > >>>> +{ > >>>> + struct irq_domain *msi_domain; > >>>> + > >>>> + if (pch_msi_handle[0]) > >>>> + return 0; > >>>> + > >>>> + pch_msi_handle[0] = irq_domain_alloc_named_fwnode("msipic-v2"); > >>>> + if (!pch_msi_handle[0]) { > >>>> + pr_err("Unable to allocate domain handle\n"); > >>>> + kfree(pch_msi_handle[0]); > >>>> + return -ENOMEM; > >>>> + } > >>>> + > >>>> + msi_domain = pci_msi_create_irq_domain(pch_msi_handle[0], > >>>> + &pch_msi_domain_info_v2, > >>>> + parent); > >>>> + if (!msi_domain) { > >>>> + pr_err("Failed to create PCI MSI domain\n"); > >>>> + kfree(pch_msi_handle[0]); > >>>> + return -ENOMEM; > >>>> + } > >>>> + > >>>> + pr_info("IRQ domain MSIPIC-V2 init done.\n"); > >>>> + return 0; > >>>> +} > >>>> #endif > >>>> -- > >>>> 2.20.1 > >>>> > >>>> > >> >