在2022年6月1日六月 上午10:59,Huacai Chen写道: > Add some misc common routines for LoongArch, including: asm-offsets > routines, futex functions, i/o memory access functions, frame-buffer > functions, procfs information display, etc. > > Reviewed-by: WANG Xuerui <git@xxxxxxxxxx> > Signed-off-by: Huacai Chen <chenhuacai@xxxxxxxxxxx> Reviewed-by: Jiaxun Yang <jiaxun.yang@xxxxxxxxxxx> > --- > arch/loongarch/include/asm/asm-offsets.h | 5 + > arch/loongarch/include/asm/fb.h | 23 ++ > arch/loongarch/include/asm/futex.h | 107 ++++++++++ > arch/loongarch/include/asm/io.h | 129 ++++++++++++ > arch/loongarch/kernel/asm-offsets.c | 254 +++++++++++++++++++++++ > arch/loongarch/kernel/io.c | 94 +++++++++ > arch/loongarch/kernel/proc.c | 122 +++++++++++ > 7 files changed, 734 insertions(+) > create mode 100644 arch/loongarch/include/asm/asm-offsets.h > create mode 100644 arch/loongarch/include/asm/fb.h > create mode 100644 arch/loongarch/include/asm/futex.h > create mode 100644 arch/loongarch/include/asm/io.h > create mode 100644 arch/loongarch/kernel/asm-offsets.c > create mode 100644 arch/loongarch/kernel/io.c > create mode 100644 arch/loongarch/kernel/proc.c > > diff --git a/arch/loongarch/include/asm/asm-offsets.h > b/arch/loongarch/include/asm/asm-offsets.h > new file mode 100644 > index 000000000000..d9ad88d293e7 > --- /dev/null > +++ b/arch/loongarch/include/asm/asm-offsets.h > @@ -0,0 +1,5 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#include <generated/asm-offsets.h> > diff --git a/arch/loongarch/include/asm/fb.h > b/arch/loongarch/include/asm/fb.h > new file mode 100644 > index 000000000000..3116bde8772d > --- /dev/null > +++ b/arch/loongarch/include/asm/fb.h > @@ -0,0 +1,23 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#ifndef _ASM_FB_H_ > +#define _ASM_FB_H_ > + > +#include <linux/fb.h> > +#include <linux/fs.h> > +#include <asm/page.h> > + > +static inline void fb_pgprotect(struct file *file, struct > vm_area_struct *vma, > + unsigned long off) > +{ > + vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); > +} > + > +static inline int fb_is_primary_device(struct fb_info *info) > +{ > + return 0; > +} > + > +#endif /* _ASM_FB_H_ */ > diff --git a/arch/loongarch/include/asm/futex.h > b/arch/loongarch/include/asm/futex.h > new file mode 100644 > index 000000000000..b27d55f92db7 > --- /dev/null > +++ b/arch/loongarch/include/asm/futex.h > @@ -0,0 +1,107 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#ifndef _ASM_FUTEX_H > +#define _ASM_FUTEX_H > + > +#include <linux/futex.h> > +#include <linux/uaccess.h> > +#include <asm/barrier.h> > +#include <asm/compiler.h> > +#include <asm/errno.h> > + > +#define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ > +{ \ > + __asm__ __volatile__( \ > + "1: ll.w %1, %4 # __futex_atomic_op\n" \ > + " " insn " \n" \ > + "2: sc.w $t0, %2 \n" \ > + " beq $t0, $zero, 1b \n" \ > + "3: \n" \ > + " .section .fixup,\"ax\" \n" \ > + "4: li.w %0, %6 \n" \ > + " b 3b \n" \ > + " .previous \n" \ > + " .section __ex_table,\"a\" \n" \ > + " "__UA_ADDR "\t1b, 4b \n" \ > + " "__UA_ADDR "\t2b, 4b \n" \ > + " .previous \n" \ > + : "=r" (ret), "=&r" (oldval), \ > + "=ZC" (*uaddr) \ > + : "0" (0), "ZC" (*uaddr), "Jr" (oparg), \ > + "i" (-EFAULT) \ > + : "memory", "t0"); \ > +} > + > +static inline int > +arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user > *uaddr) > +{ > + int oldval = 0, ret = 0; > + > + pagefault_disable(); > + > + switch (op) { > + case FUTEX_OP_SET: > + __futex_atomic_op("move $t0, %z5", ret, oldval, uaddr, oparg); > + break; > + case FUTEX_OP_ADD: > + __futex_atomic_op("add.w $t0, %1, %z5", ret, oldval, uaddr, oparg); > + break; > + case FUTEX_OP_OR: > + __futex_atomic_op("or $t0, %1, %z5", ret, oldval, uaddr, oparg); > + break; > + case FUTEX_OP_ANDN: > + __futex_atomic_op("and $t0, %1, %z5", ret, oldval, uaddr, ~oparg); > + break; > + case FUTEX_OP_XOR: > + __futex_atomic_op("xor $t0, %1, %z5", ret, oldval, uaddr, oparg); > + break; > + default: > + ret = -ENOSYS; > + } > + > + pagefault_enable(); > + > + if (!ret) > + *oval = oldval; > + > + return ret; > +} > + > +static inline int > +futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, u32 > oldval, u32 newval) > +{ > + int ret = 0; > + u32 val = 0; > + > + if (!access_ok(uaddr, sizeof(u32))) > + return -EFAULT; > + > + __asm__ __volatile__( > + "# futex_atomic_cmpxchg_inatomic \n" > + "1: ll.w %1, %3 \n" > + " bne %1, %z4, 3f \n" > + " or $t0, %z5, $zero \n" > + "2: sc.w $t0, %2 \n" > + " beq $zero, $t0, 1b \n" > + "3: \n" > + " .section .fixup,\"ax\" \n" > + "4: li.d %0, %6 \n" > + " b 3b \n" > + " .previous \n" > + " .section __ex_table,\"a\" \n" > + " "__UA_ADDR "\t1b, 4b \n" > + " "__UA_ADDR "\t2b, 4b \n" > + " .previous \n" > + : "+r" (ret), "=&r" (val), "=" GCC_OFF_SMALL_ASM() (*uaddr) > + : GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oldval), "Jr" (newval), > + "i" (-EFAULT) > + : "memory", "t0"); > + > + *uval = val; > + > + return ret; > +} > + > +#endif /* _ASM_FUTEX_H */ > diff --git a/arch/loongarch/include/asm/io.h > b/arch/loongarch/include/asm/io.h > new file mode 100644 > index 000000000000..884599739b36 > --- /dev/null > +++ b/arch/loongarch/include/asm/io.h > @@ -0,0 +1,129 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#ifndef _ASM_IO_H > +#define _ASM_IO_H > + > +#define ARCH_HAS_IOREMAP_WC > + > +#include <linux/compiler.h> > +#include <linux/kernel.h> > +#include <linux/types.h> > + > +#include <asm/addrspace.h> > +#include <asm/bug.h> > +#include <asm/byteorder.h> > +#include <asm/cpu.h> > +#include <asm/page.h> > +#include <asm/pgtable-bits.h> > +#include <asm/string.h> > + > +/* > + * On LoongArch, I/O ports mappring is following: > + * > + * | .... | > + * |-----------------------| > + * | pci io ports(64K~32M) | > + * |-----------------------| > + * | isa io ports(0 ~16K) | > + * PCI_IOBASE ->|-----------------------| > + * | .... | > + */ > +#define PCI_IOBASE ((void __iomem *)(vm_map_base + (2 * PAGE_SIZE))) > +#define PCI_IOSIZE SZ_32M > +#define ISA_IOSIZE SZ_16K > +#define IO_SPACE_LIMIT (PCI_IOSIZE - 1) > + > +/* > + * Change "struct page" to physical address. > + */ > +#define page_to_phys(page) ((phys_addr_t)page_to_pfn(page) << > PAGE_SHIFT) > + > +extern void __init __iomem *early_ioremap(u64 phys_addr, unsigned long > size); > +extern void __init early_iounmap(void __iomem *addr, unsigned long > size); > + > +#define early_memremap early_ioremap > +#define early_memunmap early_iounmap > + > +static inline void __iomem *ioremap_prot(phys_addr_t offset, unsigned > long size, > + unsigned long prot_val) > +{ > + if (prot_val == _CACHE_CC) > + return (void __iomem *)(unsigned long)(CACHE_BASE + offset); > + else > + return (void __iomem *)(unsigned long)(UNCACHE_BASE + offset); > +} > + > +/* > + * ioremap - map bus memory into CPU space > + * @offset: bus address of the memory > + * @size: size of the resource to map > + * > + * ioremap performs a platform specific sequence of operations to > + * make bus memory CPU accessible via the readb/readw/readl/writeb/ > + * writew/writel functions and the other mmio helpers. The returned > + * address is not guaranteed to be usable directly as a virtual > + * address. > + */ > +#define ioremap(offset, size) \ > + ioremap_prot((offset), (size), _CACHE_SUC) > + > +/* > + * ioremap_wc - map bus memory into CPU space > + * @offset: bus address of the memory > + * @size: size of the resource to map > + * > + * ioremap_wc performs a platform specific sequence of operations to > + * make bus memory CPU accessible via the readb/readw/readl/writeb/ > + * writew/writel functions and the other mmio helpers. The returned > + * address is not guaranteed to be usable directly as a virtual > + * address. > + * > + * This version of ioremap ensures that the memory is marked uncachable > + * but accelerated by means of write-combining feature. It is > specifically > + * useful for PCIe prefetchable windows, which may vastly improve a > + * communications performance. If it was determined on boot stage, what > + * CPU CCA doesn't support WUC, the method shall fall-back to the > + * _CACHE_SUC option (see cpu_probe() method). > + */ > +#define ioremap_wc(offset, size) \ > + ioremap_prot((offset), (size), _CACHE_WUC) > + > +/* > + * ioremap_cache - map bus memory into CPU space > + * @offset: bus address of the memory > + * @size: size of the resource to map > + * > + * ioremap_cache performs a platform specific sequence of operations to > + * make bus memory CPU accessible via the readb/readw/readl/writeb/ > + * writew/writel functions and the other mmio helpers. The returned > + * address is not guaranteed to be usable directly as a virtual > + * address. > + * > + * This version of ioremap ensures that the memory is marked cachable > by > + * the CPU. Also enables full write-combining. Useful for some > + * memory-like regions on I/O busses. > + */ > +#define ioremap_cache(offset, size) \ > + ioremap_prot((offset), (size), _CACHE_CC) > + > +static inline void iounmap(const volatile void __iomem *addr) > +{ > +} > + > +#define mmiowb() asm volatile ("dbar 0" ::: "memory") > + > +/* > + * String version of I/O memory access operations. > + */ > +extern void __memset_io(volatile void __iomem *dst, int c, size_t > count); > +extern void __memcpy_toio(volatile void __iomem *to, const void *from, > size_t count); > +extern void __memcpy_fromio(void *to, const volatile void __iomem > *from, size_t count); > +#define memset_io(c, v, l) __memset_io((c), (v), (l)) > +#define memcpy_fromio(a, c, l) __memcpy_fromio((a), (c), (l)) > +#define memcpy_toio(c, a, l) __memcpy_toio((c), (a), (l)) > + > +#include <asm-generic/io.h> > + > +#endif /* _ASM_IO_H */ > diff --git a/arch/loongarch/kernel/asm-offsets.c > b/arch/loongarch/kernel/asm-offsets.c > new file mode 100644 > index 000000000000..3531e3c60a6e > --- /dev/null > +++ b/arch/loongarch/kernel/asm-offsets.c > @@ -0,0 +1,254 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * asm-offsets.c: Calculate pt_regs and task_struct offsets. > + * > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#include <linux/types.h> > +#include <linux/sched.h> > +#include <linux/mm.h> > +#include <linux/kbuild.h> > +#include <linux/suspend.h> > +#include <asm/cpu-info.h> > +#include <asm/ptrace.h> > +#include <asm/processor.h> > + > +void output_ptreg_defines(void) > +{ > + COMMENT("LoongArch pt_regs offsets."); > + OFFSET(PT_R0, pt_regs, regs[0]); > + OFFSET(PT_R1, pt_regs, regs[1]); > + OFFSET(PT_R2, pt_regs, regs[2]); > + OFFSET(PT_R3, pt_regs, regs[3]); > + OFFSET(PT_R4, pt_regs, regs[4]); > + OFFSET(PT_R5, pt_regs, regs[5]); > + OFFSET(PT_R6, pt_regs, regs[6]); > + OFFSET(PT_R7, pt_regs, regs[7]); > + OFFSET(PT_R8, pt_regs, regs[8]); > + OFFSET(PT_R9, pt_regs, regs[9]); > + OFFSET(PT_R10, pt_regs, regs[10]); > + OFFSET(PT_R11, pt_regs, regs[11]); > + OFFSET(PT_R12, pt_regs, regs[12]); > + OFFSET(PT_R13, pt_regs, regs[13]); > + OFFSET(PT_R14, pt_regs, regs[14]); > + OFFSET(PT_R15, pt_regs, regs[15]); > + OFFSET(PT_R16, pt_regs, regs[16]); > + OFFSET(PT_R17, pt_regs, regs[17]); > + OFFSET(PT_R18, pt_regs, regs[18]); > + OFFSET(PT_R19, pt_regs, regs[19]); > + OFFSET(PT_R20, pt_regs, regs[20]); > + OFFSET(PT_R21, pt_regs, regs[21]); > + OFFSET(PT_R22, pt_regs, regs[22]); > + OFFSET(PT_R23, pt_regs, regs[23]); > + OFFSET(PT_R24, pt_regs, regs[24]); > + OFFSET(PT_R25, pt_regs, regs[25]); > + OFFSET(PT_R26, pt_regs, regs[26]); > + OFFSET(PT_R27, pt_regs, regs[27]); > + OFFSET(PT_R28, pt_regs, regs[28]); > + OFFSET(PT_R29, pt_regs, regs[29]); > + OFFSET(PT_R30, pt_regs, regs[30]); > + OFFSET(PT_R31, pt_regs, regs[31]); > + OFFSET(PT_CRMD, pt_regs, csr_crmd); > + OFFSET(PT_PRMD, pt_regs, csr_prmd); > + OFFSET(PT_EUEN, pt_regs, csr_euen); > + OFFSET(PT_ECFG, pt_regs, csr_ecfg); > + OFFSET(PT_ESTAT, pt_regs, csr_estat); > + OFFSET(PT_ERA, pt_regs, csr_era); > + OFFSET(PT_BVADDR, pt_regs, csr_badvaddr); > + OFFSET(PT_ORIG_A0, pt_regs, orig_a0); > + DEFINE(PT_SIZE, sizeof(struct pt_regs)); > + BLANK(); > +} > + > +void output_task_defines(void) > +{ > + COMMENT("LoongArch task_struct offsets."); > + OFFSET(TASK_STATE, task_struct, __state); > + OFFSET(TASK_THREAD_INFO, task_struct, stack); > + OFFSET(TASK_FLAGS, task_struct, flags); > + OFFSET(TASK_MM, task_struct, mm); > + OFFSET(TASK_PID, task_struct, pid); > + DEFINE(TASK_STRUCT_SIZE, sizeof(struct task_struct)); > + BLANK(); > +} > + > +void output_thread_info_defines(void) > +{ > + COMMENT("LoongArch thread_info offsets."); > + OFFSET(TI_TASK, thread_info, task); > + OFFSET(TI_FLAGS, thread_info, flags); > + OFFSET(TI_TP_VALUE, thread_info, tp_value); > + OFFSET(TI_CPU, thread_info, cpu); > + OFFSET(TI_PRE_COUNT, thread_info, preempt_count); > + OFFSET(TI_REGS, thread_info, regs); > + DEFINE(_THREAD_SIZE, THREAD_SIZE); > + DEFINE(_THREAD_MASK, THREAD_MASK); > + DEFINE(_IRQ_STACK_SIZE, IRQ_STACK_SIZE); > + DEFINE(_IRQ_STACK_START, IRQ_STACK_START); > + BLANK(); > +} > + > +void output_thread_defines(void) > +{ > + COMMENT("LoongArch specific thread_struct offsets."); > + OFFSET(THREAD_REG01, task_struct, thread.reg01); > + OFFSET(THREAD_REG03, task_struct, thread.reg03); > + OFFSET(THREAD_REG22, task_struct, thread.reg22); > + OFFSET(THREAD_REG23, task_struct, thread.reg23); > + OFFSET(THREAD_REG24, task_struct, thread.reg24); > + OFFSET(THREAD_REG25, task_struct, thread.reg25); > + OFFSET(THREAD_REG26, task_struct, thread.reg26); > + OFFSET(THREAD_REG27, task_struct, thread.reg27); > + OFFSET(THREAD_REG28, task_struct, thread.reg28); > + OFFSET(THREAD_REG29, task_struct, thread.reg29); > + OFFSET(THREAD_REG30, task_struct, thread.reg30); > + OFFSET(THREAD_REG31, task_struct, thread.reg31); > + OFFSET(THREAD_CSRCRMD, task_struct, > + thread.csr_crmd); > + OFFSET(THREAD_CSRPRMD, task_struct, > + thread.csr_prmd); > + OFFSET(THREAD_CSREUEN, task_struct, > + thread.csr_euen); > + OFFSET(THREAD_CSRECFG, task_struct, > + thread.csr_ecfg); > + > + OFFSET(THREAD_SCR0, task_struct, thread.scr0); > + OFFSET(THREAD_SCR1, task_struct, thread.scr1); > + OFFSET(THREAD_SCR2, task_struct, thread.scr2); > + OFFSET(THREAD_SCR3, task_struct, thread.scr3); > + > + OFFSET(THREAD_EFLAGS, task_struct, thread.eflags); > + > + OFFSET(THREAD_FPU, task_struct, thread.fpu); > + > + OFFSET(THREAD_BVADDR, task_struct, \ > + thread.csr_badvaddr); > + OFFSET(THREAD_ECODE, task_struct, \ > + thread.error_code); > + OFFSET(THREAD_TRAPNO, task_struct, thread.trap_nr); > + BLANK(); > +} > + > +void output_thread_fpu_defines(void) > +{ > + OFFSET(THREAD_FPR0, loongarch_fpu, fpr[0]); > + OFFSET(THREAD_FPR1, loongarch_fpu, fpr[1]); > + OFFSET(THREAD_FPR2, loongarch_fpu, fpr[2]); > + OFFSET(THREAD_FPR3, loongarch_fpu, fpr[3]); > + OFFSET(THREAD_FPR4, loongarch_fpu, fpr[4]); > + OFFSET(THREAD_FPR5, loongarch_fpu, fpr[5]); > + OFFSET(THREAD_FPR6, loongarch_fpu, fpr[6]); > + OFFSET(THREAD_FPR7, loongarch_fpu, fpr[7]); > + OFFSET(THREAD_FPR8, loongarch_fpu, fpr[8]); > + OFFSET(THREAD_FPR9, loongarch_fpu, fpr[9]); > + OFFSET(THREAD_FPR10, loongarch_fpu, fpr[10]); > + OFFSET(THREAD_FPR11, loongarch_fpu, fpr[11]); > + OFFSET(THREAD_FPR12, loongarch_fpu, fpr[12]); > + OFFSET(THREAD_FPR13, loongarch_fpu, fpr[13]); > + OFFSET(THREAD_FPR14, loongarch_fpu, fpr[14]); > + OFFSET(THREAD_FPR15, loongarch_fpu, fpr[15]); > + OFFSET(THREAD_FPR16, loongarch_fpu, fpr[16]); > + OFFSET(THREAD_FPR17, loongarch_fpu, fpr[17]); > + OFFSET(THREAD_FPR18, loongarch_fpu, fpr[18]); > + OFFSET(THREAD_FPR19, loongarch_fpu, fpr[19]); > + OFFSET(THREAD_FPR20, loongarch_fpu, fpr[20]); > + OFFSET(THREAD_FPR21, loongarch_fpu, fpr[21]); > + OFFSET(THREAD_FPR22, loongarch_fpu, fpr[22]); > + OFFSET(THREAD_FPR23, loongarch_fpu, fpr[23]); > + OFFSET(THREAD_FPR24, loongarch_fpu, fpr[24]); > + OFFSET(THREAD_FPR25, loongarch_fpu, fpr[25]); > + OFFSET(THREAD_FPR26, loongarch_fpu, fpr[26]); > + OFFSET(THREAD_FPR27, loongarch_fpu, fpr[27]); > + OFFSET(THREAD_FPR28, loongarch_fpu, fpr[28]); > + OFFSET(THREAD_FPR29, loongarch_fpu, fpr[29]); > + OFFSET(THREAD_FPR30, loongarch_fpu, fpr[30]); > + OFFSET(THREAD_FPR31, loongarch_fpu, fpr[31]); > + > + OFFSET(THREAD_FCSR, loongarch_fpu, fcsr); > + OFFSET(THREAD_FCC, loongarch_fpu, fcc); > + OFFSET(THREAD_VCSR, loongarch_fpu, vcsr); > + BLANK(); > +} > + > +void output_mm_defines(void) > +{ > + COMMENT("Size of struct page"); > + DEFINE(STRUCT_PAGE_SIZE, sizeof(struct page)); > + BLANK(); > + COMMENT("Linux mm_struct offsets."); > + OFFSET(MM_USERS, mm_struct, mm_users); > + OFFSET(MM_PGD, mm_struct, pgd); > + OFFSET(MM_CONTEXT, mm_struct, context); > + BLANK(); > + DEFINE(_PGD_T_SIZE, sizeof(pgd_t)); > + DEFINE(_PMD_T_SIZE, sizeof(pmd_t)); > + DEFINE(_PTE_T_SIZE, sizeof(pte_t)); > + BLANK(); > + DEFINE(_PGD_T_LOG2, PGD_T_LOG2); > +#ifndef __PAGETABLE_PMD_FOLDED > + DEFINE(_PMD_T_LOG2, PMD_T_LOG2); > +#endif > + DEFINE(_PTE_T_LOG2, PTE_T_LOG2); > + BLANK(); > + DEFINE(_PGD_ORDER, PGD_ORDER); > +#ifndef __PAGETABLE_PMD_FOLDED > + DEFINE(_PMD_ORDER, PMD_ORDER); > +#endif > + DEFINE(_PTE_ORDER, PTE_ORDER); > + BLANK(); > + DEFINE(_PMD_SHIFT, PMD_SHIFT); > + DEFINE(_PGDIR_SHIFT, PGDIR_SHIFT); > + BLANK(); > + DEFINE(_PTRS_PER_PGD, PTRS_PER_PGD); > + DEFINE(_PTRS_PER_PMD, PTRS_PER_PMD); > + DEFINE(_PTRS_PER_PTE, PTRS_PER_PTE); > + BLANK(); > + DEFINE(_PAGE_SHIFT, PAGE_SHIFT); > + DEFINE(_PAGE_SIZE, PAGE_SIZE); > + BLANK(); > +} > + > +void output_sc_defines(void) > +{ > + COMMENT("Linux sigcontext offsets."); > + OFFSET(SC_REGS, sigcontext, sc_regs); > + OFFSET(SC_PC, sigcontext, sc_pc); > + BLANK(); > +} > + > +void output_signal_defines(void) > +{ > + COMMENT("Linux signal numbers."); > + DEFINE(_SIGHUP, SIGHUP); > + DEFINE(_SIGINT, SIGINT); > + DEFINE(_SIGQUIT, SIGQUIT); > + DEFINE(_SIGILL, SIGILL); > + DEFINE(_SIGTRAP, SIGTRAP); > + DEFINE(_SIGIOT, SIGIOT); > + DEFINE(_SIGABRT, SIGABRT); > + DEFINE(_SIGFPE, SIGFPE); > + DEFINE(_SIGKILL, SIGKILL); > + DEFINE(_SIGBUS, SIGBUS); > + DEFINE(_SIGSEGV, SIGSEGV); > + DEFINE(_SIGSYS, SIGSYS); > + DEFINE(_SIGPIPE, SIGPIPE); > + DEFINE(_SIGALRM, SIGALRM); > + DEFINE(_SIGTERM, SIGTERM); > + DEFINE(_SIGUSR1, SIGUSR1); > + DEFINE(_SIGUSR2, SIGUSR2); > + DEFINE(_SIGCHLD, SIGCHLD); > + DEFINE(_SIGPWR, SIGPWR); > + DEFINE(_SIGWINCH, SIGWINCH); > + DEFINE(_SIGURG, SIGURG); > + DEFINE(_SIGIO, SIGIO); > + DEFINE(_SIGSTOP, SIGSTOP); > + DEFINE(_SIGTSTP, SIGTSTP); > + DEFINE(_SIGCONT, SIGCONT); > + DEFINE(_SIGTTIN, SIGTTIN); > + DEFINE(_SIGTTOU, SIGTTOU); > + DEFINE(_SIGVTALRM, SIGVTALRM); > + DEFINE(_SIGPROF, SIGPROF); > + DEFINE(_SIGXCPU, SIGXCPU); > + DEFINE(_SIGXFSZ, SIGXFSZ); > + BLANK(); > +} > diff --git a/arch/loongarch/kernel/io.c b/arch/loongarch/kernel/io.c > new file mode 100644 > index 000000000000..cb85bda5a6ad > --- /dev/null > +++ b/arch/loongarch/kernel/io.c > @@ -0,0 +1,94 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#include <linux/export.h> > +#include <linux/types.h> > +#include <linux/io.h> > + > +/* > + * Copy data from IO memory space to "real" memory space. > + */ > +void __memcpy_fromio(void *to, const volatile void __iomem *from, > size_t count) > +{ > + while (count && !IS_ALIGNED((unsigned long)from, 8)) { > + *(u8 *)to = __raw_readb(from); > + from++; > + to++; > + count--; > + } > + > + while (count >= 8) { > + *(u64 *)to = __raw_readq(from); > + from += 8; > + to += 8; > + count -= 8; > + } > + > + while (count) { > + *(u8 *)to = __raw_readb(from); > + from++; > + to++; > + count--; > + } > +} > +EXPORT_SYMBOL(__memcpy_fromio); > + > +/* > + * Copy data from "real" memory space to IO memory space. > + */ > +void __memcpy_toio(volatile void __iomem *to, const void *from, size_t > count) > +{ > + while (count && !IS_ALIGNED((unsigned long)to, 8)) { > + __raw_writeb(*(u8 *)from, to); > + from++; > + to++; > + count--; > + } > + > + while (count >= 8) { > + __raw_writeq(*(u64 *)from, to); > + from += 8; > + to += 8; > + count -= 8; > + } > + > + while (count) { > + __raw_writeb(*(u8 *)from, to); > + from++; > + to++; > + count--; > + } > +} > +EXPORT_SYMBOL(__memcpy_toio); > + > +/* > + * "memset" on IO memory space. > + */ > +void __memset_io(volatile void __iomem *dst, int c, size_t count) > +{ > + u64 qc = (u8)c; > + > + qc |= qc << 8; > + qc |= qc << 16; > + qc |= qc << 32; > + > + while (count && !IS_ALIGNED((unsigned long)dst, 8)) { > + __raw_writeb(c, dst); > + dst++; > + count--; > + } > + > + while (count >= 8) { > + __raw_writeq(qc, dst); > + dst += 8; > + count -= 8; > + } > + > + while (count) { > + __raw_writeb(c, dst); > + dst++; > + count--; > + } > +} > +EXPORT_SYMBOL(__memset_io); > diff --git a/arch/loongarch/kernel/proc.c b/arch/loongarch/kernel/proc.c > new file mode 100644 > index 000000000000..d25592a29196 > --- /dev/null > +++ b/arch/loongarch/kernel/proc.c > @@ -0,0 +1,122 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited > + */ > +#include <linux/delay.h> > +#include <linux/kernel.h> > +#include <linux/sched.h> > +#include <linux/seq_file.h> > +#include <asm/bootinfo.h> > +#include <asm/cpu.h> > +#include <asm/cpu-features.h> > +#include <asm/idle.h> > +#include <asm/processor.h> > +#include <asm/time.h> > + > +/* > + * No lock; only written during early bootup by CPU 0. > + */ > +static RAW_NOTIFIER_HEAD(proc_cpuinfo_chain); > + > +int __ref register_proc_cpuinfo_notifier(struct notifier_block *nb) > +{ > + return raw_notifier_chain_register(&proc_cpuinfo_chain, nb); > +} > + > +int proc_cpuinfo_notifier_call_chain(unsigned long val, void *v) > +{ > + return raw_notifier_call_chain(&proc_cpuinfo_chain, val, v); > +} > + > +static int show_cpuinfo(struct seq_file *m, void *v) > +{ > + unsigned long n = (unsigned long) v - 1; > + unsigned int version = cpu_data[n].processor_id & 0xff; > + unsigned int fp_version = cpu_data[n].fpu_vers; > + struct proc_cpuinfo_notifier_args proc_cpuinfo_notifier_args; > + > + /* > + * For the first processor also print the system type > + */ > + if (n == 0) > + seq_printf(m, "system type\t\t: %s\n\n", get_system_type()); > + > + seq_printf(m, "processor\t\t: %ld\n", n); > + seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package); > + seq_printf(m, "core\t\t\t: %d\n", cpu_data[n].core); > + seq_printf(m, "CPU Family\t\t: %s\n", __cpu_family[n]); > + seq_printf(m, "Model Name\t\t: %s\n", __cpu_full_name[n]); > + seq_printf(m, "CPU Revision\t\t: 0x%02x\n", version); > + seq_printf(m, "FPU Revision\t\t: 0x%02x\n", fp_version); > + seq_printf(m, "CPU MHz\t\t\t: %llu.%02llu\n", > + cpu_clock_freq / 1000000, (cpu_clock_freq / 10000) % 100); > + seq_printf(m, "BogoMIPS\t\t: %llu.%02llu\n", > + (lpj_fine * cpu_clock_freq / const_clock_freq) / (500000/HZ), > + ((lpj_fine * cpu_clock_freq / const_clock_freq) / (5000/HZ)) % > 100); > + seq_printf(m, "TLB Entries\t\t: %d\n", cpu_data[n].tlbsize); > + seq_printf(m, "Address Sizes\t\t: %d bits physical, %d bits > virtual\n", > + cpu_pabits + 1, cpu_vabits + 1); > + > + seq_printf(m, "ISA\t\t\t:"); > + if (cpu_has_loongarch32) > + seq_printf(m, " loongarch32"); > + if (cpu_has_loongarch64) > + seq_printf(m, " loongarch64"); > + seq_printf(m, "\n"); > + > + seq_printf(m, "Features\t\t:"); > + if (cpu_has_cpucfg) seq_printf(m, " cpucfg"); > + if (cpu_has_lam) seq_printf(m, " lam"); > + if (cpu_has_ual) seq_printf(m, " ual"); > + if (cpu_has_fpu) seq_printf(m, " fpu"); > + if (cpu_has_lsx) seq_printf(m, " lsx"); > + if (cpu_has_lasx) seq_printf(m, " lasx"); > + if (cpu_has_complex) seq_printf(m, " complex"); > + if (cpu_has_crypto) seq_printf(m, " crypto"); > + if (cpu_has_lvz) seq_printf(m, " lvz"); > + if (cpu_has_lbt_x86) seq_printf(m, " lbt_x86"); > + if (cpu_has_lbt_arm) seq_printf(m, " lbt_arm"); > + if (cpu_has_lbt_mips) seq_printf(m, " lbt_mips"); > + seq_printf(m, "\n"); > + > + seq_printf(m, "Hardware Watchpoint\t: %s", > + cpu_has_watch ? "yes, " : "no\n"); > + if (cpu_has_watch) { > + seq_printf(m, "iwatch count: %d, dwatch count: %d\n", > + cpu_data[n].watch_ireg_count, cpu_data[n].watch_dreg_count); > + } > + > + proc_cpuinfo_notifier_args.m = m; > + proc_cpuinfo_notifier_args.n = n; > + > + raw_notifier_call_chain(&proc_cpuinfo_chain, 0, > + &proc_cpuinfo_notifier_args); > + > + seq_printf(m, "\n"); > + > + return 0; > +} > + > +static void *c_start(struct seq_file *m, loff_t *pos) > +{ > + unsigned long i = *pos; > + > + return i < NR_CPUS ? (void *)(i + 1) : NULL; > +} > + > +static void *c_next(struct seq_file *m, void *v, loff_t *pos) > +{ > + ++*pos; > + return c_start(m, pos); > +} > + > +static void c_stop(struct seq_file *m, void *v) > +{ > +} > + > +const struct seq_operations cpuinfo_op = { > + .start = c_start, > + .next = c_next, > + .stop = c_stop, > + .show = show_cpuinfo, > +}; > -- > 2.27.0 -- - Jiaxun