Re: [PATCH V9 17/24] LoongArch: Add some library functions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sat, Apr 30, 2022 at 5:23 PM Huacai Chen <chenhuacai@xxxxxxxxxxx> wrote:
>
> This patch adds some library functions for LoongArch, including: delay,
> memset, memcpy, memmove, copy_user, strncpy_user, strnlen_user and tlb
> dump functions.
>
> Signed-off-by: Huacai Chen <chenhuacai@xxxxxxxxxxx>
> ---
>  arch/loongarch/include/asm/delay.h  |  26 +++++++
>  arch/loongarch/include/asm/string.h |  17 +++++
>  arch/loongarch/lib/clear_user.S     |  43 +++++++++++
>  arch/loongarch/lib/copy_user.S      |  47 ++++++++++++
>  arch/loongarch/lib/delay.c          |  43 +++++++++++
>  arch/loongarch/lib/dump_tlb.c       | 111 ++++++++++++++++++++++++++++
>  arch/loongarch/lib/memcpy.S         |  32 ++++++++
>  arch/loongarch/lib/memmove.S        |  45 +++++++++++
>  arch/loongarch/lib/memset.S         |  30 ++++++++
>  9 files changed, 394 insertions(+)
>  create mode 100644 arch/loongarch/include/asm/delay.h
>  create mode 100644 arch/loongarch/include/asm/string.h
>  create mode 100644 arch/loongarch/lib/clear_user.S
>  create mode 100644 arch/loongarch/lib/copy_user.S
>  create mode 100644 arch/loongarch/lib/delay.c
>  create mode 100644 arch/loongarch/lib/dump_tlb.c
>  create mode 100644 arch/loongarch/lib/memcpy.S
>  create mode 100644 arch/loongarch/lib/memmove.S
>  create mode 100644 arch/loongarch/lib/memset.S
>
> diff --git a/arch/loongarch/include/asm/delay.h b/arch/loongarch/include/asm/delay.h
> new file mode 100644
> index 000000000000..016b3aca65cb
> --- /dev/null
> +++ b/arch/loongarch/include/asm/delay.h
> @@ -0,0 +1,26 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#ifndef _ASM_DELAY_H
> +#define _ASM_DELAY_H
> +
> +#include <linux/param.h>
> +
> +extern void __delay(unsigned long loops);
> +extern void __ndelay(unsigned long ns);
> +extern void __udelay(unsigned long us);
> +
> +#define ndelay(ns) __ndelay(ns)
> +#define udelay(us) __udelay(us)
> +
> +/* make sure "usecs *= ..." in udelay do not overflow. */
> +#if HZ >= 1000
> +#define MAX_UDELAY_MS  1
> +#elif HZ <= 200
> +#define MAX_UDELAY_MS  5
> +#else
> +#define MAX_UDELAY_MS  (1000 / HZ)
> +#endif
> +
> +#endif /* _ASM_DELAY_H */
> diff --git a/arch/loongarch/include/asm/string.h b/arch/loongarch/include/asm/string.h
> new file mode 100644
> index 000000000000..7b29cc9c70aa
> --- /dev/null
> +++ b/arch/loongarch/include/asm/string.h
> @@ -0,0 +1,17 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#ifndef _ASM_STRING_H
> +#define _ASM_STRING_H
> +
> +#define __HAVE_ARCH_MEMSET
> +extern void *memset(void *__s, int __c, size_t __count);
> +
> +#define __HAVE_ARCH_MEMCPY
> +extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
> +
> +#define __HAVE_ARCH_MEMMOVE
> +extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
> +
> +#endif /* _ASM_STRING_H */
> diff --git a/arch/loongarch/lib/clear_user.S b/arch/loongarch/lib/clear_user.S
> new file mode 100644
> index 000000000000..b8168d22ac80
> --- /dev/null
> +++ b/arch/loongarch/lib/clear_user.S
> @@ -0,0 +1,43 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asm.h>
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +.macro fixup_ex from, to, offset, fix
> +.if \fix
> +       .section .fixup, "ax"
> +\to:   addi.d  v0, a1, \offset
> +       jr      ra
> +       .previous
> +.endif
> +       .section __ex_table, "a"
> +       PTR     \from\()b, \to\()b
> +       .previous
> +.endm
> +
> +/*
> + * unsigned long __clear_user(void *addr, size_t size)
> + *
> + * a0: addr
> + * a1: size
> + */
> +SYM_FUNC_START(__clear_user)
> +       beqz    a1, 2f
> +
> +1:     st.b    zero, a0, 0
> +       addi.d  a0, a0, 1
> +       addi.d  a1, a1, -1
> +       bgt     a1, zero, 1b
> +
> +2:     move    v0, a1
> +       jr      ra
> +
> +       fixup_ex 1, 3, 0, 1
> +SYM_FUNC_END(__clear_user)
> +
> +EXPORT_SYMBOL(__clear_user)
> diff --git a/arch/loongarch/lib/copy_user.S b/arch/loongarch/lib/copy_user.S
> new file mode 100644
> index 000000000000..43ed26304954
> --- /dev/null
> +++ b/arch/loongarch/lib/copy_user.S
> @@ -0,0 +1,47 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asm.h>
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +.macro fixup_ex from, to, offset, fix
> +.if \fix
> +       .section .fixup, "ax"
> +\to:   addi.d  v0, a2, \offset
> +       jr      ra
> +       .previous
> +.endif
> +       .section __ex_table, "a"
> +       PTR     \from\()b, \to\()b
> +       .previous
> +.endm
> +
> +/*
> + * unsigned long __copy_user(void *to, const void *from, size_t n)
> + *
> + * a0: to
> + * a1: from
> + * a2: n
> + */
> +SYM_FUNC_START(__copy_user)
> +       beqz    a2, 3f
> +
> +1:     ld.b    t0, a1, 0
> +2:     st.b    t0, a0, 0
> +       addi.d  a0, a0, 1
> +       addi.d  a1, a1, 1
> +       addi.d  a2, a2, -1
> +       bgt     a2, zero, 1b
> +
> +3:     move    v0, a2
> +       jr      ra
> +
> +       fixup_ex 1, 4, 0, 1
> +       fixup_ex 2, 4, 0, 0
> +SYM_FUNC_END(__copy_user)
> +
> +EXPORT_SYMBOL(__copy_user)
> diff --git a/arch/loongarch/lib/delay.c b/arch/loongarch/lib/delay.c
> new file mode 100644
> index 000000000000..5d856694fcfe
> --- /dev/null
> +++ b/arch/loongarch/lib/delay.c
> @@ -0,0 +1,43 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#include <linux/delay.h>
> +#include <linux/export.h>
> +#include <linux/smp.h>
> +#include <linux/timex.h>
> +
> +#include <asm/compiler.h>
> +#include <asm/processor.h>
> +
> +void __delay(unsigned long cycles)
> +{
> +       u64 t0 = get_cycles();
> +
> +       while ((unsigned long)(get_cycles() - t0) < cycles)
> +               cpu_relax();
> +}
> +EXPORT_SYMBOL(__delay);
> +
> +/*
> + * Division by multiplication: you don't have to worry about
> + * loss of precision.
> + *
> + * Use only for very small delays ( < 1 msec). Should probably use a
> + * lookup table, really, as the multiplications take much too long with
> + * short delays.  This is a "reasonable" implementation, though (and the
> + * first constant multiplications gets optimized away if the delay is
> + * a constant)
> + */
> +
> +void __udelay(unsigned long us)
> +{
> +       __delay((us * 0x000010c7ull * HZ * lpj_fine) >> 32);
> +}
> +EXPORT_SYMBOL(__udelay);
> +
> +void __ndelay(unsigned long ns)
> +{
> +       __delay((ns * 0x00000005ull * HZ * lpj_fine) >> 32);
> +}
> +EXPORT_SYMBOL(__ndelay);
> diff --git a/arch/loongarch/lib/dump_tlb.c b/arch/loongarch/lib/dump_tlb.c
> new file mode 100644
> index 000000000000..cda2c6bc7f09
> --- /dev/null
> +++ b/arch/loongarch/lib/dump_tlb.c
> @@ -0,0 +1,111 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + *
> + * Derived from MIPS:
> + * Copyright (C) 1994, 1995 by Waldorf Electronics, written by Ralf Baechle.
> + * Copyright (C) 1999 by Silicon Graphics, Inc.
> + */
> +#include <linux/kernel.h>
> +#include <linux/mm.h>
> +
> +#include <asm/loongarch.h>
> +#include <asm/page.h>
> +#include <asm/pgtable.h>
> +#include <asm/tlb.h>
> +
> +void dump_tlb_regs(void)
> +{
> +       const int field = 2 * sizeof(unsigned long);
> +
> +       pr_info("Index    : %0x\n", read_csr_tlbidx());
> +       pr_info("PageSize : %0x\n", read_csr_pagesize());
> +       pr_info("EntryHi  : %0*llx\n", field, read_csr_entryhi());
> +       pr_info("EntryLo0 : %0*llx\n", field, read_csr_entrylo0());
> +       pr_info("EntryLo1 : %0*llx\n", field, read_csr_entrylo1());
> +}
> +
> +static void dump_tlb(int first, int last)
> +{
> +       unsigned long s_entryhi, entryhi, asid;
> +       unsigned long long entrylo0, entrylo1, pa;
> +       unsigned int index;
> +       unsigned int s_index, s_asid;
> +       unsigned int pagesize, c0, c1, i;
> +       unsigned long asidmask = cpu_asid_mask(&current_cpu_data);
> +       int pwidth = 11;
> +       int vwidth = 11;
> +       int asidwidth = DIV_ROUND_UP(ilog2(asidmask) + 1, 4);
> +
> +       s_entryhi = read_csr_entryhi();
> +       s_index = read_csr_tlbidx();
> +       s_asid = read_csr_asid();
> +
> +       for (i = first; i <= last; i++) {
> +               write_csr_index(i);
> +               tlb_read();
> +               pagesize = read_csr_pagesize();
> +               entryhi  = read_csr_entryhi();
> +               entrylo0 = read_csr_entrylo0();
> +               entrylo1 = read_csr_entrylo1();
> +               index = read_csr_tlbidx();
> +               asid = read_csr_asid();
> +
> +               /* EHINV bit marks entire entry as invalid */
> +               if (index & CSR_TLBIDX_EHINV)
> +                       continue;
> +               /*
> +                * ASID takes effect in absence of G (global) bit.
> +                */
> +               if (!((entrylo0 | entrylo1) & ENTRYLO_G) &&
> +                   asid != s_asid)
> +                       continue;
> +
> +               /*
> +                * Only print entries in use
> +                */
> +               pr_info("Index: %2d pgsize=%x ", i, (1 << pagesize));
> +
> +               c0 = (entrylo0 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
> +               c1 = (entrylo1 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
> +
> +               pr_cont("va=%0*lx asid=%0*lx",
> +                       vwidth, (entryhi & ~0x1fffUL), asidwidth, asid & asidmask);
> +
> +               /* NR/NX are in awkward places, so mask them off separately */
> +               pa = entrylo0 & ~(ENTRYLO_NR | ENTRYLO_NX);
> +               pa = pa & PAGE_MASK;
> +               pr_cont("\n\t[");
> +               pr_cont("ri=%d xi=%d ",
> +                       (entrylo0 & ENTRYLO_NR) ? 1 : 0,
> +                       (entrylo0 & ENTRYLO_NX) ? 1 : 0);
> +               pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d plv=%lld] [",
> +                       pwidth, pa, c0,
> +                       (entrylo0 & ENTRYLO_D) ? 1 : 0,
> +                       (entrylo0 & ENTRYLO_V) ? 1 : 0,
> +                       (entrylo0 & ENTRYLO_G) ? 1 : 0,
> +                       (entrylo0 & ENTRYLO_PLV) >> ENTRYLO_PLV_SHIFT);
> +               /* NR/NX are in awkward places, so mask them off separately */
> +               pa = entrylo1 & ~(ENTRYLO_NR | ENTRYLO_NX);
> +               pa = pa & PAGE_MASK;
> +               pr_cont("ri=%d xi=%d ",
> +                       (entrylo1 & ENTRYLO_NR) ? 1 : 0,
> +                       (entrylo1 & ENTRYLO_NX) ? 1 : 0);
> +               pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d plv=%lld]\n",
> +                       pwidth, pa, c1,
> +                       (entrylo1 & ENTRYLO_D) ? 1 : 0,
> +                       (entrylo1 & ENTRYLO_V) ? 1 : 0,
> +                       (entrylo1 & ENTRYLO_G) ? 1 : 0,
> +                       (entrylo1 & ENTRYLO_PLV) >> ENTRYLO_PLV_SHIFT);
> +       }
> +       pr_info("\n");
> +
> +       write_csr_entryhi(s_entryhi);
> +       write_csr_tlbidx(s_index);
> +       write_csr_asid(s_asid);
> +}
> +
> +void dump_tlb_all(void)
> +{
> +       dump_tlb(0, current_cpu_data.tlbsize - 1);
> +}
> diff --git a/arch/loongarch/lib/memcpy.S b/arch/loongarch/lib/memcpy.S
> new file mode 100644
> index 000000000000..d53f1148d26b
> --- /dev/null
> +++ b/arch/loongarch/lib/memcpy.S
> @@ -0,0 +1,32 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +/*
> + * void *memcpy(void *dst, const void *src, size_t n)
> + *
> + * a0: dst
> + * a1: src
> + * a2: n
> + */
> +SYM_FUNC_START(memcpy)
> +       move    a3, a0
> +       beqz    a2, 2f
> +
> +1:     ld.b    t0, a1, 0
> +       st.b    t0, a0, 0
> +       addi.d  a0, a0, 1
> +       addi.d  a1, a1, 1
> +       addi.d  a2, a2, -1
> +       bgt     a2, zero, 1b
> +
> +2:     move    v0, a3
> +       jr      ra
> +SYM_FUNC_END(memcpy)
> +
> +EXPORT_SYMBOL(memcpy)
> diff --git a/arch/loongarch/lib/memmove.S b/arch/loongarch/lib/memmove.S
> new file mode 100644
> index 000000000000..18907d83a83b
> --- /dev/null
> +++ b/arch/loongarch/lib/memmove.S
> @@ -0,0 +1,45 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +/*
> + * void *rmemcpy(void *dst, const void *src, size_t n)
> + *
> + * a0: dst
> + * a1: src
> + * a2: n
> + */
> +SYM_FUNC_START(rmemcpy)
> +       move    a3, a0
> +       beqz    a2, 2f
> +
> +       add.d   a0, a0, a2
> +       add.d   a1, a1, a2
> +
> +1:     ld.b    t0, a1, -1
> +       st.b    t0, a0, -1
> +       addi.d  a0, a0, -1
> +       addi.d  a1, a1, -1
> +       addi.d  a2, a2, -1
> +       bgt     a2, zero, 1b
> +
> +2:     move    v0, a3
> +       jr      ra
> +SYM_FUNC_END(rmemcpy)
Why not directly use:

lib/string.c:
#ifndef __HAVE_ARCH_MEMCPY
/**
 * memcpy - Copy one area of memory to another
 * @dest: Where to copy to
 * @src: Where to copy from
 * @count: The size of the area.
 *
 * You should not use this function to access IO space, use memcpy_toio()
 * or memcpy_fromio() instead.
 */
void *memcpy(void *dest, const void *src, size_t count)
{
        char *tmp = dest;
        const char *s = src;

        while (count--)
                *tmp++ = *s++;
        return dest;
}
EXPORT_SYMBOL(memcpy);
#endif

Do you want to try a C's string implementation?
https://lore.kernel.org/linux-csky/202204051450.UN2k1raL-lkp@xxxxxxxxx/T/#Z2e.:..:20220404142354.2792428-1-guoren::40kernel.org:1arch:csky:lib:string.c

> +
> +SYM_FUNC_START(memmove)
> +       blt     a0, a1, 1f      /* dst < src, memcpy */
> +       blt     a1, a0, 2f      /* src < dst, rmemcpy */
> +       jr      ra              /* dst == src, return */
> +
> +1:     b       memcpy
> +
> +2:     b       rmemcpy
> +SYM_FUNC_END(memmove)
> +
> +EXPORT_SYMBOL(memmove)
> diff --git a/arch/loongarch/lib/memset.S b/arch/loongarch/lib/memset.S
> new file mode 100644
> index 000000000000..3fc3e7da5263
> --- /dev/null
> +++ b/arch/loongarch/lib/memset.S
> @@ -0,0 +1,30 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +/*
> + * void *memset(void *s, int c, size_t n)
> + *
> + * a0: s
> + * a1: c
> + * a2: n
> + */
> +SYM_FUNC_START(memset)
> +       move    a3, a0
> +       beqz    a2, 2f
> +
> +1:     st.b    a1, a0, 0
> +       addi.d  a0, a0, 1
> +       addi.d  a2, a2, -1
> +       bgt     a2, zero, 1b
> +
> +2:     move    v0, a3
> +       jr      ra
> +SYM_FUNC_END(memset)
> +
> +EXPORT_SYMBOL(memset)
> --
> 2.27.0
>


-- 
Best Regards
 Guo Ren

ML: https://lore.kernel.org/linux-csky/



[Index of Archives]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Video 4 Linux]     [Device Mapper]     [Linux Resources]

  Powered by Linux