Signed-off-by: Guo Ren <ren_guo@xxxxxxxxx> --- arch/csky/abiv1/inc/abi/cacheflush.h | 40 +++++++ arch/csky/abiv1/src/cacheflush.c | 32 ++++++ arch/csky/abiv2/inc/abi/cacheflush.h | 9 ++ arch/csky/include/asm/barrier.h | 14 +++ arch/csky/include/asm/cache.h | 32 ++++++ arch/csky/include/asm/cacheflush.h | 9 ++ arch/csky/include/asm/dma-mapping.h | 17 +++ arch/csky/include/asm/io.h | 23 ++++ arch/csky/include/asm/segment.h | 18 +++ arch/csky/include/asm/tlb.h | 24 ++++ arch/csky/include/asm/tlbflush.h | 22 ++++ arch/csky/include/uapi/asm/cachectl.h | 13 +++ arch/csky/mm/cachev1.c | 146 ++++++++++++++++++++++++ arch/csky/mm/cachev2.c | 90 +++++++++++++++ arch/csky/mm/syscache.c | 50 ++++++++ arch/csky/mm/tlb.c | 207 ++++++++++++++++++++++++++++++++++ 16 files changed, 746 insertions(+) create mode 100644 arch/csky/abiv1/inc/abi/cacheflush.h create mode 100644 arch/csky/abiv1/src/cacheflush.c create mode 100644 arch/csky/abiv2/inc/abi/cacheflush.h create mode 100644 arch/csky/include/asm/barrier.h create mode 100644 arch/csky/include/asm/cache.h create mode 100644 arch/csky/include/asm/cacheflush.h create mode 100644 arch/csky/include/asm/dma-mapping.h create mode 100644 arch/csky/include/asm/io.h create mode 100644 arch/csky/include/asm/segment.h create mode 100644 arch/csky/include/asm/tlb.h create mode 100644 arch/csky/include/asm/tlbflush.h create mode 100644 arch/csky/include/uapi/asm/cachectl.h create mode 100644 arch/csky/mm/cachev1.c create mode 100644 arch/csky/mm/cachev2.c create mode 100644 arch/csky/mm/syscache.c create mode 100644 arch/csky/mm/tlb.c diff --git a/arch/csky/abiv1/inc/abi/cacheflush.h b/arch/csky/abiv1/inc/abi/cacheflush.h new file mode 100644 index 0000000..94c31b4 --- /dev/null +++ b/arch/csky/abiv1/inc/abi/cacheflush.h @@ -0,0 +1,40 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ABI_CSKY_CACHEFLUSH_H +#define __ABI_CSKY_CACHEFLUSH_H + +#include <linux/compiler.h> +#include <asm/string.h> +#include <asm/cache.h> + +#define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1 +extern void flush_dcache_page(struct page *); + +#define flush_cache_mm(mm) cache_wbinv_all() +#define flush_cache_page(vma,page,pfn) cache_wbinv_all() +#define flush_cache_dup_mm(mm) cache_wbinv_all() +#define flush_icache_page(vma, page) icache_inv_all() + +#define flush_cache_range(mm,start,end) cache_wbinv_range(start, end) +#define flush_cache_vmap(start, end) cache_wbinv_range(start, end) +#define flush_cache_vunmap(start, end) cache_wbinv_range(start, end) +#define flush_icache_range(start, end) icache_inv_range(start, end) + +#define copy_from_user_page(vma, page, vaddr, dst, src, len) \ +do{ \ + cache_wbinv_all(); \ + memcpy(dst, src, len); \ + icache_inv_all(); \ +}while(0) + +#define copy_to_user_page(vma, page, vaddr, dst, src, len) \ +do{ \ + cache_wbinv_all(); \ + memcpy(dst, src, len); \ +}while(0) + +#define flush_dcache_mmap_lock(mapping) do{}while(0) +#define flush_dcache_mmap_unlock(mapping) do{}while(0) + +#endif /* __ABI_CSKY_CACHEFLUSH_H */ + diff --git a/arch/csky/abiv1/src/cacheflush.c b/arch/csky/abiv1/src/cacheflush.c new file mode 100644 index 0000000..caa095a --- /dev/null +++ b/arch/csky/abiv1/src/cacheflush.c @@ -0,0 +1,32 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#include <linux/kernel.h> +#include <linux/mm.h> +#include <linux/fs.h> +#include <linux/syscalls.h> +#include <linux/spinlock.h> +#include <asm/page.h> +#include <asm/cache.h> +#include <asm/cacheflush.h> +#include <asm/cachectl.h> +#include <abi/reg_ops.h> + +void flush_dcache_page(struct page *page) +{ + struct address_space *mapping = page_mapping(page); + unsigned long addr; + + if (mapping && !mapping_mapped(mapping)) { + set_bit(PG_arch_1, &(page)->flags); + return; + } + + /* + * We could delay the flush for the !page_mapping case too. But that + * case is for exec env/arg pages and those are %99 certainly going to + * get faulted into the tlb (and thus flushed) anyways. + */ + addr = (unsigned long) page_address(page); + dcache_wbinv_range(addr, addr + PAGE_SIZE); +} + diff --git a/arch/csky/abiv2/inc/abi/cacheflush.h b/arch/csky/abiv2/inc/abi/cacheflush.h new file mode 100644 index 0000000..f5a1b18 --- /dev/null +++ b/arch/csky/abiv2/inc/abi/cacheflush.h @@ -0,0 +1,9 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ABI_CSKY_CACHEFLUSH_H +#define __ABI_CSKY_CACHEFLUSH_H + +#include <asm-generic/cacheflush.h> + +#endif /* __ABI_CSKY_CACHEFLUSH_H */ + diff --git a/arch/csky/include/asm/barrier.h b/arch/csky/include/asm/barrier.h new file mode 100644 index 0000000..beea9a0 --- /dev/null +++ b/arch/csky/include/asm/barrier.h @@ -0,0 +1,14 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_BARRIER_H +#define __ASM_CSKY_BARRIER_H + +#ifndef __ASSEMBLY__ + +#define nop() asm volatile ("nop") +#define mb() asm volatile ("sync":::"memory") + +#include <asm-generic/barrier.h> + +#endif /* __ASSEMBLY__ */ +#endif /* __ASM_CSKY_BARRIER_H */ diff --git a/arch/csky/include/asm/cache.h b/arch/csky/include/asm/cache.h new file mode 100644 index 0000000..9742347 --- /dev/null +++ b/arch/csky/include/asm/cache.h @@ -0,0 +1,32 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_CACHE_H +#define __ASM_CSKY_CACHE_H + +/* bytes per L1 cache line */ +#define L1_CACHE_SHIFT CONFIG_L1_CACHE_SHIFT + +#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) + +#define ARCH_DMA_MINALIGN L1_CACHE_BYTES + +#ifndef __ASSEMBLY__ + +void dcache_wb_line(unsigned long start); + +void icache_inv_range(unsigned long start, unsigned long end); +void icache_inv_all(void); + +void dcache_wb_range(unsigned long start, unsigned long end); +void dcache_wbinv_range(unsigned long start, unsigned long end); +void dcache_inv_range(unsigned long start, unsigned long end); +void dcache_wbinv_all(void); + +void cache_wbinv_range(unsigned long start, unsigned long end); +void cache_wbinv_all(void); + +void dma_wbinv_range(unsigned long start, unsigned long end); +void dma_wb_range(unsigned long start, unsigned long end); + +#endif +#endif /* __ASM_CSKY_CACHE_H */ diff --git a/arch/csky/include/asm/cacheflush.h b/arch/csky/include/asm/cacheflush.h new file mode 100644 index 0000000..6a4233b --- /dev/null +++ b/arch/csky/include/asm/cacheflush.h @@ -0,0 +1,9 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_CACHEFLUSH_H +#define __ASM_CSKY_CACHEFLUSH_H + +#include <abi/cacheflush.h> + +#endif /* __ASM_CSKY_CACHEFLUSH_H */ + diff --git a/arch/csky/include/asm/dma-mapping.h b/arch/csky/include/asm/dma-mapping.h new file mode 100644 index 0000000..f901e2d --- /dev/null +++ b/arch/csky/include/asm/dma-mapping.h @@ -0,0 +1,17 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_DMA_MAPPING_H +#define __ASM_DMA_MAPPING_H + +extern struct dma_map_ops csky_dma_map_ops; + +#ifdef COMPAT_KERNEL_4_9 +static inline struct dma_map_ops *get_dma_ops(struct device *dev) +#else +static inline struct dma_map_ops *get_arch_dma_ops(struct bus_type *bus) +#endif +{ + return &csky_dma_map_ops; +} + +#endif /* __ASM_DMA_MAPPING_H */ diff --git a/arch/csky/include/asm/io.h b/arch/csky/include/asm/io.h new file mode 100644 index 0000000..fcb2142 --- /dev/null +++ b/arch/csky/include/asm/io.h @@ -0,0 +1,23 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_IO_H +#define __ASM_CSKY_IO_H + +#include <abi/pgtable-bits.h> +#include <linux/types.h> +#include <linux/version.h> + +extern void __iomem *ioremap(phys_addr_t offset, size_t size); + +extern void iounmap(void *addr); + +extern int remap_area_pages(unsigned long address, phys_addr_t phys_addr, + size_t size, unsigned long flags); + +#define ioremap_nocache(phy, sz) ioremap(phy, sz) +#define ioremap_wc ioremap_nocache +#define ioremap_wt ioremap_nocache + +#include <asm-generic/io.h> + +#endif /* __ASM_CSKY_IO_H */ diff --git a/arch/csky/include/asm/segment.h b/arch/csky/include/asm/segment.h new file mode 100644 index 0000000..cf1440c --- /dev/null +++ b/arch/csky/include/asm/segment.h @@ -0,0 +1,18 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_SEGMENT_H +#define __ASM_CSKY_SEGMENT_H + +typedef struct { + unsigned long seg; +} mm_segment_t; + +#define KERNEL_DS ((mm_segment_t) { 0xFFFFFFFF }) +#define get_ds() KERNEL_DS + +#define USER_DS ((mm_segment_t) { 0x80000000UL }) +#define get_fs() (current_thread_info()->addr_limit) +#define set_fs(x) (current_thread_info()->addr_limit = (x)) +#define segment_eq(a,b) ((a).seg == (b).seg) + +#endif /* __ASM_CSKY_SEGMENT_H */ diff --git a/arch/csky/include/asm/tlb.h b/arch/csky/include/asm/tlb.h new file mode 100644 index 0000000..428e2b2 --- /dev/null +++ b/arch/csky/include/asm/tlb.h @@ -0,0 +1,24 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_TLB_H +#define __ASM_CSKY_TLB_H + +#include <asm/cacheflush.h> + +#define tlb_start_vma(tlb, vma) \ + do { \ + if (!tlb->fullmm) \ + cache_wbinv_range(vma->vm_start, vma->vm_end); \ + } while (0) + +#define tlb_end_vma(tlb, vma) \ + do { \ + if (!tlb->fullmm) \ + cache_wbinv_range(vma->vm_start, vma->vm_end); \ + } while (0) + +#define tlb_flush(tlb) flush_tlb_mm((tlb)->mm) + +#include <asm-generic/tlb.h> + +#endif /* __ASM_CSKY_TLB_H */ diff --git a/arch/csky/include/asm/tlbflush.h b/arch/csky/include/asm/tlbflush.h new file mode 100644 index 0000000..d44cb07 --- /dev/null +++ b/arch/csky/include/asm/tlbflush.h @@ -0,0 +1,22 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_TLBFLUSH_H +#define __ASM_TLBFLUSH_H +/* + * TLB flushing: + * + * - flush_tlb_all() flushes all processes TLB entries + * - flush_tlb_mm(mm) flushes the specified mm context TLB entries + * - flush_tlb_page(vma, vmaddr) flushes one page + * - flush_tlb_range(vma, start, end) flushes a range of pages + * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages + */ +extern void flush_tlb_all(void); +extern void flush_tlb_mm(struct mm_struct *mm); +extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); +extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); +extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); + +extern void flush_tlb_one(unsigned long vaddr); + +#endif diff --git a/arch/csky/include/uapi/asm/cachectl.h b/arch/csky/include/uapi/asm/cachectl.h new file mode 100644 index 0000000..38a1a28 --- /dev/null +++ b/arch/csky/include/uapi/asm/cachectl.h @@ -0,0 +1,13 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#ifndef __ASM_CSKY_CACHECTL_H +#define __ASM_CSKY_CACHECTL_H + +/* + * See "man cacheflush" + */ +#define ICACHE (1<<0) +#define DCACHE (1<<1) +#define BCACHE (ICACHE|DCACHE) + +#endif /* __ASM_CSKY_CACHECTL_H */ diff --git a/arch/csky/mm/cachev1.c b/arch/csky/mm/cachev1.c new file mode 100644 index 0000000..797ed72 --- /dev/null +++ b/arch/csky/mm/cachev1.c @@ -0,0 +1,146 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#include <linux/spinlock.h> +#include <asm/cache.h> + +/* for L1-cache */ +#define INS_CACHE (1 << 0) +#define DATA_CACHE (1 << 1) +#define CACHE_INV (1 << 4) +#define CACHE_CLR (1 << 5) +#define CACHE_OMS (1 << 6) +#define CACHE_ITS (1 << 7) +#define CACHE_LICF (1 << 31) + +/* for L2-cache */ +#define CR22_LEVEL_SHIFT (1) +#define CR22_SET_SHIFT (7) +#define CR22_WAY_SHIFT (30) +#define CR22_WAY_SHIFT_L2 (29) + +static DEFINE_SPINLOCK(cache_lock); + +#define cache_op_line(i, value) \ + asm volatile( \ + "mtcr %0, cr22\n" \ + "mtcr %1, cr17\n" \ + ::"r"(i), "r"(value)) + +#define cache_op_line_atomic(i, value) \ + asm volatile( \ + "idly4\n" \ + "mtcr %0, cr22\n" \ + "mtcr %1, cr17\n" \ + "sync\n" \ + ::"r"(i), "r"(value)) + +#define CCR2_L2E (1 << 3) +static void cache_op_all(unsigned int value, unsigned int l2) +{ + asm volatile( + "mtcr %0, cr17\n" + ::"r"(value | CACHE_CLR)); + + asm volatile("sync\n"); + + if (l2 && (mfcr_ccr2() & CCR2_L2E)) { + asm volatile( + "mtcr %0, cr24\n" + "sync\n" + ::"r"(value)); + } +} + +static void cache_op_range( + unsigned int start, + unsigned int end, + unsigned int value, + unsigned int l2) +{ + unsigned long i, flags; + unsigned int val = value | CACHE_CLR | CACHE_OMS; + bool l2_sync; + + if (unlikely((end - start) >= PAGE_SIZE) || + unlikely(start < PAGE_OFFSET) || + unlikely(start >= PAGE_OFFSET + LOWMEM_LIMIT)) { + cache_op_all(value, l2); + return; + } + + if ((mfcr_ccr2() & CCR2_L2E) && l2) + l2_sync = 1; + else + l2_sync = 0; + + spin_lock_irqsave(&cache_lock, flags); + for(i = start; i < end; i += L1_CACHE_BYTES) { + cache_op_line(i, val); + if (l2_sync) { + asm volatile( + "sync\n"); + asm volatile( + "mtcr %0, cr24\n" + ::"r"(val)); + } + } + spin_unlock_irqrestore(&cache_lock, flags); + + asm volatile("sync\n"); +} + +void inline dcache_wb_line(unsigned long start) +{ + cache_op_line_atomic(start, DATA_CACHE|CACHE_CLR); +} + +void icache_inv_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, INS_CACHE|CACHE_INV, 0); +} + +void icache_inv_all(void) +{ + cache_op_all(INS_CACHE|CACHE_INV, 0); +} + +void dcache_wb_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, DATA_CACHE|CACHE_CLR, 0); +} + +void dcache_wbinv_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, DATA_CACHE|CACHE_CLR|CACHE_INV, 0); +} + +void dcache_inv_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, DATA_CACHE|CACHE_INV, 0); +} + +void dcache_wbinv_all(void) +{ + cache_op_all(DATA_CACHE|CACHE_CLR|CACHE_INV, 0); +} + +void cache_wbinv_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, INS_CACHE|DATA_CACHE|CACHE_CLR|CACHE_INV, 0); +} + +void cache_wbinv_all(void) +{ + cache_op_all(INS_CACHE|DATA_CACHE|CACHE_CLR|CACHE_INV, 0); +} + +void dma_wbinv_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, DATA_CACHE|CACHE_CLR|CACHE_INV, 1); +} + +void dma_wb_range(unsigned long start, unsigned long end) +{ + cache_op_range(start, end, DATA_CACHE|CACHE_INV, 1); +} + diff --git a/arch/csky/mm/cachev2.c b/arch/csky/mm/cachev2.c new file mode 100644 index 0000000..6414461 --- /dev/null +++ b/arch/csky/mm/cachev2.c @@ -0,0 +1,90 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#include <asm/cache.h> + +void inline dcache_wb_line(unsigned long start) +{ + asm volatile("dcache.cval1 %0\n"::"r"(start)); + asm volatile("sync.is\n"); +} + +void icache_inv_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) + asm volatile("icache.iva %0\n"::"r"(i)); + asm volatile("sync.is\n"); +} + +void icache_inv_all(void) +{ + asm volatile("icache.ialls\n"); + asm volatile("sync.is\n"); +} + +void dcache_wb_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) + asm volatile("dcache.cval1 %0\n"::"r"(i)); + asm volatile("sync.is\n"); +} + +void dcache_wbinv_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) { + asm volatile("dcache.cval1 %0\n"::"r"(i)); + asm volatile("dcache.iva %0\n"::"r"(i)); + } + asm volatile("sync.is\n"); +} + +void dcache_inv_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) + asm volatile("dcache.iva %0\n"::"r"(i)); + asm volatile("sync.is\n"); +} + +void dcache_wbinv_all(void) +{ + asm volatile("dcache.ciall\n"); + asm volatile("sync.is\n"); +} + +void cache_wbinv_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) { + asm volatile("dcache.cval1 %0\n"::"r"(i)); + asm volatile("dcache.iva %0\n"::"r"(i)); + asm volatile("icache.iva %0\n"::"r"(i)); + } + asm volatile("sync.is\n"); +} + +void cache_wbinv_all(void) +{ + asm volatile("dcache.ciall\n"); + asm volatile("icache.ialls\n"); + asm volatile("sync.is\n"); +} + +void dma_wbinv_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) + asm volatile("dcache.civa %0\n"::"r"(i)); + asm volatile("sync.is\n"); +} + +void dma_wb_range(unsigned long start, unsigned long end) +{ + unsigned long i; + for(i=start; i<end; i+=L1_CACHE_BYTES) + asm volatile("dcache.cva %0\n"::"r"(i)); + asm volatile("sync.is\n"); +} + diff --git a/arch/csky/mm/syscache.c b/arch/csky/mm/syscache.c new file mode 100644 index 0000000..5eeb1bd --- /dev/null +++ b/arch/csky/mm/syscache.c @@ -0,0 +1,50 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#include <linux/syscalls.h> +#include <asm/page.h> +#include <asm/cache.h> +#include <asm/cachectl.h> + +SYSCALL_DEFINE3(cacheflush, + void __user *, addr, + unsigned long, bytes, + int, cache) +{ + switch(cache) { + case ICACHE: + icache_inv_all(); + break; + case DCACHE: + dcache_wbinv_all(); + break; + case BCACHE: + cache_wbinv_all(); + break; + default: + return -EINVAL; + } + + return 0; +} + +void __update_cache(struct vm_area_struct *vma, unsigned long address, + pte_t pte) +{ + unsigned long addr; + struct page *page; + unsigned long pfn; + + pfn = pte_pfn(pte); + if (unlikely(!pfn_valid(pfn))) + return; + + page = pfn_to_page(pfn); + addr = (unsigned long) page_address(page); + + if (vma->vm_flags & VM_EXEC || + pages_do_alias(addr, address & PAGE_MASK)) + dcache_wbinv_all(); + + clear_bit(PG_arch_1, &(page)->flags); +} + diff --git a/arch/csky/mm/tlb.c b/arch/csky/mm/tlb.c new file mode 100644 index 0000000..4103464 --- /dev/null +++ b/arch/csky/mm/tlb.c @@ -0,0 +1,207 @@ +// SPDX-License-Identifier: GPL-2.0 +// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. +#include <linux/init.h> +#include <linux/sched.h> +#include <linux/mm.h> + +#include <asm/setup.h> +#include <asm/mmu_context.h> +#include <linux/module.h> +#include <asm/pgtable.h> +#include <abi/ckmmu.h> + +#define CSKY_TLB_SIZE CONFIG_CPU_TLB_SIZE + +void flush_tlb_all(void) +{ + tlb_invalid_all(); +} + +void flush_tlb_mm(struct mm_struct *mm) +{ + int cpu = smp_processor_id(); + + if (cpu_context(cpu, mm) != 0) + drop_mmu_context(mm,cpu); +} + +#define restore_asid_inv_utlb(oldpid, newpid) \ +do { \ + if((oldpid & ASID_MASK) == newpid) \ + write_mmu_entryhi(oldpid +1); \ + write_mmu_entryhi(oldpid); \ +} while(0) + +void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, + unsigned long end) +{ + struct mm_struct *mm = vma->vm_mm; + int cpu = smp_processor_id(); + + if (cpu_context(cpu, mm) != 0) { + unsigned long size, flags; + + local_irq_save(flags); + size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; + size = (size + 1) >> 1; + if (size <= CSKY_TLB_SIZE/2) { + start &= (PAGE_MASK << 1); + end += ((PAGE_SIZE << 1) - 1); + end &= (PAGE_MASK << 1); +#ifdef CONFIG_CPU_HAS_TLBI + while (start < end) { + asm volatile("tlbi.va %0"::"r"(start):); + start += (PAGE_SIZE << 1); + } +#else + { + int oldpid = read_mmu_entryhi(); + int newpid = cpu_asid(cpu, mm); + + while (start < end) { + int idx; + + write_mmu_entryhi(start | newpid); + start += (PAGE_SIZE << 1); + tlb_probe(); + idx = read_mmu_index(); + if (idx >= 0) + tlb_invalid_indexed(); + } + restore_asid_inv_utlb(oldpid, newpid); + } +#endif + } else { + drop_mmu_context(mm, cpu); + } + local_irq_restore(flags); + } +} + +void flush_tlb_kernel_range(unsigned long start, unsigned long end) +{ + unsigned long size, flags; + + local_irq_save(flags); + size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; + if (size <= CSKY_TLB_SIZE) { + start &= (PAGE_MASK << 1); + end += ((PAGE_SIZE << 1) - 1); + end &= (PAGE_MASK << 1); +#ifdef CONFIG_CPU_HAS_TLBI + while (start < end) { + asm volatile("tlbi.va %0"::"r"(start):); + start += (PAGE_SIZE << 1); + } +#else + { + int oldpid = read_mmu_entryhi(); + while (start < end) { + int idx; + write_mmu_entryhi(start); + start += (PAGE_SIZE << 1); + tlb_probe(); + idx = read_mmu_index(); + if (idx >= 0) + tlb_invalid_indexed(); + } + restore_asid_inv_utlb(oldpid, 0); + } +#endif + } else + flush_tlb_all(); + + local_irq_restore(flags); +} + +void flush_tlb_page(struct vm_area_struct *vma, unsigned long page) +{ + int cpu = smp_processor_id(); + + if (!vma || cpu_context(cpu, vma->vm_mm) != 0) { + page &= (PAGE_MASK << 1); + +#ifdef CONFIG_CPU_HAS_TLBI + asm volatile("tlbi.va %0"::"r"(page):); +#else + { + int newpid, oldpid, idx; + unsigned long flags; + local_irq_save(flags); + newpid = cpu_asid(cpu, vma->vm_mm); + oldpid = read_mmu_entryhi(); + write_mmu_entryhi(page | newpid); + tlb_probe(); + idx = read_mmu_index(); + if(idx >= 0) + tlb_invalid_indexed(); + + restore_asid_inv_utlb(oldpid, newpid); + local_irq_restore(flags); + } +#endif + } +} + +/* + * Remove one kernel space TLB entry. This entry is assumed to be marked + * global so we don't do the ASID thing. + */ +void flush_tlb_one(unsigned long page) +{ + page &= (PAGE_MASK << 1); + +#ifdef CONFIG_CPU_HAS_TLBI + asm volatile("tlbi.va %0"::"r"(page):); +#else + { + int idx, oldpid; + unsigned long flags; + oldpid = read_mmu_entryhi(); + + local_irq_save(flags); + page = page | (oldpid & 0xff); + write_mmu_entryhi(page); + tlb_probe(); + idx = read_mmu_index(); + if (idx >= 0) + tlb_invalid_indexed(); + restore_asid_inv_utlb(oldpid, oldpid); + local_irq_restore(flags); + } +#endif +} + +EXPORT_SYMBOL(flush_tlb_one); + +/* show current 32 jtlbs */ +void show_jtlb_table(void) +{ + unsigned long flags; + int entryhi, entrylo0, entrylo1; + int entry; + int oldpid; + + local_irq_save(flags); + entry = 0; + printk("\n\n\n"); + + oldpid = read_mmu_entryhi(); + while (entry < CSKY_TLB_SIZE) + { + write_mmu_index(entry); + tlb_read(); + entryhi = read_mmu_entryhi(); + entrylo0 = read_mmu_entrylo0(); + entrylo0 = entrylo0; + entrylo1 = read_mmu_entrylo1(); + entrylo1 = entrylo1; + printk("jtlb[%d]: entryhi - 0x%x; entrylo0 - 0x%x;" + " entrylo1 - 0x%x\n", + entry, entryhi, entrylo0, entrylo1); + entry++; + } + write_mmu_entryhi(oldpid); + local_irq_restore(flags); +} + -- 2.7.4