On Mon, Sep 03, 2012 at 05:32:45PM +0900, Minchan Kim wrote: > Don't we need initialization? > > diff --git a/mm/rmap.c b/mm/rmap.c > index 0f3b7cd..c0f3221 100644 > --- a/mm/rmap.c > +++ b/mm/rmap.c > @@ -416,6 +416,9 @@ static void anon_vma_ctor(void *data) > > mutex_init(&anon_vma->mutex); > atomic_set(&anon_vma->refcount, 0); > +#ifdef CONFIG_SWAP > + atomic_set(&anon_vma->swapra_miss, 0); > +#endif > INIT_LIST_HEAD(&anon_vma->head); > } Sorry about this silly problem. I'm wondering why I didn't notice it, maybe because only tested random swap after move swapra_miss to anon_vma. Subject: swap: add a simple random read swapin detection The swapin readahead does a blind readahead regardless if the swapin is sequential. This is ok for harddisk and random read, because read big size has no penality in harddisk, and if the readahead pages are garbage, they can be reclaimed fastly. But for SSD, big size read is more expensive than small size read. If readahead pages are garbage, such readahead only has overhead. This patch addes a simple random read detection like what file mmap readahead does. If random read is detected, swapin readahead will be skipped. This improves a lot for a swap workload with random IO in a fast SSD. I run anonymous mmap write micro benchmark, which will triger swapin/swapout. runtime changes with path randwrite harddisk -38.7% seqwrite harddisk -1.1% randwrite SSD -46.9% seqwrite SSD +0.3% For both harddisk and SSD, the randwrite swap workload run time is reduced significant. sequential write swap workload hasn't chanage. Interesting is the randwrite harddisk test is improved too. This might be because swapin readahead need allocate extra memory, which further tights memory pressure, so more swapout/swapin. This patch depends on readahead-fault-retry-breaks-mmap-file-read-random-detection.patch V2->V3: move swapra_miss to 'struct anon_vma' as suggested by Konstantin. V1->V2: 1. Move the swap readahead accounting to separate functions as suggested by Riel. 2. Enable the logic only with CONFIG_SWAP enabled as suggested by Minchan. Signed-off-by: Shaohua Li <shli@xxxxxxxxxxxx> Acked-by: Rik van Riel <riel@xxxxxxxxxx> --- include/linux/rmap.h | 3 ++ mm/internal.h | 52 +++++++++++++++++++++++++++++++++++++++++++++++++++ mm/memory.c | 3 +- mm/rmap.c | 3 ++ mm/shmem.c | 1 mm/swap_state.c | 6 +++++ 6 files changed, 67 insertions(+), 1 deletion(-) Index: linux/mm/swap_state.c =================================================================== --- linux.orig/mm/swap_state.c 2012-08-29 16:13:00.912112140 +0800 +++ linux/mm/swap_state.c 2012-08-30 18:28:24.678315187 +0800 @@ -20,6 +20,7 @@ #include <linux/page_cgroup.h> #include <asm/pgtable.h> +#include "internal.h" /* * swapper_space is a fiction, retained to simplify the path through @@ -379,6 +380,10 @@ struct page *swapin_readahead(swp_entry_ unsigned long mask = (1UL << page_cluster) - 1; struct blk_plug plug; + swap_cache_miss(vma); + if (swap_cache_skip_readahead(vma)) + goto skip; + /* Read a page_cluster sized and aligned cluster around offset. */ start_offset = offset & ~mask; end_offset = offset | mask; @@ -397,5 +402,6 @@ struct page *swapin_readahead(swp_entry_ blk_finish_plug(&plug); lru_add_drain(); /* Push any new pages onto the LRU now */ +skip: return read_swap_cache_async(entry, gfp_mask, vma, addr); } Index: linux/mm/memory.c =================================================================== --- linux.orig/mm/memory.c 2012-08-29 16:13:00.920112040 +0800 +++ linux/mm/memory.c 2012-08-30 13:32:05.425830660 +0800 @@ -2953,7 +2953,8 @@ static int do_swap_page(struct mm_struct ret = VM_FAULT_HWPOISON; delayacct_clear_flag(DELAYACCT_PF_SWAPIN); goto out_release; - } + } else if (!(flags & FAULT_FLAG_TRIED)) + swap_cache_hit(vma); locked = lock_page_or_retry(page, mm, flags); Index: linux/mm/internal.h =================================================================== --- linux.orig/mm/internal.h 2012-08-29 16:13:00.932111888 +0800 +++ linux/mm/internal.h 2012-09-03 15:16:30.566299444 +0800 @@ -12,6 +12,7 @@ #define __MM_INTERNAL_H #include <linux/mm.h> +#include <linux/rmap.h> void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *start_vma, unsigned long floor, unsigned long ceiling); @@ -356,3 +357,54 @@ extern unsigned long vm_mmap_pgoff(struc unsigned long, unsigned long); extern void set_pageblock_order(void); + +/* + * Unnecessary readahead harms performance. 1. for SSD, big size read is more + * expensive than small size read, so extra unnecessary read only has overhead. + * For harddisk, this overhead doesn't exist. 2. unnecessary readahead will + * allocate extra memroy, which further tights memory pressure, so more + * swapout/swapin. + * These adds a simple swap random access detection. In swap page fault, if + * page is found in swap cache, decrease an account of vma, otherwise we need + * do sync swapin and the account is increased. Optionally swapin will do + * readahead if the counter is below a threshold. + */ +#ifdef CONFIG_SWAP +#define SWAPRA_MISS_THRESHOLD (100) +#define SWAPRA_MAX_MISS ((SWAPRA_MISS_THRESHOLD) * 10) +static inline void swap_cache_hit(struct vm_area_struct *vma) +{ + if (vma && vma->anon_vma) + atomic_dec_if_positive(&vma->anon_vma->swapra_miss); +} + +static inline void swap_cache_miss(struct vm_area_struct *vma) +{ + if (!vma || !vma->anon_vma) + return; + if (atomic_read(&vma->anon_vma->swapra_miss) < SWAPRA_MAX_MISS) + atomic_inc(&vma->anon_vma->swapra_miss); +} + +static inline int swap_cache_skip_readahead(struct vm_area_struct *vma) +{ + if (!vma || !vma->anon_vma) + return 0; + return atomic_read(&vma->anon_vma->swapra_miss) > + SWAPRA_MISS_THRESHOLD; +} +#else +static inline void swap_cache_hit(struct vm_area_struct *vma) +{ +} + +static inline void swap_cache_miss(struct vm_area_struct *vma) +{ +} + +static inline int swap_cache_skip_readahead(struct vm_area_struct *vma) +{ + return 0; +} + +#endif Index: linux/include/linux/rmap.h =================================================================== --- linux.orig/include/linux/rmap.h 2012-06-01 10:10:31.686394463 +0800 +++ linux/include/linux/rmap.h 2012-08-30 18:10:12.256048781 +0800 @@ -35,6 +35,9 @@ struct anon_vma { * anon_vma if they are the last user on release */ atomic_t refcount; +#ifdef CONFIG_SWAP + atomic_t swapra_miss; +#endif /* * NOTE: the LSB of the head.next is set by Index: linux/mm/shmem.c =================================================================== --- linux.orig/mm/shmem.c 2012-08-06 16:00:45.465441525 +0800 +++ linux/mm/shmem.c 2012-08-30 18:10:51.755553250 +0800 @@ -933,6 +933,7 @@ static struct page *shmem_swapin(swp_ent pvma.vm_pgoff = index + info->vfs_inode.i_ino; pvma.vm_ops = NULL; pvma.vm_policy = spol; + pvma.anon_vma = NULL; return swapin_readahead(swap, gfp, &pvma, 0); } Index: linux/mm/rmap.c =================================================================== --- linux.orig/mm/rmap.c 2012-06-01 10:10:31.706394210 +0800 +++ linux/mm/rmap.c 2012-09-03 19:42:15.454127265 +0800 @@ -416,6 +416,9 @@ static void anon_vma_ctor(void *data) mutex_init(&anon_vma->mutex); atomic_set(&anon_vma->refcount, 0); +#ifdef CONFIG_SWAP + atomic_set(&anon_vma->swapra_miss, 0); +#endif INIT_LIST_HEAD(&anon_vma->head); } -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>