On Thu, Jan 26, 2017 at 02:57:50PM +0300, Kirill A. Shutemov wrote: > Most of work happans on head page. Only when we need to do copy data to > userspace we find relevant subpage. > > We are still limited by PAGE_SIZE per iteration. Lifting this limitation > would require some more work. Now that I debugged that bit of my brain, here's a more sensible suggestion. > @@ -1886,6 +1886,7 @@ static ssize_t do_generic_file_read(struct file *filp, loff_t *ppos, > if (unlikely(page == NULL)) > goto no_cached_page; > } > + page = compound_head(page); > if (PageReadahead(page)) { > page_cache_async_readahead(mapping, > ra, filp, page, We're going backwards and forwards a lot between subpages and page heads. I'd like to see us do this: static inline struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, int fgp_flags, gfp_t cache_gfp_mask) { struct page *page = pagecache_get_head(mapping, offset, fgp_flags, cache_gfp_mask); return page ? find_subpage(page, offset) : NULL; } static inline struct page *find_get_head(struct address_space *mapping, pgoff_t offset) { return pagecache_get_head(mapping, offset, 0, 0); } and then we can switch do_generic_file_read() to call find_get_head(), eliminating the conversion back and forth between subpages and head pages. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>