On Thu, Dec 14, 2017 at 09:04:56PM -0800, Dave Hansen wrote: > > I've got some additions to the selftests and a fix where we pass FOLL_* > flags around a bit more instead of just 'write'. I'll get those out as > soon as I do a bit more testing. Try the below; I have more in the works, but this already fixes a whole bunch of obvious fail and should fix the case I described. The thing is, you should _never_ return NULL for an access error, that's complete crap. You should also not blindly change every pte_write() test to pte_access_permitted(), that's also wrong, because then you're missing the read-access tests. Basically you need to very carefully audit each and every p??_access_permitted() call; they're currently mostly wrong. --- a/mm/gup.c +++ b/mm/gup.c @@ -66,7 +66,7 @@ static int follow_pfn_pte(struct vm_area */ static inline bool can_follow_write_pte(pte_t pte, unsigned int flags) { - return pte_access_permitted(pte, WRITE) || + return pte_write(pte) || ((flags & FOLL_FORCE) && (flags & FOLL_COW) && pte_dirty(pte)); } @@ -153,6 +153,11 @@ static struct page *follow_page_pte(stru } if (flags & FOLL_GET) { + if (!pte_access_permitted(pte, !!(flags & FOLL_WRITE))) { + page = ERR_PTR(-EFAULT); + goto out; + } + get_page(page); /* drop the pgmap reference now that we hold the page */ @@ -244,6 +249,15 @@ static struct page *follow_pmd_mask(stru pmd_migration_entry_wait(mm, pmd); goto retry; } + + if (flags & FOLL_GET) { + if (!pmd_access_permitted(*pmd, !!(flags & FOLL_WRITE))) { + page = ERR_PTR(-EFAULT); + spin_unlock(ptr); + return page; + } + } + if (pmd_devmap(*pmd)) { ptl = pmd_lock(mm, pmd); page = follow_devmap_pmd(vma, address, pmd, flags); @@ -326,6 +340,15 @@ static struct page *follow_pud_mask(stru return page; return no_page_table(vma, flags); } + + if (flags & FOLL_GET) { + if (!pud_access_permitted(*pud, !!(flags & FOLL_WRITE))) { + page = ERR_PTR(-EFAULT); + spin_unlock(ptr); + return page; + } + } + if (pud_devmap(*pud)) { ptl = pud_lock(mm, pud); page = follow_devmap_pud(vma, address, pud, flags); --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -870,9 +870,6 @@ struct page *follow_devmap_pmd(struct vm */ WARN_ONCE(flags & FOLL_COW, "mm: In follow_devmap_pmd with FOLL_COW set"); - if (!pmd_access_permitted(*pmd, flags & FOLL_WRITE)) - return NULL; - if (pmd_present(*pmd) && pmd_devmap(*pmd)) /* pass */; else @@ -1012,9 +1009,6 @@ struct page *follow_devmap_pud(struct vm assert_spin_locked(pud_lockptr(mm, pud)); - if (!pud_access_permitted(*pud, flags & FOLL_WRITE)) - return NULL; - if (pud_present(*pud) && pud_devmap(*pud)) /* pass */; else @@ -1386,7 +1380,7 @@ int do_huge_pmd_wp_page(struct vm_fault */ static inline bool can_follow_write_pmd(pmd_t pmd, unsigned int flags) { - return pmd_access_permitted(pmd, WRITE) || + return pmd_write(pmd) || ((flags & FOLL_FORCE) && (flags & FOLL_COW) && pmd_dirty(pmd)); } -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>