Re: [PATCH v3 14/16] mm: support GUP-triggered unsharing of anonymous pages

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 3/29/22 18:04, David Hildenbrand wrote:
> Whenever GUP currently ends up taking a R/O pin on an anonymous page that
> might be shared -- mapped R/O and !PageAnonExclusive() -- any write fault
> on the page table entry will end up replacing the mapped anonymous page
> due to COW, resulting in the GUP pin no longer being consistent with the
> page actually mapped into the page table.
> 
> The possible ways to deal with this situation are:
>  (1) Ignore and pin -- what we do right now.
>  (2) Fail to pin -- which would be rather surprising to callers and
>      could break user space.
>  (3) Trigger unsharing and pin the now exclusive page -- reliable R/O
>      pins.
> 
> We want to implement 3) because it provides the clearest semantics and
> allows for checking in unpin_user_pages() and friends for possible BUGs:
> when trying to unpin a page that's no longer exclusive, clearly
> something went very wrong and might result in memory corruptions that
> might be hard to debug. So we better have a nice way to spot such
> issues.
> 
> To implement 3), we need a way for GUP to trigger unsharing:
> FAULT_FLAG_UNSHARE. FAULT_FLAG_UNSHARE is only applicable to R/O mapped
> anonymous pages and resembles COW logic during a write fault. However, in
> contrast to a write fault, GUP-triggered unsharing will, for example, still
> maintain the write protection.
> 
> Let's implement FAULT_FLAG_UNSHARE by hooking into the existing write fault
> handlers for all applicable anonymous page types: ordinary pages, THP and
> hugetlb.
> 
> * If FAULT_FLAG_UNSHARE finds a R/O-mapped anonymous page that has been
>   marked exclusive in the meantime by someone else, there is nothing to do.
> * If FAULT_FLAG_UNSHARE finds a R/O-mapped anonymous page that's not
>   marked exclusive, it will try detecting if the process is the exclusive
>   owner. If exclusive, it can be set exclusive similar to reuse logic
>   during write faults via page_move_anon_rmap() and there is nothing
>   else to do; otherwise, we either have to copy and map a fresh,
>   anonymous exclusive page R/O (ordinary pages, hugetlb), or split the
>   THP.
> 
> This commit is heavily based on patches by Andrea.
> 
> Co-developed-by: Andrea Arcangeli <aarcange@xxxxxxxxxx>
> Signed-off-by: Andrea Arcangeli <aarcange@xxxxxxxxxx>
> Signed-off-by: David Hildenbrand <david@xxxxxxxxxx>

Acked-by: Vlastimil Babka <vbabka@xxxxxxx>

Modulo a nit and suspected logical bug below.

<snip>

> @@ -3072,6 +3082,7 @@ static vm_fault_t wp_page_copy(struct vm_fault *vmf)
>  		 * mmu page tables (such as kvm shadow page tables), we want the
>  		 * new page to be mapped directly into the secondary page table.
>  		 */
> +		BUG_ON(unshare && pte_write(entry));
>  		set_pte_at_notify(mm, vmf->address, vmf->pte, entry);
>  		update_mmu_cache(vma, vmf->address, vmf->pte);
>  		if (old_page) {
> @@ -3121,7 +3132,7 @@ static vm_fault_t wp_page_copy(struct vm_fault *vmf)
>  			free_swap_cache(old_page);
>  		put_page(old_page);
>  	}
> -	return page_copied ? VM_FAULT_WRITE : 0;
> +	return page_copied && !unshare ? VM_FAULT_WRITE : 0;

Could be just me but I would prefer (page_copied && !unshare) as I rarely
see these operators together like this to remember their relative priority
very well.

>  oom_free_new:
>  	put_page(new_page);
>  oom:

<snip>

> @@ -4515,8 +4550,11 @@ static inline vm_fault_t create_huge_pmd(struct vm_fault *vmf)
>  /* `inline' is required to avoid gcc 4.1.2 build error */
>  static inline vm_fault_t wp_huge_pmd(struct vm_fault *vmf)
>  {
> +	const bool unshare = vmf->flags & FAULT_FLAG_UNSHARE;
> +
>  	if (vma_is_anonymous(vmf->vma)) {
> -		if (userfaultfd_huge_pmd_wp(vmf->vma, vmf->orig_pmd))
> +		if (unlikely(unshare) &&

Is this condition flipped, should it be "likely(!unshare)"? As the similar
code in do_wp_page() does.

> +		    userfaultfd_huge_pmd_wp(vmf->vma, vmf->orig_pmd))
>  			return handle_userfault(vmf, VM_UFFD_WP);
>  		return do_huge_pmd_wp_page(vmf);
>  	}
> @@ -4651,10 +4689,11 @@ static vm_fault_t handle_pte_fault(struct vm_fault *vmf)
>  		update_mmu_tlb(vmf->vma, vmf->address, vmf->pte);
>  		goto unlock;
>  	}
> -	if (vmf->flags & FAULT_FLAG_WRITE) {
> +	if (vmf->flags & (FAULT_FLAG_WRITE|FAULT_FLAG_UNSHARE)) {
>  		if (!pte_write(entry))
>  			return do_wp_page(vmf);
> -		entry = pte_mkdirty(entry);
> +		else if (likely(vmf->flags & FAULT_FLAG_WRITE))
> +			entry = pte_mkdirty(entry);
>  	}
>  	entry = pte_mkyoung(entry);
>  	if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry,




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux