Re: [PATCH v14 07/20] mm/thp: narrow lru locking

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Jul 10, 2020 at 04:23:35PM +0800, Alex Shi wrote:
> 
> 
> 在 2020/7/9 下午11:48, Kirill A. Shutemov 写道:
> > On Mon, Jul 06, 2020 at 09:52:34PM -0700, Hugh Dickins wrote:
> >> On Mon, 6 Jul 2020, Matthew Wilcox wrote:
> >>> On Mon, Jul 06, 2020 at 05:15:09PM +0800, Alex Shi wrote:
> >>>> Hi Kirill & Johannes & Matthew,
> >>
> >> Adding Kirill, who was in patch's Cc list but not mail's Cc list.
> >>
> >> I asked Alex to direct this one particularly to Kirill and Johannes
> >> and Matthew because (and I regret that the commit message still does
> >> not make this at all clear) this patch changes the lock ordering:
> >> which for years has been lru_lock outside memcg move_lock outside
> >> i_pages lock, but here inverted to lru_lock inside i_pages lock.
> >>
> >> I don't see a strong reason to have them one way round or the other,
> >> and think Alex is right that they can safely be reversed here: but
> >> he doesn't actually give any reason for doing so (if cleanup, then
> >> I think the cleanup should have been taken further), and no reason
> >> for doing so as part of this series.
> > 
> > I've looked around and changing order of lru_lock wrt. i_pages lock seems
> > safe. I don't have much experience with memcg move_lock.
> 
> Hi Kirill,
> 
> Thanks for response!
> mem_cgroup_move_account(page) could not reach here since 2 blocks,
> 1, isolate_lru_page() before it will take page from lru, this compete for
>    page reclaim path, list non-null.
> 
> 2, try_lock_page in it, will guard split_huge_page(), !list.
> 
> > 
> > Alex, if you are going ahead with the patch, please document the locking
> > order. We have some locking orders listed at the beginning of filemap.c
> > and rmap.c.
> 
> Thanks for reminder!
> Hugh Dickins did this in above 2 files at the end of patchset, any comments?
> 
> diff --git a/mm/filemap.c b/mm/filemap.c
> index f0ae9a6308cb..1b42aaae4d3e 100644
> --- a/mm/filemap.c
> +++ b/mm/filemap.c
> @@ -101,8 +101,8 @@
>   *    ->swap_lock		(try_to_unmap_one)
>   *    ->private_lock		(try_to_unmap_one)
>   *    ->i_pages lock		(try_to_unmap_one)
> - *    ->pgdat->lru_lock		(follow_page->mark_page_accessed)
> - *    ->pgdat->lru_lock		(check_pte_range->isolate_lru_page)
> + *    ->lruvec->lru_lock	(follow_page->mark_page_accessed)
> + *    ->lruvec->lru_lock	(check_pte_range->isolate_lru_page)
>   *    ->private_lock		(page_remove_rmap->set_page_dirty)
>   *    ->i_pages lock		(page_remove_rmap->set_page_dirty)
>   *    bdi.wb->list_lock		(page_remove_rmap->set_page_dirty)
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index d5e56be42f21..926d7d95dc1d 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -3057,7 +3057,7 @@ void __memcg_kmem_uncharge_page(struct page *page, int order)
>  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
>  
>  /*
> - * Because tail pages are not marked as "used", set it. We're under
> + * Because tail pages are not marked as "used", set it. Don't need
>   * lruvec->lru_lock and migration entries setup in all page mappings.
>   */
>  void mem_cgroup_split_huge_fixup(struct page *head)
> diff --git a/mm/rmap.c b/mm/rmap.c
> index 5fe2dedce1fc..7fbc382e6f9e 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -28,12 +28,12 @@
>   *           hugetlb_fault_mutex (hugetlbfs specific page fault mutex)
>   *           anon_vma->rwsem
>   *             mm->page_table_lock or pte_lock
> - *               pgdat->lru_lock (in mark_page_accessed, isolate_lru_page)
>   *               swap_lock (in swap_duplicate, swap_info_get)
>   *                 mmlist_lock (in mmput, drain_mmlist and others)
>   *                 mapping->private_lock (in __set_page_dirty_buffers)
> - *                   mem_cgroup_{begin,end}_page_stat (memcg->move_lock)
> + *                   lock_page_memcg move_lock (in __set_page_dirty_buffers)
>   *                     i_pages lock (widely used)
> + *                       lock_page_lruvec_irq lruvec->lru_lock

I think it has to be
			    lruvec->lru_lock (in lock_page_lruvec_irq)

No?

>   *                 inode->i_lock (in set_page_dirty's __mark_inode_dirty)
>   *                 bdi.wb->list_lock (in set_page_dirty's __mark_inode_dirty)
>   *                   sb_lock (within inode_lock in fs/fs-writeback.c)
> 
> > 
> > local_irq_disable() also deserves a comment.
> > 
> 
> yes, I will add a comment for this. Do you mind give reviewed-by for this patch?

Reviewed-by: Kirill A. Shutemov <kirill.shutemov@xxxxxxxxxxxxxxx>

-- 
 Kirill A. Shutemov





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux