On Fri, Jul 10, 2020 at 04:23:35PM +0800, Alex Shi wrote: > > > 在 2020/7/9 下午11:48, Kirill A. Shutemov 写道: > > On Mon, Jul 06, 2020 at 09:52:34PM -0700, Hugh Dickins wrote: > >> On Mon, 6 Jul 2020, Matthew Wilcox wrote: > >>> On Mon, Jul 06, 2020 at 05:15:09PM +0800, Alex Shi wrote: > >>>> Hi Kirill & Johannes & Matthew, > >> > >> Adding Kirill, who was in patch's Cc list but not mail's Cc list. > >> > >> I asked Alex to direct this one particularly to Kirill and Johannes > >> and Matthew because (and I regret that the commit message still does > >> not make this at all clear) this patch changes the lock ordering: > >> which for years has been lru_lock outside memcg move_lock outside > >> i_pages lock, but here inverted to lru_lock inside i_pages lock. > >> > >> I don't see a strong reason to have them one way round or the other, > >> and think Alex is right that they can safely be reversed here: but > >> he doesn't actually give any reason for doing so (if cleanup, then > >> I think the cleanup should have been taken further), and no reason > >> for doing so as part of this series. > > > > I've looked around and changing order of lru_lock wrt. i_pages lock seems > > safe. I don't have much experience with memcg move_lock. > > Hi Kirill, > > Thanks for response! > mem_cgroup_move_account(page) could not reach here since 2 blocks, > 1, isolate_lru_page() before it will take page from lru, this compete for > page reclaim path, list non-null. > > 2, try_lock_page in it, will guard split_huge_page(), !list. > > > > > Alex, if you are going ahead with the patch, please document the locking > > order. We have some locking orders listed at the beginning of filemap.c > > and rmap.c. > > Thanks for reminder! > Hugh Dickins did this in above 2 files at the end of patchset, any comments? > > diff --git a/mm/filemap.c b/mm/filemap.c > index f0ae9a6308cb..1b42aaae4d3e 100644 > --- a/mm/filemap.c > +++ b/mm/filemap.c > @@ -101,8 +101,8 @@ > * ->swap_lock (try_to_unmap_one) > * ->private_lock (try_to_unmap_one) > * ->i_pages lock (try_to_unmap_one) > - * ->pgdat->lru_lock (follow_page->mark_page_accessed) > - * ->pgdat->lru_lock (check_pte_range->isolate_lru_page) > + * ->lruvec->lru_lock (follow_page->mark_page_accessed) > + * ->lruvec->lru_lock (check_pte_range->isolate_lru_page) > * ->private_lock (page_remove_rmap->set_page_dirty) > * ->i_pages lock (page_remove_rmap->set_page_dirty) > * bdi.wb->list_lock (page_remove_rmap->set_page_dirty) > diff --git a/mm/memcontrol.c b/mm/memcontrol.c > index d5e56be42f21..926d7d95dc1d 100644 > --- a/mm/memcontrol.c > +++ b/mm/memcontrol.c > @@ -3057,7 +3057,7 @@ void __memcg_kmem_uncharge_page(struct page *page, int order) > #ifdef CONFIG_TRANSPARENT_HUGEPAGE > > /* > - * Because tail pages are not marked as "used", set it. We're under > + * Because tail pages are not marked as "used", set it. Don't need > * lruvec->lru_lock and migration entries setup in all page mappings. > */ > void mem_cgroup_split_huge_fixup(struct page *head) > diff --git a/mm/rmap.c b/mm/rmap.c > index 5fe2dedce1fc..7fbc382e6f9e 100644 > --- a/mm/rmap.c > +++ b/mm/rmap.c > @@ -28,12 +28,12 @@ > * hugetlb_fault_mutex (hugetlbfs specific page fault mutex) > * anon_vma->rwsem > * mm->page_table_lock or pte_lock > - * pgdat->lru_lock (in mark_page_accessed, isolate_lru_page) > * swap_lock (in swap_duplicate, swap_info_get) > * mmlist_lock (in mmput, drain_mmlist and others) > * mapping->private_lock (in __set_page_dirty_buffers) > - * mem_cgroup_{begin,end}_page_stat (memcg->move_lock) > + * lock_page_memcg move_lock (in __set_page_dirty_buffers) > * i_pages lock (widely used) > + * lock_page_lruvec_irq lruvec->lru_lock I think it has to be lruvec->lru_lock (in lock_page_lruvec_irq) No? > * inode->i_lock (in set_page_dirty's __mark_inode_dirty) > * bdi.wb->list_lock (in set_page_dirty's __mark_inode_dirty) > * sb_lock (within inode_lock in fs/fs-writeback.c) > > > > > local_irq_disable() also deserves a comment. > > > > yes, I will add a comment for this. Do you mind give reviewed-by for this patch? Reviewed-by: Kirill A. Shutemov <kirill.shutemov@xxxxxxxxxxxxxxx> -- Kirill A. Shutemov