On Fri 03-11-17 00:46:18, Hugh Dickins wrote: > On Thu, 2 Nov 2017, Michal Hocko wrote: > > From: Michal Hocko <mhocko@xxxxxxxx> > > > > syzkaller has reported the following lockdep splat > > ====================================================== > > WARNING: possible circular locking dependency detected > > 4.13.0-next-20170911+ #19 Not tainted > > ------------------------------------------------------ > > syz-executor5/6914 is trying to acquire lock: > > (cpu_hotplug_lock.rw_sem){++++}, at: [<ffffffff818c1b3e>] get_online_cpus include/linux/cpu.h:126 [inline] > > (cpu_hotplug_lock.rw_sem){++++}, at: [<ffffffff818c1b3e>] lru_add_drain_all+0xe/0x20 mm/swap.c:729 > > > > but task is already holding lock: > > (&sb->s_type->i_mutex_key#9){++++}, at: [<ffffffff818fbef7>] inode_lock include/linux/fs.h:712 [inline] > > (&sb->s_type->i_mutex_key#9){++++}, at: [<ffffffff818fbef7>] shmem_add_seals+0x197/0x1060 mm/shmem.c:2768 > > > > more details [1] and dependencies explained [2]. The problem seems to be > > the usage of lru_add_drain_all from shmem_wait_for_pins. While the lock > > dependency is subtle as hell and we might want to make lru_add_drain_all > > less dependent on the hotplug locks the usage of lru_add_drain_all seems > > dubious here. The whole function cares only about radix tree tags, page > > count and page mapcount. None of those are touched from the draining > > context. So it doesn't make much sense to drain pcp caches. Moreover > > this looks like a wrong thing to do because it basically induces > > unpredictable latency to the call because draining is not for free > > (especially on larger machines with many cpus). > > > > Let's simply drop the call to lru_add_drain_all to address both issues. > > > > [1] http://lkml.kernel.org/r/089e0825eec8955c1f055c83d476@xxxxxxxxxx > > [2] http://lkml.kernel.org/r/http://lkml.kernel.org/r/20171030151009.ip4k7nwan7muouca@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx > > > > Cc: David Herrmann <dh.herrmann@xxxxxxxxx> > > Cc: Hugh Dickins <hughd@xxxxxxxxxx> > > Signed-off-by: Michal Hocko <mhocko@xxxxxxxx> > > NAK. shmem_wait_for_pins() is waiting for temporary pins on the pages > to go away, and using lru_add_drain_all() in the usual way, to lower > the refcount of pages temporarily pinned in a pagevec somewhere. Page > count is touched by draining pagevecs: I'm surprised to see you say > that it isn't - or have pagevec page references been eliminated by > a recent commit that I missed? I must be missing something here. __pagevec_lru_add_fn merely about moving the page into the appropriate LRU list, pagevec_move_tail only rotates, lru_deactivate_file_fn moves from active to inactive LRUs, lru_lazyfree_fn moves from anon to file LRUs and activate_page_drain just moves to the active list. None of those operations touch the page count AFAICS. So I would agree that some pages might be pinned outside of the LRU (lru_add_pvec) and thus unreclaimable but does this really matter. Or what else I am missing? -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>