2017-07-26 21:44 GMT+08:00 Michal Hocko <mhocko@xxxxxxxxxx>: > On Wed 26-07-17 21:07:42, Wenwei Tao wrote: >> From: Wenwei Tao <wenwei.tww@xxxxxxxxxxxxxxx> >> >> By removing the child cgroup while the parent cgroup is >> under reclaim, we could trigger the following kernel panic >> on kernel 3.10: >> ------------------------------------------------------------------------ >> kernel BUG at kernel/cgroup.c:893! >> invalid opcode: 0000 [#1] SMP >> CPU: 1 PID: 22477 Comm: kworker/1:1 Not tainted 3.10.107 #1 >> Workqueue: cgroup_destroy css_dput_fn >> task: ffff8817959a5780 ti: ffff8817e8886000 task.ti: ffff8817e8886000 >> RIP: 0010:[<ffffffff810cd6e0>] [<ffffffff810cd6e0>] >> cgroup_diput+0xc0/0xf0 >> RSP: 0000:ffff8817e8887da0 EFLAGS: 00010246 >> RAX: 0000000000000000 RBX: ffff8817a5dd5d40 RCX: dead000000000200 >> RDX: 0000000000000000 RSI: ffff8817973a6910 RDI: ffff8817f54c2a00 >> RBP: ffff8817e8887dc8 R08: ffff8817a5dd5dd0 R09: df9fb35794b01820 >> R10: df9fb35794b01820 R11: 00007fa95b1efcda R12: ffff8817a5dd5d9c >> R13: ffff8817f38b3a40 R14: ffff8817973a6910 R15: ffff8817973a6910 >> FS: 0000000000000000(0000) GS:ffff88181f220000(0000) >> knlGS:0000000000000000 >> CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >> CR2: 00007fa6e6234000 CR3: 000000179f19d000 CR4: 00000000000407e0 >> DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 >> DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400 >> Stack: >> ffff8817a5dd5d40 ffff8817a5dd5d9c ffff8817f38b3a40 ffff8817973a6910 >> 0000000000000040 ffff8817e8887df8 ffffffff811b37c2 ffff8817fa23c000 >> ffff8817f57dbb80 ffff88181f232ac0 ffff88181f237500 ffff8817e8887e10 >> Call Trace: >> [<ffffffff811b37c2>] dput+0x1a2/0x2f0 >> [<ffffffff810cbacc>] cgroup_dput.isra.21+0x1c/0x30 >> [<ffffffff810cbafd>] css_dput_fn+0x1d/0x20 >> [<ffffffff81078ebc>] process_one_work+0x17c/0x460 >> [<ffffffff81079b66>] worker_thread+0x116/0x3b0 >> [<ffffffff81079a50>] ? manage_workers.isra.25+0x290/0x290 >> [<ffffffff81080330>] kthread+0xc0/0xd0 >> [<ffffffff81080270>] ? insert_kthread_work+0x40/0x40 >> [<ffffffff815b1e08>] ret_from_fork+0x58/0x90 >> [<ffffffff81080270>] ? insert_kthread_work+0x40/0x40 >> Code: 41 5e 41 5f 5d c3 0f 1f 44 00 00 48 8b 7f 78 48 8b 07 a8 01 74 15 >> 48 81 c7 30 01 00 00 48 c7 c6 a0 a7 0c 81 e8 b2 83 02 00 eb c8 <0f> 0b >> 49 8b 4e 18 48 c7 c2 7e f1 7a 81 be 85 03 00 00 48 c7 c7 >> RIP [<ffffffff810cd6e0>] cgroup_diput+0xc0/0xf0 >> RSP <ffff8817e8887da0> >> ---[ end trace 85eeea5212c44f51 ]--- >> ------------------------------------------------------------------------ >> >> I think there is a css double put in mem_cgroup_iter. Under reclaim, >> we call mem_cgroup_iter the first time with prev == NULL, and we get >> last_visited memcg from per zone's reclaim_iter then call __mem_cgroup_iter_next >> try to get next alive memcg, __mem_cgroup_iter_next could return NULL >> if last_visited is already the last one so we put the last_visited's >> memcg css and continue to the next while loop, this time we might not >> do css_tryget(&last_visited->css) if the dead_count is changed, but >> we still do css_put(&last_visited->css), we put it twice, this could >> trigger the BUG_ON at kernel/cgroup.c:893. > > Yes, I guess your are right and I suspect that this has been silently > fixed by 519ebea3bf6d ("mm: memcontrol: factor out reclaim iterator > loading and updating"). I think a more appropriate fix is would be. > Are you able to reproduce and re-test it? > --- Yes, I think this commit can fix this issue, and I backport this commit to 3.10.107 kernel and cannot reproduce this issue. I guess this commit might need to be backported to 3.10.y stable kernel. > diff --git a/mm/memcontrol.c b/mm/memcontrol.c > index 437ae2cbe102..0848ec05c12a 100644 > --- a/mm/memcontrol.c > +++ b/mm/memcontrol.c > @@ -1224,6 +1224,8 @@ struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root, > if (last_visited && last_visited != root && > !css_tryget(&last_visited->css)) > last_visited = NULL; > + } else { > + last_visited = true; > } > } > > -- > Michal Hocko > SUSE Labs -- To unsubscribe from this list: send the line "unsubscribe cgroups" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html