On Fri 09-01-15 21:14:00, Johannes Weiner wrote: > The initialization code for the per-cpu charge stock and the soft > limit tree is compact enough to inline it into mem_cgroup_init(). > > Signed-off-by: Johannes Weiner <hannes@xxxxxxxxxxx> Acked-by: Michal Hocko <mhocko@xxxxxxx> > --- > mm/memcontrol.c | 57 ++++++++++++++++++++++++--------------------------------- > 1 file changed, 24 insertions(+), 33 deletions(-) > > diff --git a/mm/memcontrol.c b/mm/memcontrol.c > index aad254b30708..f66bb8f83ac9 100644 > --- a/mm/memcontrol.c > +++ b/mm/memcontrol.c > @@ -2137,17 +2137,6 @@ static void drain_local_stock(struct work_struct *dummy) > clear_bit(FLUSHING_CACHED_CHARGE, &stock->flags); > } > > -static void __init memcg_stock_init(void) > -{ > - int cpu; > - > - for_each_possible_cpu(cpu) { > - struct memcg_stock_pcp *stock = > - &per_cpu(memcg_stock, cpu); > - INIT_WORK(&stock->work, drain_local_stock); > - } > -} > - > /* > * Cache charges(val) to local per_cpu area. > * This will be consumed by consume_stock() function, later. > @@ -4516,26 +4505,6 @@ struct mem_cgroup *parent_mem_cgroup(struct mem_cgroup *memcg) > } > EXPORT_SYMBOL(parent_mem_cgroup); > > -static void __init mem_cgroup_soft_limit_tree_init(void) > -{ > - struct mem_cgroup_tree_per_node *rtpn; > - struct mem_cgroup_tree_per_zone *rtpz; > - int node, zone; > - > - for_each_node(node) { > - rtpn = kzalloc_node(sizeof(*rtpn), GFP_KERNEL, node); > - BUG_ON(!rtpn); > - > - soft_limit_tree.rb_tree_per_node[node] = rtpn; > - > - for (zone = 0; zone < MAX_NR_ZONES; zone++) { > - rtpz = &rtpn->rb_tree_per_zone[zone]; > - rtpz->rb_root = RB_ROOT; > - spin_lock_init(&rtpz->lock); > - } > - } > -} > - > static struct cgroup_subsys_state * __ref > mem_cgroup_css_alloc(struct cgroup_subsys_state *parent_css) > { > @@ -5927,10 +5896,32 @@ void mem_cgroup_migrate(struct page *oldpage, struct page *newpage, > */ > static int __init mem_cgroup_init(void) > { > + int cpu, nid; > + > hotcpu_notifier(memcg_cpu_hotplug_callback, 0); > + > + for_each_possible_cpu(cpu) > + INIT_WORK(&per_cpu_ptr(&memcg_stock, cpu)->work, > + drain_local_stock); > + > + for_each_node(nid) { > + struct mem_cgroup_tree_per_node *rtpn; > + int zone; > + > + rtpn = kzalloc_node(sizeof(*rtpn), GFP_KERNEL, nid); > + > + for (zone = 0; zone < MAX_NR_ZONES; zone++) { > + struct mem_cgroup_tree_per_zone *rtpz; > + > + rtpz = &rtpn->rb_tree_per_zone[zone]; > + rtpz->rb_root = RB_ROOT; > + spin_lock_init(&rtpz->lock); > + } > + soft_limit_tree.rb_tree_per_node[nid] = rtpn; > + } > + > enable_swap_cgroup(); > - mem_cgroup_soft_limit_tree_init(); > - memcg_stock_init(); > + > return 0; > } > subsys_initcall(mem_cgroup_init); > -- > 2.2.0 > -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>