Making memcg numa's scanning information update by schedule_work(). Now, memcg's numa information is updated under a thread doing memory reclaim. It's not very heavy weight now. But upcoming updates around numa scanning will add more works. This patch makes the update be done by schedule_work() and reduce latency caused by this updates. Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@xxxxxxxxxxxxxx> --- mm/memcontrol.c | 42 ++++++++++++++++++++++++++++++------------ 1 file changed, 30 insertions(+), 12 deletions(-) Index: mmotm-0710/mm/memcontrol.c =================================================================== --- mmotm-0710.orig/mm/memcontrol.c +++ mmotm-0710/mm/memcontrol.c @@ -284,6 +284,7 @@ struct mem_cgroup { nodemask_t scan_nodes; atomic_t numainfo_events; atomic_t numainfo_updating; + struct work_struct numainfo_update_work; #endif /* * Should the accounting and control be hierarchical, per subtree? @@ -1551,6 +1552,23 @@ static bool test_mem_cgroup_node_reclaim } #if MAX_NUMNODES > 1 +static void mem_cgroup_numainfo_update_work(struct work_struct *work) +{ + struct mem_cgroup *memcg; + int nid; + + memcg = container_of(work, struct mem_cgroup, numainfo_update_work); + + memcg->scan_nodes = node_states[N_HIGH_MEMORY]; + for_each_node_mask(nid, node_states[N_HIGH_MEMORY]) { + if (!test_mem_cgroup_node_reclaimable(memcg, nid, false)) + node_clear(nid, memcg->scan_nodes); + } + atomic_set(&memcg->numainfo_updating, 0); + css_put(&memcg->css); +} + + /* * Always updating the nodemask is not very good - even if we have an empty * list or the wrong list here, we can start from some node and traverse all @@ -1559,7 +1577,6 @@ static bool test_mem_cgroup_node_reclaim */ static void mem_cgroup_may_update_nodemask(struct mem_cgroup *mem) { - int nid; /* * numainfo_events > 0 means there was at least NUMAINFO_EVENTS_TARGET * pagein/pageout changes since the last update. @@ -1568,18 +1585,9 @@ static void mem_cgroup_may_update_nodema return; if (atomic_inc_return(&mem->numainfo_updating) > 1) return; - - /* make a nodemask where this memcg uses memory from */ - mem->scan_nodes = node_states[N_HIGH_MEMORY]; - - for_each_node_mask(nid, node_states[N_HIGH_MEMORY]) { - - if (!test_mem_cgroup_node_reclaimable(mem, nid, false)) - node_clear(nid, mem->scan_nodes); - } - atomic_set(&mem->numainfo_events, 0); - atomic_set(&mem->numainfo_updating, 0); + css_get(&mem->css); + schedule_work(&mem->numainfo_update_work); } /* @@ -1652,6 +1660,12 @@ bool mem_cgroup_reclaimable(struct mem_c return false; } +static void mem_cgroup_numascan_init(struct mem_cgroup *memcg) +{ + INIT_WORK(&memcg->numainfo_update_work, + mem_cgroup_numainfo_update_work); +} + #else int mem_cgroup_select_victim_node(struct mem_cgroup *mem) { @@ -1662,6 +1676,9 @@ bool mem_cgroup_reclaimable(struct mem_c { return test_mem_cgroup_node_reclaimable(mem, 0, noswap); } +static void mem_cgroup_numascan_init(struct mem_cgroup *memcg) +{ +} #endif static void __mem_cgroup_record_scanstat(unsigned long *stats, @@ -5032,6 +5049,7 @@ mem_cgroup_create(struct cgroup_subsys * mem->move_charge_at_immigrate = 0; mutex_init(&mem->thresholds_lock); spin_lock_init(&mem->scanstat.lock); + mem_cgroup_numascan_init(mem); return &mem->css; free_out: __mem_cgroup_free(mem); -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>