From: Srikar Dronamraju <srikar@xxxxxxxxxxxxxxxxxx> commit 78a9c54649ea220065aad9902460a1d137c7eafd upstream. Simple rename of the 'numabalancing_enabled' variable to 'sched_numa_balancing'. No functional changes. Suggested-by: Ingo Molnar <mingo@xxxxxxxxxx> Signed-off-by: Srikar Dronamraju <srikar@xxxxxxxxxxxxxxxxxx> Signed-off-by: Peter Zijlstra (Intel) <peterz@xxxxxxxxxxxxx> Cc: Linus Torvalds <torvalds@xxxxxxxxxxxxxxxxxxxx> Cc: Mel Gorman <mgorman@xxxxxxx> Cc: Mike Galbraith <efault@xxxxxx> Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx> Cc: Rik van Riel <riel@xxxxxxxxxx> Cc: Thomas Gleixner <tglx@xxxxxxxxxxxxx> Link: http://lkml.kernel.org/r/1439290813-6683-2-git-send-email-srikar@xxxxxxxxxxxxxxxxxx Signed-off-by: Ingo Molnar <mingo@xxxxxxxxxx> Signed-off-by: Mel Gorman <mgorman@xxxxxxxxxxxxxxxxxxx> --- kernel/sched/core.c | 6 +++--- kernel/sched/fair.c | 4 ++-- kernel/sched/sched.h | 6 +++--- 3 files changed, 8 insertions(+), 8 deletions(-) diff --git a/kernel/sched/core.c b/kernel/sched/core.c index bcd214e4b4d6..98594051d6d5 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -2124,11 +2124,11 @@ void set_numabalancing_state(bool enabled) sched_feat_set("NO_NUMA"); } #else -__read_mostly bool numabalancing_enabled; +__read_mostly bool sched_numa_balancing; void set_numabalancing_state(bool enabled) { - numabalancing_enabled = enabled; + sched_numa_balancing = enabled; } #endif /* CONFIG_SCHED_DEBUG */ @@ -2138,7 +2138,7 @@ int sysctl_numa_balancing(struct ctl_table *table, int write, { struct ctl_table t; int err; - int state = numabalancing_enabled; + int state = sched_numa_balancing; if (write && !capable(CAP_SYS_ADMIN)) return -EPERM; diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 9a5e60fe721a..33d02d6a3cf4 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -2069,7 +2069,7 @@ void task_numa_fault(int last_cpupid, int mem_node, int pages, int flags) int local = !!(flags & TNF_FAULT_LOCAL); int priv; - if (!numabalancing_enabled) + if (!sched_numa_balancing) return; /* for example, ksmd faulting in a user's mm */ @@ -7811,7 +7811,7 @@ static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) entity_tick(cfs_rq, se, queued); } - if (numabalancing_enabled) + if (sched_numa_balancing) task_tick_numa(rq, curr); } diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index 6d2a119c7ad9..252cfbb78348 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h @@ -1006,13 +1006,13 @@ extern struct static_key sched_feat_keys[__SCHED_FEAT_NR]; #ifdef CONFIG_NUMA_BALANCING #define sched_feat_numa(x) sched_feat(x) #ifdef CONFIG_SCHED_DEBUG -#define numabalancing_enabled sched_feat_numa(NUMA) +#define sched_numa_balancing sched_feat_numa(NUMA) #else -extern bool numabalancing_enabled; +extern bool sched_numa_balancing; #endif /* CONFIG_SCHED_DEBUG */ #else #define sched_feat_numa(x) (0) -#define numabalancing_enabled (0) +#define sched_numa_balancing (0) #endif /* CONFIG_NUMA_BALANCING */ static inline u64 global_rt_period(void) -- 2.6.4 -- To unsubscribe from this list: send the line "unsubscribe stable" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html