* Arun R Bharadwaj <arun@xxxxxxxxxxxxxxxxxx> [2009-03-16 16:40:45]: This patch migrates all non pinned timers and hrtimers to the current idle load balancer, from all the idle CPUs. Timers firing on busy CPUs are not migrated. Signed-off-by: Arun R Bharadwaj <arun@xxxxxxxxxxxxxxxxxx> --- include/linux/sched.h | 9 +++++++++ kernel/hrtimer.c | 12 +++++++++++- kernel/sched.c | 5 +++++ kernel/timer.c | 13 ++++++++++++- 4 files changed, 37 insertions(+), 2 deletions(-) Index: linux.trees.git/kernel/timer.c =================================================================== --- linux.trees.git.orig/kernel/timer.c +++ linux.trees.git/kernel/timer.c @@ -37,6 +37,7 @@ #include <linux/delay.h> #include <linux/tick.h> #include <linux/kallsyms.h> +#include <linux/sched.h> #include <asm/uaccess.h> #include <asm/unistd.h> @@ -606,7 +607,7 @@ __mod_timer(struct timer_list *timer, un { struct tvec_base *base, *new_base; unsigned long flags; - int ret; + int ret, current_cpu, preferred_cpu; ret = 0; @@ -627,6 +628,16 @@ __mod_timer(struct timer_list *timer, un new_base = __get_cpu_var(tvec_bases); + current_cpu = smp_processor_id(); + preferred_cpu = get_nohz_load_balancer(); + if (get_sysctl_timer_migration() && idle_cpu(current_cpu) && + !pinned && preferred_cpu != -1) { + new_base = per_cpu(tvec_bases, preferred_cpu); + timer_set_base(timer, new_base); + timer->expires = expires; + internal_add_timer(new_base, timer); + goto out_unlock; + } if (base != new_base) { /* * We are trying to schedule the timer on the local CPU. Index: linux.trees.git/kernel/hrtimer.c =================================================================== --- linux.trees.git.orig/kernel/hrtimer.c +++ linux.trees.git/kernel/hrtimer.c @@ -43,6 +43,8 @@ #include <linux/seq_file.h> #include <linux/err.h> #include <linux/debugobjects.h> +#include <linux/sched.h> +#include <linux/timer.h> #include <asm/uaccess.h> @@ -198,8 +200,16 @@ switch_hrtimer_base(struct hrtimer *time { struct hrtimer_clock_base *new_base; struct hrtimer_cpu_base *new_cpu_base; + int current_cpu, preferred_cpu; + + current_cpu = smp_processor_id(); + preferred_cpu = get_nohz_load_balancer(); + if (get_sysctl_timer_migration() && !pinned && preferred_cpu != -1 + && idle_cpu(current_cpu)) + new_cpu_base = &per_cpu(hrtimer_bases, preferred_cpu); + else + new_cpu_base = &__get_cpu_var(hrtimer_bases); - new_cpu_base = &__get_cpu_var(hrtimer_bases); new_base = &new_cpu_base->clock_base[base->index]; if (base != new_base) { Index: linux.trees.git/include/linux/sched.h =================================================================== --- linux.trees.git.orig/include/linux/sched.h +++ linux.trees.git/include/linux/sched.h @@ -265,6 +265,7 @@ static inline int select_nohz_load_balan } #endif +extern int get_nohz_load_balancer(void); /* * Only dump TASK_* tasks. (0 for all tasks) */ @@ -1769,6 +1770,14 @@ int sched_nr_latency_handler(struct ctl_ struct file *file, void __user *buffer, size_t *length, loff_t *ppos); #endif +static inline int get_sysctl_timer_migration(void) +{ +#ifdef CONFIG_SCHED_DEBUG + return sysctl_timer_migration; +#else + return 1; +#endif +} extern unsigned int sysctl_sched_rt_period; extern int sysctl_sched_rt_runtime; Index: linux.trees.git/kernel/sched.c =================================================================== --- linux.trees.git.orig/kernel/sched.c +++ linux.trees.git/kernel/sched.c @@ -4009,6 +4009,11 @@ static struct { .load_balancer = ATOMIC_INIT(-1), }; +int get_nohz_load_balancer(void) +{ + return atomic_read(&nohz.load_balancer); +} + /* * This routine will try to nominate the ilb (idle load balancing) * owner among the cpus whose ticks are stopped. ilb owner will do the idle _______________________________________________ linux-pm mailing list linux-pm@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linux-foundation.org/mailman/listinfo/linux-pm