Dear RT Folks, I'm pleased to announce the 3.4.97-rt121 stable release. You can get this release via the git tree at: git://git.kernel.org/pub/scm/linux/kernel/git/rt/linux-stable-rt.git branch: v3.4-rt Head SHA1: 819dd8a63540f5e696a4b5c2d7aba657d9e86dce Or to build 3.4.97-rt121 directly, the following patches should be applied: http://www.kernel.org/pub/linux/kernel/v3.x/linux-3.4.tar.xz http://www.kernel.org/pub/linux/kernel/v3.x/patch-3.4.97.xz http://www.kernel.org/pub/linux/kernel/projects/rt/3.4/patch-3.4.97-rt121.patch.xz You can also build from 3.4.97-rt120 by applying the incremental patch: http://www.kernel.org/pub/linux/kernel/projects/rt/3.4/incr/patch-3.4.97-rt120-rt121.patch.xz Enjoy, -- Steve Changes from v3.4.97-rt120: --- Steven Rostedt (1): sched: Do not clear PF_NO_SETAFFINITY flag in select_fallback_rq() Steven Rostedt (Red Hat) (1): Linux 3.4.97-rt121 Thomas Gleixner (1): workqueue: Prevent deadlock/stall on RT Zhao Hongjiang (1): hrtimer:fix the miss of hrtimer_peek_ahead_timers in nort code ---- kernel/hrtimer.c | 6 +++++- kernel/sched/core.c | 13 +++++-------- kernel/workqueue.c | 41 +++++++++++++++++++++++++++++++++++------ localversion-rt | 2 +- 4 files changed, 46 insertions(+), 16 deletions(-) --------------------------- diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 3f7d3b5b0717..afb62c7cc494 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c @@ -1457,7 +1457,11 @@ static int hrtimer_rt_defer(struct hrtimer *timer) #else -static inline void hrtimer_rt_run_pending(void) { } +static inline void hrtimer_rt_run_pending(void) +{ + hrtimer_peek_ahead_timers(); +} + static inline int hrtimer_rt_defer(struct hrtimer *timer) { return 0; } #endif diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 1199054adb2e..5ba55a8b26e0 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -1325,12 +1325,6 @@ out: } } - /* - * Clear PF_THREAD_BOUND, otherwise we wreckage - * migrate_disable/enable. See optimization for - * PF_THREAD_BOUND tasks there. - */ - p->flags &= ~PF_THREAD_BOUND; return dest_cpu; } @@ -3580,9 +3574,8 @@ need_resched: static inline void sched_submit_work(struct task_struct *tsk) { - if (!tsk->state || tsk_is_pi_blocked(tsk)) + if (!tsk->state) return; - /* * If a worker went to sleep, notify and ask workqueue whether * it wants to wake up a task to maintain concurrency. @@ -3592,6 +3585,10 @@ static inline void sched_submit_work(struct task_struct *tsk) if (tsk->flags & PF_WQ_WORKER && !tsk->saved_state) wq_worker_sleeping(tsk); + + if (tsk_is_pi_blocked(tsk)) + return; + /* * If we are going to sleep and we have plugged IO queued, * make sure to submit it to avoid deadlocks. diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 4d21bfdc1637..653d7fccb762 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -319,6 +319,31 @@ static inline int __next_wq_cpu(int cpu, const struct cpumask *mask, (cpu) < WORK_CPU_NONE; \ (cpu) = __next_wq_cpu((cpu), cpu_possible_mask, (wq))) +#ifdef CONFIG_PREEMPT_RT_BASE +static inline void rt_lock_idle_list(struct global_cwq *gcwq) +{ + preempt_disable(); +} +static inline void rt_unlock_idle_list(struct global_cwq *gcwq) +{ + preempt_enable(); +} +static inline void sched_lock_idle_list(struct global_cwq *gcwq) { } +static inline void sched_unlock_idle_list(struct global_cwq *gcwq) { } +#else +static inline void rt_lock_idle_list(struct global_cwq *gcwq) { } +static inline void rt_unlock_idle_list(struct global_cwq *gcwq) { } +static inline void sched_lock_idle_list(struct global_cwq *gcwq) +{ + spin_lock_irq(&gcwq->lock); +} +static inline void sched_unlock_idle_list(struct global_cwq *gcwq) +{ + spin_unlock_irq(&gcwq->lock); +} +#endif + + #ifdef CONFIG_DEBUG_OBJECTS_WORK static struct debug_obj_descr work_debug_descr; @@ -650,10 +675,16 @@ static struct worker *first_worker(struct global_cwq *gcwq) */ static void wake_up_worker(struct global_cwq *gcwq) { - struct worker *worker = first_worker(gcwq); + struct worker *worker; + + rt_lock_idle_list(gcwq); + + worker = first_worker(gcwq); if (likely(worker)) wake_up_process(worker->task); + + rt_unlock_idle_list(gcwq); } /** @@ -696,7 +727,6 @@ void wq_worker_sleeping(struct task_struct *task) cpu = smp_processor_id(); gcwq = get_gcwq(cpu); - spin_lock_irq(&gcwq->lock); /* * The counterpart of the following dec_and_test, implied mb, * worklist not empty test sequence is in insert_work(). @@ -704,11 +734,10 @@ void wq_worker_sleeping(struct task_struct *task) */ if (atomic_dec_and_test(get_gcwq_nr_running(cpu)) && !list_empty(&gcwq->worklist)) { - worker = first_worker(gcwq); - if (worker) - wake_up_process(worker->task); + sched_lock_idle_list(gcwq); + wake_up_worker(gcwq); + sched_unlock_idle_list(gcwq); } - spin_unlock_irq(&gcwq->lock); } /** diff --git a/localversion-rt b/localversion-rt index fff72aad59ab..8188d85cffe0 100644 --- a/localversion-rt +++ b/localversion-rt @@ -1 +1 @@ --rt120 +-rt121 -- To unsubscribe from this list: send the line "unsubscribe linux-rt-users" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html