Add sysctl knobs just for easier debugging/testing, to tune the maximum batch size, maximum time to wait before flush, and turning off the feature entirely. Signed-off-by: Joel Fernandes (Google) <joel@xxxxxxxxxxxxxxxxx> --- include/linux/sched/sysctl.h | 4 ++++ kernel/rcu/lazy.c | 12 ++++++++++-- kernel/sysctl.c | 23 +++++++++++++++++++++++ 3 files changed, 37 insertions(+), 2 deletions(-) diff --git a/include/linux/sched/sysctl.h b/include/linux/sched/sysctl.h index c19dd5a2c05c..55ffc61beed1 100644 --- a/include/linux/sched/sysctl.h +++ b/include/linux/sched/sysctl.h @@ -16,6 +16,10 @@ enum { sysctl_hung_task_timeout_secs = 0 }; extern unsigned int sysctl_sched_child_runs_first; +extern unsigned int sysctl_rcu_lazy; +extern unsigned int sysctl_rcu_lazy_batch; +extern unsigned int sysctl_rcu_lazy_jiffies; + enum sched_tunable_scaling { SCHED_TUNABLESCALING_NONE, SCHED_TUNABLESCALING_LOG, diff --git a/kernel/rcu/lazy.c b/kernel/rcu/lazy.c index 55e406cfc528..0af9fb67c92b 100644 --- a/kernel/rcu/lazy.c +++ b/kernel/rcu/lazy.c @@ -12,6 +12,10 @@ // How much to wait before flushing? #define MAX_LAZY_JIFFIES 10000 +unsigned int sysctl_rcu_lazy_batch = MAX_LAZY_BATCH; +unsigned int sysctl_rcu_lazy_jiffies = MAX_LAZY_JIFFIES; +unsigned int sysctl_rcu_lazy = 1; + // We cast lazy_rcu_head to rcu_head and back. This keeps the API simple while // allowing us to use lockless list node in the head. Also, we use BUILD_BUG_ON // later to ensure that rcu_head and lazy_rcu_head are of the same size. @@ -49,6 +53,10 @@ void call_rcu_lazy(struct rcu_head *head_rcu, rcu_callback_t func) struct lazy_rcu_head *head = (struct lazy_rcu_head *)head_rcu; struct rcu_lazy_pcp *rlp; + if (!sysctl_rcu_lazy) { + return call_rcu(head_rcu, func); + } + preempt_disable(); rlp = this_cpu_ptr(&rcu_lazy_pcp_ins); preempt_enable(); @@ -67,11 +75,11 @@ void call_rcu_lazy(struct rcu_head *head_rcu, rcu_callback_t func) llist_add(&head->llist_node, &rlp->head); // Flush queue if too big - if (atomic_inc_return(&rlp->count) >= MAX_LAZY_BATCH) { + if (atomic_inc_return(&rlp->count) >= sysctl_rcu_lazy_batch) { lazy_rcu_flush_cpu(rlp); } else { if (!delayed_work_pending(&rlp->work)) { - schedule_delayed_work(&rlp->work, MAX_LAZY_JIFFIES); + schedule_delayed_work(&rlp->work, sysctl_rcu_lazy_jiffies); } } } diff --git a/kernel/sysctl.c b/kernel/sysctl.c index 5ae443b2882e..2ba830ca71ec 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c @@ -1659,6 +1659,29 @@ static struct ctl_table kern_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, +#ifdef CONFIG_RCU_LAZY + { + .procname = "rcu_lazy", + .data = &sysctl_rcu_lazy, + .maxlen = sizeof(unsigned int), + .mode = 0644, + .proc_handler = proc_dointvec, + }, + { + .procname = "rcu_lazy_batch", + .data = &sysctl_rcu_lazy_batch, + .maxlen = sizeof(unsigned int), + .mode = 0644, + .proc_handler = proc_dointvec, + }, + { + .procname = "rcu_lazy_jiffies", + .data = &sysctl_rcu_lazy_jiffies, + .maxlen = sizeof(unsigned int), + .mode = 0644, + .proc_handler = proc_dointvec, + }, +#endif #ifdef CONFIG_SCHEDSTATS { .procname = "sched_schedstats", -- 2.36.0.550.gb090851708-goog