> On Nov 30, 2022, at 11:55 AM, Boqun Feng <boqun.feng@xxxxxxxxx> wrote: > > On Wed, Nov 30, 2022 at 04:39:02PM +0800, Pingfan Liu wrote: >> ssp->srcu_cb_mutex is introduced to allow the other srcu state machine >> to advance as soon as possible. But according to the implement of >> workqueue, the same work_struct is serialized and can not run >> concurrently in fact. >> >> Quoting from Documentation/core-api/workqueue.rst >> " >> Non-reentrance Conditions >> ========================= >> >> Workqueue guarantees that a work item cannot be re-entrant if the following >> conditions hold after a work item gets queued: >> >> 1. The work function hasn't been changed. >> 2. No one queues the work item to another workqueue. >> 3. The work item hasn't been reinitiated. >> " >> >> To allow the concurrence to some extent, it can be achieved by changing >> the work function to break the conditions. As a result, when >> srcu_gp_end() releases srcu_gp_mutex, a new state machine can begin. >> >> Signed-off-by: Pingfan Liu <kernelfans@xxxxxxxxx> >> Cc: Lai Jiangshan <jiangshanlai@xxxxxxxxx> >> Cc: "Paul E. McKenney" <paulmck@xxxxxxxxxx> >> Cc: Frederic Weisbecker <frederic@xxxxxxxxxx> >> Cc: Josh Triplett <josh@xxxxxxxxxxxxxxxx> >> Cc: Steven Rostedt <rostedt@xxxxxxxxxxx> >> Cc: Mathieu Desnoyers <mathieu.desnoyers@xxxxxxxxxxxx> >> Cc: "Zhang, Qiang1" <qiang1.zhang@xxxxxxxxx> >> To: rcu@xxxxxxxxxxxxxxx >> --- >> kernel/rcu/srcutree.c | 19 +++++++++++++++++++ >> 1 file changed, 19 insertions(+) >> >> diff --git a/kernel/rcu/srcutree.c b/kernel/rcu/srcutree.c >> index 1c304fec89c0..56dd9bb2c8b8 100644 >> --- a/kernel/rcu/srcutree.c >> +++ b/kernel/rcu/srcutree.c >> @@ -75,6 +75,7 @@ static bool __read_mostly srcu_init_done; >> static void srcu_invoke_callbacks(struct work_struct *work); >> static void srcu_reschedule(struct srcu_struct *ssp, unsigned long delay); >> static void process_srcu(struct work_struct *work); >> +static void process_srcu_wrap(struct work_struct *work); >> static void srcu_delay_timer(struct timer_list *t); >> >> /* Wrappers for lock acquisition and release, see raw_spin_lock_rcu_node(). */ >> @@ -763,6 +764,11 @@ static void srcu_gp_end(struct srcu_struct *ssp) >> cbdelay = 0; >> >> WRITE_ONCE(ssp->srcu_last_gp_end, ktime_get_mono_fast_ns()); >> + /* Change work func so work can be concurrent */ >> + if (ssp->work.work.func == process_srcu_wrap) >> + ssp->work.work.func = process_srcu; >> + else >> + ssp->work.work.func = process_srcu_wrap; > > This looks really hacky ;-) It would be good that workqueue has an API > to allow "resetting" a work. > > Do you have any number of the potential performance improvement? Agreed. Another question - have you verified the workqueue behavior after this change (say using tracing), regardless of what the workqueue documentation says? Thanks, - Joel > > Regards, > Boqun > >> rcu_seq_end(&ssp->srcu_gp_seq); >> gpseq = rcu_seq_current(&ssp->srcu_gp_seq); >> if (ULONG_CMP_LT(ssp->srcu_gp_seq_needed_exp, gpseq)) >> @@ -1637,6 +1643,19 @@ static void process_srcu(struct work_struct *work) >> srcu_reschedule(ssp, curdelay); >> } >> >> +/* >> + * The ssp->work is expected to be concurrent to some extent, but the current >> + * workqueue does not support the concurrence on the same work. (Refer to the >> + * section "Non-reentrance Conditions" in the file workqueue.rst) >> + * Resolving it by changing the work func. >> + * >> + * Prevent compilering from optimizing out it. >> + */ >> +static __used void process_srcu_wrap(struct work_struct *work) >> +{ >> + process_srcu(work); >> +} >> + >> void srcutorture_get_gp_data(enum rcutorture_type test_type, >> struct srcu_struct *ssp, int *flags, >> unsigned long *gp_seq) >> -- >> 2.31.1 >>