On Thu, Nov 05, 2020 at 10:55:51AM -0800, Paul E. McKenney wrote: > On Tue, Nov 03, 2020 at 09:26:03AM -0500, Joel Fernandes (Google) wrote: > > Memory barriers are needed when updating the full length of the > > segcblist, however it is not fully clearly why one is needed before and > > after. This patch therefore adds additional comments to the function > > header to explain it. > > > > Signed-off-by: Joel Fernandes (Google) <joel@xxxxxxxxxxxxxxxxx> > > Looks good, thank you! As always, I could not resist the urge to > do a bit of wordsmithing, so that the queued commit is as shown > below. Please let me know if I messed anything up. > Thanx, Paul > > ------------------------------------------------------------------------ > > commit 7dac7adefcae7558b3a85a16f51186d621623733 > Author: Joel Fernandes (Google) <joel@xxxxxxxxxxxxxxxxx> > Date: Tue Nov 3 09:26:03 2020 -0500 > > rcu/segcblist: Add additional comments to explain smp_mb() > > One counter-intuitive property of RCU is the fact that full memory > barriers are needed both before and after updates to the full > (non-segmented) length. This patch therefore helps to assist the > reader's intuition by adding appropriate comments. > > [ paulmck: Wordsmithing. ] > Signed-off-by: Joel Fernandes (Google) <joel@xxxxxxxxxxxxxxxxx> > Signed-off-by: Paul E. McKenney <paulmck@xxxxxxxxxx> > > diff --git a/kernel/rcu/rcu_segcblist.c b/kernel/rcu/rcu_segcblist.c > index bb246d8..b6dda7c 100644 > --- a/kernel/rcu/rcu_segcblist.c > +++ b/kernel/rcu/rcu_segcblist.c > @@ -94,17 +94,77 @@ static void rcu_segcblist_set_len(struct rcu_segcblist *rsclp, long v) > * field to disagree with the actual number of callbacks on the structure. > * This increase is fully ordered with respect to the callers accesses > * both before and after. > + * > + * So why on earth is a memory barrier required both before and after > + * the update to the ->len field??? > + * > + * The reason is that rcu_barrier() locklessly samples each CPU's ->len > + * field, and if a given CPU's field is zero, avoids IPIing that CPU. > + * This can of course race with both queuing and invoking of callbacks. > + * Failng to correctly handle either of these races could result in > + * rcu_barrier() failing to IPI a CPU that actually had callbacks queued > + * which rcu_barrier() was obligated to wait on. And if rcu_barrier() > + * failed to wait on such a callback, unloading certain kernel modules > + * would result in calls to functions whose code was no longer present in > + * the kernel, for but one example. > + * > + * Therefore, ->len transitions from 1->0 and 0->1 have to be carefully > + * ordered with respect with both list modifications and the rcu_barrier(). > + * > + * The queuing case is CASE 1 and the invoking case is CASE 2. > + * > + * CASE 1: Suppose that CPU 0 has no callbacks queued, but invokes > + * call_rcu() just as CPU 1 invokes rcu_barrier(). CPU 0's ->len field > + * will transition from 0->1, which is one of the transitions that must be > + * handled carefully. Without the full memory barriers before the ->len > + * update and at the beginning of rcu_barrier(), the following could happen: > + * > + * CPU 0 CPU 1 > + * > + * call_rcu(). > + * rcu_barrier() sees ->len as 0. > + * set ->len = 1. > + * rcu_barrier() does nothing. > + * module is unloaded. > + * callback invokes unloaded function! > + * > + * With the full barriers, any case where rcu_barrier() sees ->len as 0 will > + * have unambiguously preceded the return from the racing call_rcu(), which > + * means that this call_rcu() invocation is OK to not wait on. After all, > + * you are supposed to make sure that any problematic call_rcu() invocations > + * happen before the rcu_barrier(). Unfortunately, I did not understand your explanation. To me the barrier *before* the setting of length is needed on CPU0 only for 1->0 transition (Dequeue). Where as in your example above, it is for enqueue. This was case 1 in my patch: + * To illustrate the problematic scenario to avoid: + * P0 (what P1 sees) P1 + * set len = 0 + * rcu_barrier sees len as 0 + * dequeue from list + * rcu_barrier does nothing. + * Here, P1 should see the transition of 1->0 *after* the CB is dequeued. Which means you needed a memory barrier *before* the setting of len from 1->0 and *after* the dequeue. IOW, rcu_barrier should 'see' the memory ordering as: 1. dequeue 2. set len from 1 -> 0. For the enqueue case, it is the reverse, rcu_barrier should see: 1. set len from 0 -> 1 2. enqueue Either way, the point I think I was trying to make is that the length should always be seen as non-zero if the list is non-empty. Basically, the rcu_barrier() should always not do the fast-path if the list is non-empty. Worst-case it might do the slow-path when it is not necessary, but it should never do the fast-path when it was not supposed to. Thoughts? thanks, - Joel > + * > + * > + * CASE 2: Suppose that CPU 0 is invoking its last callback just as CPU 1 invokes > + * rcu_barrier(). CPU 0's ->len field will transition from 1->0, which is one > + * of the transitions that must be handled carefully. Without the full memory > + * barriers after the ->len update and at the end of rcu_barrier(), the following > + * could happen: > + * > + * CPU 0 CPU 1 > + * > + * start invoking last callback > + * set ->len = 0 (reordered) > + * rcu_barrier() sees ->len as 0 > + * rcu_barrier() does nothing. > + * module is unloaded > + * callback executing after unloaded! > + * > + * With the full barriers, any case where rcu_barrier() sees ->len as 0 > + * will be fully ordered after the completion of the callback function, > + * so that the module unloading operation is completely safe. > + * > */ > void rcu_segcblist_add_len(struct rcu_segcblist *rsclp, long v) > { > #ifdef CONFIG_RCU_NOCB_CPU > - smp_mb__before_atomic(); /* Up to the caller! */ > + smp_mb__before_atomic(); // Read header comment above. > atomic_long_add(v, &rsclp->len); > - smp_mb__after_atomic(); /* Up to the caller! */ > + smp_mb__after_atomic(); // Read header comment above. > #else > - smp_mb(); /* Up to the caller! */ > + smp_mb(); // Read header comment above. > WRITE_ONCE(rsclp->len, rsclp->len + v); > - smp_mb(); /* Up to the caller! */ > + smp_mb(); // Read header comment above. > #endif > } >