Adding some additional RCU experts on CC. On Fri, Feb 23, 2024 at 03:22:13PM +0800, Ze Gao wrote: > Dear Paul, > > I'm reading this article [1] by you and have doubts about how these > memory guarantees are provided in Linux. > > As quoted from [1], RCU needs to make sure: > > >1. Each CPU that has an RCU read-side critical section that begins before synchronize_rcu() >starts is guaranteed to execute a full memory barrier between the time that the RCU read-side >critical section ends and the time that synchronize_rcu() returns. Without this guarantee, a pre-> existing RCU read-side critical section might hold a reference to the newly removed struct foo > after the kfree() on line 14 of remove_gp_synchronous(). > > >2. Each CPU that has an RCU read-side critical section that ends after synchronize_rcu() >returns is guaranteed to execute a full memory barrier between the time that synchronize_rcu() >begins and the time that the RCU read-side critical section begins. Without this guarantee, a >later RCU read-side critical section running after the kfree() on line 14 of >remove_gp_synchronous() might later run do_something_gp() and find the newly deleted >struct foo. > > FWIW, I can understand the necessity for smp_mb() for both cases you > have posted in the quick quiz. But I'm really curious about how > Tree-RCU with !CONFIG_PREEMPT provides such > guarantees where obviously smp_mb() cannot be provided by > rcu_read_{lock, unlock}. > > Also, I see you've answered another related question in a different > quiz about how RCU infers quiescent states, After digging for a > while, I try to make the best guess of the answers so here is my > understanding: > > 1) both Guarantee #1 and #2 are to make sure any loads/stores by > updater are properly propagated to newly arriving readers so to avoid > > 2) according to 1), so in cases where rcu_read_{lock, unlock} > generates no code, the smp_mb() is actually provided by where the CPU > reports a quiescent state. IOW, Tree-RCU with !CONFIG_PREEMPT provides > Guarantee #1 and #2 by extending the grace period requested by > synchronize_rcu(), so in both following cases you mentioned, smp_mb() > can be provided by accessing rcu_node structure's ->lock field > mentioned in [2]. So far so good. > Case A: > CPU 1: rcu_read_lock() > CPU 1: q = rcu_dereference(gp); /* Very likely to return p. */ > CPU 0: list_del_rcu(p); > CPU 0: synchronize_rcu() starts. > CPU 1: do_something_with(q->a); /* No smp_mb(), so might happen after > kfree(). */ > CPU 1: rcu_read_unlock() Here CPU 1 must report its quiescent state to the RCU core code. This involves locking and memory barriers that ensure that the rcu_read_lock() is seen by all to precede the return from synchronize_rcu(). > CPU 0: synchronize_rcu() returns. > CPU 0: kfree(p); > > Case B: > CPU 0: list_del_rcu(p); > CPU 0: synchronize_rcu() starts. If synchronize_rcu() cannot prove that it started before a given rcu_read_lock(), it must assume that the rcu_read_lock() was there first. This is mediated by the RCU core code that sees that the grace period started. So if CPU 1 sees the grace period as having started before the rcu_read_lock(), then the grace period will not wait for the corresponding rcu_read_unlock(). Otherwise, it will wait, avoiding the situation shown below. Again, with locks and memory barriers providing ordering. > CPU 1: rcu_read_lock() > CPU 1: q = rcu_dereference(gp); /* Might return p if no memory barrier. */ > CPU 0: synchronize_rcu() returns. > CPU 0: kfree(p); > CPU 1: do_something_with(q->a); /* Boom!!! */ > CPU 1: rcu_read_unlock() > > I should've studied the code to find the answer, but It may take years > to know the details :). > (no kidding given the large codebase and its complicacy). So I'm > being bold to directly write to you for help. Please forgive me for > being reckless. > > Appreciate your excellent docs on this topic and look forward to your > comments to clear my doubts. As you say, the code is non-trivial. Something about needing to scale to systems having thousands of CPUs, conserve energy on battery-powered systems, tolerate CPUs coming and going (for example, in suspend/resume), help to provide deep sub-millisecond real-time latencies, work properly when being used before it is initialized during kernel boot, and so on. So [3] is the summary I wrote up to communicate how this works (sadly, all 648 lines of it), with [4] being a graphical summary of how things work. Studying these carefully over a period of time has proven quite helpful to some people. Section 4 of [5] gives an overview, but with less detail. It might be a good place to get a running start for digging through [3][4]. We have a bunch of low-level RCU design documents publicly available [6], which might also be helpful. Or you might wish to take a look at a simpler RCU implementation, which faces the same issues, but which is easier to get one's head around. Userspace RCU is such an implementation (and there are many others [8]), and its web page [8] points to a great deal of documentation, perhaps most notably the February 2012 IEEE TPDS paper. This is not easy stuff, and I encourage you to keep at it! Thanx, Paul > Regards, > -- Ze > --- > [1]: https://dri.freedesktop.org/docs/drm/RCU/Design/Requirements/Requirements.html#memory-barrier-guarantees > [2]: https://dri.freedesktop.org/docs/drm/RCU/Design/Memory-Ordering/Tree-RCU-Memory-Ordering.html#what-is-tree-rcu-s-grace-period-memory-ordering-guarantee [3]: https://dri.freedesktop.org/docs/drm/RCU/Design/Memory-Ordering/Tree-RCU-Memory-Ordering.html#what-is-tree-rcu-s-grace-period-memory-ordering-guarantee [4] https://www.kernel.org/doc/Documentation/RCU/Design/Memory-Ordering/Tree-RCU-Diagram.html [5] https://github.com/michaliskok/rcu/blob/master/rcupaper.pdf [6] https://docs.google.com/document/d/1GCdQC8SDbb54W1shjEXqGZ0Rq8a6kIeYutdSIajfpLA/edit?usp=sharing [7] https://kernel.org/pub/linux/kernel/people/paulmck/perfbook/perfbook.html Section 9.5.5 has a list of uses and implementations. [8] https://liburcu.org