On Mon, Aug 29, 2022 at 01:31:31PM -0700, Paul E. McKenney wrote: > On Mon, Aug 29, 2022 at 09:46:22PM +0200, Frederic Weisbecker wrote: > > It's really sad that this is the first time I hear about that. I've been working > > on this code for years now without this usecase in mind. And yet it's fundamental. > > > > I asked several times around about other usecases of rcu_nocbs than nohz_full= > > and nobody reported that. I can hardly even google a significant link > > between power saving and rcu_nocbs= > > > > If this is really used that way for a long time then it's a cruel disconnection > > between users and developers. > > Knowing me, you probably asked about rcu_nocbs and I probably thought > you were asking about nohz_full. :-/ Can't remember but no big deal, now we know about it and we can move forward with that in mind. > > > > > 2) NOCB implies performance issues. > > > > > > Which kinds of? There is slightly worse boot times, but I'm guessing that's do > > > with the extra scheduling overhead of the extra threads which is usually not a > > > problem except that RCU is used in the critical path of boot up (on ChromeOS). > > > > I never measured it myself but executing callbacks on another CPUs, with > > context switches and locking can only involve significant performance issues if callbacks > > are frequent. So it's a tradeoff between power and performance. > > It has indeed been a problem for some workloads in the past. But I don't > know of any recent measurements. And NOCB has gotten at least somewhat > faster over the years. I should try a comparison on a simple kernel build someday. > > > > > 3) We are mixing up two very different things in a single list of callbacks: > > > > lazy callbacks and flooding callbacks, as a result we are adding lots of > > > > off-topic corner cases all around: > > > > * a seperate lazy len field to struct rcu_cblist whose purpose is much more > > > > general than just bypass/lazy > > > > * "lazy" specialized parameters to general purpose cblist management > > > > functions > > > > > > I think just 1 or 2 functions have a new lazy param. It didn't seem too > > > intrusive to me. > > > > What bothers me is that struct cblist has a general purpose and we are adding a field > > and a parameter that is relevant to only one specialized user. > > This does sound like a bad idea, now that you mention it. Joel, if > this is still in place, can it be moved near the rcu_data structure's > bypass-related fields? > > And by the way, thank you for reviewing this patch series! I'll go into a deeper review if we proceed. > > > > So here is a proposal: how about forgetting NOCB for now and instead add a new > > > > RCU_LAZY_TAIL segment in the struct rcu_segcblist right after RCU_NEXT_TAIL? > > > > Then ignore that segment until some timer expiry has been met or the CPU is > > > > known to be busy? Probably some tiny bits need to be tweaked in segcblist > > > > management functions but probably not that much. And also make sure that entrain() > > > > queues to RCU_LAZY_TAIL. > > > > > > > > Then the only difference in the case of NOCB is that we add a new timer to the > > > > nocb group leader instead of a local timer in !NOCB. > > > > > > It sounds reasonable, but I'll go with Paul on the usecase argument - who would > > > actually care about lazy CBs outside of power, and would those guys ever use > > > !NO_CB if they cared about power / battery? > > > > _Everybody_ cares about power. Those who don't yet will very soon ;-) > > Apparently not enough to use CONFIG_RCU_FAST_NO_HZ. Though to be fair, > that option had its own performance issues. And it would not reduce > grace periods anywhere near as much as call_rcu_lazy(). But the problem > was that last I checked on server workloads, the callbacks were mostly > those that could not reasonably be lazy. Right, but like I said, even servers can sometimes find a moment to think about their good life. > > And given the numbers you provided with your measurements, I bet this will > > be significant with !NOCB as well. This is not only delaying callbacks execution, > > this also reduces the frequency of grace periods, and that impact should be > > quite visible. > > > > Note I'm not stricly opposed to the current approach. But I can't say I'm > > comfortable with it. > > > > Can we do a simple test? Would it be possible to affine every rcuo%c/%d kthread > > to the corresponding CPU%d? For example affine rcuop/1 to CPU 1, rcuop/2 to > > CPU2, etc... And then relaunch your measurements on top of that? > > > > The point is that having the callback kthreads affined to their corresponding > > CPUs should elude the power saving advantages of rcu_nocbs=, back to roughly > > a !NOCB behaviour powerwise (except we have context switches). If you find good > > numbers with this setup then you'll find good numbers with !NOCB. > > Another test would be to look at which callbacks are being invoked > on each grace period. We have to have a substantial number of grace > periods having all lazy callbacks before call_rcu_lazy() has any chance > of helping. This would need to happen on a server platform because > Android and ChromeOS data might or might not carry over. Also that yes. Thanks.