On Tue, Mar 7, 2023 at 12:19 PM Frederic Weisbecker <frederic@xxxxxxxxxx> wrote: > > On Tue, Mar 07, 2023 at 08:41:17AM -0500, Joel Fernandes wrote: > > On Tue, Mar 7, 2023 at 8:01 AM Frederic Weisbecker <frederic@xxxxxxxxxx> wrote: > > > > > > On Fri, Mar 03, 2023 at 09:38:51PM +0000, Joel Fernandes (Google) wrote: > > > > On many systems, a great deal of boot (in userspace) happens after the > > > > kernel thinks the boot has completed. It is difficult to determine if > > > > the system has really booted from the kernel side. Some features like > > > > lazy-RCU can risk slowing down boot time if, say, a callback has been > > > > added that the boot synchronously depends on. Further expedited callbacks > > > > can get unexpedited way earlier than it should be, thus slowing down > > > > boot (as shown in the data below). > > > > > > > > For these reasons, this commit adds a config option > > > > 'CONFIG_RCU_BOOT_END_DELAY' and a boot parameter rcupdate.boot_end_delay. > > > > Userspace can also make RCU's view of the system as booted, by writing the > > > > time in milliseconds to: /sys/module/rcupdate/parameters/rcu_boot_end_delay > > > > Or even just writing a value of 0 to this sysfs node. > > > > However, under no circumstance will the boot be allowed to end earlier > > > > than just before init is launched. > > > > > > > > The default value of CONFIG_RCU_BOOT_END_DELAY is chosen as 15s. This > > > > suites ChromeOS and also a PREEMPT_RT system below very well, which need > > > > no config or parameter changes, and just a simple application of this patch. A > > > > system designer can also choose a specific value here to keep RCU from marking > > > > boot completion. As noted earlier, RCU's perspective of the system as booted > > > > will not be marker until at least rcu_boot_end_delay milliseconds have passed > > > > or an update is made via writing a small value (or 0) in milliseconds to: > > > > /sys/module/rcupdate/parameters/rcu_boot_end_delay. > > > > > > > > One side-effect of this patch is, there is a risk that a real-time workload > > > > launched just after the kernel boots will suffer interruptions due to expedited > > > > RCU, which previous ended just before init was launched. However, to mitigate > > > > such an issue (however unlikely), the user should either tune > > > > CONFIG_RCU_BOOT_END_DELAY to a smaller value than 15 seconds or write a value > > > > of 0 to /sys/module/rcupdate/parameters/rcu_boot_end_delay, once userspace > > > > boots, and before launching the real-time workload. > > > > > > > > Qiuxu also noted impressive boot-time improvements with earlier version > > > > of patch. An excerpt from the data he shared: > > > > > > > > 1) Testing environment: > > > > OS : CentOS Stream 8 (non-RT OS) > > > > Kernel : v6.2 > > > > Machine : Intel Cascade Lake server (2 sockets, each with 44 logical threads) > > > > Qemu args : -cpu host -enable-kvm, -smp 88,threads=2,sockets=2, … > > > > > > > > 2) OS boot time definition: > > > > The time from the start of the kernel boot to the shell command line > > > > prompt is shown from the console. [ Different people may have > > > > different OS boot time definitions. ] > > > > > > > > 3) Measurement method (very rough method): > > > > A timer in the kernel periodically prints the boot time every 100ms. > > > > As soon as the shell command line prompt is shown from the console, > > > > we record the boot time printed by the timer, then the printed boot > > > > time is the OS boot time. > > > > > > > > 4) Measured OS boot time (in seconds) > > > > a) Measured 10 times w/o this patch: > > > > 8.7s, 8.4s, 8.6s, 8.2s, 9.0s, 8.7s, 8.8s, 9.3s, 8.8s, 8.3s > > > > The average OS boot time was: ~8.7s > > > > > > > > b) Measure 10 times w/ this patch: > > > > 8.5s, 8.2s, 7.6s, 8.2s, 8.7s, 8.2s, 7.8s, 8.2s, 9.3s, 8.4s > > > > The average OS boot time was: ~8.3s. > > > > > > > > Tested-by: Qiuxu Zhuo <qiuxu.zhuo@xxxxxxxxx> > > > > Signed-off-by: Joel Fernandes (Google) <joel@xxxxxxxxxxxxxxxxx> > > > > > > I still don't really like that: > > > > > > 1) It feels like we are curing a symptom for which we don't know the cause. > > > Which RCU write side caller is the source of this slow boot? Some tracepoints > > > reporting the wait duration within synchronize_rcu() calls between the end of > > > the kernel boot and the end of userspace boot may be helpful. > > > > Just to clarify (and I feel we discussed this recently) -- there is no > > callback I am aware of right now causing a slow boot. The reason for > > doing this is we don't have such issues in the future; so it is a > > protection. Note the repeated call outs to the scsi callback and also > > the rcu_barrier() issue previously fixed. Further, we already see > > slight improvements in boot times with disabling lazy during boot (its > > not much but its there). Yes, we should fix issues instead of hiding > > them - but we also would like to improve the user experience -- just > > like we disable lazy and expedited during suspend. > > > > So what is the problem that you really have with this patch even with > > data showing improvements? I actually wanted a mechanism like this > > from the beginning and was trying to get Intel to write the patch, but > > I ended up writing it. > > Let's put it another way: kernel boot is mostly code that won't execute > again. User boot (or rather the kernel part of it) OTOH is code that is > subject to be repeated again. > > A lot of the kernel boot code is __init code that will execute only once. > And there it makes sense to force hurry and expedited because we may easily > miss something and after all this all happens only once, also there is no > interference with userspace, etc... > > User boot OTOH use common kernel code: syscalls, signal, files, etc... And that > code will be called also after the boot. > > So if there is something slowing down user boot, there are some good chances > that this thing slows down userspace in general. > > Therefore we need to know exactly what's going on because the problem may be > bigger than what you observe on boot. These are good points. It motivates me to dig further, as we may be setting ourselves up for longer term problems for shorter term gains otherwise. I am thinking I finish my debugobjects patch soon which adds metadata to callbacks and expose the details via debugfs, and provide it to Qiuxu and ChromeOS folks to run and study the boot time. > > > 2) The kernel boot was already covered before this patch so this is about > > > userspace code calling into the kernel. Is that piece of code also called > > > after the boot? In that case are we missing a conversion from > > > synchronize_rcu() to synchronize_rcu_expedited() somewhere? Because then > > > the problem is more general than just boot. > > > > > > This needs to be analyzed first and if it happens that the issue really > > > needs to be fixed with telling the kernel that userspace has completed > > > booting, eg: because the problem is not in a few callsites that need conversion > > > to expedited but instead in the accumulation of lots of calls that should stay > > > as is: > > > > There is no such callback I am aware off that needs such a conversion > > and I don't think that will help give any guarantees because there is > > no preventing someone from adding a callback that synchronously slows > > boot. The approach here is to put a protection. However, I will do > > some more investigations into what else may be slowing things as I do > > hold a lot of weight for your words! :) > > Kernel boot is already handled and userspace boot can not add a new RCU callback. Right, so that is in line with your point about userspace slowing down even after boot, if I am not mistaken. > > > 3) This arbitrary timeout looks dangerous to me as latency sensitive code > > > may run right after the boot. Either you choose a value that is too low > > > and you miss the optimization or the value is too high and you may break > > > things. > > > > So someone is presenting a timing sensitive workload within 15 seconds > > of boot? Please provide some evidence of that. > > I have no idea, there are billions of computers running out there, it's a disaster... Haha... Linux success sounds like a nice problem to have. ;-) > > The only evidence right now is on the plus side even for the RT system. > > Right it's improving the boot of an RT system, doesn't mean it's not breaking > post boot of others. True. However, I still feel a protection in the future would make sense in general after we finish these investigations. > > > 4) This should be fixed the way you did: > > > a) a kernel parameter like you did > > > b) The init process (systemd?) tells the kernel when it judges that userspace > > > has completed booting. > > > c) Make these interfaces more generic, maybe that information will be useful > > > outside RCU. For example the kernel parameter should be > > > "user_booted_reported" and the sysfs (should be sysctl?): > > > kernel.user_booted = 1 > > > d) But yuck, this means we must know if the init process supports that... > > > > > > For these reasons, let's make sure we know exactly what is going on first. > > > > I can investigate this more and get back to you. > > > > One of the challenges is getting boot tracing working properly. > > Systems do weird things like turning off tracing during boot and/or > > clearing trace buffers. > > Just compare the average and total duration of all synchronize_rcu() calls > (before and after forcing expedited) between launching initand userspace boot > completion. Sure there will be noise but if a difference can be measured before > and after your patch, then a difference might be measureable on tracing as > well... Well of course tracing can induce subtle things... But let's try at > least, we want to know what we are fixing here. You mean using function graph tracer? For the synchronize_rcu() stuff, I'll have to defer to the Qiuxu to try tracing synchronize_rcu() on his PREEMPT_RT system since I don't have access to that system. I can try to provide a patch that will make tracing that easier, but that will be a few days probably as I'm traveling... On ChromeOS we are seeing slight improvements with this patch (though it is not clear whether it is statistically significant). So I have to dig deeper what is going on there. - Joel