Am 01.12.2023 um 19:35 schrieb Johannes Berg:
[I guess we should keep the CCs so other see it]
Looking at the stuck check it will be bogus in simulations.
True.
You might as well ifdef that instead.
If a simulation is running insert the entropy regardless and do not compute the derivatives used in the check.
Actually you mostly don't want anything inserted in that case, so it's
not bad to skip it.
I was mostly thinking this might be better than adding a completely
unrelated ifdef. Also I guess in real systems with a bad implementation
of random_get_entropy(), the second/third derivates might be
constant/zero for quite a while, so may be better to abort?
Maybe dump question: could we simply implement a timex.h for UM which
delegates in non-timetravel mode to the x86 variant and otherwise pull
some randomness from the host or from a file/pipe configurable from the
UML commandline for random_get_entropy()?
I would say, if the random jitter is truly deterministic for a
simulation, that seems to be good enough.
Said that, it would be nice to be able to configure all random sources
to pull entropy from some file that we are able to configure from the
command line, but that is a different topic.
In any case, I couldn't figure out any way to not configure this into
the kernel when any kind of crypto is also in ...
johannes