On Thu, Sep 19, 2019 at 1:45 PM Alexander E. Patrakov <patrakov@xxxxxxxxx> wrote: > > This already resembles in-kernel haveged (except that it doesn't credit > entropy), and Willy Tarreau said "collect the small entropy where it is, > period" today. So, too many people touched upon the topic in one day, > and therefore I'll bite. I'm one of the people who aren't entirely convinced by the jitter entropy - I definitely believe it exists, I just am not necessarily convinced about the actual entropy calculations. So while I do think we should take things like the cycle counter into account just because I think it's a a useful way to force some noise, I am *not* a huge fan of the jitter entropy driver either, because of the whole "I'm not convinced about the amount of entropy". The whole "third order time difference" thing would make sense if the time difference was some kind of smooth function - which it is at a macro level. But at a micro level, I could easily see the time difference having some very simple pattern - say that your cycle counter isn't really cycle-granular, and the load takes 5.33 "cycles" and you see a time difference pattern of (5, 5, 6, 5, 5, 6, ...). No real entropy at all there, it is 100% reliable. At a macro level, that's a very smooth curve, and you'd say "ok, time difference is 5.3333 (repeating)". But that's not what the jitter entropy code does. It just does differences of differences. And that completely non-random pattern has a first-order difference of 0, 1, 1, 0, 1, 1.. and a second order of 1, 0, 1, 1, 0, and so on forever. So the "jitter entropy" logic will assign that completely repeatable thing entropy, because the delta difference doesn't ever go away. Maybe I misread it. We used to (we still do, but we used to too) do that same third-order delta difference ourselves for the interrupt timing entropy estimation in add_timer_randomness(). But I think it's more valid with something that likely has more noise (interrupt timing really _should_ be noisy). It's not clear that the jitterentropy load really has all that much noise. That said, I'm _also_ not a fan of the user mode models - they happen too late anyway for some users, and as you say, it leaves us open to random (heh) user mode distribution choices that may be more or less broken. I would perhaps be willing to just put my foot down, and say "ok, we'll solve the 'getrandom(0)' issue by just saying that if that blocks too much, we'll do the jitter entropy thing". Making absolutely nobody happy, but working in practice. And maybe encouraging the people who don't like jitter entropy to use GRND_SECURE instead. Linus