On Wed, Feb 16, 2022 at 5:05 AM Greg KH <gregkh@xxxxxxxxxxxxxxxxxxx> wrote: > > On Tue, Feb 15, 2022 at 04:32:48PM -0500, Brian Geffon wrote: > > On Tue, Feb 15, 2022 at 2:45 PM Greg KH <gregkh@xxxxxxxxxxxxxxxxxxx> wrote: > > > > > > On Tue, Feb 15, 2022 at 11:22:33AM -0800, Brian Geffon wrote: > > > > When eagerly switching PKRU in switch_fpu_finish() it checks that > > > > current is not a kernel thread as kernel threads will never use PKRU. > > > > It's possible that this_cpu_read_stable() on current_task > > > > (ie. get_current()) is returning an old cached value. To resolve this > > > > reference next_p directly rather than relying on current. > > > > > > > > As written it's possible when switching from a kernel thread to a > > > > userspace thread to observe a cached PF_KTHREAD flag and never restore > > > > the PKRU. And as a result this issue only occurs when switching > > > > from a kernel thread to a userspace thread, switching from a non kernel > > > > thread works perfectly fine because all that is considered in that > > > > situation are the flags from some other non kernel task and the next fpu > > > > is passed in to switch_fpu_finish(). > > > > > > > > This behavior only exists between 5.2 and 5.13 when it was fixed by a > > > > rewrite decoupling PKRU from xstate, in: > > > > commit 954436989cc5 ("x86/fpu: Remove PKRU handling from switch_fpu_finish()") > > > > > > > > Unfortunately backporting the fix from 5.13 is probably not realistic as > > > > it's part of a 60+ patch series which rewrites most of the PKRU handling. > > > > > > > > Fixes: 0cecca9d03c9 ("x86/fpu: Eager switch PKRU state") > > > > Signed-off-by: Brian Geffon <bgeffon@xxxxxxxxxx> > > > > Signed-off-by: Willis Kung <williskung@xxxxxxxxxx> > > > > Tested-by: Willis Kung <williskung@xxxxxxxxxx> > > > > Cc: <stable@xxxxxxxxxxxxxxx> # v5.4.x > > > > Cc: <stable@xxxxxxxxxxxxxxx> # v5.10.x > > > > --- > > > > arch/x86/include/asm/fpu/internal.h | 13 ++++++++----- > > > > arch/x86/kernel/process_32.c | 6 ++---- > > > > arch/x86/kernel/process_64.c | 6 ++---- > > > > 3 files changed, 12 insertions(+), 13 deletions(-) > > > > > > So this is ONLY for 5.4.y and 5.10.y? I'm really really loath to take > > > non-upstream changes as 95% of the time (really) it goes wrong. > > > > That's correct, this bug was introduced in 5.2 and that code was > > completely refactored in 5.13 indirectly fixing it. > Hi Greg, > What series of commits contain that work? This is the series, https://lore.kernel.org/all/20210623120127.327154589@xxxxxxxxxxxxx/, it does quite a bit of cleanup to correct the larger problem of having pkru merged into xstate. > And again, why not just take them? What is wrong with that if this is > such a big issue? Anything is possible I suppose but looking through the series it seems that it's not going to apply cleanly so we're going to end up with something that, if we made it work, would look very different and would touch a much larger cross section of code. If the concern here is risk of things going wrong, attempting to pull back or cherry-pick parts of this series seems riskier. However, if we don't attempt to pull back all those patches I will likely be proposing at least one more small patch for 5.4 and 5.10 to fix PKRU in these kernels because right now it's broken, particularly on AMD processors as Dave mentioned. > > > > How was this tested, and what do the maintainers of this subsystem > > > think? And will you be around to fix the bugs in this when they are > > > found? > > > > This has been trivial to reproduce, I've used a small repro which I've > > put here: https://gist.github.com/bgaff/9f8cbfc8dd22e60f9492e4f0aff8f04f > > , I also was able to reproduce this using the protection_keys self > > tests on a 11th Gen Core i5-1135G7. I'm happy to commit to addressing > > any bugs that may appear. I'll see what the maintainers say, but there > > is also a smaller fix that just involves using this_cpu_read() in > > switch_fpu_finish() for this specific issue, although that approach > > isn't as clean. > > Can you add the test to the in-kernel tests so that we make sure it is > fixed and never comes back? I'm already able to reproduce it with the kernel selftests. To be honest, I'm not sure why this hasn't been reported yet. I could be doing something horribly wrong. But it seems the likely reason is that my compiler is doing what it's allowed to do, which is optimize the load of current_task. current -> get_current() -> this_cpu_read_stable(current_task) is allowed to read a cached value. Perhaps gcc is just not taking advantage of that optimization, I'm not sure. But writing to current_task and then immediately reading it back via this_cpu_read_stable() can be problematic for this reason, and the disassembled code shows that this happening. Brian > > thanks, > > greg k-h