* Eric Biggers <ebiggers@xxxxxxxxxx> wrote: > From: Eric Biggers <ebiggers@xxxxxxxxxx> > > Currently kernel-mode FPU is not always usable in softirq context on > x86, since softirqs can nest inside a kernel-mode FPU section in task > context, and nested use of kernel-mode FPU is not supported. > > Therefore, x86 SIMD-optimized code that can be called in softirq context > has to sometimes fall back to non-SIMD code. There are two options for > the fallback, both of which are pretty terrible: > > (a) Use a scalar fallback. This can be 10-100x slower than vectorized > code because it cannot use specialized instructions like AES, SHA, > or carryless multiplication. > > (b) Execute the request asynchronously using a kworker. In other > words, use the "crypto SIMD helper" in crypto/simd.c. > > Currently most of the x86 en/decryption code (skcipher and aead > algorithms) uses option (b), since this avoids the slow scalar fallback > and it is easier to wire up. But option (b) is still really bad for its > own reasons: > > - Punting the request to a kworker is bad for performance too. > > - It forces the algorithm to be marked as asynchronous > (CRYPTO_ALG_ASYNC), preventing it from being used by crypto API > users who request a synchronous algorithm. That's another huge > performance problem, which is especially unfortunate for users who > don't even do en/decryption in softirq context. > > - It makes all en/decryption operations take a detour through > crypto/simd.c. That involves additional checks and an additional > indirect call, which slow down en/decryption for *everyone*. > > Fortunately, the skcipher and aead APIs are only usable in task and > softirq context in the first place. Thus, if kernel-mode FPU were to > be reliably usable in softirq context, no fallback would be needed. > Indeed, other architectures such as arm, arm64, and riscv have > already done this. > > Therefore, this patch updates x86 accordingly to reliably support > kernel-mode FPU in softirqs. > > This is done by just disabling softirq processing in kernel-mode FPU > sections (when hardirqs are not already disabled), as that prevents the > nesting that was problematic. > > This will delay some softirqs slightly, but only ones that would have > otherwise been nested inside a task context kernel-mode FPU section. > Any such softirqs would have taken the slow fallback path before if they > tried to do any en/decryption. Now these softirqs will just run at the > end of the task context kernel-mode FPU section (since local_bh_enable() > runs pending softirqs) and will no longer take the slow fallback path. > > Alternatives considered: > > - Make kernel-mode FPU sections fully preemptible. This would require > growing task_struct by another struct fpstate which is more than 2K. So that's something that will probably happen once the kernel is built using APX anyway? > - Make softirqs save/restore the kernel-mode FPU state to a per-CPU > struct fpstate when nested use is detected. Somewhat interesting, but > seems unnecessary when a simpler solution exists. So: > void kernel_fpu_begin_mask(unsigned int kfpu_mask) > { > - preempt_disable(); > + if (!irqs_disabled()) > + fpregs_lock(); > + if (!irqs_disabled()) > + fpregs_unlock(); So why is the irqs_disabled() check needed here? (On x86 it can be a bit expensive at times, because the IRQ flag has to be loaded, including all flags, so basically it's a soft synchronization point of a sort.) Ie. why cannot we simply do a local_bh_disable()/enable() pair (on !RT), ie. fpregs_lock()/fpregs_unlock()? local_bh_disable() is very similar in cost to preempt_disable(), both are increasing the preempt_count. Thanks, Ingo