On Sat, Jun 03, 2023 at 08:22:27AM -0700, Chang S. Bae wrote: > == Key Handle Restriction == > > A key handle may be encoded with some restrictions. It's unclear what this means. Please avoid passive tense and the word "may" like this. I think you mean something like "The AES-KL instruction set supports selecting key usage restrictions at key handle creation time." > Restrict every handle only available in kernel mode via setkey(). I think you mean something like "Restrict all key handles created by the kernel to kernel mode use only." Can you also mention why you are doing this? I suppose it might as well be done, but I'm not seeing how it would actually matter. What other sorts of key usage restrictions does AES-KL support? Are any other ones useful here? > Subsequently the key handle could be corrupted or fail with handle > restrictions. Then, encrypt()/decrypt() returns -EINVAL. Aren't these scenarios actually impossible? At least without memory corruption. > == Userspace Exposition == > > Some hardware implementations may have some performance penalties. Likewise, please avoid vague statements like this. This makes it unclear whether this is something that happens in the real world or whether it's just theoretical. You indeed have actual benchmark results that show that AES-KL is much slower than AES-NI on current CPUs, right? > But, for disk encryption, storage bandwidth may be the bottleneck before > encryption bandwidth. Again, please try to be less vague. E.g. "With a slow storage device, storage bandwidth is the bottleneck, even if disk encryption is enabled..." > Thus, advertise it with a unique name 'xts-aes-aeskl' in /proc/crypto while > not replacing AES-NI under the generic name 'xts(aes)' with a lower priority. The above sentence seems to say that xts-aes-aeskl does *not* have a lower priority than xts-aes-aesni. But actually it does. > Then, the performance is unlikely better than 64-bit which has already a gap > vs. AES-NI. I don't understand what this sentence is trying to say. > diff --git a/arch/x86/crypto/Kconfig b/arch/x86/crypto/Kconfig > index 9bbfd01cfa2f..658adfd7aebf 100644 > --- a/arch/x86/crypto/Kconfig > +++ b/arch/x86/crypto/Kconfig > @@ -2,6 +2,11 @@ > > menu "Accelerated Cryptographic Algorithms for CPU (x86)" > > +config AS_HAS_KEYLOCKER > + def_bool $(as-instr,encodekey256 %eax$(comma)%eax) > + help > + Supported by binutils >= 2.36 and LLVM integrated assembler >= V12 It looks like arch/x86/Kconfig.assembler would be a better place for this. > diff --git a/arch/x86/crypto/aeskl-intel_asm.S b/arch/x86/crypto/aeskl-intel_asm.S > new file mode 100644 > index 000000000000..61addc61dd4e > --- /dev/null > +++ b/arch/x86/crypto/aeskl-intel_asm.S > @@ -0,0 +1,552 @@ > +/* SPDX-License-Identifier: GPL-2.0-or-later */ > +/* > + * Implement AES algorithm using AES Key Locker instructions. > + * > + * Most code is based from the AES-NI implementation, aesni-intel_asm.S > + * > + */ > + > +#include <linux/linkage.h> > +#include <linux/cfi_types.h> > +#include <asm/errno.h> > +#include <asm/inst.h> > +#include <asm/frame.h> > +#include "aes-helper_asm.S" > + > +.text > + > +#define STATE1 %xmm0 > +#define STATE2 %xmm1 > +#define STATE3 %xmm2 > +#define STATE4 %xmm3 > +#define STATE5 %xmm4 > +#define STATE6 %xmm5 > +#define STATE7 %xmm6 > +#define STATE8 %xmm7 > +#define STATE STATE1 > + > +#define IV %xmm9 > +#define KEY %xmm10 > +#define INC %xmm13 > + > +#define IN1 %xmm8 > +#define IN IN1 Why do both IN1 and IN exist? Shouldn't there just be IN? > + > +#define AREG %rax Shouldn't %rax just be hardcoded? > +#define HANDLEP %rdi This should be called CTX, to match the function prototypes. > +#define UKEYP OUTP This should be called IN_KEY, to match the function prototypes. > +#define GF128MUL_MASK %xmm11 > + > +/* > + * int __aeskl_setkey(struct crypto_aes_ctx *ctx, const u8 *in_key, unsigned int key_len) > + */ > +SYM_FUNC_START(__aeskl_setkey) > + FRAME_BEGIN > + movl %edx, 480(HANDLEP) > + movdqu (UKEYP), STATE1 > + mov $1, %eax > + cmp $16, %dl > + je .Lsetkey_128 > + > + movdqu 0x10(UKEYP), STATE2 > + encodekey256 %eax, %eax > + movdqu STATE4, 0x30(HANDLEP) > + jmp .Lsetkey_end > +.Lsetkey_128: > + encodekey128 %eax, %eax > + > +.Lsetkey_end: > + movdqu STATE1, (HANDLEP) > + movdqu STATE2, 0x10(HANDLEP) > + movdqu STATE3, 0x20(HANDLEP) The moves to the ctx should use movdqa, since it is aligned. > + > + xor AREG, AREG > + FRAME_END > + RET > +SYM_FUNC_END(__aeskl_setkey) This function always returns 0, so it really should return void. > +/* > + * int __aeskl_enc(const void *ctx, u8 *dst, const u8 *src) > + */ > +SYM_FUNC_START(__aeskl_enc) > + FRAME_BEGIN > + movdqu (INP), STATE > + movl 480(HANDLEP), KLEN > + > + cmp $16, KLEN > + je .Lenc_128 > + aesenc256kl (HANDLEP), STATE > + jz .Lenc_err > + jmp .Lenc_noerr > +.Lenc_128: > + aesenc128kl (HANDLEP), STATE > + jz .Lenc_err > + > +.Lenc_noerr: > + xor AREG, AREG > + jmp .Lenc_end > +.Lenc_err: > + mov $(-EINVAL), AREG > +.Lenc_end: > + movdqu STATE, (OUTP) > + FRAME_END > + RET > +SYM_FUNC_END(__aeskl_enc) In the common case (successful AES-256 encryption) this is executing 'jmp' twice. I think the code should be rearranged to eliminate these jmps. > +/* > + * int __aeskl_xts_encrypt(const struct crypto_aes_ctx *ctx, u8 *dst, > + * const u8 *src, unsigned int len, le128 *iv) > + */ > +SYM_FUNC_START(__aeskl_xts_encrypt) __aeskl_xts_encrypt() and __aeskl_xts_decrypt() are very similar. To reduce code duplication, can you consider generating them from a macro that takes an argument that indicates whether it is encrypt or decrypt? > +static int aeskl_setkey(struct crypto_tfm *tfm, void *raw_ctx, const u8 *in_key, > + unsigned int keylen) > +{ > + struct crypto_aes_ctx *ctx = (struct crypto_aes_ctx *)raw_ctx; > + int err; > + > + if (!crypto_simd_usable()) > + return -EBUSY; > + > + if (keylen != AES_KEYSIZE_128 && keylen != AES_KEYSIZE_192 && > + keylen != AES_KEYSIZE_256) > + return -EINVAL; > + > + kernel_fpu_begin(); > + if (unlikely(keylen == AES_KEYSIZE_192)) { > + pr_warn_once("AES-KL does not support 192-bit key. Use AES-NI.\n"); > + err = aesni_set_key(ctx, in_key, keylen); > + } else { > + if (!valid_keylocker()) > + err = -ENODEV; > + else > + err = __aeskl_setkey(ctx, in_key, keylen); > + } > + kernel_fpu_end(); > + > + return err; > +} [...] > + .cra_ctxsize = XTS_AES_CTX_SIZE, [...] Something that your AES-KL code does that's a bit ugly is that it abuses 'struct crypto_aes_ctx' to store a Keylocker key handle instead of the actual AES key schedule which the struct is supposed to be for. The proper way to represent that would be to make the tfm context for xts-aes-aeskl be a union of crypto_aes_ctx and a Keylocker specific context. If you don't do that and instead keep the proposed workaround, then please add a comment somewhere that very clearly explains how the struct is being used. Above aeskl_setkey() or above .cra_ctxsize might be a good place. > diff --git a/arch/x86/crypto/aesni-intel_glue.h b/arch/x86/crypto/aesni-intel_glue.h > new file mode 100644 > index 000000000000..5b1919f49efe > --- /dev/null > +++ b/arch/x86/crypto/aesni-intel_glue.h > @@ -0,0 +1,16 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > + > +/* > + * Support for Intel AES-NI instructions. This file contains function > + * prototypes to be referenced for other AES implementations > + */ It would be helpful if this comment was more concrete, like "These are AES-NI functions that are used by the AES-KL code as a fallback when it is given a 192-bit key. Key Locker does not support 192-bit keys." - Eric