On 05/26/2017, 08:54 AM, Jiri Slaby wrote: > On 05/19/2017, 11:35 PM, Josh Poimboeuf wrote: >> https://github.com/jpoimboe/linux/blob/undwarf/arch/x86/kernel/unwind_undwarf.c > > JFYI, it crashes in sha1_transform_avx due to crypto changes. You > perhaps missed that this beast uses ebp (not rbp) register for > computations. I had to do: > > --- a/arch/x86/crypto/sha1_ssse3_asm.S > +++ b/arch/x86/crypto/sha1_ssse3_asm.S > @@ -37,7 +37,7 @@ > #define REG_A %ecx > #define REG_B %esi > #define REG_C %edi > -#define REG_D %ebp > +#define REG_D %r12d > #define REG_E %edx > > #define REG_T1 %eax > @@ -74,6 +74,7 @@ > SYM_FUNC_START(\name) > > push %rbx > + push %r12 > push %rbp > > mov %rsp, %rbp > @@ -99,6 +100,7 @@ > rep stosq > > leaveq # deallocate workspace > + pop %r12 > pop %rbx > ret > > > I am afraid there are more of these, e.g. in aesni-intel_asm.S. aesni-intel_asm.S is OK -- only untouched x86_32 part uses ebp. But sha1_avx2_x86_64_asm.S is not. They use *all* usable registers including ebp in the computations hidden behind the SHA1_PIPELINED_MAIN_BODY macro. The only work around I can see is to push rbp/pop rbp around the computation as it used to do with rbx: --- a/arch/x86/crypto/sha1_avx2_x86_64_asm.S +++ b/arch/x86/crypto/sha1_avx2_x86_64_asm.S @@ -636,6 +636,7 @@ _loop3: /* Align stack */ mov %rsp, %rbp and $~(0x20-1), %rsp + push %rbp sub $RESERVE_STACK, %rsp avx2_zeroupper @@ -661,6 +662,7 @@ _loop3: avx2_zeroupper add $RESERVE_STACK, %rsp + pop %rbp leaveq pop %r15 regards, -- js suse labs -- To unsubscribe from this list: send the line "unsubscribe live-patching" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html