Traps enable interrupts conditionally but rely on the ASM return code to disable them again. That results in redundant interrupt disable and trace calls. Make the trap handlers disable interrupts before returning to avoid that, which allows simplification of the ASM entry code. Originally-by: Peter Zijlstra <peterz@xxxxxxxxxxxxx> Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx> --- arch/x86/kernel/traps.c | 32 +++++++++++++++++++++----------- arch/x86/mm/fault.c | 7 +++++-- 2 files changed, 26 insertions(+), 13 deletions(-) --- a/arch/x86/kernel/traps.c +++ b/arch/x86/kernel/traps.c @@ -276,6 +276,7 @@ static void do_error_trap(struct pt_regs NOTIFY_STOP) { cond_local_irq_enable(regs); do_trap(trapnr, signr, str, regs, error_code, sicode, addr); + cond_local_irq_disable(regs); } } @@ -501,6 +502,7 @@ dotraplinkage void do_bounds(struct pt_r die("bounds", regs, error_code); } + cond_local_irq_disable(regs); return; exit_trap: @@ -512,6 +514,7 @@ dotraplinkage void do_bounds(struct pt_r * time.. */ do_trap(X86_TRAP_BR, SIGSEGV, "bounds", regs, error_code, 0, NULL); + cond_local_irq_disable(regs); } dotraplinkage void @@ -525,19 +528,19 @@ do_general_protection(struct pt_regs *re if (static_cpu_has(X86_FEATURE_UMIP)) { if (user_mode(regs) && fixup_umip_exception(regs)) - return; + goto exit_trap; } if (v8086_mode(regs)) { local_irq_enable(); handle_vm86_fault((struct kernel_vm86_regs *) regs, error_code); - return; + goto exit_trap; } tsk = current; if (!user_mode(regs)) { if (fixup_exception(regs, X86_TRAP_GP, error_code, 0)) - return; + goto exit_trap; tsk->thread.error_code = error_code; tsk->thread.trap_nr = X86_TRAP_GP; @@ -549,12 +552,12 @@ do_general_protection(struct pt_regs *re */ if (!preemptible() && kprobe_running() && kprobe_fault_handler(regs, X86_TRAP_GP)) - return; + goto exit_trap; if (notify_die(DIE_GPF, desc, regs, error_code, X86_TRAP_GP, SIGSEGV) != NOTIFY_STOP) die(desc, regs, error_code); - return; + goto exit_trap; } tsk->thread.error_code = error_code; @@ -563,6 +566,8 @@ do_general_protection(struct pt_regs *re show_signal(tsk, SIGSEGV, "", desc, regs, error_code); force_sig(SIGSEGV); +exit_trap: + cond_local_irq_disable(regs); } NOKPROBE_SYMBOL(do_general_protection); @@ -783,9 +788,7 @@ dotraplinkage void do_debug(struct pt_re if (v8086_mode(regs)) { handle_vm86_trap((struct kernel_vm86_regs *) regs, error_code, X86_TRAP_DB); - cond_local_irq_disable(regs); - debug_stack_usage_dec(); - goto exit; + goto exit_irq; } if (WARN_ON_ONCE((dr6 & DR_STEP) && !user_mode(regs))) { @@ -802,6 +805,8 @@ dotraplinkage void do_debug(struct pt_re si_code = get_si_code(tsk->thread.debugreg6); if (tsk->thread.debugreg6 & (DR_STEP | DR_TRAP_BITS) || user_icebp) send_sigtrap(regs, error_code, si_code); + +exit_irq: cond_local_irq_disable(regs); debug_stack_usage_dec(); @@ -827,7 +832,7 @@ static void math_error(struct pt_regs *r if (!user_mode(regs)) { if (fixup_exception(regs, trapnr, error_code, 0)) - return; + goto exit_trap; task->thread.error_code = error_code; task->thread.trap_nr = trapnr; @@ -835,7 +840,7 @@ static void math_error(struct pt_regs *r if (notify_die(DIE_TRAP, str, regs, error_code, trapnr, SIGFPE) != NOTIFY_STOP) die(str, regs, error_code); - return; + goto exit_trap; } /* @@ -849,10 +854,12 @@ static void math_error(struct pt_regs *r si_code = fpu__exception_code(fpu, trapnr); /* Retry when we get spurious exceptions: */ if (!si_code) - return; + goto exit_trap; force_sig_fault(SIGFPE, si_code, (void __user *)uprobe_get_trap_addr(regs)); +exit_trap: + cond_local_irq_disable(regs); } dotraplinkage void do_coprocessor_error(struct pt_regs *regs, long error_code) @@ -888,6 +895,8 @@ do_device_not_available(struct pt_regs * info.regs = regs; math_emulate(&info); + + cond_local_irq_disable(regs); return; } #endif @@ -918,6 +927,7 @@ dotraplinkage void do_iret_error(struct do_trap(X86_TRAP_IRET, SIGILL, "iret exception", regs, error_code, ILL_BADSTK, (void __user *)NULL); } + local_irq_disable(); } #endif --- a/arch/x86/mm/fault.c +++ b/arch/x86/mm/fault.c @@ -1500,10 +1500,13 @@ static noinline void return; /* Was the fault on kernel-controlled part of the address space? */ - if (unlikely(fault_in_kernel_space(address))) + if (unlikely(fault_in_kernel_space(address))) { do_kern_addr_fault(regs, hw_error_code, address); - else + } else { do_user_addr_fault(regs, hw_error_code, address); + if (regs->flags & X86_EFLAGS_IF) + local_irq_disable(); + } } NOKPROBE_SYMBOL(__do_page_fault);