When the IRQs are threaded, the part of the handler that runs in interruption context can be pretty fast, as per design, while letting the slow part to run into the thread handler. In some cases, given IRQs can be triggered too fast, making it impossible for the irq_thread to be able to keep up handling every request. If two requests happen before any irq_thread handler is able to finish, no increment to threads_handled happen, causing threads_handled and threads_handled_last to be equal, which will ends up causing irqs_unhandled to be incremented in note_interrupt(). Once irqs_unhandled gets to ~100k, the IRQ line gets disabled, disrupting the device work. As of today, the only way to reset irqs_unhandled before disabling the IRQ line is to stay 100ms without having any increment to irqs_unhandled, which can be pretty hard to happen if the IRQ is very busy. On top of that, some irq_thread handlers can handle requests in batches, effectively incrementing threads_handled only once despite dealing with a lot of requests, which make the irqs_unhandled to reach 100k pretty fast if the IRQ is getting a lot of requests. This IRQ line disable bug can be easily reproduced with a serial8250 console on a PREEMPT_RT kernel: it only takes the user to print a lot of text to the console (or to ttyS0): around 300k chars should be enough. To fix this bug, reset irqs_unhandled whenever irq_thread handles at least one IRQ request. This fix makes possible to avoid disabling IRQs which irq_thread handlers can take long (while on heavy usage of the IRQ line), without losing the ability of disabling IRQs that actually get unhandled for too long. Signed-off-by: Leonardo Bras <leobras@xxxxxxxxxx> --- kernel/irq/spurious.c | 8 ++++++++ 1 file changed, 8 insertions(+) diff --git a/kernel/irq/spurious.c b/kernel/irq/spurious.c index 02b2daf074414..b60748f89738a 100644 --- a/kernel/irq/spurious.c +++ b/kernel/irq/spurious.c @@ -339,6 +339,14 @@ void note_interrupt(struct irq_desc *desc, irqreturn_t action_ret) handled |= SPURIOUS_DEFERRED; if (handled != desc->threads_handled_last) { action_ret = IRQ_HANDLED; + + /* + * If the thread handlers handle + * one IRQ reset the unhandled + * IRQ counter. + */ + desc->irqs_unhandled = 0; + /* * Note: We keep the SPURIOUS_DEFERRED * bit set. We are handling the -- 2.43.0