Hi Thomas, Sorry for the late reply. I'll apply the patches and retest in this week. Please wait a while. Thanks, Yasuaki Ishimatsu On 10/04/2017 05:04 PM, Thomas Gleixner wrote: > On Tue, 3 Oct 2017, Thomas Gleixner wrote: >> Can you please apply the debug patch below. > > I found an issue with managed interrupts when the affinity mask of an > managed interrupt spawns multiple CPUs. Explanation in the changelog > below. I'm not sure that this cures the problems you have, but at least I > could prove that it's not doing what it should do. The failure I'm seing is > fixed, but I can't test that megasas driver due to -ENOHARDWARE. > > Can you please apply the patch below on top of Linus tree and retest? > > Please send me the outputs I asked you to provide last time in any case > (success or fail). > > @block/scsi folks: Can you please run that through your tests as well? > > Thanks, > > tglx > > 8<----------------------- > Subject: genirq/cpuhotplug: Enforce affinity setting on startup of managed irqs > From: Thomas Gleixner <tglx@xxxxxxxxxxxxx> > Date: Wed, 04 Oct 2017 21:07:38 +0200 > > Managed interrupts can end up in a stale state on CPU hotplug. If the > interrupt is not targeting a single CPU, i.e. the affinity mask spawns > multiple CPUs then the following can happen: > > After boot: > > dstate: 0x01601200 > IRQD_ACTIVATED > IRQD_IRQ_STARTED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_SET > IRQD_AFFINITY_MANAGED > node: 0 > affinity: 24-31 > effectiv: 24 > pending: 0 > > After offlining CPU 31 - 24 > > dstate: 0x01a31000 > IRQD_IRQ_DISABLED > IRQD_IRQ_MASKED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_SET > IRQD_AFFINITY_MANAGED > IRQD_MANAGED_SHUTDOWN > node: 0 > affinity: 24-31 > effectiv: 24 > pending: 0 > > Now CPU 25 gets onlined again, so it should get the effective interrupt > affinity for this interruopt, but due to the x86 interrupt affinity setter > restrictions this ends up after restarting the interrupt with: > > dstate: 0x01601300 > IRQD_ACTIVATED > IRQD_IRQ_STARTED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_SET > IRQD_SETAFFINITY_PENDING > IRQD_AFFINITY_MANAGED > node: 0 > affinity: 24-31 > effectiv: 24 > pending: 24-31 > > So the interrupt is still affine to CPU 24, which was the last CPU to go > offline of that affinity set and the move to an online CPU within 24-31, > in this case 25, is pending. This mechanism is x86/ia64 specific as those > architectures cannot move interrupts from thread context and do this when > an interrupt is actually handled. So the move is set to pending. > > Whats worse is that offlining CPU 25 again results in: > > dstate: 0x01601300 > IRQD_ACTIVATED > IRQD_IRQ_STARTED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_SET > IRQD_SETAFFINITY_PENDING > IRQD_AFFINITY_MANAGED > node: 0 > affinity: 24-31 > effectiv: 24 > pending: 24-31 > > This means the interrupt has not been shut down, because the outgoing CPU > is not in the effective affinity mask, but of course nothing notices that > the effective affinity mask is pointing at an offline CPU. > > In the case of restarting a managed interrupt the move restriction does not > apply, so the affinity setting can be made unconditional. This needs to be > done _before_ the interrupt is started up as otherwise the condition for > moving it from thread context would not longer be fulfilled. > > With that change applied onlining CPU 25 after offlining 31-24 results in: > > dstate: 0x01600200 > IRQD_ACTIVATED > IRQD_IRQ_STARTED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_MANAGED > node: 0 > affinity: 24-31 > effectiv: 25 > pending: > > And after offlining CPU 25: > > dstate: 0x01a30000 > IRQD_IRQ_DISABLED > IRQD_IRQ_MASKED > IRQD_SINGLE_TARGET > IRQD_AFFINITY_MANAGED > IRQD_MANAGED_SHUTDOWN > node: 0 > affinity: 24-31 > effectiv: 25 > pending: > > which is the correct and expected result. > > To complete that, add some debug code to catch this kind of situation in > the cpu offline code and warn about interrupt chips which allow affinity > setting and do not update the effective affinity mask if that feature is > enabled. > > Reported-by: YASUAKI ISHIMATSU <yasu.isimatu@xxxxxxxxx> > Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx> > > --- > kernel/irq/chip.c | 2 +- > kernel/irq/cpuhotplug.c | 28 +++++++++++++++++++++++++++- > kernel/irq/manage.c | 17 +++++++++++++++++ > 3 files changed, 45 insertions(+), 2 deletions(-) > > --- a/kernel/irq/chip.c > +++ b/kernel/irq/chip.c > @@ -265,8 +265,8 @@ int irq_startup(struct irq_desc *desc, b > irq_setup_affinity(desc); > break; > case IRQ_STARTUP_MANAGED: > + irq_do_set_affinity(d, aff, false); > ret = __irq_startup(desc); > - irq_set_affinity_locked(d, aff, false); > break; > case IRQ_STARTUP_ABORT: > return 0; > --- a/kernel/irq/cpuhotplug.c > +++ b/kernel/irq/cpuhotplug.c > @@ -18,8 +18,34 @@ > static inline bool irq_needs_fixup(struct irq_data *d) > { > const struct cpumask *m = irq_data_get_effective_affinity_mask(d); > + unsigned int cpu = smp_processor_id(); > > - return cpumask_test_cpu(smp_processor_id(), m); > +#ifdef CONFIG_GENERIC_IRQ_EFFECTIVE_AFF_MASK > + /* > + * The cpumask_empty() check is a workaround for interrupt chips, > + * which do not implement effective affinity, but the architecture has > + * enabled the config switch. Use the general affinity mask instead. > + */ > + if (cpumask_empty(m)) > + m = irq_data_get_affinity_mask(d); > + > + /* > + * Sanity check. If the mask is not empty when excluding the outgoing > + * CPU then it must contain at least one online CPU. The outgoing CPU > + * has been removed from the online mask already. > + */ > + if (cpumask_any_but(m, cpu) < nr_cpu_ids && > + cpumask_any_and(m, cpu_online_mask) >= nr_cpu_ids) { > + /* > + * If this happens then there was a missed IRQ fixup at some > + * point. Warn about it and enforce fixup. > + */ > + pr_warn("Eff. affinity %*pbl of IRQ %u contains only offline CPUs after offlining CPU %u\n", > + cpumask_pr_args(m), d->irq, cpu); > + return true; > + } > +#endif > + return cpumask_test_cpu(cpu, m); > } > > static bool migrate_one_irq(struct irq_desc *desc) > --- a/kernel/irq/manage.c > +++ b/kernel/irq/manage.c > @@ -168,6 +168,19 @@ void irq_set_thread_affinity(struct irq_ > set_bit(IRQTF_AFFINITY, &action->thread_flags); > } > > +static void irq_validate_effective_affinity(struct irq_data *data) > +{ > +#ifdef CONFIG_GENERIC_IRQ_EFFECTIVE_AFF_MASK > + const struct cpumask *m = irq_data_get_effective_affinity_mask(data); > + struct irq_chip *chip = irq_data_get_irq_chip(data); > + > + if (!cpumask_empty(m)) > + return; > + pr_warn_once("irq_chip %s did not update eff. affinity mask of irq %u\n", > + chip->name, data->irq); > +#endif > +} > + > int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, > bool force) > { > @@ -175,12 +188,16 @@ int irq_do_set_affinity(struct irq_data > struct irq_chip *chip = irq_data_get_irq_chip(data); > int ret; > > + if (!chip || !chip->irq_set_affinity) > + return -EINVAL; > + > ret = chip->irq_set_affinity(data, mask, force); > switch (ret) { > case IRQ_SET_MASK_OK: > case IRQ_SET_MASK_OK_DONE: > cpumask_copy(desc->irq_common_data.affinity, mask); > case IRQ_SET_MASK_OK_NOCOPY: > + irq_validate_effective_affinity(data); > irq_set_thread_affinity(desc); > ret = 0; > } >