Re: [bug report] x86/split_lock: Make life miserable for split lockers

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



+Joerg

On 11/28/23 11:40, Tony Luck wrote:
On Tue, Nov 28, 2023 at 04:12:24PM +0300, Dan Carpenter wrote:
Hello Tony Luck,

The patch b041b525dab9: "x86/split_lock: Make life miserable for
split lockers" from Mar 10, 2022 (linux-next), leads to the following
Smatch static checker warning:

	arch/x86/kernel/cpu/intel.c:1179 split_lock_warn()
	warn: sleeping in atomic context

arch/x86/kernel/cpu/intel.c
     1158 static void split_lock_warn(unsigned long ip)
     1159 {
     1160         struct delayed_work *work;
     1161         int cpu;
     1162
     1163         if (!current->reported_split_lock)
     1164                 pr_warn_ratelimited("#AC: %s/%d took a split_lock trap at address: 0x%lx\n",
     1165                                     current->comm, current->pid, ip);
     1166         current->reported_split_lock = 1;
     1167
     1168         if (sysctl_sld_mitigate) {
     1169                 /*
     1170                  * misery factor #1:
     1171                  * sleep 10ms before trying to execute split lock.
     1172                  */
     1173                 if (msleep_interruptible(10) > 0)
     1174                         return;
     1175                 /*
     1176                  * Misery factor #2:
     1177                  * only allow one buslocked disabled core at a time.
     1178                  */
--> 1179                 if (down_interruptible(&buslock_sem) == -EINTR)
     1180                         return;
     1181                 work = &sl_reenable_unlock;
     1182         } else {
     1183                 work = &sl_reenable;
     1184         }
     1185
     1186         cpu = get_cpu();
     1187         schedule_delayed_work_on(cpu, work, 2);
     1188
     1189         /* Disable split lock detection on this CPU to make progress */
     1190         sld_update_msr(false);
     1191         put_cpu();
     1192 }

The call tree is:

kernel_exc_vmm_communication() <- disables preempt
-> vc_raw_handle_exception()
    -> vc_forward_exception()
       -> exc_alignment_check()
          -> __exc_alignment_check()
             -> handle_user_split_lock()
                -> split_lock_warn()

I think maybe the mismatch is that kernel_exc_vmm_communication() calls
irqentry_nmi_enter(regs); which disable preemption but exc_alignment_check()
does local_irq_enable() which doesn't enable it.

I think we need some arch/x86/kernel/sev.c expertise to explain the
preemption requirements in that stack trace. Adding Tom Lendacky.

Adding Joerg as the original developer of this code.

I believe that irqentry_nmi_enter() is used so that we are ensured that the kernel can't be interrupted while using the per-CPU GHCB when entered from kernel-mode in order to avoid nested #VCs (except for an NMI). Joerg might have further insights since there was a lot of discussion around these changes.

I'm not sure if is possible, but I wonder if irqentry_nmi_exit() can be issued before forwarding the exception - or even delay forwarding the exception until after irqentry_nmi_exit().

Thanks,
Tom


Also why does arch/x86 not have a dedicated mailing list?

Good question. X86 was once the default architecture. So everything went to
linux-kernel@xxxxxxxxxxxxxxx. I'll add that to Cc: for this. But maybe
it's time for an x86 specific list?

regards,
dan carpenter

-Tony




[Index of Archives]     [Kernel Development]     [Kernel Announce]     [Kernel Newbies]     [Linux Networking Development]     [Share Photos]     [IDE]     [Security]     [Git]     [Netfilter]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Device Mapper]

  Powered by Linux