Re: cgroup2 freezer and kvm_vm_worker_thread()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 10/29/24 01:07, Tejun Heo wrote:
Hello,

Luca is reporting that cgroups which have kvm instances inside never
complete freezing. This can be trivially reproduced:

   root@test ~# mkdir /sys/fs/cgroup/test
   root@test ~# echo $fish_pid > /sys/fs/cgroup/test/cgroup.procs
   root@test ~# qemu-system-x86_64 --nographic -enable-kvm

and in another terminal:

   root@test ~# echo 1 > /sys/fs/cgroup/test/cgroup.freeze
   root@test ~# cat /sys/fs/cgroup/test/cgroup.events
   populated 1
   frozen 0
   root@test ~# for i in (cat /sys/fs/cgroup/test/cgroup.threads); echo $i; cat /proc/$i/stack; end
   2070
   [<0>] do_freezer_trap+0x42/0x70
   [<0>] get_signal+0x4da/0x870
   [<0>] arch_do_signal_or_restart+0x1a/0x1c0
   [<0>] syscall_exit_to_user_mode+0x73/0x120
   [<0>] do_syscall_64+0x87/0x140
   [<0>] entry_SYSCALL_64_after_hwframe+0x76/0x7e
   2159
   [<0>] do_freezer_trap+0x42/0x70
   [<0>] get_signal+0x4da/0x870
   [<0>] arch_do_signal_or_restart+0x1a/0x1c0
   [<0>] syscall_exit_to_user_mode+0x73/0x120
   [<0>] do_syscall_64+0x87/0x140
   [<0>] entry_SYSCALL_64_after_hwframe+0x76/0x7e
   2160
   [<0>] do_freezer_trap+0x42/0x70
   [<0>] get_signal+0x4da/0x870
   [<0>] arch_do_signal_or_restart+0x1a/0x1c0
   [<0>] syscall_exit_to_user_mode+0x73/0x120
   [<0>] do_syscall_64+0x87/0x140
   [<0>] entry_SYSCALL_64_after_hwframe+0x76/0x7e
   2161
   [<0>] kvm_nx_huge_page_recovery_worker+0xea/0x680
   [<0>] kvm_vm_worker_thread+0x8f/0x2b0
   [<0>] kthread+0xe8/0x110
   [<0>] ret_from_fork+0x33/0x40
   [<0>] ret_from_fork_asm+0x1a/0x30
   2164
   [<0>] do_freezer_trap+0x42/0x70
   [<0>] get_signal+0x4da/0x870
   [<0>] arch_do_signal_or_restart+0x1a/0x1c0
   [<0>] syscall_exit_to_user_mode+0x73/0x120
   [<0>] do_syscall_64+0x87/0x140
   [<0>] entry_SYSCALL_64_after_hwframe+0x76/0x7e

The cgroup freezing happens in the signal delivery path but
kvm_vm_worker_thread() thread never call into the signal delivery path while
joining non-root cgroups, so they never get frozen. Because the cgroup
freezer determines whether a given cgroup is frozen by comparing the number
of frozen threads to the total number of threads in the cgroup, the cgroup
never becomes frozen and users waiting for the state transition may hang
indefinitely.

There are two paths that we can take:

1. Make kvm_vm_worker_thread() call into signal delivery path.
    io_wq_worker() is in a similar boat and handles signal delivery and can
    be frozen and trapped like regular threads.

For the freezing part, would this be anything more than

fdiff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c
index d16ce8174ed6..b7b6a1c1b6a4 100644
--- a/arch/x86/kvm/mmu/mmu.c
+++ b/arch/x86/kvm/mmu/mmu.c
@@ -47,6 +47,7 @@
 #include <linux/kern_levels.h>
 #include <linux/kstrtox.h>
 #include <linux/kthread.h>
+#include <linux/freezer.h>
 #include <linux/wordpart.h>
#include <asm/page.h>
@@ -7429,22 +7430,27 @@ static long get_nx_huge_page_recovery_timeout(u64 start_time)
 static int kvm_nx_huge_page_recovery_worker(struct kvm *kvm, uintptr_t data)
 {
 	u64 start_time;
-	long remaining_time;
+	u64 end_time;
+
+	set_freezable();
while (true) {
 		start_time = get_jiffies_64();
-		remaining_time = get_nx_huge_page_recovery_timeout(start_time);
+		end_time = start_time + get_nx_huge_page_recovery_timeout(start_time);
- set_current_state(TASK_INTERRUPTIBLE);
-		while (!kthread_should_stop() && remaining_time > 0) {
-			schedule_timeout(remaining_time);
-			remaining_time = get_nx_huge_page_recovery_timeout(start_time);
+		for (;;) {
 			set_current_state(TASK_INTERRUPTIBLE);
+			if (kthread_freezable_should_stop(NULL))
+				break;
+			start_time = get_jiffies_64();
+			if ((s64)(end_time - start_time) <= 0)
+				break;
+			schedule_timeout(end_time - start_time);
 		}
set_current_state(TASK_RUNNING); - if (kthread_should_stop())
+		if (kthread_freezable_should_stop(NULL))
 			return 0;
kvm_recover_nx_huge_pages(kvm);

(untested beyond compilation).

I'm not sure if the KVM worker thread should process signals.  We want it
to take the CPU time it uses from the guest, but otherwise it's not running
on behalf of userspace in the way that io_wq_worker() is.

Paolo





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]     [Monitors]

  Powered by Linux