Hi KarimAllah, On 10.07.19 17:24, Raslan, KarimAllah wrote: > On Mon, 2019-07-08 at 22:39 +0200, Jan Kiszka wrote: >> Hi all, >> >> it seems the "new" KVM_SET_NESTED_STATE interface has some remaining >> robustness issues. > > I would be very interested to learn about any more robustness issues that you > are seeing. > >> The most urgent one: With the help of latest QEMU >> master that uses this interface, you can easily crash the host. You just >> need to start qemu-system-x86 -enable-kvm in L1 and then hard-reset L1. >> The host CPU that ran this will stall, the system will freeze soon. > > Just to confirm, you start an L2 guest using qemu inside an L1-guest and then > hard-reset the L1 guest? Exactly. > > Are you running any special workload in L2 or L1 when you reset? Also how Nope. It is a standard (though rather oldish) userland in L1, just running a more recent kernel 5.2. > exactly are you doing this "hard reset"? system_reset from the monitor or "reset" from QEMU window menu. > > (sorry just tried this in my setup and I did not see any problem but my setup > is slightly different, so just ruling out obvious stuff). > If it helps, I can share privately a guest image that was built via https://github.com/siemens/jailhouse-images which exposes the reset issue after starting Jailhouse (instead of qemu-system-x86_64 - though that should "work" as well, just not tested yet). It's about 70M packed. Host-wise, 5.2.0 + QEMU master should do. I can also provide you the .config if needed. >> >> I've also seen a pattern with my Jailhouse test VM where I seems to get >> stuck in a loop between L1 and L2: >> >> qemu-system-x86-6660 [007] 398.691401: kvm_nested_vmexit: rip 7fa9ee5224e4 reason IO_INSTRUCTION info1 5658000b info2 0 int_info 0 int_info_err 0 >> qemu-system-x86-6660 [007] 398.691402: kvm_fpu: unload >> qemu-system-x86-6660 [007] 398.691403: kvm_userspace_exit: reason KVM_EXIT_IO (2) >> qemu-system-x86-6660 [007] 398.691440: kvm_fpu: load >> qemu-system-x86-6660 [007] 398.691441: kvm_pio: pio_read at 0x5658 size 4 count 1 val 0x4 >> qemu-system-x86-6660 [007] 398.691443: kvm_mmu_get_page: existing sp gfn 3a22e 1/4 q3 direct --x !pge !nxe root 6 sync >> qemu-system-x86-6660 [007] 398.691444: kvm_entry: vcpu 3 >> qemu-system-x86-6660 [007] 398.691475: kvm_exit: reason IO_INSTRUCTION rip 0x7fa9ee5224e4 info 5658000b 0 >> qemu-system-x86-6660 [007] 398.691476: kvm_nested_vmexit: rip 7fa9ee5224e4 reason IO_INSTRUCTION info1 5658000b info2 0 int_info 0 int_info_err 0 >> qemu-system-x86-6660 [007] 398.691477: kvm_fpu: unload >> qemu-system-x86-6660 [007] 398.691478: kvm_userspace_exit: reason KVM_EXIT_IO (2) >> qemu-system-x86-6660 [007] 398.691526: kvm_fpu: load >> qemu-system-x86-6660 [007] 398.691527: kvm_pio: pio_read at 0x5658 size 4 count 1 val 0x4 >> qemu-system-x86-6660 [007] 398.691529: kvm_mmu_get_page: existing sp gfn 3a22e 1/4 q3 direct --x !pge !nxe root 6 sync >> qemu-system-x86-6660 [007] 398.691530: kvm_entry: vcpu 3 >> qemu-system-x86-6660 [007] 398.691533: kvm_exit: reason IO_INSTRUCTION rip 0x7fa9ee5224e4 info 5658000b 0 >> qemu-system-x86-6660 [007] 398.691534: kvm_nested_vmexit: rip 7fa9ee5224e4 reason IO_INSTRUCTION info1 5658000b info2 0 int_info 0 int_info_err 0 >> >> These issues disappear when going from ebbfef2f back to 6cfd7639 (both >> with build fixes) in QEMU. > > This is the QEMU that you are using in L0 to launch an L1 guest, right? or are > you still referring to the QEMU mentioned above? This scenario is similar but still a bit different than the above. Yes, same L0 image and host QEMU here (and the traces were taken on the host, obviously), but the workload is now as follows: - boot L1 Linux - enable Jailhouse inside L1 - move the mouse over the graphical desktop of L2, ie. the former L1 Linux (Jailhouse is now L1) - the L1/L2 guests enter the loop above while trying to read from the vmmouse port Jan -- Siemens AG, Corporate Technology, CT RDA IOT SES-DE Corporate Competence Center Embedded Linux