There is a new requirement for s390 secure execution guests that the hypervisor ensures all AP queues are reset and disassociated from the KVM guest before the secure configuration is torn down. It is the responsibility of the vfio_ap device driver to handle this. Prior to commit ("vfio: remove VFIO_GROUP_NOTIFY_SET_KVM"), the driver reset all AP queues passed through to a KVM guest when notified that the KVM pointer was being set to NULL. Subsequently, the AP queues are only reset when the fd for the mediated device used to pass the queues through to the guest is closed (the vfio_ap_mdev_close_device() callback). This is not a problem when userspace is well-behaved and uses the KVM_DEV_VFIO_GROUP_DEL attribute to remove the VFIO group; however, if userspace for some reason does not close the mdev fd, a secure execution guest will tear down its configuration before the AP queues are reset because the teardown is done in the kvm_arch_destroy_vm function which is invoked prior to vm_destroy_devices. This patch proposes a simple solution; rather than introducing a new notifier into vfio or callback into KVM, what aoubt reversing the order in which the kvm_arch_destroy_vm and kvm_destroy_devices are called. In some very limited testing (i.e., the automated regression tests for the vfio_ap device driver) this did not seem to cause any problems. The question remains, is there a good technical reason why the VM is destroyed before the devices it is using? This is not intuitive, so this is a request for comments on this proposed patch. The assumption here is that the medev fd will get closed when the devices are destroyed. Signed-off-by: Tony Krowiak <akrowiak@xxxxxxxxxxxxx> --- virt/kvm/kvm_main.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index a49df8988cd6..edaf2918be9b 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -1248,8 +1248,8 @@ static void kvm_destroy_vm(struct kvm *kvm) #else kvm_flush_shadow_all(kvm); #endif - kvm_arch_destroy_vm(kvm); kvm_destroy_devices(kvm); + kvm_arch_destroy_vm(kvm); for (i = 0; i < KVM_ADDRESS_SPACE_NUM; i++) { kvm_free_memslots(kvm, &kvm->__memslots[i][0]); kvm_free_memslots(kvm, &kvm->__memslots[i][1]); -- 2.31.1