Alright, I'll send that soon. Sorry for the mixup. On Wed, Apr 17, 2013 at 8:30 AM, Gleb Natapov <gleb@xxxxxxxxxx> wrote: > On Wed, Apr 17, 2013 at 08:24:21AM -0700, Andrew Honig wrote: >> I'm happy to not add a new API and use __kvm_set_memory_region to >> unregister private memory regions, but I thought chaning the API was >> the approach you asked for when I sent a previous patch. See the end >> of: http://article.gmane.org/gmane.comp.emulators.kvm.devel/107753 >> >> Did I misunderstand your comment from 8 April? >> > Ugh, yes. My "Please send a second version" was in response to your > question "or should I send a second version of this current patch?" > Where second version is the same as the first one but uses > __kvm_set_memory_region() for slot deletion. I am sorry about the > confusion :( > >> On Wed, Apr 17, 2013 at 6:10 AM, Gleb Natapov <gleb@xxxxxxxxxx> wrote: >> > On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote: >> >> >> >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c >> >> when a VM is created and destroyed. >> >> >> >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for >> >> architecture specific reasons. It currently allocates the pages on behalf >> >> of user space, but has no way of cleanly freeing that memory while the >> >> user space process is still running. For user space processes that want >> >> more control over that memory, this patch allows user space to provide the >> >> memory that KVM uses. >> >> >> >> Signed-off-by: Andrew Honig <ahonig@xxxxxxxxxx> >> > I thought we agreed on not adding new API and using >> > __kvm_set_memory_region() to unregister private memory regions. >> > >> > >> >> --- >> >> Documentation/virtual/kvm/api.txt | 8 +++++++ >> >> arch/arm/kvm/arm.c | 6 +++++ >> >> arch/ia64/kvm/kvm-ia64.c | 6 +++++ >> >> arch/powerpc/kvm/powerpc.c | 6 +++++ >> >> arch/s390/kvm/kvm-s390.c | 6 +++++ >> >> arch/x86/include/asm/kvm_host.h | 7 ++++++ >> >> arch/x86/kvm/svm.c | 8 +++++++ >> >> arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- >> >> arch/x86/kvm/x86.c | 12 ++++++++-- >> >> include/linux/kvm_host.h | 2 ++ >> >> virt/kvm/kvm_main.c | 2 ++ >> >> 11 files changed, 105 insertions(+), 5 deletions(-) >> >> >> >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt >> >> index 119358d..aa18cac 100644 >> >> --- a/Documentation/virtual/kvm/api.txt >> >> +++ b/Documentation/virtual/kvm/api.txt >> >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr >> >> be identical. This allows large pages in the guest to be backed by large >> >> pages in the host. >> >> >> >> +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture >> >> +specific reasons. Calling this ioctl with the special memslot >> >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for >> >> +that memory. If that memslot is not set before creating VCPUs for a VM then >> >> +kvm will allocate the memory on behalf of user space, but userspace will not >> >> +be able to free that memory. User space should treat this memory as opaque >> >> +and not modify it. >> >> + >> >> The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and >> >> KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of >> >> writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to >> >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >> >> index 5a93698..ac52f14 100644 >> >> --- a/arch/arm/kvm/arm.c >> >> +++ b/arch/arm/kvm/arm.c >> >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c >> >> index ad3126a..570dd97 100644 >> >> --- a/arch/ia64/kvm/kvm-ia64.c >> >> +++ b/arch/ia64/kvm/kvm-ia64.c >> >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c >> >> index 934413c..6e3843b 100644 >> >> --- a/arch/powerpc/kvm/powerpc.c >> >> +++ b/arch/powerpc/kvm/powerpc.c >> >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return kvmppc_core_create_memslot(slot, npages); >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c >> >> index 4cf35a0..a97f495 100644 >> >> --- a/arch/s390/kvm/kvm-s390.c >> >> +++ b/arch/s390/kvm/kvm-s390.c >> >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> /* Section: memory related */ >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h >> >> index 4979778..7215817 100644 >> >> --- a/arch/x86/include/asm/kvm_host.h >> >> +++ b/arch/x86/include/asm/kvm_host.h >> >> @@ -37,6 +37,7 @@ >> >> /* memory slots that are not exposed to userspace */ >> >> #define KVM_PRIVATE_MEM_SLOTS 3 >> >> #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) >> >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 >> >> >> >> #define KVM_MMIO_SIZE 16 >> >> >> >> @@ -553,6 +554,9 @@ struct kvm_arch { >> >> struct page *ept_identity_pagetable; >> >> bool ept_identity_pagetable_done; >> >> gpa_t ept_identity_map_addr; >> >> + unsigned long ept_ptr; >> >> + unsigned long apic_ptr; >> >> + unsigned long tss_ptr; >> >> >> >> unsigned long irq_sources_bitmap; >> >> s64 kvmclock_offset; >> >> @@ -640,6 +644,9 @@ struct kvm_x86_ops { >> >> bool (*cpu_has_accelerated_tpr)(void); >> >> void (*cpuid_update)(struct kvm_vcpu *vcpu); >> >> >> >> + int (*set_private_memory)(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem); >> >> + >> >> /* Create, but do not attach this VCPU */ >> >> struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); >> >> void (*vcpu_free)(struct kvm_vcpu *vcpu); >> >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c >> >> index e1b1ce2..3cc4e56 100644 >> >> --- a/arch/x86/kvm/svm.c >> >> +++ b/arch/x86/kvm/svm.c >> >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) >> >> return 0; >> >> } >> >> >> >> +static int svm_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) >> >> { >> >> struct vcpu_svm *svm; >> >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { >> >> .hardware_disable = svm_hardware_disable, >> >> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, >> >> >> >> + .set_private_memory = svm_set_private_memory, >> >> + >> >> .vcpu_create = svm_create_vcpu, >> >> .vcpu_free = svm_free_vcpu, >> >> .vcpu_reset = svm_vcpu_reset, >> >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c >> >> index 6667042..796ac07 100644 >> >> --- a/arch/x86/kvm/vmx.c >> >> +++ b/arch/x86/kvm/vmx.c >> >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) >> >> kvm_userspace_mem.flags = 0; >> >> kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; >> >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + if (kvm->arch.apic_ptr) { >> >> + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> >> + } else { >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + } >> >> + >> >> if (r) >> >> goto out; >> >> >> >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) >> >> kvm_userspace_mem.guest_phys_addr = >> >> kvm->arch.ept_identity_map_addr; >> >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + if (kvm->arch.ept_ptr) { >> >> + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> >> + } else { >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + } >> >> + >> >> if (r) >> >> goto out; >> >> >> >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) >> >> .flags = 0, >> >> }; >> >> >> >> - ret = kvm_set_memory_region(kvm, &tss_mem, false); >> >> + if (kvm->arch.tss_ptr) { >> >> + tss_mem.userspace_addr = kvm->arch.tss_ptr; >> >> + ret = kvm_set_memory_region(kvm, &tss_mem, true); >> >> + } else { >> >> + ret = kvm_set_memory_region(kvm, &tss_mem, false); >> >> + } >> >> + >> >> if (ret) >> >> return ret; >> >> kvm->arch.tss_addr = addr; >> >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) >> >> vmx_complete_interrupts(vmx); >> >> } >> >> >> >> +static int vmx_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + /* >> >> + * Early sanity checking so userspace gets an error message during >> >> + * memory setup and not when trying to use this memory. >> >> + * Checks to see if the memory is valid are performed later when >> >> + * the memory is used. >> >> + */ >> >> + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || >> >> + mem->memory_size & (PAGE_SIZE - 1) || >> >> + mem->memory_size < PAGE_SIZE * 5) >> >> + return -EINVAL; >> >> + >> >> + kvm->arch.ept_ptr = mem->userspace_addr; >> >> + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; >> >> + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; >> >> + >> >> + return 0; >> >> +} >> >> + >> >> static void vmx_free_vcpu(struct kvm_vcpu *vcpu) >> >> { >> >> struct vcpu_vmx *vmx = to_vmx(vcpu); >> >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { >> >> .hardware_disable = hardware_disable, >> >> .cpu_has_accelerated_tpr = report_flexpriority, >> >> >> >> + .set_private_memory = vmx_set_private_memory, >> >> + >> >> .vcpu_create = vmx_create_vcpu, >> >> .vcpu_free = vmx_free_vcpu, >> >> .vcpu_reset = vmx_vcpu_reset, >> >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c >> >> index e172132..7045d0a 100644 >> >> --- a/arch/x86/kvm/x86.c >> >> +++ b/arch/x86/kvm/x86.c >> >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) >> >> kvm_free_pit(kvm); >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return kvm_x86_ops->set_private_memory(kvm, mem); >> >> +} >> >> + >> >> void kvm_arch_destroy_vm(struct kvm *kvm) >> >> { >> >> kvm_iommu_unmap_guest(kvm); >> >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> * Only private memory slots need to be mapped here since >> >> * KVM_SET_MEMORY_REGION ioctl is no longer supported. >> >> */ >> >> - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { >> >> + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && >> >> + !user_alloc) { >> >> unsigned long userspace_addr; >> >> >> >> /* >> >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, >> >> >> >> int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; >> >> >> >> - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { >> >> + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && >> >> + !user_alloc) { >> >> int ret; >> >> >> >> ret = vm_munmap(old.userspace_addr, >> >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h >> >> index c139582..f441d1f 100644 >> >> --- a/include/linux/kvm_host.h >> >> +++ b/include/linux/kvm_host.h >> >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, >> >> void kvm_arch_free_memslot(struct kvm_memory_slot *free, >> >> struct kvm_memory_slot *dont); >> >> int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem); >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c >> >> index f18013f..5372225 100644 >> >> --- a/virt/kvm/kvm_main.c >> >> +++ b/virt/kvm/kvm_main.c >> >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, >> >> kvm_userspace_memory_region *mem, >> >> bool user_alloc) >> >> { >> >> + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) >> >> + return kvm_arch_set_private_memory(kvm, mem); >> >> if (mem->slot >= KVM_USER_MEM_SLOTS) >> >> return -EINVAL; >> >> return kvm_set_memory_region(kvm, mem, user_alloc); >> >> -- >> >> 1.7.10.4 >> >> >> >> -- >> >> To unsubscribe from this list: send the line "unsubscribe kvm" in >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> > >> > -- >> > Gleb. > > -- > Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html