On Tue, Jan 04, 2022 at 05:43:50PM +0000, Sean Christopherson wrote: > On Fri, Dec 31, 2021, Chao Peng wrote: > > On Tue, Dec 28, 2021 at 09:48:08PM +0000, Sean Christopherson wrote: > > >KVM handles > > > reverse engineering the memslot to get the offset and whatever else it needs. > > > notify_fallocate() and other callbacks are unchanged, though they probably can > > > drop the inode. > > > > > > E.g. likely with bad math and handwaving on the overlap detection: > > > > > > int kvm_private_fd_fallocate_range(void *owner, pgoff_t start, pgoff_t end) > > > { > > > struct kvm_memory_slot *slot = owner; > > > struct kvm_gfn_range gfn_range = { > > > .slot = slot, > > > .start = (start - slot->private_offset) >> PAGE_SHIFT, > > > .end = (end - slot->private_offset) >> PAGE_SHIFT, > > > .may_block = true, > > > }; > > > > > > if (!has_overlap(slot, start, end)) > > > return 0; > > > > > > gfn_range.end = min(gfn_range.end, slot->base_gfn + slot->npages); > > > > > > kvm_unmap_gfn_range(slot->kvm, &gfn_range); > > > return 0; > > > } > > > > I understand this KVM side handling, but again one fd can have multiple > > memslots. How shmem decides to notify which memslot from a list of > > memslots when it invokes the notify_fallocate()? Or just notify all > > the possible memslots then let KVM to check? > > Heh, yeah, those are the two choices. :-) > > Either the backing store needs to support registering callbacks for specific, > arbitrary ranges, or it needs to invoke all registered callbacks. Invoking all > callbacks has my vote; it's much simpler to implement and is unlikely to incur > meaningful overhead. _Something_ has to find the overlapping ranges, that cost > doesn't magically go away if it's pushed into the backing store. > > Note, invoking all notifiers is also aligned with the mmu_notifier behavior. Sounds a good reason. Then shmem side only needs to maintain a list of users. Chao