On Mon, Jun 14, 2021 at 10:05:18AM +0100, Steven Price wrote: > I realise there are still open questions[1] around the performance of > this series (the 'big lock', tag_sync_lock, introduced in the first > patch). But there should be no impact on non-MTE workloads and until we > get real MTE-enabled hardware it's hard to know whether there is a need > for something more sophisticated or not. Peter Collingbourne's patch[3] > to clear the tags at page allocation time should hide more of the impact > for non-VM cases. So the remaining concern is around VM startup which > could be effectively serialised through the lock. [...] > [1]: https://lore.kernel.org/r/874ke7z3ng.wl-maz%40kernel.org Start-up, VM resume, migration could be affected by this lock, basically any time you fault a page into the guest. As you said, for now it should be fine as long as the hardware doesn't support MTE or qemu doesn't enable MTE in guests. But the problem won't go away. We have a partial solution with an array of locks to mitigate against this but there's still the question of whether we should actually bother for something that's unlikely to happen in practice: MAP_SHARED memory in guests (ignoring the stage 1 case for now). If MAP_SHARED in guests is not a realistic use-case, we have the vma in user_mem_abort() and if the VM_SHARED flag is set together with MTE enabled for guests, we can reject the mapping. We can discuss the stage 1 case separately from this series. -- Catalin _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/mailman/listinfo/kvmarm