Re: [PATCH v4 3/6] KVM: arm64: Enable ring-based dirty memory tracking

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Peter and Marc,

On 9/28/22 4:21 AM, Peter Xu wrote:
On Tue, Sep 27, 2022 at 01:32:07PM -0400, Marc Zyngier wrote:
On Tue, 27 Sep 2022 12:02:52 -0400,
Peter Xu <peterx@xxxxxxxxxx> wrote:

[...]


Any decision made on how to tackle with the GIC status dirty bits?

Which dirty bits? Are you talking of the per-RD pending bits?

Gavin found that some dirty pfn path may not have vcpu context for aarch64
offlist.

Borrowing Gavin's trace dump:

   el0t_64_sync
   el0t_64_sync_handler
   el0_svc
   do_el0_svc
   __arm64_sys_ioctl
   kvm_device_ioctl
   vgic_its_set_attr
   vgic_its_save_tables_v0
   kvm_write_guest
   __kvm_write_guest_page
   mark_page_dirty_in_slot

With current code it'll trigger the warning in mark_page_dirty_in_slot.

An userspace approach is doable by setting these pages as always dirty in
userspace (QEMU), but even if so IIUC we'll need to drop the warning
message in mark_page_dirty_in_slot() then we take no-vcpu dirty as no-op
and expected.

I'll leave the details to Gavin.


Thanks to Peter for bringing the issue to here for further discussion. I'm
slowly approaching the idea/design to resolve the issue. If Marc agrees, It
wouldn't stop us to merge the series since the newly introduced capability
is only used by kvm/selftests/dirty_log_test. It means more changes are needed
by QEMU in order to enable the feature there through KVM_CAP_DIRTY_LOG_RING_ACQ_REL.
I can post follow-up patches to fix the VGIC/ITS issue.

We had some internal discussion about the mentioned issue. It's all about the
various VGIC/ITS tables, listed as below. In QEMU, those tables are requested
to be saved to memory for migration or shutdown. Unfortunately, there is no
running vcpu in this particular path and the corresponding dirty bits are
dropped with warning messages in mark_page_dirty_in_slot().

   ---> Command to save VGIC/ITS tables
        group: KVM_DEV_ARM_VGIC_GRP_CTRL
        attr : KVM_DEV_ARM_ITS_SAVE_TABLES

   ---> VGIC/ITS tables to be saved
        Device Table                     // REG_GITS_BASER_0
        Interrupt Translation Tables     // GITS_CMD_MAPD
        Collection Table                 // REG_GITS_BASER_1

   ---> QEMU's backtraces, triggering the issue

        main                                thread_start
        qemu_main                           start_thread
        qemu_cleanup                        qemu_thread_start
        vm_shutdown                         migration_thread
        do_vm_stop                          migration_iteration_run
        vm_state_notify                     migration_completion
        vm_change_state_handler             vm_stop_force_state
                                            do_vm_stop
                                            vm_state_notify
                                            vm_change_state_handler          // In hw/intc/arm_gicv3_its_kvm.c

I have rough idea as below. It's appreciated if you can comment before I'm
going a head for the prototype. The overall idea is to introduce another
dirty ring for KVM (kvm-dirty-ring). It's updated and visited separately
to dirty ring for vcpu (vcpu-dirty-ring).

   - When the various VGIC/ITS table base addresses are specified, kvm-dirty-ring
     entries are added to mark those pages as 'always-dirty'. In mark_page_dirty_in_slot(),
     those 'always-dirty' pages will be skipped, no entries pushed to vcpu-dirty-ring.

   - Similar to vcpu-dirty-ring, kvm-dirty-ring is accessed from userspace through
     mmap(kvm->fd). However, there won't have similar reset interface. It means
     'struct kvm_dirty_gfn::flags' won't track any information as we do for
     vcpu-dirty-ring. In this regard, kvm-dirty-ring is purely shared buffer to
     advertise 'always-dirty' pages from host to userspace.
- For QEMU, shutdown/suspend/resume cases won't be concerning us any more. The
     only concerned case is migration. When the migration is about to complete,
     kvm-dirty-ring entries are fetched and the dirty bits are updated to global
     dirty page bitmap and RAMBlock's dirty page bitmap. For this, I'm still reading
     the code to find the best spot to do it.

Thanks,
Gavin



_______________________________________________
kvmarm mailing list
kvmarm@xxxxxxxxxxxxxxxxxxxxx
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm



[Index of Archives]     [Linux KVM]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux