On Wed, Dec 29, 2010 at 03:18:13PM +0800, Sheng Yang wrote: > On Tuesday 28 December 2010 20:26:13 Avi Kivity wrote: > > On 12/22/2010 10:44 AM, Sheng Yang wrote: > > > Then we can support mask bit operation of assigned devices now. > > > > > > > > > @@ -3817,14 +3819,16 @@ static int > > > emulator_write_emulated_onepage(unsigned long addr, > > > > > > mmio: > > > trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, *(u64 *)val); > > > > > > + r = vcpu_mmio_write(vcpu, gpa, bytes, val); > > > > > > /* > > > > > > * Is this MMIO handled locally? > > > */ > > > > > > - if (!vcpu_mmio_write(vcpu, gpa, bytes, val)) > > > + if (!r) > > > > > > return X86EMUL_CONTINUE; > > > > > > vcpu->mmio_needed = 1; > > > > > > - vcpu->run->exit_reason = KVM_EXIT_MMIO; > > > + vcpu->run->exit_reason = (r == -ENOTSYNC) ? > > > + KVM_EXIT_MSIX_ROUTING_UPDATE : KVM_EXIT_MMIO; > > > > This isn't very pretty, exit_reason should be written in > > vcpu_mmio_write(). I guess we can refactor it later. > > Sure. > > > > > +#define KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV (1<< 0) > > > + > > > +#define KVM_MSIX_MMIO_TYPE_BASE_TABLE (1<< 8) > > > +#define KVM_MSIX_MMIO_TYPE_BASE_PBA (1<< 9) > > > + > > > +#define KVM_MSIX_MMIO_TYPE_DEV_MASK 0x00ff > > > +#define KVM_MSIX_MMIO_TYPE_BASE_MASK 0xff00 > > > > Any explanation of these? > > I chose to use assigned device id instead of one specific table id, because every > device should got at most one MSI MMIO(the same should applied to vfio device as > well), and if we use specific table ID, we need way to associate with the device > anyway, to perform mask/unmask or other operation. So I think it's better to use > device ID here directly. Table id will be needed to make things work for emulated devices. My idea was this: we have the device id in kvm_assigned_msix_entry already. Just put table id and entry number in kvm_irq_routing_entry (create a new gsi type for this). The result will also work for irqfd because these are mapped to gsi. > And for the table and pba address, it's due to the mapping in userspace may know > the guest MSI-X table address and PBA address at different time(due to different > BAR, refer to the code in assigned_dev_iomem_map() of qemu). So I purposed this > API to allow each of them can be passed to kernel space individually. > > > > > +struct kvm_msix_mmio_user { > > > + __u32 dev_id; > > > + __u16 type; > > > + __u16 max_entries_nr; > > > + __u64 base_addr; > > > + __u64 base_va; > > > + __u64 flags; > > > + __u64 reserved[4]; > > > +}; > > > + > > > > > > > > > +int kvm_assigned_device_update_msix_mask_bit(struct kvm *kvm, > > > + int assigned_dev_id, int entry, u32 flag) > > > +{ > > > > Need a better name for 'flag' (and make it a bool). > > > > > + int r = -EFAULT; > > > + struct kvm_assigned_dev_kernel *adev; > > > + int i; > > > + > > > + if (!irqchip_in_kernel(kvm)) > > > + return r; > > > + > > > + mutex_lock(&kvm->lock); > > > + adev = kvm_find_assigned_dev(&kvm->arch.assigned_dev_head, > > > + assigned_dev_id); > > > + if (!adev) > > > + goto out; > > > + > > > + for (i = 0; i< adev->entries_nr; i++) > > > + if (adev->host_msix_entries[i].entry == entry) { > > > + if (flag) > > > + disable_irq_nosync( > > > + adev->host_msix_entries[i].vector); > > > + else > > > + enable_irq(adev->host_msix_entries[i].vector); > > > + r = 0; > > > + break; > > > + } > > > +out: > > > + mutex_unlock(&kvm->lock); > > > + return r; > > > +} > > > > > > @@ -1988,6 +2008,12 @@ static int kvm_dev_ioctl_create_vm(void) > > > > > > return r; > > > > > > } > > > > > > #endif > > > > > > + r = kvm_register_msix_mmio_dev(kvm); > > > + if (r< 0) { > > > + kvm_put_kvm(kvm); > > > + return r; > > > + } > > > > Shouldn't this be part of individual KVM_REGISTER_MSIX_MMIO calls? > > In fact this MMIO device is more like global one for the VM, not for every > devices. It should handle all MMIO from all MSI-X enabled devices, so I put it in > the VM init/destroy process. > > > > +static int msix_table_mmio_read(struct kvm_io_device *this, gpa_t addr, > > > int len, + void *val) > > > +{ > > > + struct kvm_msix_mmio_dev *mmio_dev = > > > + container_of(this, struct kvm_msix_mmio_dev, table_dev); > > > + struct kvm_msix_mmio *mmio; > > > + int idx, ret = 0, entry, offset, r; > > > + > > > + mutex_lock(&mmio_dev->lock); > > > + idx = get_mmio_table_index(mmio_dev, addr, len); > > > + if (idx< 0) { > > > + ret = -EOPNOTSUPP; > > > + goto out; > > > + } > > > + if ((addr& 0x3) || (len != 4&& len != 8)) > > > + goto out; > > > > What about (addr & 4) && (len == 8)? Is it supported? It may cross entry > > boundaries. > > Should not supported. But I haven't found words on the PCI spec for it. So I > didn't add this check. IMPLEMENTATION NOTE MSI-X Memory Space Structures in Read/Write Memory .... For all accesses to MSI-X Table and MSI-X PBA fields, software must use aligned full DWORD or aligned full QWORD transactions; otherwise, the result is undefined. > > > > > + mmio =&mmio_dev->mmio[idx]; > > > + > > > + entry = (addr - mmio->table_base_addr) / PCI_MSIX_ENTRY_SIZE; > > > + offset = addr& 0xf; > > > + r = copy_from_user(val, (void *)(mmio->table_base_va + > > > + entry * PCI_MSIX_ENTRY_SIZE + offset), len); > > > > > > > > > + if (r) > > > + goto out; > > > +out: > > > + mutex_unlock(&mmio_dev->lock); > > > + return ret; > > > +} > > > + > > > +static int msix_table_mmio_write(struct kvm_io_device *this, gpa_t addr, > > > + int len, const void *val) > > > +{ > > > + struct kvm_msix_mmio_dev *mmio_dev = > > > + container_of(this, struct kvm_msix_mmio_dev, table_dev); > > > + struct kvm_msix_mmio *mmio; > > > + int idx, entry, offset, ret = 0, r = 0; > > > + gpa_t entry_base; > > > + u32 old_ctrl, new_ctrl; > > > + > > > + mutex_lock(&mmio_dev->lock); > > > + idx = get_mmio_table_index(mmio_dev, addr, len); > > > + if (idx< 0) { > > > + ret = -EOPNOTSUPP; > > > + goto out; > > > + } > > > + if ((addr& 0x3) || (len != 4&& len != 8)) > > > + goto out; > > > + mmio =&mmio_dev->mmio[idx]; > > > + entry = (addr - mmio->table_base_addr) / PCI_MSIX_ENTRY_SIZE; > > > + entry_base = mmio->table_base_va + entry * PCI_MSIX_ENTRY_SIZE; > > > + offset = addr& 0xF; > > > + > > > + if (copy_from_user(&old_ctrl, > > > + entry_base + PCI_MSIX_ENTRY_VECTOR_CTRL, > > > + sizeof old_ctrl)) > > > + goto out; > > > > get_user() is easier. > > > > > + > > > + /* No allow writing to other fields when entry is unmasked */ > > > + if (!(old_ctrl& PCI_MSIX_ENTRY_CTRL_MASKBIT)&& > > > + offset != PCI_MSIX_ENTRY_VECTOR_CTRL) > > > + goto out; > > > + > > > + if (copy_to_user(entry_base + offset, val, len)) > > > + goto out; > > > > > > + > > > + if (copy_from_user(&new_ctrl, > > > + entry_base + PCI_MSIX_ENTRY_VECTOR_CTRL, > > > + sizeof new_ctrl)) > > > + goto out; > > > > put_user() > > > > > + > > > + if ((offset< PCI_MSIX_ENTRY_VECTOR_CTRL&& len == 4) || > > > + (offset< PCI_MSIX_ENTRY_DATA&& len == 8)) > > > + ret = -ENOTSYNC; > > > + if (old_ctrl == new_ctrl) > > > + goto out; > > > + if (!(old_ctrl& PCI_MSIX_ENTRY_CTRL_MASKBIT)&& > > > + (new_ctrl& PCI_MSIX_ENTRY_CTRL_MASKBIT)) > > > + r = update_msix_mask_bit(mmio_dev->kvm, mmio, entry, 1); > > > + else if ((old_ctrl& PCI_MSIX_ENTRY_CTRL_MASKBIT)&& > > > + !(new_ctrl& PCI_MSIX_ENTRY_CTRL_MASKBIT)) > > > + r = update_msix_mask_bit(mmio_dev->kvm, mmio, entry, 0); > > > + if (r || ret) > > > + ret = -ENOTSYNC; > > > +out: > > > + mutex_unlock(&mmio_dev->lock); > > > + return ret; > > > +} > > > > blank line... > > > > > +static const struct kvm_io_device_ops msix_mmio_table_ops = { > > > + .read = msix_table_mmio_read, > > > + .write = msix_table_mmio_write, > > > +}; > > > + > > > ++ > > > +int kvm_vm_ioctl_register_msix_mmio(struct kvm *kvm, > > > + struct kvm_msix_mmio_user *mmio_user) > > > +{ > > > + struct kvm_msix_mmio_dev *mmio_dev =&kvm->msix_mmio_dev; > > > + struct kvm_msix_mmio *mmio = NULL; > > > + int r = 0, i; > > > + > > > + mutex_lock(&mmio_dev->lock); > > > + for (i = 0; i< mmio_dev->mmio_nr; i++) { > > > + if (mmio_dev->mmio[i].dev_id == mmio_user->dev_id&& > > > + (mmio_dev->mmio[i].type& KVM_MSIX_MMIO_TYPE_DEV_MASK) == > > > + (mmio_user->type& KVM_MSIX_MMIO_TYPE_DEV_MASK)) { > > > + mmio =&mmio_dev->mmio[i]; > > > + if (mmio->max_entries_nr != mmio_user->max_entries_nr) { > > > + r = -EINVAL; > > > + goto out; > > > + } > > > + break; > > > + } > > > + } > > > + if (!mmio) { > > > + if (mmio_dev->mmio_nr == KVM_MSIX_MMIO_MAX) { > > > + r = -ENOSPC; > > > + goto out; > > > + } > > > + mmio =&mmio_dev->mmio[mmio_dev->mmio_nr]; > > > + mmio_dev->mmio_nr++; > > > + } > > > + mmio->max_entries_nr = mmio_user->max_entries_nr; > > > > Sanity check to avoid overflow. > > > > > + mmio->dev_id = mmio_user->dev_id; > > > + mmio->flags = mmio_user->flags; > > > > Check for unsupported bits (all of them at present?) > > > > > + if ((mmio_user->type& KVM_MSIX_MMIO_TYPE_DEV_MASK) == > > > + KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV) > > > + mmio->type = KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV; > > > + > > > + if ((mmio_user->type& KVM_MSIX_MMIO_TYPE_BASE_MASK) == > > > + KVM_MSIX_MMIO_TYPE_BASE_TABLE) { > > > + mmio->type |= KVM_MSIX_MMIO_TYPE_BASE_TABLE; > > > + mmio->table_base_addr = mmio_user->base_addr; > > > + mmio->table_base_va = mmio_user->base_va; > > > > Check for va in kernel space. > > > > > + } else if ((mmio_user->type& KVM_MSIX_MMIO_TYPE_BASE_MASK) == > > > + KVM_MSIX_MMIO_TYPE_BASE_PBA) { > > > + mmio->type |= KVM_MSIX_MMIO_TYPE_BASE_PBA; > > > + mmio->pba_base_addr = mmio_user->base_addr; > > > + mmio->pba_base_va = mmio_user->base_va; > > > + } > > > +out: > > > + mutex_unlock(&mmio_dev->lock); > > > + return r; > > > +} > > > + > > > + > > > > In all, looks reasonable. I'd like to see documentation for it, and > > review from the pci people. Alex, mst? Some general comments: PBA isn't supported in this version, which is OK, but let's not add a capability until it is, and let's not try to guess what the interface will look like. I think keeping PBA in userspace will be hard because it needs to be modified from interrupt context. Removing the PBA stub will make the interface simpler. > Would add the API document soon. > > -- > regards > Yang, Sheng -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html