On Mon, 2015-10-05 at 08:33 +0000, Bhushan Bharat wrote: > > > > -----Original Message----- > > From: Alex Williamson [mailto:alex.williamson@xxxxxxxxxx] > > Sent: Saturday, October 03, 2015 4:17 AM > > To: Bhushan Bharat-R65777 <Bharat.Bhushan@xxxxxxxxxxxxx> > > Cc: kvmarm@xxxxxxxxxxxxxxxxxxxxx; kvm@xxxxxxxxxxxxxxx; > > christoffer.dall@xxxxxxxxxx; eric.auger@xxxxxxxxxx; pranavkumar@xxxxxxxxxx; > > marc.zyngier@xxxxxxx; will.deacon@xxxxxxx > > Subject: Re: [RFC PATCH 6/6] arm-smmu: Allow to set iommu mapping for > > MSI > > > > On Wed, 2015-09-30 at 20:26 +0530, Bharat Bhushan wrote: > > > Finally ARM SMMU declare that iommu-mapping for MSI-pages are not set > > > automatically and it should be set explicitly. > > > > > > Signed-off-by: Bharat Bhushan <Bharat.Bhushan@xxxxxxxxxxxxx> > > > --- > > > drivers/iommu/arm-smmu.c | 7 ++++++- > > > 1 file changed, 6 insertions(+), 1 deletion(-) > > > > > > diff --git a/drivers/iommu/arm-smmu.c b/drivers/iommu/arm-smmu.c > > index > > > a3956fb..9d37e72 100644 > > > --- a/drivers/iommu/arm-smmu.c > > > +++ b/drivers/iommu/arm-smmu.c > > > @@ -1401,13 +1401,18 @@ static int arm_smmu_domain_get_attr(struct > > iommu_domain *domain, > > > enum iommu_attr attr, void *data) { > > > struct arm_smmu_domain *smmu_domain = > > to_smmu_domain(domain); > > > + struct iommu_domain_msi_maps *msi_maps; > > > > > > switch (attr) { > > > case DOMAIN_ATTR_NESTING: > > > *(int *)data = (smmu_domain->stage == > > ARM_SMMU_DOMAIN_NESTED); > > > return 0; > > > case DOMAIN_ATTR_MSI_MAPPING: > > > - /* Dummy handling added */ > > > + msi_maps = data; > > > + > > > + msi_maps->automap = false; > > > + msi_maps->override_automap = true; > > > + > > > return 0; > > > default: > > > return -ENODEV; > > > > In previous discussions I understood one of the problems you were trying to > > solve was having a limited number of MSI banks and while you may be able > > to get isolated MSI banks for some number of users, it wasn't unlimited and > > sharing may be required. I don't see any of that addressed in this series. > > That problem was on PowerPC. Infact there were two problems, one which MSI bank to be used and second how to create iommu-mapping for device assigned to userspace. > First problem was PowerPC specific and that will be solved separately. > For second problem, earlier I tried to added a couple of MSI specific ioctls and you suggested (IIUC) that we should have a generic reserved-iova type of API and then we can map MSI bank using reserved-iova and this will not require involvement of user-space. > > > > > Also, the management of reserved IOVAs vs MSI addresses looks really > > dubious to me. How does your platform pick an MSI address and what are > > we breaking by covertly changing it? We seem to be masking over at the > > VFIO level, where there should be lower level interfaces doing the right thing > > when we configure MSI on the device. > > Yes, In my understanding the right solution should be: > 1) VFIO driver should know what physical-msi-address will be used for devices in an iommu-group. > I did not find an generic API, on PowerPC I added some function in ffrescale msi-driver and called from vfio-iommu-fsl-pamu.c (not yet upstreamed). > 2) VFIO driver should know what IOVA to be used for creating iommu-mapping (VFIO APIs patch of this patch series) > 3) VFIO driver will create the iommu-mapping using (1) and (2) > 4) VFIO driver should be able to tell the msi-driver that for a given device it should use different IOVA. So when composing the msi message (for the devices is the given iommu-group) it should use that programmed iova as MSI-address. This interface also needed to be developed. > > I was not sure of which approach we should take. The current approach in the patch is simple to develop so I went ahead to take input but I agree this does not look very good. > What do you think, should drop this approach and work out the approach as described above. I'm certainly not interested in applying an maintaining an interim solution that isn't the right one. It seems like VFIO is too involved in this process in your example. On x86 we have per vector isolation and the only thing we're missing is reporting back of the region used by MSI vectors as reserved IOVA space (but it's standard on x86, so an x86 VM user will never use it for IOVA). In your model, the MSI IOVA space is programmable, but it has page granularity (I assume). Therefore we shouldn't be sharing that page with anyone. That seems to suggest we need to allocate a page of vector space from the host kernel, setup the IOVA mapping, and then the host kernel should know to only allocate MSI vectors for these devices from that pre-allocated page. Otherwise we need to call the interrupts unsafe, like we do on x86 without interrupt remapping. Thanks, Alex -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html