On Thu, Oct 05, 2017 at 11:08:12AM +0100, Jean-Philippe Brucker wrote: > Hi Jordan, > > On 04/10/17 20:43, Jordan Crouse wrote: > > Trying to start back up the conversation about multiple address > > spaces for IOMMU devices. If you will remember Jean-Philippe posted > > some patches back in February for SVM on arm-smmu-v3. > > > > For quite some time the downstream Snapdragon kernels have supported > > something we call "per-process" page tables for the GPU. As with full SVM > > this involves creating a virtual address space per task but unlike SVM > > we don't automatically share the page table from the CPU. Instead we > > want to create a new page table and explicitly map/unmap address ranges > > into it. We provide the physical address of the page table to the GPU and > > it goes through the mechanics of programming the TTBR0 and invalidating > > the TLB when starts executing a submission for a given task. > > Why does the GPU need the pgd? Does it implement its own MMU specifically > for process contexts? I understand you don't use PASIDs/SSIDs to isolate > process PT but context switch instead? The GPU uses the same SSID for all transactions. On the Snapdragon the GPU has access to some of the context bank registers for the IOMMU. The kernel driver writes the address of the pagetable for the subsequent submission into a special opcode. When the GPU starts executing that submission it executes a complicated process of stalling the bus and writing the physical address of the new pagetable directly to the TTBR0 register. It is as messy as it sounds but it works given the restrictions on hardware. > > As with all things IOMMU this discussion needs to be split into two parts - > > the API and the implementation. I want to focus on the generic API for this > > email. Shortly after Jean-Philippe posted his patches I sent out a rough > > prototype of how the downstream solution worked [1]: > > > > +-----------------+ +------------------+ > > | "master" domain | ---> | "dynamic" domain | > > +-----------------+ \ +------------------+ > > \ > > \ +------------------+ > > - | "dynamic" domain | > > +------------------+ > > I also considered using hierarchical domains in my first prototype, but it > didn't seem to fit the IOMMU API. In the RFC that I intend to post this > week, I propose an iommu_process structure for everything process related. > > I'm not sure if my new proposal fits your model since I didn't intend > iommu_process to be controllable externally with an IOMMU map/unmap > interface (the meat of the bind/unbind API is really page table sharing). > In v2 bind/unbind still only returns a PASID, not the process structure, > but I'll Cc you so we can work something out. I saw your CC today - I'll look and see what you've come up with. > > Given a "master" domain (created in the normal way) we can create any number > > of "dynamic" domains which share the same configuration as the master (table > > format, context bank, quirks, etc). When the dynamic domain is allocated/ > > attached it creates a new page table - for all intents and purposes this is > > a "real" domain except that it doesn't actually touch the hardware. We can > > use this domain with iommu_map() / iommu_unmap() as usual and then pass the > > physical address (acquired through a IOMMU domain attribute) to the GPU and > > everything works. > > > > The main goal for this approach was to try to use the iommu API instead of > > teaching the GPU driver how to deal with several generations of page table > > formats and IOMMU devices. Shoehorning it into the domain struct was a > > path of least resistance that has served Snapdragon well but it was never > > really anything we considered to be a generic solution. > > > > In the SVM patches, Jean-Philippe introduces iommu_bind_task(): > > https://patchwork.codeaurora.org/patch/184777/. Given a struct task and > > a bit of other stuff it goes off and does SVM magic. > > > > My proposal would be to extend this slightly and return a iommu_task > > struct from iommu_bind_task: > > > > struct iommu_task *iommu_bind_task(struct device *dev, strut task_strut *task, > > int *pasid, int flags, void *priv); > > Since the GPU driver acts as a proxy and the PT are not shared, I suppose > you don't need the task_struct at all? Or maybe just for cleaning up on > process exit? Right - we just use it as a handy unique token. Jordan -- The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum, a Linux Foundation Collaborative Project -- To unsubscribe from this list: send the line "unsubscribe linux-arm-msm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html