> -----Original Message----- > From: Brost, Matthew <matthew.brost@xxxxxxxxx> > Sent: Thursday, January 2, 2025 7:15 AM > To: Upadhyay, Tejas <tejas.upadhyay@xxxxxxxxx> > Cc: intel-xe@xxxxxxxxxxxxxxxxxxxxx; dri-devel@xxxxxxxxxxxxxxxxxxxxx; > christian.koenig@xxxxxxx; Auld, Matthew <matthew.auld@xxxxxxxxx>; > Mrozek, Michal <michal.mrozek@xxxxxxxxx> > Subject: Re: [PATCH V6] drm/xe/mmap: Add mmap support for PCI memory > barrier > > On Wed, Dec 18, 2024 at 06:26:28PM +0530, Tejas Upadhyay wrote: > > In order to avoid having userspace to use MI_MEM_FENCE, we are adding > > a mechanism for userspace to generate a PCI memory barrier with low > > overhead (avoiding IOCTL call as well as writing to VRAM will adds > > some overhead). > > > > This is implemented by memory-mapping a page as uncached that is > > backed by MMIO on the dGPU and thus allowing userspace to do memory > > write to the page without invoking an IOCTL. > > We are selecting the MMIO so that it is not accessible from the PCI > > bus so that the MMIO writes themselves are ignored, but the PCI memory > > barrier will still take action as the MMIO filtering will happen after > > the memory barrier effect. > > > > When we detect special defined offset in mmap(), We are mapping 4K > > page which contains the last of page of doorbell MMIO range to > > userspace for same purpose. > > > > For user to query special offset we are adding special flag in > > mmap_offset ioctl which needs to be passed as follows, struct > > drm_xe_gem_mmap_offset mmo = { > > .handle = 0, /* this must be 0 */ > > .flags = DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER, > > }; > > igt_ioctl(fd, DRM_IOCTL_XE_GEM_MMAP_OFFSET, &mmo); map = > mmap(NULL, > > size, PROT_WRITE, MAP_SHARED, fd, mmo); > > > > Note: Test coverage for this is added by IGT > > https://patchwork.freedesktop.org/series/140368/ here. > > UMD implementing test, once PR is ready will attach with > > this patch. > > > > V6(MAuld) > > - Move physical mmap to fault handler > > - Modify kernel-doc and attach UMD PR when ready > > V5(MAuld) > > - Return invalid early in case of non 4K PAGE_SIZE > > - Format kernel-doc and add note for 4K PAGE_SIZE HW limit > > V4(MAuld) > > - Add kernel-doc for uapi change > > - Restrict page size to 4K > > V3(MAuld) > > - Remove offset defination from UAPI to be able to change later > > - Edit commit message for special flag addition > > V2(MAuld) > > - Add fault handler with dummy page to handle unplug device > > - Add Build check for special offset to be below normal start page > > - Test d3hot, mapping seems to be valid in d3hot as well > > - Add more info to commit message > > > > Test-with: 20241118113606.1490397-1-tejas.upadhyay@xxxxxxxxx > > Cc: Matthew Auld <matthew.auld@xxxxxxxxx> > > Cc: Michal Mrozek <michal.mrozek@xxxxxxxxx> > > Reviewed-by: Matthew Auld <matthew.auld@xxxxxxxxx> > > Signed-off-by: Tejas Upadhyay <tejas.upadhyay@xxxxxxxxx> > > Do we need a query to see if the Xe driver has this uAPI? Currently, I have though of using below to check driver support for this UAPI, +static bool is_pci_membarrier_supported(int fd) +{ + struct drm_xe_gem_mmap_offset mmo = { + .handle = 0, + .flags = DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER, + }; + + return (igt_ioctl(fd, DRM_IOCTL_XE_GEM_MMAP_OFFSET, &mmo) == 0); +} IGT already using this. Please let me know if it makes sense. Tejas > > Matt > > > --- > > drivers/gpu/drm/xe/xe_bo.c | 16 ++++- > > drivers/gpu/drm/xe/xe_bo.h | 2 + > > drivers/gpu/drm/xe/xe_device.c | 103 > ++++++++++++++++++++++++++++++++- > > include/uapi/drm/xe_drm.h | 29 +++++++++- > > 4 files changed, 147 insertions(+), 3 deletions(-) > > > > diff --git a/drivers/gpu/drm/xe/xe_bo.c b/drivers/gpu/drm/xe/xe_bo.c > > index e6c896ad5602..86f2c00a0afd 100644 > > --- a/drivers/gpu/drm/xe/xe_bo.c > > +++ b/drivers/gpu/drm/xe/xe_bo.c > > @@ -2263,9 +2263,23 @@ int xe_gem_mmap_offset_ioctl(struct > drm_device *dev, void *data, > > XE_IOCTL_DBG(xe, args->reserved[0] || args->reserved[1])) > > return -EINVAL; > > > > - if (XE_IOCTL_DBG(xe, args->flags)) > > + if (XE_IOCTL_DBG(xe, args->flags & > > + ~DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER)) > > return -EINVAL; > > > > + if (args->flags & DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER) { > > + if (XE_IOCTL_DBG(xe, args->handle)) > > + return -EINVAL; > > + > > + if (XE_IOCTL_DBG(xe, PAGE_SIZE > SZ_4K)) > > + return -EINVAL; > > + > > + BUILD_BUG_ON(((XE_PCI_BARRIER_MMAP_OFFSET >> > XE_PTE_SHIFT) + > > + SZ_4K) >= DRM_FILE_PAGE_OFFSET_START); > > + args->offset = XE_PCI_BARRIER_MMAP_OFFSET; > > + return 0; > > + } > > + > > gem_obj = drm_gem_object_lookup(file, args->handle); > > if (XE_IOCTL_DBG(xe, !gem_obj)) > > return -ENOENT; > > diff --git a/drivers/gpu/drm/xe/xe_bo.h b/drivers/gpu/drm/xe/xe_bo.h > > index d9386ab03140..04995c5ced32 100644 > > --- a/drivers/gpu/drm/xe/xe_bo.h > > +++ b/drivers/gpu/drm/xe/xe_bo.h > > @@ -75,6 +75,8 @@ > > > > #define XE_BO_PROPS_INVALID (-1) > > > > +#define XE_PCI_BARRIER_MMAP_OFFSET (0x50 << XE_PTE_SHIFT) > > + > > struct sg_table; > > > > struct xe_bo *xe_bo_alloc(void); > > diff --git a/drivers/gpu/drm/xe/xe_device.c > > b/drivers/gpu/drm/xe/xe_device.c index 7f021ec5f8e7..bacfeec9dc0b > > 100644 > > --- a/drivers/gpu/drm/xe/xe_device.c > > +++ b/drivers/gpu/drm/xe/xe_device.c > > @@ -232,12 +232,113 @@ static long xe_drm_compat_ioctl(struct file > > *file, unsigned int cmd, unsigned lo #define xe_drm_compat_ioctl NULL > > #endif > > > > +static void barrier_open(struct vm_area_struct *vma) { > > + drm_dev_get(vma->vm_private_data); > > +} > > + > > +static void barrier_close(struct vm_area_struct *vma) { > > + drm_dev_put(vma->vm_private_data); > > +} > > + > > +static void barrier_release_dummy_page(struct drm_device *dev, void > > +*res) { > > + struct page *dummy_page = (struct page *)res; > > + > > + __free_page(dummy_page); > > +} > > + > > +static vm_fault_t barrier_fault(struct vm_fault *vmf) { > > + struct drm_device *dev = vmf->vma->vm_private_data; > > + struct vm_area_struct *vma = vmf->vma; > > + vm_fault_t ret = VM_FAULT_NOPAGE; > > + pgprot_t prot; > > + int idx; > > + > > + prot = vm_get_page_prot(vma->vm_flags); > > + > > + if (drm_dev_enter(dev, &idx)) { > > + unsigned long pfn; > > + > > +#define LAST_DB_PAGE_OFFSET 0x7ff001 > > + pfn = PHYS_PFN(pci_resource_start(to_pci_dev(dev->dev), 0) > + > > + LAST_DB_PAGE_OFFSET); > > + ret = vmf_insert_pfn_prot(vma, vma->vm_start, pfn, > > + pgprot_noncached(prot)); > > + drm_dev_exit(idx); > > + } else { > > + struct page *page; > > + > > + /* Allocate new dummy page to map all the VA range in this > VMA to it*/ > > + page = alloc_page(GFP_KERNEL | __GFP_ZERO); > > + if (!page) > > + return VM_FAULT_OOM; > > + > > + /* Set the page to be freed using drmm release action */ > > + if (drmm_add_action_or_reset(dev, > barrier_release_dummy_page, page)) > > + return VM_FAULT_OOM; > > + > > + ret = vmf_insert_pfn_prot(vma, vma->vm_start, > page_to_pfn(page), > > + prot); > > + } > > + > > + return ret; > > +} > > + > > +static const struct vm_operations_struct vm_ops_barrier = { > > + .open = barrier_open, > > + .close = barrier_close, > > + .fault = barrier_fault, > > +}; > > + > > +static int xe_pci_barrier_mmap(struct file *filp, > > + struct vm_area_struct *vma) { > > + struct drm_file *priv = filp->private_data; > > + struct drm_device *dev = priv->minor->dev; > > + > > + if (vma->vm_end - vma->vm_start > SZ_4K) > > + return -EINVAL; > > + > > + if (is_cow_mapping(vma->vm_flags)) > > + return -EINVAL; > > + > > + if (vma->vm_flags & (VM_READ | VM_EXEC)) > > + return -EINVAL; > > + > > + vm_flags_clear(vma, VM_MAYREAD | VM_MAYEXEC); > > + vm_flags_set(vma, VM_PFNMAP | VM_DONTEXPAND | > VM_DONTDUMP | VM_IO); > > + vma->vm_ops = &vm_ops_barrier; > > + vma->vm_private_data = dev; > > + drm_dev_get(vma->vm_private_data); > > + > > + return 0; > > +} > > + > > +static int xe_mmap(struct file *filp, struct vm_area_struct *vma) { > > + struct drm_file *priv = filp->private_data; > > + struct drm_device *dev = priv->minor->dev; > > + > > + if (drm_dev_is_unplugged(dev)) > > + return -ENODEV; > > + > > + switch (vma->vm_pgoff) { > > + case XE_PCI_BARRIER_MMAP_OFFSET >> XE_PTE_SHIFT: > > + return xe_pci_barrier_mmap(filp, vma); > > + } > > + > > + return drm_gem_mmap(filp, vma); > > +} > > + > > static const struct file_operations xe_driver_fops = { > > .owner = THIS_MODULE, > > .open = drm_open, > > .release = drm_release_noglobal, > > .unlocked_ioctl = xe_drm_ioctl, > > - .mmap = drm_gem_mmap, > > + .mmap = xe_mmap, > > .poll = drm_poll, > > .read = drm_read, > > .compat_ioctl = xe_drm_compat_ioctl, diff --git > > a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h index > > f62689ca861a..cac607a30f6d 100644 > > --- a/include/uapi/drm/xe_drm.h > > +++ b/include/uapi/drm/xe_drm.h > > @@ -811,6 +811,32 @@ struct drm_xe_gem_create { > > > > /** > > * struct drm_xe_gem_mmap_offset - Input of > > &DRM_IOCTL_XE_GEM_MMAP_OFFSET > > + * > > + * The @flags can be: > > + * - %DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER - For user to query > > + special offset > > + * for use in mmap ioctl. Writing to the returned mmap address will > > + generate a > > + * PCI memory barrier with low overhead (avoiding IOCTL call as well > > + as writing > > + * to VRAM which would also add overhead), acting like an > > + MI_MEM_FENCE > > + * instruction. > > + * > > + * Note: The mmap size can be at most 4K, due to HW limitations. As > > + a result > > + * this interface is only supported on CPU architectures that > > + support 4K page > > + * size. The mmap_offset ioctl will detect this and gracefully > > + return an > > + * error, where userspace is expected to have a different fallback > > + method for > > + * triggering a barrier. > > + * > > + * Roughly the usage would be as follows: > > + * > > + * .. code-block:: C > > + * > > + * struct drm_xe_gem_mmap_offset mmo = { > > + * .handle = 0, // must be set to 0 > > + * .flags = DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER, > > + * }; > > + * > > + * err = ioctl(fd, DRM_IOCTL_XE_GEM_MMAP_OFFSET, &mmo); > > + * map = mmap(NULL, size, PROT_WRITE, MAP_SHARED, fd, mmo.offset); > > + * map[i] = 0xdeadbeaf; // issue barrier > > */ > > struct drm_xe_gem_mmap_offset { > > /** @extensions: Pointer to the first extension struct, if any */ @@ > > -819,7 +845,8 @@ struct drm_xe_gem_mmap_offset { > > /** @handle: Handle for the object being mapped. */ > > __u32 handle; > > > > - /** @flags: Must be zero */ > > +#define DRM_XE_MMAP_OFFSET_FLAG_PCI_BARRIER (1 << 0) > > + /** @flags: Flags */ > > __u32 flags; > > > > /** @offset: The fake offset to use for subsequent mmap call */ > > -- > > 2.34.1 > >