Re: [RFC] drm/vgem: Don't use get_page in fault handler.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am 07.07.20 um 10:59 schrieb Daniel Vetter:
On Tue, Jul 7, 2020 at 9:27 AM Lepton Wu <ytht.net@xxxxxxxxx> wrote:
For pages which are allocated in ttm with transparent huge pages,
tail pages have zero as reference count. The current vgem code use
get_page on them and it will trigger BUG when release_pages get called
on those pages later.

Here I attach the minimal code to trigger this bug on a qemu VM which
enables virtio gpu (card1) and vgem (card0). BTW, since the upstream
virtio gpu has changed to use drm gem and moved away from ttm. So we
have to use an old kernel like 5.4 to reproduce it. But I guess
same thing could happen for a real GPU if the real GPU use similar code
path to allocate pages from ttm. I am not sure about two things: first, do we
need to fix this? will a real GPU hit this code path? Second, suppose this
need to be fixed, should this be fixed in ttm side or vgem side?  If we remove
"huge_flags &= ~__GFP_COMP" from ttm_get_pages, then get_page in vgem works
fine. But it's there for fix another bug:
https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fbugs.freedesktop.org%2Fshow_bug.cgi%3Fid%3D103138&amp;data=02%7C01%7Cchristian.koenig%40amd.com%7Cfc0831be8fd848abfd8908d82254266d%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637297092113216357&amp;sdata=S%2BtLJyB8mqyyE%2BSIjbfHM6FGFuFjrfI50tahpaoJ3wU%3D&amp;reserved=0
It could also be "fixed" with this patch. But I am really not sure if this is
the way to go.
For imported dma-buf objects, vgem should not handle this itself, but
forward to the exporter through the dma_buf_mmap stuff. We have
helpers for this all now, probably just not wired up correctly. Trying
to ensure that all combinations of mmap code across all drivers work
the same doesn't work.

Yes, Daniel is right what vgem does here is fundamentally broken in many ways.

First of all nobody should ever call get_page() on the pages TTM has allocated. Those pages come out of a block box and should not be touched by anybody.

Second the DMA-buf doesn't even need to have pages backing them in the case of P2P.

And third you mess up the coherency with this massively. E.g. no correct cache flushes any more.

Regards,
Christian.

Caveat: I might not be understanding exactly what's going on here :-)
-Daniel

Here is the code to reproduce this bug:

unsigned int WIDTH = 1024;
unsigned int HEIGHT = 513;
unsigned int size = WIDTH * HEIGHT * 4;

int work(int vfd, int dfd, int handle) {
         int ret;
         struct drm_prime_handle hf = {.handle =  handle };
         ret = ioctl(dfd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &hf);
         fprintf(stderr, "fd is %d\n", hf.fd);
         hf.flags = DRM_CLOEXEC | DRM_RDWR;
         ret = ioctl(vfd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &hf);
         fprintf(stderr, "new handle is %d\n", hf.handle);
         struct drm_mode_map_dumb map = {.handle = hf.handle };
         ret = ioctl(vfd, DRM_IOCTL_MODE_MAP_DUMB, &map);
         fprintf(stderr, "need map at offset %lld\n", map.offset);
         unsigned char * ptr = mmap(NULL, size, PROT_READ|PROT_WRITE, MAP_SHARED, vfd,
                           map.offset);
         memset(ptr, 2, size);
         munmap(ptr, size);
}

int main()
{
         int vfd = open("/dev/dri/card0", O_RDWR); // vgem
         int dfd = open("/dev/dri/card1", O_RDWR); // virtio gpu

         int ret;
         struct drm_mode_create_dumb ct = {};

         ct.height = HEIGHT;
         ct.width = WIDTH;
         ct.bpp = 32;
         ret = ioctl(dfd, DRM_IOCTL_MODE_CREATE_DUMB, &ct);
         work(vfd, dfd, ct.handle);
         fprintf(stderr, "done\n");
}

Signed-off-by: Lepton Wu <ytht.net@xxxxxxxxx>
---
  drivers/gpu/drm/vgem/vgem_drv.c | 7 ++-----
  1 file changed, 2 insertions(+), 5 deletions(-)

diff --git a/drivers/gpu/drm/vgem/vgem_drv.c b/drivers/gpu/drm/vgem/vgem_drv.c
index ec1a8ebb6f1b..be3d97e29804 100644
--- a/drivers/gpu/drm/vgem/vgem_drv.c
+++ b/drivers/gpu/drm/vgem/vgem_drv.c
@@ -87,9 +87,8 @@ static vm_fault_t vgem_gem_fault(struct vm_fault *vmf)

         mutex_lock(&obj->pages_lock);
         if (obj->pages) {
-               get_page(obj->pages[page_offset]);
-               vmf->page = obj->pages[page_offset];
-               ret = 0;
+               ret = vmf_insert_pfn(vmf->vma, vmf->address,
+                                    page_to_pfn(obj->pages[page_offset]));
         }
         mutex_unlock(&obj->pages_lock);
         if (ret) {
@@ -263,7 +262,6 @@ static struct drm_ioctl_desc vgem_ioctls[] = {

  static int vgem_mmap(struct file *filp, struct vm_area_struct *vma)
  {
-       unsigned long flags = vma->vm_flags;
         int ret;

         ret = drm_gem_mmap(filp, vma);
@@ -273,7 +271,6 @@ static int vgem_mmap(struct file *filp, struct vm_area_struct *vma)
         /* Keep the WC mmaping set by drm_gem_mmap() but our pages
          * are ordinary and not special.
          */
-       vma->vm_flags = flags | VM_DONTEXPAND | VM_DONTDUMP;
         return 0;
  }

--
2.27.0.212.ge8ba1cc988-goog

_______________________________________________
dri-devel mailing list
dri-devel@xxxxxxxxxxxxxxxxxxxxx
https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.freedesktop.org%2Fmailman%2Flistinfo%2Fdri-devel&amp;data=02%7C01%7Cchristian.koenig%40amd.com%7Cfc0831be8fd848abfd8908d82254266d%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637297092113216357&amp;sdata=C1sO1pUj0wnga%2BodcDRl9gA%2FNpIGO368KeQUAFyWK2g%3D&amp;reserved=0



_______________________________________________
dri-devel mailing list
dri-devel@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/dri-devel



[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux