[PATCH 11/17] drm/gem_object: Rename struct drm_gem_object::dev to drm

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



"dev" is usually used for struct device pointers. Using it for struct
drm_device pointers is irritating (for me at least). Replace it
by "drm" and adapt all users.

Signed-off-by: Uwe Kleine-König <u.kleine-koenig@xxxxxxxxxxxxxx>
---
 drivers/accel/ivpu/ivpu_gem.c                 |  2 +-
 drivers/accel/ivpu/ivpu_gem.h                 |  2 +-
 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c    |  4 +-
 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c   |  4 +-
 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c       |  2 +-
 drivers/gpu/drm/armada/armada_gem.c           |  6 +--
 drivers/gpu/drm/drm_gem.c                     | 14 +++---
 drivers/gpu/drm/drm_gem_dma_helper.c          | 11 ++---
 drivers/gpu/drm/drm_gem_shmem_helper.c        | 39 +++++++++--------
 drivers/gpu/drm/drm_gem_vram_helper.c         |  4 +-
 drivers/gpu/drm/drm_prime.c                   | 12 +++---
 drivers/gpu/drm/etnaviv/etnaviv_gem.c         | 18 ++++----
 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c   |  2 +-
 drivers/gpu/drm/exynos/exynos_drm_gem.c       | 15 ++++---
 drivers/gpu/drm/gma500/gem.c                  |  6 +--
 drivers/gpu/drm/i915/display/intel_dpt.c      |  2 +-
 drivers/gpu/drm/i915/display/intel_fb.c       |  6 +--
 .../gpu/drm/i915/display/intel_frontbuffer.c  | 18 ++++----
 .../drm/i915/display/skl_universal_plane.c    |  2 +-
 drivers/gpu/drm/i915/gem/i915_gem_clflush.c   |  2 +-
 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c    |  8 ++--
 drivers/gpu/drm/i915/gem/i915_gem_domain.c    |  6 +--
 drivers/gpu/drm/i915/gem/i915_gem_internal.c  |  2 +-
 drivers/gpu/drm/i915/gem/i915_gem_mman.c      | 18 ++++----
 drivers/gpu/drm/i915/gem/i915_gem_object.c    | 26 +++++------
 drivers/gpu/drm/i915/gem/i915_gem_object.h    |  2 +-
 drivers/gpu/drm/i915/gem/i915_gem_pages.c     |  8 ++--
 drivers/gpu/drm/i915/gem/i915_gem_phys.c      | 12 +++---
 drivers/gpu/drm/i915/gem/i915_gem_shmem.c     |  4 +-
 drivers/gpu/drm/i915/gem/i915_gem_stolen.c    |  8 ++--
 drivers/gpu/drm/i915/gem/i915_gem_tiling.c    |  8 ++--
 drivers/gpu/drm/i915/gem/i915_gem_ttm.c       | 24 +++++------
 drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c  |  4 +-
 drivers/gpu/drm/i915/gem/i915_gem_userptr.c   | 10 ++---
 .../gpu/drm/i915/gem/selftests/huge_pages.c   |  4 +-
 .../i915/gem/selftests/i915_gem_client_blt.c  |  2 +-
 .../drm/i915/gem/selftests/i915_gem_context.c |  4 +-
 .../drm/i915/gem/selftests/i915_gem_dmabuf.c  |  4 +-
 .../drm/i915/gem/selftests/i915_gem_migrate.c |  2 +-
 .../drm/i915/gem/selftests/i915_gem_mman.c    |  8 ++--
 drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c  |  2 +-
 drivers/gpu/drm/i915/gt/shmem_utils.c         |  2 +-
 drivers/gpu/drm/i915/gvt/dmabuf.c             |  2 +-
 drivers/gpu/drm/i915/i915_gem.c               | 12 +++---
 drivers/gpu/drm/i915/i915_gem_gtt.c           |  6 +--
 drivers/gpu/drm/i915/i915_vma.c               | 10 ++---
 drivers/gpu/drm/lima/lima_gem.c               |  4 +-
 drivers/gpu/drm/loongson/lsdc_gem.c           | 12 +++---
 drivers/gpu/drm/loongson/lsdc_ttm.c           |  6 +--
 drivers/gpu/drm/mediatek/mtk_drm_gem.c        |  6 +--
 drivers/gpu/drm/msm/msm_gem.c                 | 42 +++++++++---------
 drivers/gpu/drm/msm/msm_gem_prime.c           |  2 +-
 drivers/gpu/drm/nouveau/nouveau_bo.c          |  2 +-
 drivers/gpu/drm/nouveau/nouveau_prime.c       |  2 +-
 drivers/gpu/drm/omapdrm/omap_gem.c            | 43 ++++++++++---------
 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c     |  4 +-
 drivers/gpu/drm/panfrost/panfrost_gem.c       |  2 +-
 drivers/gpu/drm/panfrost/panfrost_mmu.c       |  4 +-
 drivers/gpu/drm/qxl/qxl_gem.c                 |  2 +-
 drivers/gpu/drm/qxl/qxl_object.c              |  4 +-
 drivers/gpu/drm/qxl/qxl_object.h              |  2 +-
 drivers/gpu/drm/qxl/qxl_release.c             |  2 +-
 drivers/gpu/drm/qxl/qxl_ttm.c                 |  2 +-
 drivers/gpu/drm/radeon/radeon_prime.c         |  2 +-
 drivers/gpu/drm/rockchip/rockchip_drm_gem.c   | 23 +++++-----
 drivers/gpu/drm/tegra/gem.c                   | 21 ++++-----
 drivers/gpu/drm/ttm/ttm_bo_util.c             |  2 +-
 drivers/gpu/drm/ttm/ttm_bo_vm.c               |  4 +-
 drivers/gpu/drm/v3d/v3d_bo.c                  |  4 +-
 drivers/gpu/drm/v3d/v3d_mmu.c                 |  4 +-
 drivers/gpu/drm/vc4/vc4_bo.c                  | 20 ++++-----
 drivers/gpu/drm/vc4/vc4_validate.c            |  2 +-
 drivers/gpu/drm/vc4/vc4_validate_shaders.c    |  2 +-
 drivers/gpu/drm/virtio/virtgpu_gem.c          |  6 +--
 drivers/gpu/drm/virtio/virtgpu_object.c       |  4 +-
 drivers/gpu/drm/virtio/virtgpu_prime.c        |  6 +--
 drivers/gpu/drm/virtio/virtgpu_vram.c         |  8 ++--
 drivers/gpu/drm/vmwgfx/vmwgfx_gem.c           |  3 +-
 drivers/gpu/drm/xen/xen_drm_front.c           |  4 +-
 drivers/gpu/drm/xen/xen_drm_front_gem.c       |  2 +-
 include/drm/drm_gem.h                         |  4 +-
 81 files changed, 319 insertions(+), 312 deletions(-)

diff --git a/drivers/accel/ivpu/ivpu_gem.c b/drivers/accel/ivpu/ivpu_gem.c
index 52b339aefadc..75b78e63acc3 100644
--- a/drivers/accel/ivpu/ivpu_gem.c
+++ b/drivers/accel/ivpu/ivpu_gem.c
@@ -464,7 +464,7 @@ static struct sg_table *ivpu_bo_get_sg_table(struct drm_gem_object *obj)
 	if (ret)
 		return ERR_PTR(ret);
 
-	return drm_prime_pages_to_sg(obj->dev, bo->pages, npages);
+	return drm_prime_pages_to_sg(obj->drm, bo->pages, npages);
 }
 
 static vm_fault_t ivpu_vm_fault(struct vm_fault *vmf)
diff --git a/drivers/accel/ivpu/ivpu_gem.h b/drivers/accel/ivpu/ivpu_gem.h
index 6b0ceda5f253..ae6252513fbb 100644
--- a/drivers/accel/ivpu/ivpu_gem.h
+++ b/drivers/accel/ivpu/ivpu_gem.h
@@ -99,7 +99,7 @@ static inline pgprot_t ivpu_bo_pgprot(struct ivpu_bo *bo, pgprot_t prot)
 
 static inline struct ivpu_device *ivpu_bo_to_vdev(struct ivpu_bo *bo)
 {
-	return to_ivpu_device(bo->base.dev);
+	return to_ivpu_device(bo->base.drm);
 }
 
 static inline void *ivpu_to_cpu_addr(struct ivpu_bo *bo, u32 vpu_addr)
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
index b4fcad0e62f7..8f7fee850cde 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
@@ -524,11 +524,11 @@ int amdgpu_amdkfd_get_dmabuf_info(struct amdgpu_device *adev, int dma_buf_fd,
 		goto out_put;
 
 	obj = dma_buf->priv;
-	if (obj->dev->driver != adev_to_drm(adev)->driver)
+	if (obj->drm->driver != adev_to_drm(adev)->driver)
 		/* Can't handle buffers from different drivers */
 		goto out_put;
 
-	adev = drm_to_adev(obj->dev);
+	adev = drm_to_adev(obj->drm);
 	bo = gem_to_amdgpu_bo(obj);
 	if (!(bo->preferred_domains & (AMDGPU_GEM_DOMAIN_VRAM |
 				    AMDGPU_GEM_DOMAIN_GTT)))
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
index 12210598e5b8..9461cc8c3257 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
@@ -168,7 +168,7 @@ static struct sg_table *amdgpu_dma_buf_map(struct dma_buf_attachment *attach,
 
 	switch (bo->tbo.resource->mem_type) {
 	case TTM_PL_TT:
-		sgt = drm_prime_pages_to_sg(obj->dev,
+		sgt = drm_prime_pages_to_sg(obj->drm,
 					    bo->tbo.ttm->pages,
 					    bo->tbo.ttm->num_pages);
 		if (IS_ERR(sgt))
@@ -438,7 +438,7 @@ struct drm_gem_object *amdgpu_gem_prime_import(struct drm_device *dev,
 
 	if (dma_buf->ops == &amdgpu_dmabuf_ops) {
 		obj = dma_buf->priv;
-		if (obj->dev == dev) {
+		if (obj->drm == dev) {
 			/*
 			 * Importing dmabuf exported from out own gem increases
 			 * refcount on gem itself instead of f_count of dmabuf.
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
index 7da871972a8e..b3340d080e0b 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
@@ -48,7 +48,7 @@ static const struct drm_gem_object_funcs amdgpu_gem_object_funcs;
 static vm_fault_t amdgpu_gem_fault(struct vm_fault *vmf)
 {
 	struct ttm_buffer_object *bo = vmf->vma->vm_private_data;
-	struct drm_device *ddev = bo->base.dev;
+	struct drm_device *ddev = bo->base.drm;
 	vm_fault_t ret;
 	int idx;
 
diff --git a/drivers/gpu/drm/armada/armada_gem.c b/drivers/gpu/drm/armada/armada_gem.c
index 26d10065d534..9d1768da4c14 100644
--- a/drivers/gpu/drm/armada/armada_gem.c
+++ b/drivers/gpu/drm/armada/armada_gem.c
@@ -41,7 +41,7 @@ static size_t roundup_gem_size(size_t size)
 void armada_gem_free_object(struct drm_gem_object *obj)
 {
 	struct armada_gem_object *dobj = drm_to_armada_gem(obj);
-	struct armada_private *priv = drm_to_armada_dev(obj->dev);
+	struct armada_private *priv = drm_to_armada_dev(obj->drm);
 
 	DRM_DEBUG_DRIVER("release obj %p\n", dobj);
 
@@ -493,7 +493,7 @@ armada_gem_prime_export(struct drm_gem_object *obj, int flags)
 	exp_info.flags = O_RDWR;
 	exp_info.priv = obj;
 
-	return drm_gem_dmabuf_export(obj->dev, &exp_info);
+	return drm_gem_dmabuf_export(obj->drm, &exp_info);
 }
 
 struct drm_gem_object *
@@ -504,7 +504,7 @@ armada_gem_prime_import(struct drm_device *dev, struct dma_buf *buf)
 
 	if (buf->ops == &armada_gem_prime_dmabuf_ops) {
 		struct drm_gem_object *obj = buf->priv;
-		if (obj->dev == dev) {
+		if (obj->drm == dev) {
 			/*
 			 * Importing our own dmabuf(s) increases the
 			 * refcount on the gem object itself.
diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c
index c18686f434d4..ce9a8a3d281c 100644
--- a/drivers/gpu/drm/drm_gem.c
+++ b/drivers/gpu/drm/drm_gem.c
@@ -154,7 +154,7 @@ void drm_gem_private_object_init(struct drm_device *dev,
 {
 	BUG_ON((size & (PAGE_SIZE - 1)) != 0);
 
-	obj->dev = dev;
+	obj->drm = dev;
 	obj->filp = NULL;
 
 	kref_init(&obj->refcount);
@@ -195,7 +195,7 @@ EXPORT_SYMBOL(drm_gem_private_object_fini);
  */
 static void drm_gem_object_handle_free(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	/* Remove any name for this object */
 	if (obj->name) {
@@ -216,7 +216,7 @@ static void drm_gem_object_exported_dma_buf_free(struct drm_gem_object *obj)
 static void
 drm_gem_object_handle_put_unlocked(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	bool final = false;
 
 	if (WARN_ON(READ_ONCE(obj->handle_count) == 0))
@@ -355,7 +355,7 @@ drm_gem_handle_create_tail(struct drm_file *file_priv,
 			   struct drm_gem_object *obj,
 			   u32 *handlep)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	u32 handle;
 	int ret;
 
@@ -422,7 +422,7 @@ int drm_gem_handle_create(struct drm_file *file_priv,
 			  struct drm_gem_object *obj,
 			  u32 *handlep)
 {
-	mutex_lock(&obj->dev->object_name_lock);
+	mutex_lock(&obj->drm->object_name_lock);
 
 	return drm_gem_handle_create_tail(file_priv, obj, handlep);
 }
@@ -442,7 +442,7 @@ EXPORT_SYMBOL(drm_gem_handle_create);
 void
 drm_gem_free_mmap_offset(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	drm_vma_offset_remove(dev->vma_offset_manager, &obj->vma_node);
 }
@@ -468,7 +468,7 @@ EXPORT_SYMBOL(drm_gem_free_mmap_offset);
 int
 drm_gem_create_mmap_offset_size(struct drm_gem_object *obj, size_t size)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	return drm_vma_offset_add(dev->vma_offset_manager, &obj->vma_node,
 				  size / PAGE_SIZE);
diff --git a/drivers/gpu/drm/drm_gem_dma_helper.c b/drivers/gpu/drm/drm_gem_dma_helper.c
index 870b90b78bc4..ff328191cfa5 100644
--- a/drivers/gpu/drm/drm_gem_dma_helper.c
+++ b/drivers/gpu/drm/drm_gem_dma_helper.c
@@ -234,11 +234,12 @@ void drm_gem_dma_free(struct drm_gem_dma_object *dma_obj)
 		drm_prime_gem_destroy(gem_obj, dma_obj->sgt);
 	} else if (dma_obj->vaddr) {
 		if (dma_obj->map_noncoherent)
-			dma_free_noncoherent(gem_obj->dev->dev, dma_obj->base.size,
+			dma_free_noncoherent(gem_obj->drm->dev,
+					     dma_obj->base.size,
 					     dma_obj->vaddr, dma_obj->dma_addr,
 					     DMA_TO_DEVICE);
 		else
-			dma_free_wc(gem_obj->dev->dev, dma_obj->base.size,
+			dma_free_wc(gem_obj->drm->dev, dma_obj->base.size,
 				    dma_obj->vaddr, dma_obj->dma_addr);
 	}
 
@@ -428,7 +429,7 @@ struct sg_table *drm_gem_dma_get_sg_table(struct drm_gem_dma_object *dma_obj)
 	if (!sgt)
 		return ERR_PTR(-ENOMEM);
 
-	ret = dma_get_sgtable(obj->dev->dev, sgt, dma_obj->vaddr,
+	ret = dma_get_sgtable(obj->drm->dev, sgt, dma_obj->vaddr,
 			      dma_obj->dma_addr, obj->size);
 	if (ret < 0)
 		goto out;
@@ -535,11 +536,11 @@ int drm_gem_dma_mmap(struct drm_gem_dma_object *dma_obj, struct vm_area_struct *
 	if (dma_obj->map_noncoherent) {
 		vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
 
-		ret = dma_mmap_pages(dma_obj->base.dev->dev,
+		ret = dma_mmap_pages(dma_obj->base.drm->dev,
 				     vma, vma->vm_end - vma->vm_start,
 				     virt_to_page(dma_obj->vaddr));
 	} else {
-		ret = dma_mmap_wc(dma_obj->base.dev->dev, vma, dma_obj->vaddr,
+		ret = dma_mmap_wc(dma_obj->base.drm->dev, vma, dma_obj->vaddr,
 				  dma_obj->dma_addr,
 				  vma->vm_end - vma->vm_start);
 	}
diff --git a/drivers/gpu/drm/drm_gem_shmem_helper.c b/drivers/gpu/drm/drm_gem_shmem_helper.c
index a783d2245599..7c0478f06558 100644
--- a/drivers/gpu/drm/drm_gem_shmem_helper.c
+++ b/drivers/gpu/drm/drm_gem_shmem_helper.c
@@ -144,10 +144,10 @@ void drm_gem_shmem_free(struct drm_gem_shmem_object *shmem)
 	} else {
 		dma_resv_lock(shmem->base.resv, NULL);
 
-		drm_WARN_ON(obj->dev, shmem->vmap_use_count);
+		drm_WARN_ON(obj->drm, shmem->vmap_use_count);
 
 		if (shmem->sgt) {
-			dma_unmap_sgtable(obj->dev->dev, shmem->sgt,
+			dma_unmap_sgtable(obj->drm->dev, shmem->sgt,
 					  DMA_BIDIRECTIONAL, 0);
 			sg_free_table(shmem->sgt);
 			kfree(shmem->sgt);
@@ -155,7 +155,7 @@ void drm_gem_shmem_free(struct drm_gem_shmem_object *shmem)
 		if (shmem->pages)
 			drm_gem_shmem_put_pages(shmem);
 
-		drm_WARN_ON(obj->dev, shmem->pages_use_count);
+		drm_WARN_ON(obj->drm, shmem->pages_use_count);
 
 		dma_resv_unlock(shmem->base.resv);
 	}
@@ -177,7 +177,7 @@ static int drm_gem_shmem_get_pages(struct drm_gem_shmem_object *shmem)
 
 	pages = drm_gem_get_pages(obj);
 	if (IS_ERR(pages)) {
-		drm_dbg_kms(obj->dev, "Failed to get pages (%ld)\n",
+		drm_dbg_kms(obj->drm, "Failed to get pages (%ld)\n",
 			    PTR_ERR(pages));
 		shmem->pages_use_count = 0;
 		return PTR_ERR(pages);
@@ -210,7 +210,7 @@ void drm_gem_shmem_put_pages(struct drm_gem_shmem_object *shmem)
 
 	dma_resv_assert_held(shmem->base.resv);
 
-	if (drm_WARN_ON_ONCE(obj->dev, !shmem->pages_use_count))
+	if (drm_WARN_ON_ONCE(obj->drm, !shmem->pages_use_count))
 		return;
 
 	if (--shmem->pages_use_count > 0)
@@ -261,7 +261,7 @@ int drm_gem_shmem_pin(struct drm_gem_shmem_object *shmem)
 	struct drm_gem_object *obj = &shmem->base;
 	int ret;
 
-	drm_WARN_ON(obj->dev, obj->import_attach);
+	drm_WARN_ON(obj->drm, obj->import_attach);
 
 	ret = dma_resv_lock_interruptible(shmem->base.resv, NULL);
 	if (ret)
@@ -284,7 +284,7 @@ void drm_gem_shmem_unpin(struct drm_gem_shmem_object *shmem)
 {
 	struct drm_gem_object *obj = &shmem->base;
 
-	drm_WARN_ON(obj->dev, obj->import_attach);
+	drm_WARN_ON(obj->drm, obj->import_attach);
 
 	dma_resv_lock(shmem->base.resv, NULL);
 	drm_gem_shmem_unpin_locked(shmem);
@@ -316,7 +316,7 @@ int drm_gem_shmem_vmap(struct drm_gem_shmem_object *shmem,
 	if (obj->import_attach) {
 		ret = dma_buf_vmap(obj->import_attach->dmabuf, map);
 		if (!ret) {
-			if (drm_WARN_ON(obj->dev, map->is_iomem)) {
+			if (drm_WARN_ON(obj->drm, map->is_iomem)) {
 				dma_buf_vunmap(obj->import_attach->dmabuf, map);
 				return -EIO;
 			}
@@ -346,7 +346,7 @@ int drm_gem_shmem_vmap(struct drm_gem_shmem_object *shmem,
 	}
 
 	if (ret) {
-		drm_dbg_kms(obj->dev, "Failed to vmap pages, error %d\n", ret);
+		drm_dbg_kms(obj->drm, "Failed to vmap pages, error %d\n", ret);
 		goto err_put_pages;
 	}
 
@@ -384,7 +384,7 @@ void drm_gem_shmem_vunmap(struct drm_gem_shmem_object *shmem,
 	} else {
 		dma_resv_assert_held(shmem->base.resv);
 
-		if (drm_WARN_ON_ONCE(obj->dev, !shmem->vmap_use_count))
+		if (drm_WARN_ON_ONCE(obj->drm, !shmem->vmap_use_count))
 			return;
 
 		if (--shmem->vmap_use_count > 0)
@@ -440,11 +440,11 @@ EXPORT_SYMBOL(drm_gem_shmem_madvise);
 void drm_gem_shmem_purge(struct drm_gem_shmem_object *shmem)
 {
 	struct drm_gem_object *obj = &shmem->base;
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	dma_resv_assert_held(shmem->base.resv);
 
-	drm_WARN_ON(obj->dev, !drm_gem_shmem_is_purgeable(shmem));
+	drm_WARN_ON(obj->drm, !drm_gem_shmem_is_purgeable(shmem));
 
 	dma_unmap_sgtable(dev->dev, shmem->sgt, DMA_BIDIRECTIONAL, 0);
 	sg_free_table(shmem->sgt);
@@ -522,7 +522,7 @@ static vm_fault_t drm_gem_shmem_fault(struct vm_fault *vmf)
 	dma_resv_lock(shmem->base.resv, NULL);
 
 	if (page_offset >= num_pages ||
-	    drm_WARN_ON_ONCE(obj->dev, !shmem->pages) ||
+	    drm_WARN_ON_ONCE(obj->drm, !shmem->pages) ||
 	    shmem->madv < 0) {
 		ret = VM_FAULT_SIGBUS;
 	} else {
@@ -541,7 +541,7 @@ static void drm_gem_shmem_vm_open(struct vm_area_struct *vma)
 	struct drm_gem_object *obj = vma->vm_private_data;
 	struct drm_gem_shmem_object *shmem = to_drm_gem_shmem_obj(obj);
 
-	drm_WARN_ON(obj->dev, obj->import_attach);
+	drm_WARN_ON(obj->drm, obj->import_attach);
 
 	dma_resv_lock(shmem->base.resv, NULL);
 
@@ -550,7 +550,7 @@ static void drm_gem_shmem_vm_open(struct vm_area_struct *vma)
 	 * mmap'd, vm_open() just grabs an additional reference for the new
 	 * mm the vma is getting copied into (ie. on fork()).
 	 */
-	if (!drm_WARN_ON_ONCE(obj->dev, !shmem->pages_use_count))
+	if (!drm_WARN_ON_ONCE(obj->drm, !shmem->pages_use_count))
 		shmem->pages_use_count++;
 
 	dma_resv_unlock(shmem->base.resv);
@@ -656,9 +656,10 @@ struct sg_table *drm_gem_shmem_get_sg_table(struct drm_gem_shmem_object *shmem)
 {
 	struct drm_gem_object *obj = &shmem->base;
 
-	drm_WARN_ON(obj->dev, obj->import_attach);
+	drm_WARN_ON(obj->drm, obj->import_attach);
 
-	return drm_prime_pages_to_sg(obj->dev, shmem->pages, obj->size >> PAGE_SHIFT);
+	return drm_prime_pages_to_sg(obj->drm, shmem->pages,
+				     obj->size >> PAGE_SHIFT);
 }
 EXPORT_SYMBOL_GPL(drm_gem_shmem_get_sg_table);
 
@@ -671,7 +672,7 @@ static struct sg_table *drm_gem_shmem_get_pages_sgt_locked(struct drm_gem_shmem_
 	if (shmem->sgt)
 		return shmem->sgt;
 
-	drm_WARN_ON(obj->dev, obj->import_attach);
+	drm_WARN_ON(obj->drm, obj->import_attach);
 
 	ret = drm_gem_shmem_get_pages(shmem);
 	if (ret)
@@ -683,7 +684,7 @@ static struct sg_table *drm_gem_shmem_get_pages_sgt_locked(struct drm_gem_shmem_
 		goto err_put_pages;
 	}
 	/* Map the pages for use by the h/w. */
-	ret = dma_map_sgtable(obj->dev->dev, sgt, DMA_BIDIRECTIONAL, 0);
+	ret = dma_map_sgtable(obj->drm->dev, sgt, DMA_BIDIRECTIONAL, 0);
 	if (ret)
 		goto err_free_sgt;
 
diff --git a/drivers/gpu/drm/drm_gem_vram_helper.c b/drivers/gpu/drm/drm_gem_vram_helper.c
index fc45a65708b7..54e53ca19482 100644
--- a/drivers/gpu/drm/drm_gem_vram_helper.c
+++ b/drivers/gpu/drm/drm_gem_vram_helper.c
@@ -395,7 +395,7 @@ static int drm_gem_vram_kmap_locked(struct drm_gem_vram_object *gbo,
 static void drm_gem_vram_kunmap_locked(struct drm_gem_vram_object *gbo,
 				       struct iosys_map *map)
 {
-	struct drm_device *dev = gbo->bo.base.dev;
+	struct drm_device *dev = gbo->bo.base.drm;
 
 	if (drm_WARN_ON_ONCE(dev, !gbo->vmap_use_count))
 		return;
@@ -552,7 +552,7 @@ static void drm_gem_vram_bo_driver_evict_flags(struct drm_gem_vram_object *gbo,
 static void drm_gem_vram_bo_driver_move_notify(struct drm_gem_vram_object *gbo)
 {
 	struct ttm_buffer_object *bo = &gbo->bo;
-	struct drm_device *dev = bo->base.dev;
+	struct drm_device *dev = bo->base.drm;
 
 	if (drm_WARN_ON_ONCE(dev, gbo->vmap_use_count))
 		return;
diff --git a/drivers/gpu/drm/drm_prime.c b/drivers/gpu/drm/drm_prime.c
index 63b709a67471..f8cc88084be6 100644
--- a/drivers/gpu/drm/drm_prime.c
+++ b/drivers/gpu/drm/drm_prime.c
@@ -251,7 +251,7 @@ struct dma_buf *drm_gem_dmabuf_export(struct drm_device *dev,
 
 	drm_dev_get(dev);
 	drm_gem_object_get(obj);
-	dma_buf->file->f_mapping = obj->dev->anon_inode->i_mapping;
+	dma_buf->file->f_mapping = obj->drm->anon_inode->i_mapping;
 
 	return dma_buf;
 }
@@ -269,7 +269,7 @@ EXPORT_SYMBOL(drm_gem_dmabuf_export);
 void drm_gem_dmabuf_release(struct dma_buf *dma_buf)
 {
 	struct drm_gem_object *obj = dma_buf->priv;
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	/* drop the reference on the export fd holds */
 	drm_gem_object_put(obj);
@@ -742,14 +742,14 @@ int drm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
 	}
 
 	/* Used by drm_gem_mmap() to lookup the GEM object */
-	priv->minor = obj->dev->primary;
+	priv->minor = obj->drm->primary;
 	fil->private_data = priv;
 
 	ret = drm_vma_node_allow(&obj->vma_node, priv);
 	if (ret)
 		goto out;
 
-	ret = obj->dev->driver->fops->mmap(fil, vma);
+	ret = obj->drm->driver->fops->mmap(fil, vma);
 
 	drm_vma_node_revoke(&obj->vma_node, priv);
 out:
@@ -871,7 +871,7 @@ EXPORT_SYMBOL(drm_prime_get_contiguous_size);
 struct dma_buf *drm_gem_prime_export(struct drm_gem_object *obj,
 				     int flags)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct dma_buf_export_info exp_info = {
 		.exp_name = KBUILD_MODNAME, /* white lie for debug */
 		.owner = dev->driver->fops->owner,
@@ -911,7 +911,7 @@ struct drm_gem_object *drm_gem_prime_import_dev(struct drm_device *dev,
 
 	if (dma_buf->ops == &drm_gem_prime_dmabuf_ops) {
 		obj = dma_buf->priv;
-		if (obj->dev == dev) {
+		if (obj->drm == dev) {
 			/*
 			 * Importing dmabuf exported from our own gem increases
 			 * refcount on gem itself instead of f_count of dmabuf.
diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem.c b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
index b5f73502e3dd..95827f55ce2b 100644
--- a/drivers/gpu/drm/etnaviv/etnaviv_gem.c
+++ b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
@@ -19,7 +19,7 @@ static struct lock_class_key etnaviv_userptr_lock_class;
 
 static void etnaviv_gem_scatter_map(struct etnaviv_gem_object *etnaviv_obj)
 {
-	struct drm_device *dev = etnaviv_obj->base.dev;
+	struct drm_device *dev = etnaviv_obj->base.drm;
 	struct sg_table *sgt = etnaviv_obj->sgt;
 
 	/*
@@ -32,7 +32,7 @@ static void etnaviv_gem_scatter_map(struct etnaviv_gem_object *etnaviv_obj)
 
 static void etnaviv_gem_scatterlist_unmap(struct etnaviv_gem_object *etnaviv_obj)
 {
-	struct drm_device *dev = etnaviv_obj->base.dev;
+	struct drm_device *dev = etnaviv_obj->base.drm;
 	struct sg_table *sgt = etnaviv_obj->sgt;
 
 	/*
@@ -57,7 +57,7 @@ static void etnaviv_gem_scatterlist_unmap(struct etnaviv_gem_object *etnaviv_obj
 /* called with etnaviv_obj->lock held */
 static int etnaviv_gem_shmem_get_pages(struct etnaviv_gem_object *etnaviv_obj)
 {
-	struct drm_device *dev = etnaviv_obj->base.dev;
+	struct drm_device *dev = etnaviv_obj->base.drm;
 	struct page **p = drm_gem_get_pages(&etnaviv_obj->base);
 
 	if (IS_ERR(p)) {
@@ -99,11 +99,11 @@ struct page **etnaviv_gem_get_pages(struct etnaviv_gem_object *etnaviv_obj)
 	}
 
 	if (!etnaviv_obj->sgt) {
-		struct drm_device *dev = etnaviv_obj->base.dev;
+		struct drm_device *dev = etnaviv_obj->base.drm;
 		int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
 		struct sg_table *sgt;
 
-		sgt = drm_prime_pages_to_sg(etnaviv_obj->base.dev,
+		sgt = drm_prime_pages_to_sg(etnaviv_obj->base.drm,
 					    etnaviv_obj->pages, npages);
 		if (IS_ERR(sgt)) {
 			dev_err(dev->dev, "failed to allocate sgt: %ld\n",
@@ -205,7 +205,7 @@ int etnaviv_gem_mmap_offset(struct drm_gem_object *obj, u64 *offset)
 	/* Make it mmapable */
 	ret = drm_gem_create_mmap_offset(obj);
 	if (ret)
-		dev_err(obj->dev->dev, "could not allocate mmap offset\n");
+		dev_err(obj->drm->dev, "could not allocate mmap offset\n");
 	else
 		*offset = drm_vma_node_offset_addr(&obj->vma_node);
 
@@ -368,7 +368,7 @@ int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
 		struct drm_etnaviv_timespec *timeout)
 {
 	struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	bool write = !!(op & ETNA_PREP_WRITE);
 	int ret;
 
@@ -406,7 +406,7 @@ int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
 
 int etnaviv_gem_cpu_fini(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
 
 	if (etnaviv_obj->flags & ETNA_BO_CACHED) {
@@ -487,7 +487,7 @@ static const struct etnaviv_gem_ops etnaviv_gem_shmem_ops = {
 void etnaviv_gem_free_object(struct drm_gem_object *obj)
 {
 	struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
-	struct etnaviv_drm_private *priv = obj->dev->dev_private;
+	struct etnaviv_drm_private *priv = obj->drm->dev_private;
 	struct etnaviv_vram_mapping *mapping, *tmp;
 
 	/* object should not be active */
diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c b/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
index 3524b5811682..5bf4d1ee977f 100644
--- a/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
+++ b/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
@@ -22,7 +22,7 @@ struct sg_table *etnaviv_gem_prime_get_sg_table(struct drm_gem_object *obj)
 	if (WARN_ON(!etnaviv_obj->pages))  /* should have already pinned! */
 		return ERR_PTR(-EINVAL);
 
-	return drm_prime_pages_to_sg(obj->dev, etnaviv_obj->pages, npages);
+	return drm_prime_pages_to_sg(obj->drm, etnaviv_obj->pages, npages);
 }
 
 int etnaviv_gem_prime_vmap(struct drm_gem_object *obj, struct iosys_map *map)
diff --git a/drivers/gpu/drm/exynos/exynos_drm_gem.c b/drivers/gpu/drm/exynos/exynos_drm_gem.c
index 638ca96830e9..8ab968881110 100644
--- a/drivers/gpu/drm/exynos/exynos_drm_gem.c
+++ b/drivers/gpu/drm/exynos/exynos_drm_gem.c
@@ -24,7 +24,7 @@ static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct
 
 static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem, bool kvmap)
 {
-	struct drm_device *dev = exynos_gem->base.dev;
+	struct drm_device *dev = exynos_gem->base.drm;
 	unsigned long attr = 0;
 
 	if (exynos_gem->dma_addr) {
@@ -71,7 +71,7 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem, bool kvmap)
 
 static void exynos_drm_free_buf(struct exynos_drm_gem *exynos_gem)
 {
-	struct drm_device *dev = exynos_gem->base.dev;
+	struct drm_device *dev = exynos_gem->base.drm;
 
 	if (!exynos_gem->dma_addr) {
 		DRM_DEV_DEBUG_KMS(dev->dev, "dma_addr is invalid.\n");
@@ -100,7 +100,8 @@ static int exynos_drm_gem_handle_create(struct drm_gem_object *obj,
 	if (ret)
 		return ret;
 
-	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->dev), "gem handle = 0x%x\n", *handle);
+	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->drm), "gem handle = 0x%x\n",
+			  *handle);
 
 	/* drop reference from allocate - handle holds it now. */
 	drm_gem_object_put(obj);
@@ -112,7 +113,7 @@ void exynos_drm_gem_destroy(struct exynos_drm_gem *exynos_gem)
 {
 	struct drm_gem_object *obj = &exynos_gem->base;
 
-	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->dev), "handle count = %d\n",
+	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->drm), "handle count = %d\n",
 			  obj->handle_count);
 
 	/*
@@ -270,7 +271,7 @@ struct exynos_drm_gem *exynos_drm_gem_get(struct drm_file *filp,
 static int exynos_drm_gem_mmap_buffer(struct exynos_drm_gem *exynos_gem,
 				      struct vm_area_struct *vma)
 {
-	struct drm_device *drm_dev = exynos_gem->base.dev;
+	struct drm_device *drm_dev = exynos_gem->base.drm;
 	unsigned long vm_size;
 	int ret;
 
@@ -370,7 +371,7 @@ static int exynos_drm_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct
 
 	vm_flags_set(vma, VM_IO | VM_DONTEXPAND | VM_DONTDUMP);
 
-	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->dev), "flags = 0x%x\n",
+	DRM_DEV_DEBUG_KMS(to_dma_dev(obj->drm), "flags = 0x%x\n",
 			  exynos_gem->flags);
 
 	/* non-cachable as default. */
@@ -405,7 +406,7 @@ struct drm_gem_object *exynos_drm_gem_prime_import(struct drm_device *dev,
 struct sg_table *exynos_drm_gem_prime_get_sg_table(struct drm_gem_object *obj)
 {
 	struct exynos_drm_gem *exynos_gem = to_exynos_gem(obj);
-	struct drm_device *drm_dev = obj->dev;
+	struct drm_device *drm_dev = obj->drm;
 	struct sg_table *sgt;
 	int ret;
 
diff --git a/drivers/gpu/drm/gma500/gem.c b/drivers/gpu/drm/gma500/gem.c
index 4b7627a72637..70d74ae54d56 100644
--- a/drivers/gpu/drm/gma500/gem.c
+++ b/drivers/gpu/drm/gma500/gem.c
@@ -28,7 +28,7 @@
 int psb_gem_pin(struct psb_gem_object *pobj)
 {
 	struct drm_gem_object *obj = &pobj->base;
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
 	u32 gpu_base = dev_priv->gtt.gatt_start;
 	struct page **pages;
@@ -73,7 +73,7 @@ int psb_gem_pin(struct psb_gem_object *pobj)
 void psb_gem_unpin(struct psb_gem_object *pobj)
 {
 	struct drm_gem_object *obj = &pobj->base;
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
 	u32 gpu_base = dev_priv->gtt.gatt_start;
 	unsigned long npages;
@@ -264,7 +264,7 @@ static vm_fault_t psb_gem_fault(struct vm_fault *vmf)
 	struct drm_psb_private *dev_priv;
 
 	obj = vma->vm_private_data;	/* GEM object */
-	dev = obj->dev;
+	dev = obj->drm;
 	dev_priv = to_drm_psb_private(dev);
 
 	pobj = to_psb_gem_object(obj);
diff --git a/drivers/gpu/drm/i915/display/intel_dpt.c b/drivers/gpu/drm/i915/display/intel_dpt.c
index b70dd87ab7fb..6122d2fd2115 100644
--- a/drivers/gpu/drm/i915/display/intel_dpt.c
+++ b/drivers/gpu/drm/i915/display/intel_dpt.c
@@ -242,7 +242,7 @@ struct i915_address_space *
 intel_dpt_create(struct intel_framebuffer *fb)
 {
 	struct drm_gem_object *obj = &intel_fb_obj(&fb->base)->base;
-	struct drm_i915_private *i915 = to_i915(obj->dev);
+	struct drm_i915_private *i915 = to_i915(obj->drm);
 	struct drm_i915_gem_object *dpt_obj;
 	struct i915_address_space *vm;
 	struct i915_dpt *dpt;
diff --git a/drivers/gpu/drm/i915/display/intel_fb.c b/drivers/gpu/drm/i915/display/intel_fb.c
index 4109c894abd3..e870a46ce12a 100644
--- a/drivers/gpu/drm/i915/display/intel_fb.c
+++ b/drivers/gpu/drm/i915/display/intel_fb.c
@@ -1885,7 +1885,7 @@ static int intel_user_framebuffer_create_handle(struct drm_framebuffer *fb,
 						unsigned int *handle)
 {
 	struct drm_i915_gem_object *obj = intel_fb_obj(fb);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	if (i915_gem_object_is_userptr(obj)) {
 		drm_dbg(&i915->drm,
@@ -1920,7 +1920,7 @@ int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
 			   struct drm_i915_gem_object *obj,
 			   struct drm_mode_fb_cmd2 *mode_cmd)
 {
-	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
+	struct drm_i915_private *dev_priv = to_i915(obj->base.drm);
 	struct drm_framebuffer *fb = &intel_fb->base;
 	u32 max_stride;
 	unsigned int tiling, stride;
@@ -2089,7 +2089,7 @@ intel_user_framebuffer_create(struct drm_device *dev,
 		return ERR_PTR(-ENOENT);
 
 	/* object is backed with LMEM for discrete */
-	i915 = to_i915(obj->base.dev);
+	i915 = to_i915(obj->base.drm);
 	if (HAS_LMEM(i915) && !i915_gem_object_can_migrate(obj, INTEL_REGION_LMEM_0)) {
 		/* object is "remote", not in local memory */
 		i915_gem_object_put(obj);
diff --git a/drivers/gpu/drm/i915/display/intel_frontbuffer.c b/drivers/gpu/drm/i915/display/intel_frontbuffer.c
index 17a7aa8b28c2..40d4a1e70c96 100644
--- a/drivers/gpu/drm/i915/display/intel_frontbuffer.c
+++ b/drivers/gpu/drm/i915/display/intel_frontbuffer.c
@@ -167,7 +167,7 @@ void __intel_fb_invalidate(struct intel_frontbuffer *front,
 			   enum fb_op_origin origin,
 			   unsigned int frontbuffer_bits)
 {
-	struct drm_i915_private *i915 = to_i915(front->obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(front->obj->base.drm);
 
 	if (origin == ORIGIN_CS) {
 		spin_lock(&i915->display.fb_tracking.lock);
@@ -188,7 +188,7 @@ void __intel_fb_flush(struct intel_frontbuffer *front,
 		      enum fb_op_origin origin,
 		      unsigned int frontbuffer_bits)
 {
-	struct drm_i915_private *i915 = to_i915(front->obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(front->obj->base.drm);
 
 	if (origin == ORIGIN_CS) {
 		spin_lock(&i915->display.fb_tracking.lock);
@@ -221,14 +221,14 @@ static void frontbuffer_retire(struct i915_active *ref)
 }
 
 static void frontbuffer_release(struct kref *ref)
-	__releases(&to_i915(front->obj->base.dev)->display.fb_tracking.lock)
+	__releases(&to_i915(front->obj->base.drm)->display.fb_tracking.lock)
 {
 	struct intel_frontbuffer *front =
 		container_of(ref, typeof(*front), ref);
 	struct drm_i915_gem_object *obj = front->obj;
 	struct i915_vma *vma;
 
-	drm_WARN_ON(obj->base.dev, atomic_read(&front->bits));
+	drm_WARN_ON(obj->base.drm, atomic_read(&front->bits));
 
 	spin_lock(&obj->vma.lock);
 	for_each_ggtt_vma(vma, obj) {
@@ -238,7 +238,7 @@ static void frontbuffer_release(struct kref *ref)
 	spin_unlock(&obj->vma.lock);
 
 	RCU_INIT_POINTER(obj->frontbuffer, NULL);
-	spin_unlock(&to_i915(obj->base.dev)->display.fb_tracking.lock);
+	spin_unlock(&to_i915(obj->base.drm)->display.fb_tracking.lock);
 
 	i915_active_fini(&front->write);
 
@@ -249,7 +249,7 @@ static void frontbuffer_release(struct kref *ref)
 struct intel_frontbuffer *
 intel_frontbuffer_get(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct intel_frontbuffer *front;
 
 	front = __intel_frontbuffer_get(obj);
@@ -286,7 +286,7 @@ void intel_frontbuffer_put(struct intel_frontbuffer *front)
 {
 	kref_put_lock(&front->ref,
 		      frontbuffer_release,
-		      &to_i915(front->obj->base.dev)->display.fb_tracking.lock);
+		      &to_i915(front->obj->base.drm)->display.fb_tracking.lock);
 }
 
 /**
@@ -315,13 +315,13 @@ void intel_frontbuffer_track(struct intel_frontbuffer *old,
 	BUILD_BUG_ON(I915_MAX_PLANES > INTEL_FRONTBUFFER_BITS_PER_PIPE);
 
 	if (old) {
-		drm_WARN_ON(old->obj->base.dev,
+		drm_WARN_ON(old->obj->base.drm,
 			    !(atomic_read(&old->bits) & frontbuffer_bits));
 		atomic_andnot(frontbuffer_bits, &old->bits);
 	}
 
 	if (new) {
-		drm_WARN_ON(new->obj->base.dev,
+		drm_WARN_ON(new->obj->base.drm,
 			    atomic_read(&new->bits) & frontbuffer_bits);
 		atomic_or(frontbuffer_bits, &new->bits);
 	}
diff --git a/drivers/gpu/drm/i915/display/skl_universal_plane.c b/drivers/gpu/drm/i915/display/skl_universal_plane.c
index e4a398224b91..5c02b9b805e2 100644
--- a/drivers/gpu/drm/i915/display/skl_universal_plane.c
+++ b/drivers/gpu/drm/i915/display/skl_universal_plane.c
@@ -1855,7 +1855,7 @@ static bool skl_fb_scalable(const struct drm_framebuffer *fb)
 
 static bool bo_has_valid_encryption(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	return intel_pxp_key_check(i915->pxp, obj, false) == 0;
 }
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_clflush.c b/drivers/gpu/drm/i915/gem/i915_gem_clflush.c
index 385ffc575b48..9086f474c386 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_clflush.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_clflush.c
@@ -72,7 +72,7 @@ static struct clflush *clflush_work_create(struct drm_i915_gem_object *obj)
 bool i915_gem_clflush_object(struct drm_i915_gem_object *obj,
 			     unsigned int flags)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct clflush *clflush;
 
 	assert_object_held(obj);
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
index 1df74f7aa3dc..e1b4fe3ca5f1 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
@@ -94,7 +94,7 @@ static void i915_gem_dmabuf_vunmap(struct dma_buf *dma_buf,
 static int i915_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma)
 {
 	struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	int ret;
 
 	if (obj->base.size < vma->vm_end - vma->vm_start)
@@ -231,12 +231,12 @@ struct dma_buf *i915_gem_prime_export(struct drm_gem_object *gem_obj, int flags)
 			return ERR_PTR(ret);
 	}
 
-	return drm_gem_dmabuf_export(gem_obj->dev, &exp_info);
+	return drm_gem_dmabuf_export(gem_obj->drm, &exp_info);
 }
 
 static int i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *sgt;
 
 	assert_object_held(obj);
@@ -291,7 +291,7 @@ struct drm_gem_object *i915_gem_prime_import(struct drm_device *dev,
 	if (dma_buf->ops == &i915_dmabuf_ops) {
 		obj = dma_buf_to_obj(dma_buf);
 		/* is it from our device? */
-		if (obj->base.dev == dev &&
+		if (obj->base.drm == dev &&
 		    !I915_SELFTEST_ONLY(force_different_devices)) {
 			/*
 			 * Importing dmabuf exported from out own gem increases
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_domain.c b/drivers/gpu/drm/i915/gem/i915_gem_domain.c
index dfaaa8b66ac3..167c0272865a 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_domain.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_domain.c
@@ -22,7 +22,7 @@
 
 static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	if (IS_DGFX(i915))
 		return false;
@@ -40,7 +40,7 @@ static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj)
 
 bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	if (obj->cache_dirty)
 		return false;
@@ -430,7 +430,7 @@ i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
 				     const struct i915_gtt_view *view,
 				     unsigned int flags)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_vma *vma;
 	int ret;
 
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_internal.c b/drivers/gpu/drm/i915/gem/i915_gem_internal.c
index 6bc26b4b06b8..28e5e6f74c8e 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_internal.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_internal.c
@@ -32,7 +32,7 @@ static void internal_free_pages(struct sg_table *st)
 
 static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *st;
 	struct scatterlist *sg;
 	unsigned int npages; /* restricted by sg_alloc_table */
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_mman.c b/drivers/gpu/drm/i915/gem/i915_gem_mman.c
index 4e7a838ab7bd..89ef6b44aa03 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_mman.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_mman.c
@@ -295,7 +295,7 @@ static vm_fault_t vm_fault_gtt(struct vm_fault *vmf)
 	struct vm_area_struct *area = vmf->vma;
 	struct i915_mmap_offset *mmo = area->vm_private_data;
 	struct drm_i915_gem_object *obj = mmo->obj;
-	struct drm_device *dev = obj->base.dev;
+	struct drm_device *dev = obj->base.drm;
 	struct drm_i915_private *i915 = to_i915(dev);
 	struct intel_runtime_pm *rpm = &i915->runtime_pm;
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
@@ -524,7 +524,7 @@ void __i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj)
  */
 void i915_gem_object_release_mmap_gtt(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	intel_wakeref_t wakeref;
 
 	/*
@@ -594,7 +594,7 @@ void i915_gem_object_release_mmap_offset(struct drm_i915_gem_object *obj)
 
 		spin_unlock(&obj->mmo.lock);
 		drm_vma_node_unmap(&mmo->vma_node,
-				   obj->base.dev->anon_inode->i_mapping);
+				   obj->base.drm->anon_inode->i_mapping);
 		spin_lock(&obj->mmo.lock);
 	}
 	spin_unlock(&obj->mmo.lock);
@@ -643,7 +643,7 @@ insert_mmo(struct drm_i915_gem_object *obj, struct i915_mmap_offset *mmo)
 
 		if (pos->mmap_type == mmo->mmap_type) {
 			spin_unlock(&obj->mmo.lock);
-			drm_vma_offset_remove(obj->base.dev->vma_offset_manager,
+			drm_vma_offset_remove(obj->base.drm->vma_offset_manager,
 					      &mmo->vma_node);
 			kfree(mmo);
 			return pos;
@@ -666,7 +666,7 @@ mmap_offset_attach(struct drm_i915_gem_object *obj,
 		   enum i915_mmap_type mmap_type,
 		   struct drm_file *file)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_mmap_offset *mmo;
 	int err;
 
@@ -684,7 +684,7 @@ mmap_offset_attach(struct drm_i915_gem_object *obj,
 	mmo->mmap_type = mmap_type;
 	drm_vma_node_reset(&mmo->vma_node);
 
-	err = drm_vma_offset_add(obj->base.dev->vma_offset_manager,
+	err = drm_vma_offset_add(obj->base.drm->vma_offset_manager,
 				 &mmo->vma_node, obj->base.size / PAGE_SIZE);
 	if (likely(!err))
 		goto insert;
@@ -696,7 +696,7 @@ mmap_offset_attach(struct drm_i915_gem_object *obj,
 		goto err;
 
 	i915_gem_drain_freed_objects(i915);
-	err = drm_vma_offset_add(obj->base.dev->vma_offset_manager,
+	err = drm_vma_offset_add(obj->base.drm->vma_offset_manager,
 				 &mmo->vma_node, obj->base.size / PAGE_SIZE);
 	if (err)
 		goto err;
@@ -941,7 +941,7 @@ i915_gem_object_mmap(struct drm_i915_gem_object *obj,
 		     struct i915_mmap_offset *mmo,
 		     struct vm_area_struct *vma)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct drm_device *dev = &i915->drm;
 	struct file *anon;
 
@@ -1065,7 +1065,7 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma)
 
 int i915_gem_fb_mmap(struct drm_i915_gem_object *obj, struct vm_area_struct *vma)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct drm_device *dev = &i915->drm;
 	struct i915_mmap_offset *mmo = NULL;
 	enum i915_mmap_type mmap_type;
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c
index 97ac6fb37958..8364fcf66a13 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_object.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c
@@ -149,7 +149,7 @@ void __i915_gem_object_fini(struct drm_i915_gem_object *obj)
 void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj,
 					 unsigned int cache_level)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	obj->pat_index = i915_gem_get_pat_index(i915, cache_level);
 
@@ -177,7 +177,7 @@ void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj,
 void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj,
 				   unsigned int pat_index)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	if (obj->pat_index == pat_index)
 		return;
@@ -199,7 +199,7 @@ void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj,
 
 bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	/*
 	 * This is purely from a security perspective, so we simply don't care
@@ -290,7 +290,7 @@ void __i915_gem_free_object_rcu(struct rcu_head *head)
 {
 	struct drm_i915_gem_object *obj =
 		container_of(head, typeof(*obj), rcu);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	i915_gem_object_free(obj);
 
@@ -302,7 +302,7 @@ static void __i915_gem_object_free_mmaps(struct drm_i915_gem_object *obj)
 {
 	/* Skip serialisation and waking the device if known to be not used. */
 
-	if (obj->userfault_count && !IS_DGFX(to_i915(obj->base.dev)))
+	if (obj->userfault_count && !IS_DGFX(to_i915(obj->base.drm)))
 		i915_gem_object_release_mmap_gtt(obj);
 
 	if (!RB_EMPTY_ROOT(&obj->mmo.offsets)) {
@@ -313,7 +313,7 @@ static void __i915_gem_object_free_mmaps(struct drm_i915_gem_object *obj)
 		rbtree_postorder_for_each_entry_safe(mmo, mn,
 						     &obj->mmo.offsets,
 						     offset) {
-			drm_vma_offset_remove(obj->base.dev->vma_offset_manager,
+			drm_vma_offset_remove(obj->base.drm->vma_offset_manager,
 					      &mmo->vma_node);
 			kfree(mmo);
 		}
@@ -437,7 +437,7 @@ static void __i915_gem_free_work(struct work_struct *work)
 static void i915_gem_free_object(struct drm_gem_object *gem_obj)
 {
 	struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	GEM_BUG_ON(i915_gem_object_is_framebuffer(obj));
 
@@ -528,7 +528,7 @@ static bool object_has_mappable_iomem(struct drm_i915_gem_object *obj)
 {
 	GEM_BUG_ON(!i915_gem_object_has_iomem(obj));
 
-	if (IS_DGFX(to_i915(obj->base.dev)))
+	if (IS_DGFX(to_i915(obj->base.drm)))
 		return i915_ttm_resource_mappable(i915_gem_to_ttm(obj)->resource);
 
 	return true;
@@ -631,7 +631,7 @@ bool i915_gem_object_migratable(struct drm_i915_gem_object *obj)
 bool i915_gem_object_has_struct_page(const struct drm_i915_gem_object *obj)
 {
 #ifdef CONFIG_LOCKDEP
-	if (IS_DGFX(to_i915(obj->base.dev)) &&
+	if (IS_DGFX(to_i915(obj->base.drm)) &&
 	    i915_gem_object_evictable((void __force *)obj))
 		assert_object_held_shared(obj);
 #endif
@@ -650,7 +650,7 @@ bool i915_gem_object_has_struct_page(const struct drm_i915_gem_object *obj)
 bool i915_gem_object_has_iomem(const struct drm_i915_gem_object *obj)
 {
 #ifdef CONFIG_LOCKDEP
-	if (IS_DGFX(to_i915(obj->base.dev)) &&
+	if (IS_DGFX(to_i915(obj->base.drm)) &&
 	    i915_gem_object_evictable((void __force *)obj))
 		assert_object_held_shared(obj);
 #endif
@@ -677,7 +677,7 @@ bool i915_gem_object_has_iomem(const struct drm_i915_gem_object *obj)
 bool i915_gem_object_can_migrate(struct drm_i915_gem_object *obj,
 				 enum intel_region_id id)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	unsigned int num_allowed = obj->mm.n_placements;
 	struct intel_memory_region *mr;
 	unsigned int i;
@@ -779,7 +779,7 @@ int __i915_gem_object_migrate(struct drm_i915_gem_object *obj,
 			      enum intel_region_id id,
 			      unsigned int flags)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct intel_memory_region *mr;
 
 	GEM_BUG_ON(id >= INTEL_REGION_UNKNOWN);
@@ -849,7 +849,7 @@ bool i915_gem_object_needs_ccs_pages(struct drm_i915_gem_object *obj)
 	bool lmem_placement = false;
 	int i;
 
-	if (!HAS_FLAT_CCS(to_i915(obj->base.dev)))
+	if (!HAS_FLAT_CCS(to_i915(obj->base.drm)))
 		return false;
 
 	if (obj->flags & I915_BO_ALLOC_CCS_AUX)
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h b/drivers/gpu/drm/i915/gem/i915_gem_object.h
index 884a17275b3a..1ee051c25205 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_object.h
+++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h
@@ -20,7 +20,7 @@
 
 enum intel_region_id;
 
-#define obj_to_i915(obj__) to_i915((obj__)->base.dev)
+#define obj_to_i915(obj__) to_i915((obj__)->base.drm)
 
 static inline bool i915_gem_object_size_2big(u64 size)
 {
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
index 89fc8ea6bcfc..8ea50a3933f3 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
@@ -18,7 +18,7 @@
 void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
 				 struct sg_table *pages)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	unsigned long supported = RUNTIME_INFO(i915)->page_sizes;
 	bool shrinkable;
 	int i;
@@ -96,7 +96,7 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
 
 int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	int err;
 
 	assert_object_held_shared(obj);
@@ -192,7 +192,7 @@ static void unmap_object(struct drm_i915_gem_object *obj, void *ptr)
 
 static void flush_tlb_invalidate(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct intel_gt *gt = to_gt(i915);
 
 	if (!obj->mm.tlb)
@@ -411,7 +411,7 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
 		}
 
 		type = I915_MAP_WC;
-	} else if (IS_DGFX(to_i915(obj->base.dev))) {
+	} else if (IS_DGFX(to_i915(obj->base.drm))) {
 		type = I915_MAP_WB;
 	}
 
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_phys.c b/drivers/gpu/drm/i915/gem/i915_gem_phys.c
index 76efe98eaa14..e93af6786995 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_phys.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_phys.c
@@ -20,7 +20,7 @@
 static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
 {
 	struct address_space *mapping = obj->base.filp->f_mapping;
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct scatterlist *sg;
 	struct sg_table *st;
 	dma_addr_t dma;
@@ -40,7 +40,7 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
 	 * to handle all possible callers, and given typical object sizes,
 	 * the alignment of the buddy allocation will naturally match.
 	 */
-	vaddr = dma_alloc_coherent(obj->base.dev->dev,
+	vaddr = dma_alloc_coherent(obj->base.drm->dev,
 				   roundup_pow_of_two(obj->base.size),
 				   &dma, GFP_KERNEL);
 	if (!vaddr)
@@ -90,7 +90,7 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
 err_st:
 	kfree(st);
 err_pci:
-	dma_free_coherent(obj->base.dev->dev,
+	dma_free_coherent(obj->base.drm->dev,
 			  roundup_pow_of_two(obj->base.size),
 			  vaddr, dma);
 	return -ENOMEM;
@@ -136,7 +136,7 @@ i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj,
 	sg_free_table(pages);
 	kfree(pages);
 
-	dma_free_coherent(obj->base.dev->dev,
+	dma_free_coherent(obj->base.drm->dev,
 			  roundup_pow_of_two(obj->base.size),
 			  vaddr, dma);
 }
@@ -146,7 +146,7 @@ int i915_gem_object_pwrite_phys(struct drm_i915_gem_object *obj,
 {
 	void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset;
 	char __user *user_data = u64_to_user_ptr(args->data_ptr);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	int err;
 
 	err = i915_gem_object_wait(obj,
@@ -247,7 +247,7 @@ int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align)
 		return -EBUSY;
 
 	if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) {
-		drm_dbg(obj->base.dev,
+		drm_dbg(obj->base.drm,
 			"Attempting to obtain a purgeable object\n");
 		return -EFAULT;
 	}
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
index 33d5d5178103..e127e271ca6f 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
@@ -194,7 +194,7 @@ int shmem_sg_alloc_table(struct drm_i915_private *i915, struct sg_table *st,
 
 static int shmem_get_pages(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct intel_memory_region *mem = obj->mm.region;
 	struct address_space *mapping = obj->base.filp->f_mapping;
 	unsigned int max_segment = i915_sg_segment_size(i915->drm.dev);
@@ -358,7 +358,7 @@ __i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
 				struct sg_table *pages,
 				bool needs_clflush)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	GEM_BUG_ON(obj->mm.madv == __I915_MADV_PURGED);
 
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
index 3b094d36a0b0..b92faa05abda 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
@@ -621,9 +621,9 @@ i915_pages_create_for_stolen(struct drm_device *dev,
 
 static int i915_gem_object_get_pages_stolen(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *pages =
-		i915_pages_create_for_stolen(obj->base.dev,
+		i915_pages_create_for_stolen(obj->base.drm,
 					     obj->stolen->start,
 					     obj->stolen->size);
 	if (IS_ERR(pages))
@@ -642,7 +642,7 @@ static int i915_gem_object_get_pages_stolen(struct drm_i915_gem_object *obj)
 static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
 					     struct sg_table *pages)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	/* Should only be called from i915_gem_object_release_stolen() */
 
 	dbg_poison(to_gt(i915)->ggtt,
@@ -657,7 +657,7 @@ static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
 static void
 i915_gem_object_release_stolen(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct drm_mm_node *stolen = fetch_and_zero(&obj->stolen);
 
 	GEM_BUG_ON(!stolen);
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
index a049ca0b7980..754080930d3e 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
@@ -119,7 +119,7 @@ static bool
 i915_tiling_ok(struct drm_i915_gem_object *obj,
 	       unsigned int tiling, unsigned int stride)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	unsigned int tile_width;
 
 	/* Linear is always fine */
@@ -183,7 +183,7 @@ static int
 i915_gem_object_fence_prepare(struct drm_i915_gem_object *obj,
 			      int tiling_mode, unsigned int stride)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 	struct i915_vma *vma, *vn;
 	LIST_HEAD(unbind);
@@ -221,7 +221,7 @@ i915_gem_object_fence_prepare(struct drm_i915_gem_object *obj,
 
 bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 
 	return to_gt(i915)->ggtt->bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 &&
 		i915_gem_object_is_tiled(obj);
@@ -231,7 +231,7 @@ int
 i915_gem_object_set_tiling(struct drm_i915_gem_object *obj,
 			   unsigned int tiling, unsigned int stride)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_vma *vma;
 	int err;
 
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_ttm.c b/drivers/gpu/drm/i915/gem/i915_gem_ttm.c
index 9227f8146a58..ad778c7c42e3 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_ttm.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_ttm.c
@@ -310,7 +310,7 @@ static struct ttm_tt *i915_ttm_tt_create(struct ttm_buffer_object *bo,
 	__i915_refct_sgt_init(&i915_tt->cached_rsgt, bo->base.size,
 			      &tt_rsgt_ops);
 
-	i915_tt->dev = obj->base.dev->dev;
+	i915_tt->dev = obj->base.drm->dev;
 
 	return &i915_tt->ttm;
 
@@ -1042,7 +1042,7 @@ static vm_fault_t vm_fault_ttm(struct vm_fault *vmf)
 {
 	struct vm_area_struct *area = vmf->vma;
 	struct ttm_buffer_object *bo = area->vm_private_data;
-	struct drm_device *dev = bo->base.dev;
+	struct drm_device *dev = bo->base.drm;
 	struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
 	intel_wakeref_t wakeref = 0;
 	vm_fault_t ret;
@@ -1110,7 +1110,7 @@ static vm_fault_t vm_fault_ttm(struct vm_fault *vmf)
 	}
 
 	if (i915_ttm_cpu_maps_iomem(bo->resource))
-		wakeref = intel_runtime_pm_get(&to_i915(obj->base.dev)->runtime_pm);
+		wakeref = intel_runtime_pm_get(&to_i915(obj->base.drm)->runtime_pm);
 
 	if (drm_dev_enter(dev, &idx)) {
 		ret = ttm_bo_vm_fault_reserved(vmf, vmf->vma->vm_page_prot,
@@ -1129,15 +1129,15 @@ static vm_fault_t vm_fault_ttm(struct vm_fault *vmf)
 	 */
 	if (ret == VM_FAULT_NOPAGE && wakeref && !obj->userfault_count) {
 		obj->userfault_count = 1;
-		spin_lock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
-		list_add(&obj->userfault_link, &to_i915(obj->base.dev)->runtime_pm.lmem_userfault_list);
-		spin_unlock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
+		spin_lock(&to_i915(obj->base.drm)->runtime_pm.lmem_userfault_lock);
+		list_add(&obj->userfault_link, &to_i915(obj->base.drm)->runtime_pm.lmem_userfault_list);
+		spin_unlock(&to_i915(obj->base.drm)->runtime_pm.lmem_userfault_lock);
 
 		GEM_WARN_ON(!i915_ttm_cpu_maps_iomem(bo->resource));
 	}
 
 	if (wakeref & CONFIG_DRM_I915_USERFAULT_AUTOSUSPEND)
-		intel_wakeref_auto(&to_i915(obj->base.dev)->runtime_pm.userfault_wakeref,
+		intel_wakeref_auto(&to_i915(obj->base.drm)->runtime_pm.userfault_wakeref,
 				   msecs_to_jiffies_timeout(CONFIG_DRM_I915_USERFAULT_AUTOSUSPEND));
 
 	i915_ttm_adjust_lru(obj);
@@ -1146,7 +1146,7 @@ static vm_fault_t vm_fault_ttm(struct vm_fault *vmf)
 
 out_rpm:
 	if (wakeref)
-		intel_runtime_pm_put(&to_i915(obj->base.dev)->runtime_pm, wakeref);
+		intel_runtime_pm_put(&to_i915(obj->base.drm)->runtime_pm, wakeref);
 
 	return ret;
 }
@@ -1205,13 +1205,13 @@ static void i915_ttm_unmap_virtual(struct drm_i915_gem_object *obj)
 	assert_object_held_shared(obj);
 
 	if (i915_ttm_cpu_maps_iomem(bo->resource)) {
-		wakeref = intel_runtime_pm_get(&to_i915(obj->base.dev)->runtime_pm);
+		wakeref = intel_runtime_pm_get(&to_i915(obj->base.drm)->runtime_pm);
 
 		/* userfault_count is protected by obj lock and rpm wakeref. */
 		if (obj->userfault_count) {
-			spin_lock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
+			spin_lock(&to_i915(obj->base.drm)->runtime_pm.lmem_userfault_lock);
 			list_del(&obj->userfault_link);
-			spin_unlock(&to_i915(obj->base.dev)->runtime_pm.lmem_userfault_lock);
+			spin_unlock(&to_i915(obj->base.drm)->runtime_pm.lmem_userfault_lock);
 			obj->userfault_count = 0;
 		}
 	}
@@ -1221,7 +1221,7 @@ static void i915_ttm_unmap_virtual(struct drm_i915_gem_object *obj)
 	ttm_bo_unmap_virtual(i915_gem_to_ttm(obj));
 
 	if (wakeref)
-		intel_runtime_pm_put(&to_i915(obj->base.dev)->runtime_pm, wakeref);
+		intel_runtime_pm_put(&to_i915(obj->base.drm)->runtime_pm, wakeref);
 }
 
 static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = {
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
index 7078af2f8f79..87d9987be741 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
@@ -121,7 +121,7 @@ void i915_ttm_adjust_gem_after_move(struct drm_i915_gem_object *obj)
 		mem_flags = i915_ttm_cpu_maps_iomem(bo->resource) ? I915_BO_FLAG_IOMEM :
 			I915_BO_FLAG_STRUCT_PAGE;
 		mem_type = bo->resource->mem_type;
-		cache_level = i915_ttm_cache_level(to_i915(bo->base.dev), bo->resource,
+		cache_level = i915_ttm_cache_level(to_i915(bo->base.drm), bo->resource,
 						   bo->ttm);
 	}
 
@@ -473,7 +473,7 @@ __i915_ttm_move(struct ttm_buffer_object *bo,
 {
 	const bool memcpy_allowed = i915_ttm_memcpy_allowed(bo, dst_mem);
 	struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
-	struct drm_i915_private *i915 = to_i915(bo->base.dev);
+	struct drm_i915_private *i915 = to_i915(bo->base.drm);
 	struct i915_ttm_memcpy_work *copy_work = NULL;
 	struct i915_ttm_memcpy_arg _arg, *arg = &_arg;
 	struct dma_fence *fence = ERR_PTR(-EINVAL);
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
index 1d3ebdf4069b..a0a313e75af9 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
@@ -62,7 +62,7 @@ static bool i915_gem_userptr_invalidate(struct mmu_interval_notifier *mni,
 					unsigned long cur_seq)
 {
 	struct drm_i915_gem_object *obj = container_of(mni, struct drm_i915_gem_object, userptr.notifier);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	long r;
 
 	if (!mmu_notifier_range_blockable(range))
@@ -128,7 +128,7 @@ static void i915_gem_object_userptr_drop_ref(struct drm_i915_gem_object *obj)
 
 static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj)
 {
-	unsigned int max_segment = i915_sg_segment_size(obj->base.dev->dev);
+	unsigned int max_segment = i915_sg_segment_size(obj->base.drm->dev);
 	struct sg_table *st;
 	struct page **pvec;
 	unsigned int num_pages; /* limited by sg_alloc_table_from_pages_segment */
@@ -386,7 +386,7 @@ i915_gem_userptr_release(struct drm_i915_gem_object *obj)
 static int
 i915_gem_userptr_dmabuf_export(struct drm_i915_gem_object *obj)
 {
-	drm_dbg(obj->base.dev, "Exporting userptr no longer allowed\n");
+	drm_dbg(obj->base.drm, "Exporting userptr no longer allowed\n");
 
 	return -EINVAL;
 }
@@ -395,7 +395,7 @@ static int
 i915_gem_userptr_pwrite(struct drm_i915_gem_object *obj,
 			const struct drm_i915_gem_pwrite *args)
 {
-	drm_dbg(obj->base.dev, "pwrite to userptr no longer allowed\n");
+	drm_dbg(obj->base.drm, "pwrite to userptr no longer allowed\n");
 
 	return -EINVAL;
 }
@@ -404,7 +404,7 @@ static int
 i915_gem_userptr_pread(struct drm_i915_gem_object *obj,
 		       const struct drm_i915_gem_pread *args)
 {
-	drm_dbg(obj->base.dev, "pread from userptr no longer allowed\n");
+	drm_dbg(obj->base.drm, "pread from userptr no longer allowed\n");
 
 	return -EINVAL;
 }
diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
index df6c9a84252c..90778bc07d67 100644
--- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
+++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
@@ -210,7 +210,7 @@ huge_pages_object(struct drm_i915_private *i915,
 
 static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	const u64 max_len = rounddown_pow_of_two(UINT_MAX);
 	struct sg_table *st;
 	struct scatterlist *sg;
@@ -265,7 +265,7 @@ static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
 
 static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *st;
 	struct scatterlist *sg;
 	unsigned int page_size;
diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
index ff81af4c8202..ce58ec31da89 100644
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
+++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
@@ -142,7 +142,7 @@ static int prepare_blit(const struct tiled_blits *t,
 			struct blit_buffer *src,
 			struct drm_i915_gem_object *batch)
 {
-	const int ver = GRAPHICS_VER(to_i915(batch->base.dev));
+	const int ver = GRAPHICS_VER(to_i915(batch->base.drm));
 	bool use_64b_reloc = ver >= 8;
 	u32 src_pitch, dst_pitch;
 	u32 cmd, *cs;
diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
index 7021b6e9b219..4a3ad4c6f522 100644
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
+++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
@@ -476,7 +476,7 @@ static int gpu_fill(struct intel_context *ce,
 
 static int cpu_fill(struct drm_i915_gem_object *obj, u32 value)
 {
-	const bool has_llc = HAS_LLC(to_i915(obj->base.dev));
+	const bool has_llc = HAS_LLC(to_i915(obj->base.drm));
 	unsigned int need_flush;
 	unsigned long n, m;
 	int err;
@@ -941,7 +941,7 @@ emit_rpcs_query(struct drm_i915_gem_object *obj,
 		struct intel_context *ce,
 		struct i915_request **rq_out)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_request *rq;
 	struct i915_gem_ww_ctx ww;
 	struct i915_vma *batch;
diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c
index e57f9390076c..84db924e560e 100644
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c
+++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c
@@ -284,7 +284,7 @@ static int igt_dmabuf_import_same_driver(struct drm_i915_private *i915,
 		goto out_import;
 
 	/* Now try a fake an importer */
-	import_attach = dma_buf_attach(dmabuf, obj->base.dev->dev);
+	import_attach = dma_buf_attach(dmabuf, obj->base.drm->dev);
 	if (IS_ERR(import_attach)) {
 		err = PTR_ERR(import_attach);
 		goto out_import;
@@ -360,7 +360,7 @@ static int igt_dmabuf_import(void *arg)
 		goto out_dmabuf;
 	}
 
-	if (obj->base.dev != &i915->drm) {
+	if (obj->base.drm != &i915->drm) {
 		pr_err("i915_gem_prime_import created a non-i915 object!\n");
 		err = -EINVAL;
 		goto out_obj;
diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c
index a93a90b15907..97e73c1c21f2 100644
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c
+++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c
@@ -15,7 +15,7 @@
 static int igt_fill_check_buffer(struct drm_i915_gem_object *obj,
 				 bool fill)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	unsigned int i, count = obj->base.size / sizeof(u32);
 	enum i915_map_type map_type =
 		i915_coherent_map_type(i915, obj, false);
diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
index a93d8f9f8bc1..2641a532fd84 100644
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
+++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
@@ -94,7 +94,7 @@ static int check_partial_mapping(struct drm_i915_gem_object *obj,
 				 struct rnd_state *prng)
 {
 	const unsigned long npages = obj->base.size / PAGE_SIZE;
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_gtt_view view;
 	struct i915_vma *vma;
 	unsigned long offset;
@@ -188,7 +188,7 @@ static int check_partial_mappings(struct drm_i915_gem_object *obj,
 {
 	const unsigned int nreal = obj->scratch / PAGE_SIZE;
 	const unsigned long npages = obj->base.size / PAGE_SIZE;
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_vma *vma;
 	unsigned long page;
 	int err;
@@ -539,7 +539,7 @@ static int igt_smoke_tiling(void *arg)
 
 static int make_obj_busy(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct intel_engine_cs *engine;
 
 	for_each_uabi_engine(engine, i915) {
@@ -868,7 +868,7 @@ static int wc_check(struct drm_i915_gem_object *obj)
 
 static bool can_mmap(struct drm_i915_gem_object *obj, enum i915_mmap_type type)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	bool no_map;
 
 	if (obj->ops->mmap_offset)
diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c b/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
index 37d0b0fe791d..78f11da72393 100644
--- a/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
+++ b/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
@@ -818,7 +818,7 @@ i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj,
 	if (obj->bit_17 == NULL) {
 		obj->bit_17 = bitmap_zalloc(page_count, GFP_KERNEL);
 		if (obj->bit_17 == NULL) {
-			drm_err(&to_i915(obj->base.dev)->drm,
+			drm_err(&to_i915(obj->base.drm)->drm,
 				"Failed to allocate memory for bit 17 record\n");
 			return;
 		}
diff --git a/drivers/gpu/drm/i915/gt/shmem_utils.c b/drivers/gpu/drm/i915/gt/shmem_utils.c
index 449c9ed44382..e0522222e0be 100644
--- a/drivers/gpu/drm/i915/gt/shmem_utils.c
+++ b/drivers/gpu/drm/i915/gt/shmem_utils.c
@@ -33,7 +33,7 @@ struct file *shmem_create_from_data(const char *name, void *data, size_t len)
 
 struct file *shmem_create_from_object(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	enum i915_map_type map_type;
 	struct file *file;
 	void *ptr;
diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c
index 6834f9fe40cf..c6fb1a0366dc 100644
--- a/drivers/gpu/drm/i915/gvt/dmabuf.c
+++ b/drivers/gpu/drm/i915/gvt/dmabuf.c
@@ -44,7 +44,7 @@
 
 static int vgpu_gem_get_pages(struct drm_i915_gem_object *obj)
 {
-	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
+	struct drm_i915_private *dev_priv = to_i915(obj->base.drm);
 	struct intel_vgpu *vgpu;
 	struct sg_table *st;
 	struct scatterlist *sg;
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 1f65bb33dd21..bf7c85874110 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -117,7 +117,7 @@ i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
 int i915_gem_object_unbind(struct drm_i915_gem_object *obj,
 			   unsigned long flags)
 {
-	struct intel_runtime_pm *rpm = &to_i915(obj->base.dev)->runtime_pm;
+	struct intel_runtime_pm *rpm = &to_i915(obj->base.drm)->runtime_pm;
 	bool vm_trylock = !!(flags & I915_GEM_OBJECT_UNBIND_VM_TRYLOCK);
 	LIST_HEAD(still_in_list);
 	intel_wakeref_t wakeref;
@@ -305,7 +305,7 @@ static struct i915_vma *i915_gem_gtt_prepare(struct drm_i915_gem_object *obj,
 					     struct drm_mm_node *node,
 					     bool write)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 	struct i915_vma *vma;
 	struct i915_gem_ww_ctx ww;
@@ -366,7 +366,7 @@ static void i915_gem_gtt_cleanup(struct drm_i915_gem_object *obj,
 				 struct drm_mm_node *node,
 				 struct i915_vma *vma)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 
 	i915_gem_object_unpin_pages(obj);
@@ -382,7 +382,7 @@ static int
 i915_gem_gtt_pread(struct drm_i915_gem_object *obj,
 		   const struct drm_i915_gem_pread *args)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 	unsigned long remain, offset;
 	intel_wakeref_t wakeref;
@@ -545,7 +545,7 @@ static int
 i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj,
 			 const struct drm_i915_gem_pwrite *args)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 	struct intel_runtime_pm *rpm = &i915->runtime_pm;
 	unsigned long remain, offset;
@@ -908,7 +908,7 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
 			    const struct i915_gtt_view *view,
 			    u64 size, u64 alignment, u64 flags)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 	struct i915_vma *vma;
 	int ret;
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
index 7bd1861ddbdf..d1891306fa28 100644
--- a/drivers/gpu/drm/i915/i915_gem_gtt.c
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
@@ -29,7 +29,7 @@ int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
 			       struct sg_table *pages)
 {
 	do {
-		if (dma_map_sg_attrs(obj->base.dev->dev,
+		if (dma_map_sg_attrs(obj->base.drm->dev,
 				     pages->sgl, pages->nents,
 				     DMA_BIDIRECTIONAL,
 				     DMA_ATTR_SKIP_CPU_SYNC |
@@ -45,7 +45,7 @@ int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
 		 * the DMA remapper, i915_gem_shrink will return 0.
 		 */
 		GEM_BUG_ON(obj->mm.pages == pages);
-	} while (i915_gem_shrink(NULL, to_i915(obj->base.dev),
+	} while (i915_gem_shrink(NULL, to_i915(obj->base.drm),
 				 obj->base.size >> PAGE_SHIFT, NULL,
 				 I915_SHRINK_BOUND |
 				 I915_SHRINK_UNBOUND));
@@ -56,7 +56,7 @@ int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
 void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj,
 			       struct sg_table *pages)
 {
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
 
 	/* XXX This does not prevent more requests being submitted! */
diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
index ffb425ba591c..1acff7929fd2 100644
--- a/drivers/gpu/drm/i915/i915_vma.c
+++ b/drivers/gpu/drm/i915/i915_vma.c
@@ -74,14 +74,14 @@ static void vma_print_allocator(struct i915_vma *vma, const char *reason)
 	char buf[512];
 
 	if (!vma->node.stack) {
-		drm_dbg(&to_i915(vma->obj->base.dev)->drm,
+		drm_dbg(&to_i915(vma->obj->base.drm)->drm,
 			"vma.node [%08llx + %08llx] %s: unknown owner\n",
 			vma->node.start, vma->node.size, reason);
 		return;
 	}
 
 	stack_depot_snprint(vma->node.stack, buf, sizeof(buf), 0);
-	drm_dbg(&to_i915(vma->obj->base.dev)->drm,
+	drm_dbg(&to_i915(vma->obj->base.drm)->drm,
 		"vma.node [%08llx + %08llx] %s: inserted at %s\n",
 		vma->node.start, vma->node.size, reason, buf);
 }
@@ -805,7 +805,7 @@ i915_vma_insert(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
 	 * attempt to find space.
 	 */
 	if (size > end - 2 * guard) {
-		drm_dbg(&to_i915(vma->obj->base.dev)->drm,
+		drm_dbg(&to_i915(vma->obj->base.drm)->drm,
 			"Attempting to bind an object larger than the aperture: request=%llu > %s aperture=%llu\n",
 			size, flags & PIN_MAPPABLE ? "mappable" : "total", end);
 		return -ENOSPC;
@@ -985,7 +985,7 @@ intel_rotate_pages(struct intel_rotation_info *rot_info,
 		   struct drm_i915_gem_object *obj)
 {
 	unsigned int size = intel_rotation_info_size(rot_info);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *st;
 	struct scatterlist *sg;
 	int ret = -ENOMEM;
@@ -1194,7 +1194,7 @@ intel_remap_pages(struct intel_remapped_info *rem_info,
 		  struct drm_i915_gem_object *obj)
 {
 	unsigned int size = intel_remapped_info_size(rem_info);
-	struct drm_i915_private *i915 = to_i915(obj->base.dev);
+	struct drm_i915_private *i915 = to_i915(obj->base.drm);
 	struct sg_table *st;
 	struct scatterlist *sg;
 	unsigned int gtt_offset = 0;
diff --git a/drivers/gpu/drm/lima/lima_gem.c b/drivers/gpu/drm/lima/lima_gem.c
index 4f9736e5f929..280311199263 100644
--- a/drivers/gpu/drm/lima/lima_gem.c
+++ b/drivers/gpu/drm/lima/lima_gem.c
@@ -22,7 +22,7 @@ int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm)
 {
 	struct page **pages;
 	struct address_space *mapping = bo->base.base.filp->f_mapping;
-	struct device *dev = bo->base.base.dev->dev;
+	struct device *dev = bo->base.base.drm->dev;
 	size_t old_size = bo->heap_size;
 	size_t new_size = bo->heap_size ? bo->heap_size * 2 :
 		(lima_heap_init_nr_pages << PAGE_SHIFT);
@@ -150,7 +150,7 @@ static void lima_gem_free_object(struct drm_gem_object *obj)
 	struct lima_bo *bo = to_lima_bo(obj);
 
 	if (!list_empty(&bo->va))
-		dev_err(obj->dev->dev, "lima gem free bo still has va\n");
+		dev_err(obj->drm->dev, "lima gem free bo still has va\n");
 
 	drm_gem_shmem_free(&bo->base);
 }
diff --git a/drivers/gpu/drm/loongson/lsdc_gem.c b/drivers/gpu/drm/loongson/lsdc_gem.c
index 04293df2f0de..a4029fd72df9 100644
--- a/drivers/gpu/drm/loongson/lsdc_gem.c
+++ b/drivers/gpu/drm/loongson/lsdc_gem.c
@@ -54,11 +54,11 @@ static struct sg_table *lsdc_gem_prime_get_sg_table(struct drm_gem_object *obj)
 	struct ttm_tt *tt = tbo->ttm;
 
 	if (!tt) {
-		drm_err(obj->dev, "sharing a buffer without backing memory\n");
+		drm_err(obj->drm, "sharing a buffer without backing memory\n");
 		return ERR_PTR(-ENOMEM);
 	}
 
-	return drm_prime_pages_to_sg(obj->dev, tt->pages, tt->num_pages);
+	return drm_prime_pages_to_sg(obj->drm, tt->pages, tt->num_pages);
 }
 
 static void lsdc_gem_object_free(struct drm_gem_object *obj)
@@ -82,13 +82,13 @@ static int lsdc_gem_object_vmap(struct drm_gem_object *obj, struct iosys_map *ma
 
 	ret = lsdc_bo_pin(lbo, 0, NULL);
 	if (unlikely(ret)) {
-		drm_err(obj->dev, "pin %p for vmap failed\n", lbo);
+		drm_err(obj->drm, "pin %p for vmap failed\n", lbo);
 		return ret;
 	}
 
 	ret = ttm_bo_vmap(tbo, &lbo->map);
 	if (ret) {
-		drm_err(obj->dev, "ttm bo vmap failed\n");
+		drm_err(obj->drm, "ttm bo vmap failed\n");
 		lsdc_bo_unpin(lbo);
 		return ret;
 	}
@@ -107,7 +107,7 @@ static void lsdc_gem_object_vunmap(struct drm_gem_object *obj, struct iosys_map
 	struct lsdc_bo *lbo = to_lsdc_bo(tbo);
 
 	if (unlikely(!lbo->vmap_count)) {
-		drm_warn(obj->dev, "%p is not mapped\n", lbo);
+		drm_warn(obj->drm, "%p is not mapped\n", lbo);
 		return;
 	}
 
@@ -126,7 +126,7 @@ static int lsdc_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struc
 
 	ret = ttm_bo_mmap_obj(vma, tbo);
 	if (unlikely(ret)) {
-		drm_warn(obj->dev, "mmap %p failed\n", tbo);
+		drm_warn(obj->drm, "mmap %p failed\n", tbo);
 		return ret;
 	}
 
diff --git a/drivers/gpu/drm/loongson/lsdc_ttm.c b/drivers/gpu/drm/loongson/lsdc_ttm.c
index bb0c8fd43a75..8674b7d9b678 100644
--- a/drivers/gpu/drm/loongson/lsdc_ttm.c
+++ b/drivers/gpu/drm/loongson/lsdc_ttm.c
@@ -163,7 +163,7 @@ static int lsdc_bo_move(struct ttm_buffer_object *tbo,
 			struct ttm_resource *new_mem,
 			struct ttm_place *hop)
 {
-	struct drm_device *ddev = tbo->base.dev;
+	struct drm_device *ddev = tbo->base.drm;
 	struct ttm_resource *old_mem = tbo->resource;
 	struct lsdc_bo *lbo = to_lsdc_bo(tbo);
 	int ret;
@@ -254,7 +254,7 @@ static struct ttm_device_funcs lsdc_bo_driver = {
 u64 lsdc_bo_gpu_offset(struct lsdc_bo *lbo)
 {
 	struct ttm_buffer_object *tbo = &lbo->tbo;
-	struct drm_device *ddev = tbo->base.dev;
+	struct drm_device *ddev = tbo->base.drm;
 	struct ttm_resource *resource = tbo->resource;
 
 	if (unlikely(!tbo->pin_count)) {
@@ -359,7 +359,7 @@ int lsdc_bo_kmap(struct lsdc_bo *lbo)
 {
 	struct ttm_buffer_object *tbo = &lbo->tbo;
 	struct drm_gem_object *gem = &tbo->base;
-	struct drm_device *ddev = gem->dev;
+	struct drm_device *ddev = gem->drm;
 	long ret;
 	int err;
 
diff --git a/drivers/gpu/drm/mediatek/mtk_drm_gem.c b/drivers/gpu/drm/mediatek/mtk_drm_gem.c
index a25b28d3ee90..852f1cdaf31c 100644
--- a/drivers/gpu/drm/mediatek/mtk_drm_gem.c
+++ b/drivers/gpu/drm/mediatek/mtk_drm_gem.c
@@ -100,7 +100,7 @@ struct mtk_drm_gem_obj *mtk_drm_gem_create(struct drm_device *dev,
 void mtk_drm_gem_free_object(struct drm_gem_object *obj)
 {
 	struct mtk_drm_gem_obj *mtk_gem = to_mtk_gem_obj(obj);
-	struct mtk_drm_private *priv = obj->dev->dev_private;
+	struct mtk_drm_private *priv = obj->drm->dev_private;
 
 	if (mtk_gem->sg)
 		drm_prime_gem_destroy(obj, mtk_gem->sg);
@@ -151,7 +151,7 @@ static int mtk_drm_gem_object_mmap(struct drm_gem_object *obj,
 {
 	int ret;
 	struct mtk_drm_gem_obj *mtk_gem = to_mtk_gem_obj(obj);
-	struct mtk_drm_private *priv = obj->dev->dev_private;
+	struct mtk_drm_private *priv = obj->drm->dev_private;
 
 	/*
 	 * Set vm_pgoff (used as a fake buffer offset by DRM) to 0 and map the
@@ -182,7 +182,7 @@ static int mtk_drm_gem_object_mmap(struct drm_gem_object *obj,
 struct sg_table *mtk_gem_prime_get_sg_table(struct drm_gem_object *obj)
 {
 	struct mtk_drm_gem_obj *mtk_gem = to_mtk_gem_obj(obj);
-	struct mtk_drm_private *priv = obj->dev->dev_private;
+	struct mtk_drm_private *priv = obj->drm->dev_private;
 	struct sg_table *sgt;
 	int ret;
 
diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
index 635744bc4765..8866fe7297ee 100644
--- a/drivers/gpu/drm/msm/msm_gem.c
+++ b/drivers/gpu/drm/msm/msm_gem.c
@@ -22,7 +22,7 @@
 static dma_addr_t physaddr(struct drm_gem_object *obj)
 {
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	return (((dma_addr_t)msm_obj->vram_node->start) << PAGE_SHIFT) +
 			priv->vram.paddr;
 }
@@ -49,21 +49,21 @@ static bool use_pages(struct drm_gem_object *obj)
 
 static void sync_for_device(struct msm_gem_object *msm_obj)
 {
-	struct device *dev = msm_obj->base.dev->dev;
+	struct device *dev = msm_obj->base.drm->dev;
 
 	dma_map_sgtable(dev, msm_obj->sgt, DMA_BIDIRECTIONAL, 0);
 }
 
 static void sync_for_cpu(struct msm_gem_object *msm_obj)
 {
-	struct device *dev = msm_obj->base.dev->dev;
+	struct device *dev = msm_obj->base.drm->dev;
 
 	dma_unmap_sgtable(dev, msm_obj->sgt, DMA_BIDIRECTIONAL, 0);
 }
 
 static void update_lru_active(struct drm_gem_object *obj)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	GEM_WARN_ON(!msm_obj->pages);
@@ -81,7 +81,7 @@ static void update_lru_active(struct drm_gem_object *obj)
 
 static void update_lru_locked(struct drm_gem_object *obj)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	msm_gem_assert_locked(&msm_obj->base);
@@ -97,7 +97,7 @@ static void update_lru_locked(struct drm_gem_object *obj)
 
 static void update_lru(struct drm_gem_object *obj)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 
 	mutex_lock(&priv->lru.lock);
 	update_lru_locked(obj);
@@ -108,7 +108,7 @@ static void update_lru(struct drm_gem_object *obj)
 static struct page **get_pages_vram(struct drm_gem_object *obj, int npages)
 {
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	dma_addr_t paddr;
 	struct page **p;
 	int ret, i;
@@ -141,7 +141,7 @@ static struct page **get_pages(struct drm_gem_object *obj)
 	msm_gem_assert_locked(obj);
 
 	if (!msm_obj->pages) {
-		struct drm_device *dev = obj->dev;
+		struct drm_device *dev = obj->drm;
 		struct page **p;
 		int npages = obj->size >> PAGE_SHIFT;
 
@@ -158,7 +158,7 @@ static struct page **get_pages(struct drm_gem_object *obj)
 
 		msm_obj->pages = p;
 
-		msm_obj->sgt = drm_prime_pages_to_sg(obj->dev, p, npages);
+		msm_obj->sgt = drm_prime_pages_to_sg(obj->drm, p, npages);
 		if (IS_ERR(msm_obj->sgt)) {
 			void *ptr = ERR_CAST(msm_obj->sgt);
 
@@ -182,7 +182,7 @@ static struct page **get_pages(struct drm_gem_object *obj)
 static void put_pages_vram(struct drm_gem_object *obj)
 {
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 
 	spin_lock(&priv->vram.lock);
 	drm_mm_remove_node(msm_obj->vram_node);
@@ -222,15 +222,15 @@ static void put_pages(struct drm_gem_object *obj)
 static struct page **msm_gem_pin_pages_locked(struct drm_gem_object *obj,
 					      unsigned madv)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 	struct page **p;
 
 	msm_gem_assert_locked(obj);
 
 	if (GEM_WARN_ON(msm_obj->madv > madv)) {
-		DRM_DEV_ERROR(obj->dev->dev, "Invalid madv state: %u vs %u\n",
-			msm_obj->madv, madv);
+		DRM_DEV_ERROR(obj->drm->dev, "Invalid madv state: %u vs %u\n",
+			      msm_obj->madv, madv);
 		return ERR_PTR(-EBUSY);
 	}
 
@@ -323,7 +323,7 @@ static vm_fault_t msm_gem_fault(struct vm_fault *vmf)
 /** get mmap offset */
 static uint64_t mmap_offset(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	int ret;
 
 	msm_gem_assert_locked(obj);
@@ -489,7 +489,7 @@ int msm_gem_pin_vma_locked(struct drm_gem_object *obj, struct msm_gem_vma *vma)
 
 void msm_gem_unpin_locked(struct drm_gem_object *obj)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	msm_gem_assert_locked(obj);
@@ -509,7 +509,7 @@ void msm_gem_unpin_locked(struct drm_gem_object *obj)
  */
 void msm_gem_unpin_active(struct drm_gem_object *obj)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	mutex_lock(&priv->lru.lock);
@@ -778,7 +778,7 @@ void msm_gem_put_vaddr(struct drm_gem_object *obj)
  */
 int msm_gem_madvise(struct drm_gem_object *obj, unsigned madv)
 {
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	msm_gem_lock(obj);
@@ -804,8 +804,8 @@ int msm_gem_madvise(struct drm_gem_object *obj, unsigned madv)
 
 void msm_gem_purge(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
-	struct msm_drm_private *priv = obj->dev->dev_private;
+	struct drm_device *dev = obj->drm;
+	struct msm_drm_private *priv = obj->drm->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	msm_gem_assert_locked(obj);
@@ -845,7 +845,7 @@ void msm_gem_purge(struct drm_gem_object *obj)
  */
 void msm_gem_evict(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 
 	msm_gem_assert_locked(obj);
@@ -1025,7 +1025,7 @@ void msm_gem_describe_objects(struct list_head *list, struct seq_file *m)
 static void msm_gem_free_object(struct drm_gem_object *obj)
 {
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct msm_drm_private *priv = dev->dev_private;
 
 	mutex_lock(&priv->obj_lock);
diff --git a/drivers/gpu/drm/msm/msm_gem_prime.c b/drivers/gpu/drm/msm/msm_gem_prime.c
index 5f68e31a3e4e..94f587465435 100644
--- a/drivers/gpu/drm/msm/msm_gem_prime.c
+++ b/drivers/gpu/drm/msm/msm_gem_prime.c
@@ -19,7 +19,7 @@ struct sg_table *msm_gem_prime_get_sg_table(struct drm_gem_object *obj)
 	if (WARN_ON(!msm_obj->pages))  /* should have already pinned! */
 		return ERR_PTR(-ENOMEM);
 
-	return drm_prime_pages_to_sg(obj->dev, msm_obj->pages, npages);
+	return drm_prime_pages_to_sg(obj->drm, msm_obj->pages, npages);
 }
 
 int msm_gem_prime_vmap(struct drm_gem_object *obj, struct iosys_map *map)
diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c
index c2ec91cc845d..99ca043c6da4 100644
--- a/drivers/gpu/drm/nouveau/nouveau_bo.c
+++ b/drivers/gpu/drm/nouveau/nouveau_bo.c
@@ -148,7 +148,7 @@ nouveau_bo_del_ttm(struct ttm_buffer_object *bo)
 	 * If nouveau_bo_new() allocated this buffer, the GEM object was never
 	 * initialized, so don't attempt to release it.
 	 */
-	if (bo->base.dev)
+	if (bo->base.drm)
 		drm_gem_object_release(&bo->base);
 	else
 		dma_resv_fini(&bo->base._resv);
diff --git a/drivers/gpu/drm/nouveau/nouveau_prime.c b/drivers/gpu/drm/nouveau/nouveau_prime.c
index f42c2b1b0363..b90570f26350 100644
--- a/drivers/gpu/drm/nouveau/nouveau_prime.c
+++ b/drivers/gpu/drm/nouveau/nouveau_prime.c
@@ -32,7 +32,7 @@ struct sg_table *nouveau_gem_prime_get_sg_table(struct drm_gem_object *obj)
 {
 	struct nouveau_bo *nvbo = nouveau_gem_object(obj);
 
-	return drm_prime_pages_to_sg(obj->dev, nvbo->bo.ttm->pages,
+	return drm_prime_pages_to_sg(obj->drm, nvbo->bo.ttm->pages,
 				     nvbo->bo.ttm->num_pages);
 }
 
diff --git a/drivers/gpu/drm/omapdrm/omap_gem.c b/drivers/gpu/drm/omapdrm/omap_gem.c
index 6b58a5bb7b44..b2b49819484d 100644
--- a/drivers/gpu/drm/omapdrm/omap_gem.c
+++ b/drivers/gpu/drm/omapdrm/omap_gem.c
@@ -133,7 +133,7 @@ struct omap_drm_usergart {
 /** get mmap offset */
 u64 omap_gem_mmap_offset(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	int ret;
 	size_t size;
 
@@ -167,7 +167,7 @@ static void omap_gem_evict_entry(struct drm_gem_object *obj,
 		enum tiler_fmt fmt, struct omap_drm_usergart_entry *entry)
 {
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
-	struct omap_drm_private *priv = obj->dev->dev_private;
+	struct omap_drm_private *priv = obj->drm->dev_private;
 	int n = priv->usergart[fmt].height;
 	size_t size = PAGE_SIZE * n;
 	loff_t off = omap_gem_mmap_offset(obj) +
@@ -178,12 +178,12 @@ static void omap_gem_evict_entry(struct drm_gem_object *obj,
 		int i;
 		/* if stride > than PAGE_SIZE then sparse mapping: */
 		for (i = n; i > 0; i--) {
-			unmap_mapping_range(obj->dev->anon_inode->i_mapping,
+			unmap_mapping_range(obj->drm->anon_inode->i_mapping,
 					    off, PAGE_SIZE, 1);
 			off += PAGE_SIZE * m;
 		}
 	} else {
-		unmap_mapping_range(obj->dev->anon_inode->i_mapping,
+		unmap_mapping_range(obj->drm->anon_inode->i_mapping,
 				    off, size, 1);
 	}
 
@@ -194,7 +194,7 @@ static void omap_gem_evict_entry(struct drm_gem_object *obj,
 static void omap_gem_evict(struct drm_gem_object *obj)
 {
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
-	struct omap_drm_private *priv = obj->dev->dev_private;
+	struct omap_drm_private *priv = obj->drm->dev_private;
 
 	if (omap_obj->flags & OMAP_BO_TILED_MASK) {
 		enum tiler_fmt fmt = gem2fmt(omap_obj->flags);
@@ -220,7 +220,7 @@ static void omap_gem_evict(struct drm_gem_object *obj)
  */
 static int omap_gem_attach_pages(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 	struct page **pages;
 	int npages = obj->size >> PAGE_SHIFT;
@@ -238,7 +238,8 @@ static int omap_gem_attach_pages(struct drm_gem_object *obj)
 
 	pages = drm_gem_get_pages(obj);
 	if (IS_ERR(pages)) {
-		dev_err(obj->dev->dev, "could not get pages: %ld\n", PTR_ERR(pages));
+		dev_err(obj->drm->dev, "could not get pages: %ld\n",
+			PTR_ERR(pages));
 		return PTR_ERR(pages);
 	}
 
@@ -301,7 +302,7 @@ static void omap_gem_detach_pages(struct drm_gem_object *obj)
 
 	for (i = 0; i < npages; i++) {
 		if (omap_obj->dma_addrs[i])
-			dma_unmap_page(obj->dev->dev, omap_obj->dma_addrs[i],
+			dma_unmap_page(obj->drm->dev, omap_obj->dma_addrs[i],
 				       PAGE_SIZE, DMA_TO_DEVICE);
 	}
 
@@ -373,7 +374,7 @@ static vm_fault_t omap_gem_fault_2d(struct drm_gem_object *obj,
 		struct vm_area_struct *vma, struct vm_fault *vmf)
 {
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
-	struct omap_drm_private *priv = obj->dev->dev_private;
+	struct omap_drm_private *priv = obj->drm->dev_private;
 	struct omap_drm_usergart_entry *entry;
 	enum tiler_fmt fmt = gem2fmt(omap_obj->flags);
 	struct page *pages[64];  /* XXX is this too much to have on stack? */
@@ -450,7 +451,7 @@ static vm_fault_t omap_gem_fault_2d(struct drm_gem_object *obj,
 	err = tiler_pin(entry->block, pages, ARRAY_SIZE(pages), 0, true);
 	if (err) {
 		ret = vmf_error(err);
-		dev_err(obj->dev->dev, "failed to pin: %d\n", err);
+		dev_err(obj->drm->dev, "failed to pin: %d\n", err);
 		return ret;
 	}
 
@@ -647,7 +648,7 @@ int omap_gem_roll(struct drm_gem_object *obj, u32 roll)
 	int ret = 0;
 
 	if (roll > npages) {
-		dev_err(obj->dev->dev, "invalid roll: %d\n", roll);
+		dev_err(obj->drm->dev, "invalid roll: %d\n", roll);
 		return -EINVAL;
 	}
 
@@ -664,7 +665,7 @@ int omap_gem_roll(struct drm_gem_object *obj, u32 roll)
 		ret = tiler_pin(omap_obj->block, omap_obj->pages, npages,
 				roll, true);
 		if (ret)
-			dev_err(obj->dev->dev, "could not repin: %d\n", ret);
+			dev_err(obj->drm->dev, "could not repin: %d\n", ret);
 	}
 
 fail:
@@ -700,7 +701,7 @@ static inline bool omap_gem_is_cached_coherent(struct drm_gem_object *obj)
  */
 void omap_gem_cpu_sync_page(struct drm_gem_object *obj, int pgoff)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 
 	if (omap_gem_is_cached_coherent(obj))
@@ -717,7 +718,7 @@ void omap_gem_cpu_sync_page(struct drm_gem_object *obj, int pgoff)
 void omap_gem_dma_sync_buffer(struct drm_gem_object *obj,
 		enum dma_data_direction dir)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 	int i, npages = obj->size >> PAGE_SHIFT;
 	struct page **pages = omap_obj->pages;
@@ -768,7 +769,7 @@ static int omap_gem_pin_tiler(struct drm_gem_object *obj)
 
 	if (IS_ERR(block)) {
 		ret = PTR_ERR(block);
-		dev_err(obj->dev->dev, "could not remap: %d (%d)\n", ret, fmt);
+		dev_err(obj->drm->dev, "could not remap: %d (%d)\n", ret, fmt);
 		goto fail;
 	}
 
@@ -776,7 +777,7 @@ static int omap_gem_pin_tiler(struct drm_gem_object *obj)
 	ret = tiler_pin(block, omap_obj->pages, npages, omap_obj->roll, true);
 	if (ret) {
 		tiler_release(block);
-		dev_err(obj->dev->dev, "could not pin: %d\n", ret);
+		dev_err(obj->drm->dev, "could not pin: %d\n", ret);
 		goto fail;
 	}
 
@@ -805,7 +806,7 @@ static int omap_gem_pin_tiler(struct drm_gem_object *obj)
  */
 int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr)
 {
-	struct omap_drm_private *priv = obj->dev->dev_private;
+	struct omap_drm_private *priv = obj->drm->dev_private;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 	int ret = 0;
 
@@ -849,7 +850,7 @@ int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr)
  */
 static void omap_gem_unpin_locked(struct drm_gem_object *obj)
 {
-	struct omap_drm_private *priv = obj->dev->dev_private;
+	struct omap_drm_private *priv = obj->drm->dev_private;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 	int ret;
 
@@ -867,12 +868,12 @@ static void omap_gem_unpin_locked(struct drm_gem_object *obj)
 		if (priv->has_dmm) {
 			ret = tiler_unpin(omap_obj->block);
 			if (ret) {
-				dev_err(obj->dev->dev,
+				dev_err(obj->drm->dev,
 					"could not unpin pages: %d\n", ret);
 			}
 			ret = tiler_release(omap_obj->block);
 			if (ret) {
-				dev_err(obj->dev->dev,
+				dev_err(obj->drm->dev,
 					"could not release unmap: %d\n", ret);
 			}
 			omap_obj->dma_addr = 0;
@@ -1195,7 +1196,7 @@ void omap_gem_describe_objects(struct list_head *list, struct seq_file *m)
 
 static void omap_gem_free_object(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct omap_drm_private *priv = dev->dev_private;
 	struct omap_gem_object *omap_obj = to_omap_bo(obj);
 
diff --git a/drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c b/drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c
index 8e194dbc9506..f00a811d716a 100644
--- a/drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c
+++ b/drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c
@@ -92,7 +92,7 @@ struct dma_buf *omap_gem_prime_export(struct drm_gem_object *obj, int flags)
 	exp_info.priv = obj;
 	exp_info.resv = obj->resv;
 
-	return drm_gem_dmabuf_export(obj->dev, &exp_info);
+	return drm_gem_dmabuf_export(obj->drm, &exp_info);
 }
 
 /* -----------------------------------------------------------------------------
@@ -109,7 +109,7 @@ struct drm_gem_object *omap_gem_prime_import(struct drm_device *dev,
 
 	if (dma_buf->ops == &omap_dmabuf_ops) {
 		obj = dma_buf->priv;
-		if (obj->dev == dev) {
+		if (obj->drm == dev) {
 			/*
 			 * Importing dmabuf exported from out own gem increases
 			 * refcount on gem itself instead of f_count of dmabuf.
diff --git a/drivers/gpu/drm/panfrost/panfrost_gem.c b/drivers/gpu/drm/panfrost/panfrost_gem.c
index 3c812fbd126f..8c3eebcad5a5 100644
--- a/drivers/gpu/drm/panfrost/panfrost_gem.c
+++ b/drivers/gpu/drm/panfrost/panfrost_gem.c
@@ -17,7 +17,7 @@
 static void panfrost_gem_free_object(struct drm_gem_object *obj)
 {
 	struct panfrost_gem_object *bo = to_panfrost_bo(obj);
-	struct panfrost_device *pfdev = obj->dev->dev_private;
+	struct panfrost_device *pfdev = obj->drm->dev_private;
 
 	/*
 	 * Make sure the BO is no longer inserted in the shrinker list before
diff --git a/drivers/gpu/drm/panfrost/panfrost_mmu.c b/drivers/gpu/drm/panfrost/panfrost_mmu.c
index c0123d09f699..8d4e94725310 100644
--- a/drivers/gpu/drm/panfrost/panfrost_mmu.c
+++ b/drivers/gpu/drm/panfrost/panfrost_mmu.c
@@ -323,7 +323,7 @@ int panfrost_mmu_map(struct panfrost_gem_mapping *mapping)
 	struct panfrost_gem_object *bo = mapping->obj;
 	struct drm_gem_shmem_object *shmem = &bo->base;
 	struct drm_gem_object *obj = &shmem->base;
-	struct panfrost_device *pfdev = to_panfrost_device(obj->dev);
+	struct panfrost_device *pfdev = to_panfrost_device(obj->drm);
 	struct sg_table *sgt;
 	int prot = IOMMU_READ | IOMMU_WRITE;
 
@@ -348,7 +348,7 @@ void panfrost_mmu_unmap(struct panfrost_gem_mapping *mapping)
 {
 	struct panfrost_gem_object *bo = mapping->obj;
 	struct drm_gem_object *obj = &bo->base.base;
-	struct panfrost_device *pfdev = to_panfrost_device(obj->dev);
+	struct panfrost_device *pfdev = to_panfrost_device(obj->drm);
 	struct io_pgtable_ops *ops = mapping->mmu->pgtbl_ops;
 	u64 iova = mapping->mmnode.start << PAGE_SHIFT;
 	size_t len = mapping->mmnode.size << PAGE_SHIFT;
diff --git a/drivers/gpu/drm/qxl/qxl_gem.c b/drivers/gpu/drm/qxl/qxl_gem.c
index a08da0bd9098..d3b8dae3e55e 100644
--- a/drivers/gpu/drm/qxl/qxl_gem.c
+++ b/drivers/gpu/drm/qxl/qxl_gem.c
@@ -34,7 +34,7 @@ void qxl_gem_object_free(struct drm_gem_object *gobj)
 	struct qxl_device *qdev;
 	struct ttm_buffer_object *tbo;
 
-	qdev = to_qxl(gobj->dev);
+	qdev = to_qxl(gobj->drm);
 
 	qxl_surface_evict(qdev, qobj, false);
 
diff --git a/drivers/gpu/drm/qxl/qxl_object.c b/drivers/gpu/drm/qxl/qxl_object.c
index 06a58dad5f5c..4d3f5cec26ce 100644
--- a/drivers/gpu/drm/qxl/qxl_object.c
+++ b/drivers/gpu/drm/qxl/qxl_object.c
@@ -38,7 +38,7 @@ static void qxl_ttm_bo_destroy(struct ttm_buffer_object *tbo)
 	struct qxl_device *qdev;
 
 	bo = to_qxl_bo(tbo);
-	qdev = to_qxl(bo->tbo.base.dev);
+	qdev = to_qxl(bo->tbo.base.drm);
 
 	qxl_surface_evict(qdev, bo, false);
 	WARN_ON_ONCE(bo->map_count > 0);
@@ -295,7 +295,7 @@ struct qxl_bo *qxl_bo_ref(struct qxl_bo *bo)
 static int __qxl_bo_pin(struct qxl_bo *bo)
 {
 	struct ttm_operation_ctx ctx = { false, false };
-	struct drm_device *ddev = bo->tbo.base.dev;
+	struct drm_device *ddev = bo->tbo.base.drm;
 	int r;
 
 	if (bo->tbo.pin_count) {
diff --git a/drivers/gpu/drm/qxl/qxl_object.h b/drivers/gpu/drm/qxl/qxl_object.h
index 53392cb90eec..1cc9e9d2e432 100644
--- a/drivers/gpu/drm/qxl/qxl_object.h
+++ b/drivers/gpu/drm/qxl/qxl_object.h
@@ -34,7 +34,7 @@ static inline int qxl_bo_reserve(struct qxl_bo *bo)
 	r = ttm_bo_reserve(&bo->tbo, true, false, NULL);
 	if (unlikely(r != 0)) {
 		if (r != -ERESTARTSYS) {
-			struct drm_device *ddev = bo->tbo.base.dev;
+			struct drm_device *ddev = bo->tbo.base.drm;
 
 			dev_err(ddev->dev, "%p reserve failed\n", bo);
 		}
diff --git a/drivers/gpu/drm/qxl/qxl_release.c b/drivers/gpu/drm/qxl/qxl_release.c
index 368d26da0d6a..e82d959a9517 100644
--- a/drivers/gpu/drm/qxl/qxl_release.c
+++ b/drivers/gpu/drm/qxl/qxl_release.c
@@ -205,7 +205,7 @@ static int qxl_release_validate_bo(struct qxl_bo *bo)
 		return ret;
 
 	/* allocate a surface for reserved + validated buffers */
-	ret = qxl_bo_check_id(to_qxl(bo->tbo.base.dev), bo);
+	ret = qxl_bo_check_id(to_qxl(bo->tbo.base.drm), bo);
 	if (ret)
 		return ret;
 	return 0;
diff --git a/drivers/gpu/drm/qxl/qxl_ttm.c b/drivers/gpu/drm/qxl/qxl_ttm.c
index 1a82629bce3f..708ef211be07 100644
--- a/drivers/gpu/drm/qxl/qxl_ttm.c
+++ b/drivers/gpu/drm/qxl/qxl_ttm.c
@@ -129,7 +129,7 @@ static void qxl_bo_move_notify(struct ttm_buffer_object *bo,
 	if (!qxl_ttm_bo_is_qxl_bo(bo) || !bo->resource)
 		return;
 	qbo = to_qxl_bo(bo);
-	qdev = to_qxl(qbo->tbo.base.dev);
+	qdev = to_qxl(qbo->tbo.base.drm);
 
 	if (bo->resource->mem_type == TTM_PL_PRIV && qbo->surface_id)
 		qxl_surface_evict(qdev, qbo, new_mem ? true : false);
diff --git a/drivers/gpu/drm/radeon/radeon_prime.c b/drivers/gpu/drm/radeon/radeon_prime.c
index b3cfc99f4d7e..8b2342c20909 100644
--- a/drivers/gpu/drm/radeon/radeon_prime.c
+++ b/drivers/gpu/drm/radeon/radeon_prime.c
@@ -38,7 +38,7 @@ struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj)
 {
 	struct radeon_bo *bo = gem_to_radeon_bo(obj);
 
-	return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages,
+	return drm_prime_pages_to_sg(obj->drm, bo->tbo.ttm->pages,
 				     bo->tbo.ttm->num_pages);
 }
 
diff --git a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c
index b8f8b45ebf59..fa91155a7dc0 100644
--- a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c
+++ b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c
@@ -20,7 +20,7 @@
 
 static int rockchip_gem_iommu_map(struct rockchip_gem_object *rk_obj)
 {
-	struct drm_device *drm = rk_obj->base.dev;
+	struct drm_device *drm = rk_obj->base.drm;
 	struct rockchip_drm_private *private = drm->dev_private;
 	int prot = IOMMU_READ | IOMMU_WRITE;
 	ssize_t ret;
@@ -61,7 +61,7 @@ static int rockchip_gem_iommu_map(struct rockchip_gem_object *rk_obj)
 
 static int rockchip_gem_iommu_unmap(struct rockchip_gem_object *rk_obj)
 {
-	struct drm_device *drm = rk_obj->base.dev;
+	struct drm_device *drm = rk_obj->base.drm;
 	struct rockchip_drm_private *private = drm->dev_private;
 
 	iommu_unmap(private->domain, rk_obj->dma_addr, rk_obj->size);
@@ -77,7 +77,7 @@ static int rockchip_gem_iommu_unmap(struct rockchip_gem_object *rk_obj)
 
 static int rockchip_gem_get_pages(struct rockchip_gem_object *rk_obj)
 {
-	struct drm_device *drm = rk_obj->base.dev;
+	struct drm_device *drm = rk_obj->base.drm;
 	int ret, i;
 	struct scatterlist *s;
 
@@ -87,7 +87,7 @@ static int rockchip_gem_get_pages(struct rockchip_gem_object *rk_obj)
 
 	rk_obj->num_pages = rk_obj->base.size >> PAGE_SHIFT;
 
-	rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->base.dev,
+	rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->base.drm,
 					    rk_obj->pages, rk_obj->num_pages);
 	if (IS_ERR(rk_obj->sgt)) {
 		ret = PTR_ERR(rk_obj->sgt);
@@ -157,7 +157,7 @@ static int rockchip_gem_alloc_dma(struct rockchip_gem_object *rk_obj,
 				  bool alloc_kmap)
 {
 	struct drm_gem_object *obj = &rk_obj->base;
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 
 	rk_obj->dma_attrs = DMA_ATTR_WRITE_COMBINE;
 
@@ -179,7 +179,7 @@ static int rockchip_gem_alloc_buf(struct rockchip_gem_object *rk_obj,
 				  bool alloc_kmap)
 {
 	struct drm_gem_object *obj = &rk_obj->base;
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 	struct rockchip_drm_private *private = drm->dev_private;
 
 	if (private->domain)
@@ -198,7 +198,7 @@ static void rockchip_gem_free_iommu(struct rockchip_gem_object *rk_obj)
 static void rockchip_gem_free_dma(struct rockchip_gem_object *rk_obj)
 {
 	struct drm_gem_object *obj = &rk_obj->base;
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 
 	dma_free_attrs(drm->dev, obj->size, rk_obj->kvaddr, rk_obj->dma_addr,
 		       rk_obj->dma_attrs);
@@ -229,7 +229,7 @@ static int rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj,
 					    struct vm_area_struct *vma)
 {
 	struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj);
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 
 	return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr,
 			      obj->size, rk_obj->dma_attrs);
@@ -328,7 +328,7 @@ rockchip_gem_create_object(struct drm_device *drm, unsigned int size,
  */
 void rockchip_gem_free_object(struct drm_gem_object *obj)
 {
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 	struct rockchip_drm_private *private = drm->dev_private;
 	struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj);
 
@@ -426,12 +426,13 @@ int rockchip_gem_dumb_create(struct drm_file *file_priv,
 struct sg_table *rockchip_gem_prime_get_sg_table(struct drm_gem_object *obj)
 {
 	struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj);
-	struct drm_device *drm = obj->dev;
+	struct drm_device *drm = obj->drm;
 	struct sg_table *sgt;
 	int ret;
 
 	if (rk_obj->pages)
-		return drm_prime_pages_to_sg(obj->dev, rk_obj->pages, rk_obj->num_pages);
+		return drm_prime_pages_to_sg(obj->drm, rk_obj->pages,
+					     rk_obj->num_pages);
 
 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
 	if (!sgt)
diff --git a/drivers/gpu/drm/tegra/gem.c b/drivers/gpu/drm/tegra/gem.c
index a4023163493d..2cb60cafdf43 100644
--- a/drivers/gpu/drm/tegra/gem.c
+++ b/drivers/gpu/drm/tegra/gem.c
@@ -340,7 +340,7 @@ static int tegra_bo_get_pages(struct drm_device *drm, struct tegra_bo *bo)
 
 	bo->num_pages = bo->gem.size >> PAGE_SHIFT;
 
-	bo->sgt = drm_prime_pages_to_sg(bo->gem.dev, bo->pages, bo->num_pages);
+	bo->sgt = drm_prime_pages_to_sg(bo->gem.drm, bo->pages, bo->num_pages);
 	if (IS_ERR(bo->sgt)) {
 		err = PTR_ERR(bo->sgt);
 		goto put_pages;
@@ -493,7 +493,7 @@ static struct tegra_bo *tegra_bo_import(struct drm_device *drm,
 
 void tegra_bo_free_object(struct drm_gem_object *gem)
 {
-	struct tegra_drm *tegra = gem->dev->dev_private;
+	struct tegra_drm *tegra = gem->drm->dev_private;
 	struct host1x_bo_mapping *mapping, *tmp;
 	struct tegra_bo *bo = to_tegra_bo(gem);
 
@@ -502,7 +502,8 @@ void tegra_bo_free_object(struct drm_gem_object *gem)
 		if (mapping->cache)
 			host1x_bo_unpin(mapping);
 		else
-			dev_err(gem->dev->dev, "mapping %p stale for device %s\n", mapping,
+			dev_err(gem->drm->dev,
+				"mapping %p stale for device %s\n", mapping,
 				dev_name(mapping->dev));
 	}
 
@@ -514,7 +515,7 @@ void tegra_bo_free_object(struct drm_gem_object *gem)
 						  DMA_TO_DEVICE);
 		drm_prime_gem_destroy(gem, NULL);
 	} else {
-		tegra_bo_free(gem->dev, bo);
+		tegra_bo_free(gem->drm, bo);
 	}
 
 	drm_gem_object_release(gem);
@@ -578,7 +579,7 @@ int __tegra_gem_mmap(struct drm_gem_object *gem, struct vm_area_struct *vma)
 		vm_flags_clear(vma, VM_PFNMAP);
 		vma->vm_pgoff = 0;
 
-		err = dma_mmap_wc(gem->dev->dev, vma, bo->vaddr, bo->iova,
+		err = dma_mmap_wc(gem->drm->dev, vma, bo->vaddr, bo->iova,
 				  gem->size);
 		if (err < 0) {
 			drm_gem_vm_close(vma);
@@ -668,7 +669,7 @@ static int tegra_gem_prime_begin_cpu_access(struct dma_buf *buf,
 {
 	struct drm_gem_object *gem = buf->priv;
 	struct tegra_bo *bo = to_tegra_bo(gem);
-	struct drm_device *drm = gem->dev;
+	struct drm_device *drm = gem->drm;
 
 	if (bo->pages)
 		dma_sync_sgtable_for_cpu(drm->dev, bo->sgt, DMA_FROM_DEVICE);
@@ -681,7 +682,7 @@ static int tegra_gem_prime_end_cpu_access(struct dma_buf *buf,
 {
 	struct drm_gem_object *gem = buf->priv;
 	struct tegra_bo *bo = to_tegra_bo(gem);
-	struct drm_device *drm = gem->dev;
+	struct drm_device *drm = gem->drm;
 
 	if (bo->pages)
 		dma_sync_sgtable_for_device(drm->dev, bo->sgt, DMA_TO_DEVICE);
@@ -741,13 +742,13 @@ struct dma_buf *tegra_gem_prime_export(struct drm_gem_object *gem,
 	DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
 
 	exp_info.exp_name = KBUILD_MODNAME;
-	exp_info.owner = gem->dev->driver->fops->owner;
+	exp_info.owner = gem->drm->driver->fops->owner;
 	exp_info.ops = &tegra_gem_prime_dmabuf_ops;
 	exp_info.size = gem->size;
 	exp_info.flags = flags;
 	exp_info.priv = gem;
 
-	return drm_gem_dmabuf_export(gem->dev, &exp_info);
+	return drm_gem_dmabuf_export(gem->drm, &exp_info);
 }
 
 struct drm_gem_object *tegra_gem_prime_import(struct drm_device *drm,
@@ -758,7 +759,7 @@ struct drm_gem_object *tegra_gem_prime_import(struct drm_device *drm,
 	if (buf->ops == &tegra_gem_prime_dmabuf_ops) {
 		struct drm_gem_object *gem = buf->priv;
 
-		if (gem->dev == drm) {
+		if (gem->drm == drm) {
 			drm_gem_object_get(gem);
 			return gem;
 		}
diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
index fd9fd3d15101..3637af18b867 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_util.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
@@ -250,7 +250,7 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo,
 		fbo->base.base.resv = &fbo->base.base._resv;
 
 	dma_resv_init(&fbo->base.base._resv);
-	fbo->base.base.dev = NULL;
+	fbo->base.base.drm = NULL;
 	ret = dma_resv_trylock(&fbo->base.base._resv);
 	WARN_ON(!ret);
 
diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
index 4212b8c91dd4..c2e9d1f777ea 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
@@ -292,7 +292,7 @@ vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot)
 {
 	struct vm_area_struct *vma = vmf->vma;
 	struct ttm_buffer_object *bo = vma->vm_private_data;
-	struct drm_device *ddev = bo->base.dev;
+	struct drm_device *ddev = bo->base.drm;
 	vm_fault_t ret = VM_FAULT_NOPAGE;
 	unsigned long address;
 	unsigned long pfn;
@@ -323,7 +323,7 @@ vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
 	struct vm_area_struct *vma = vmf->vma;
 	pgprot_t prot;
 	struct ttm_buffer_object *bo = vma->vm_private_data;
-	struct drm_device *ddev = bo->base.dev;
+	struct drm_device *ddev = bo->base.drm;
 	vm_fault_t ret;
 	int idx;
 
diff --git a/drivers/gpu/drm/v3d/v3d_bo.c b/drivers/gpu/drm/v3d/v3d_bo.c
index 8b3229a37c6d..8fb1c38424f8 100644
--- a/drivers/gpu/drm/v3d/v3d_bo.c
+++ b/drivers/gpu/drm/v3d/v3d_bo.c
@@ -30,7 +30,7 @@
  */
 void v3d_free_object(struct drm_gem_object *obj)
 {
-	struct v3d_dev *v3d = to_v3d_dev(obj->dev);
+	struct v3d_dev *v3d = to_v3d_dev(obj->drm);
 	struct v3d_bo *bo = to_v3d_bo(obj);
 
 	v3d_mmu_remove_ptes(bo);
@@ -88,7 +88,7 @@ struct drm_gem_object *v3d_create_object(struct drm_device *dev, size_t size)
 static int
 v3d_bo_create_finish(struct drm_gem_object *obj)
 {
-	struct v3d_dev *v3d = to_v3d_dev(obj->dev);
+	struct v3d_dev *v3d = to_v3d_dev(obj->drm);
 	struct v3d_bo *bo = to_v3d_bo(obj);
 	struct sg_table *sgt;
 	int ret;
diff --git a/drivers/gpu/drm/v3d/v3d_mmu.c b/drivers/gpu/drm/v3d/v3d_mmu.c
index 5a453532901f..efea2273ab93 100644
--- a/drivers/gpu/drm/v3d/v3d_mmu.c
+++ b/drivers/gpu/drm/v3d/v3d_mmu.c
@@ -87,7 +87,7 @@ int v3d_mmu_set_page_table(struct v3d_dev *v3d)
 void v3d_mmu_insert_ptes(struct v3d_bo *bo)
 {
 	struct drm_gem_shmem_object *shmem_obj = &bo->base;
-	struct v3d_dev *v3d = to_v3d_dev(shmem_obj->base.dev);
+	struct v3d_dev *v3d = to_v3d_dev(shmem_obj->base.drm);
 	u32 page = bo->node.start;
 	u32 page_prot = V3D_PTE_WRITEABLE | V3D_PTE_VALID;
 	struct sg_dma_page_iter dma_iter;
@@ -113,7 +113,7 @@ void v3d_mmu_insert_ptes(struct v3d_bo *bo)
 
 void v3d_mmu_remove_ptes(struct v3d_bo *bo)
 {
-	struct v3d_dev *v3d = to_v3d_dev(bo->base.base.dev);
+	struct v3d_dev *v3d = to_v3d_dev(bo->base.base.drm);
 	u32 npages = bo->base.base.size >> V3D_MMU_PAGE_SHIFT;
 	u32 page;
 
diff --git a/drivers/gpu/drm/vc4/vc4_bo.c b/drivers/gpu/drm/vc4/vc4_bo.c
index 71a6f5e47817..54ff7922ae4d 100644
--- a/drivers/gpu/drm/vc4/vc4_bo.c
+++ b/drivers/gpu/drm/vc4/vc4_bo.c
@@ -132,7 +132,7 @@ static int vc4_get_user_label(struct vc4_dev *vc4, const char *name)
 static void vc4_bo_set_label(struct drm_gem_object *gem_obj, int label)
 {
 	struct vc4_bo *bo = to_vc4_bo(gem_obj);
-	struct vc4_dev *vc4 = to_vc4_dev(gem_obj->dev);
+	struct vc4_dev *vc4 = to_vc4_dev(gem_obj->drm);
 
 	lockdep_assert_held(&vc4->bo_lock);
 
@@ -166,7 +166,7 @@ static uint32_t bo_page_index(size_t size)
 static void vc4_bo_destroy(struct vc4_bo *bo)
 {
 	struct drm_gem_object *obj = &bo->base.base;
-	struct vc4_dev *vc4 = to_vc4_dev(obj->dev);
+	struct vc4_dev *vc4 = to_vc4_dev(obj->drm);
 
 	lockdep_assert_held(&vc4->bo_lock);
 
@@ -185,7 +185,7 @@ static void vc4_bo_destroy(struct vc4_bo *bo)
 
 static void vc4_bo_remove_from_cache(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 
 	lockdep_assert_held(&vc4->bo_lock);
 	list_del(&bo->unref_head);
@@ -249,7 +249,7 @@ static void vc4_bo_cache_purge(struct drm_device *dev)
 
 void vc4_bo_add_to_purgeable_pool(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 
 	if (WARN_ON_ONCE(vc4->is_vc5))
 		return;
@@ -263,7 +263,7 @@ void vc4_bo_add_to_purgeable_pool(struct vc4_bo *bo)
 
 static void vc4_bo_remove_from_purgeable_pool_locked(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 
 	if (WARN_ON_ONCE(vc4->is_vc5))
 		return;
@@ -287,7 +287,7 @@ static void vc4_bo_remove_from_purgeable_pool_locked(struct vc4_bo *bo)
 
 void vc4_bo_remove_from_purgeable_pool(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 
 	mutex_lock(&vc4->purgeable.lock);
 	vc4_bo_remove_from_purgeable_pool_locked(bo);
@@ -297,7 +297,7 @@ void vc4_bo_remove_from_purgeable_pool(struct vc4_bo *bo)
 static void vc4_bo_purge(struct drm_gem_object *obj)
 {
 	struct vc4_bo *bo = to_vc4_bo(obj);
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 
 	WARN_ON(!mutex_is_locked(&bo->madv_lock));
 	WARN_ON(bo->madv != VC4_MADV_DONTNEED);
@@ -542,7 +542,7 @@ static void vc4_bo_cache_free_old(struct drm_device *dev)
  */
 static void vc4_free_object(struct drm_gem_object *gem_bo)
 {
-	struct drm_device *dev = gem_bo->dev;
+	struct drm_device *dev = gem_bo->drm;
 	struct vc4_dev *vc4 = to_vc4_dev(dev);
 	struct vc4_bo *bo = to_vc4_bo(gem_bo);
 	struct list_head *cache_list;
@@ -619,7 +619,7 @@ static void vc4_bo_cache_time_work(struct work_struct *work)
 
 int vc4_bo_inc_usecnt(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 	int ret;
 
 	if (WARN_ON_ONCE(vc4->is_vc5))
@@ -659,7 +659,7 @@ int vc4_bo_inc_usecnt(struct vc4_bo *bo)
 
 void vc4_bo_dec_usecnt(struct vc4_bo *bo)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.drm);
 
 	if (WARN_ON_ONCE(vc4->is_vc5))
 		return;
diff --git a/drivers/gpu/drm/vc4/vc4_validate.c b/drivers/gpu/drm/vc4/vc4_validate.c
index 7dff3ca5af6b..1a279455e67b 100644
--- a/drivers/gpu/drm/vc4/vc4_validate.c
+++ b/drivers/gpu/drm/vc4/vc4_validate.c
@@ -356,7 +356,7 @@ validate_gl_shader_state(VALIDATE_ARGS)
 static int
 validate_tile_binning_config(VALIDATE_ARGS)
 {
-	struct drm_device *dev = exec->exec_bo->base.dev;
+	struct drm_device *dev = exec->exec_bo->base.drm;
 	struct vc4_dev *vc4 = to_vc4_dev(dev);
 	uint8_t flags;
 	uint32_t tile_state_size;
diff --git a/drivers/gpu/drm/vc4/vc4_validate_shaders.c b/drivers/gpu/drm/vc4/vc4_validate_shaders.c
index 9745f8810eca..a9aea68a6441 100644
--- a/drivers/gpu/drm/vc4/vc4_validate_shaders.c
+++ b/drivers/gpu/drm/vc4/vc4_validate_shaders.c
@@ -778,7 +778,7 @@ vc4_handle_branch_target(struct vc4_shader_validation_state *validation_state)
 struct vc4_validated_shader_info *
 vc4_validate_shader(struct drm_gem_dma_object *shader_obj)
 {
-	struct vc4_dev *vc4 = to_vc4_dev(shader_obj->base.dev);
+	struct vc4_dev *vc4 = to_vc4_dev(shader_obj->base.drm);
 	bool found_shader_end = false;
 	int shader_end_ip = 0;
 	uint32_t last_thread_switch_ip = -3;
diff --git a/drivers/gpu/drm/virtio/virtgpu_gem.c b/drivers/gpu/drm/virtio/virtgpu_gem.c
index 7db48d17ee3a..f7e9521dda34 100644
--- a/drivers/gpu/drm/virtio/virtgpu_gem.c
+++ b/drivers/gpu/drm/virtio/virtgpu_gem.c
@@ -117,7 +117,7 @@ int virtio_gpu_mode_dumb_mmap(struct drm_file *file_priv,
 int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
 			       struct drm_file *file)
 {
-	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
+	struct virtio_gpu_device *vgdev = obj->drm->dev_private;
 	struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
 	struct virtio_gpu_object_array *objs;
 
@@ -127,7 +127,7 @@ int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
 	/* the context might still be missing when the first ioctl is
 	 * DRM_IOCTL_MODE_CREATE_DUMB or DRM_IOCTL_PRIME_FD_TO_HANDLE
 	 */
-	virtio_gpu_create_context(obj->dev, file);
+	virtio_gpu_create_context(obj->drm, file);
 
 	objs = virtio_gpu_array_alloc(1);
 	if (!objs)
@@ -144,7 +144,7 @@ int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
 void virtio_gpu_gem_object_close(struct drm_gem_object *obj,
 				 struct drm_file *file)
 {
-	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
+	struct virtio_gpu_device *vgdev = obj->drm->dev_private;
 	struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
 	struct virtio_gpu_object_array *objs;
 
diff --git a/drivers/gpu/drm/virtio/virtgpu_object.c b/drivers/gpu/drm/virtio/virtgpu_object.c
index c7e74cf13022..ced5a9240456 100644
--- a/drivers/gpu/drm/virtio/virtgpu_object.c
+++ b/drivers/gpu/drm/virtio/virtgpu_object.c
@@ -63,7 +63,7 @@ static void virtio_gpu_resource_id_put(struct virtio_gpu_device *vgdev, uint32_t
 
 void virtio_gpu_cleanup_object(struct virtio_gpu_object *bo)
 {
-	struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private;
+	struct virtio_gpu_device *vgdev = bo->base.base.drm->dev_private;
 
 	virtio_gpu_resource_id_put(vgdev, bo->hw_res_handle);
 	if (virtio_gpu_is_shmem(bo)) {
@@ -86,7 +86,7 @@ void virtio_gpu_cleanup_object(struct virtio_gpu_object *bo)
 static void virtio_gpu_free_object(struct drm_gem_object *obj)
 {
 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
-	struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private;
+	struct virtio_gpu_device *vgdev = bo->base.base.drm->dev_private;
 
 	if (bo->created) {
 		virtio_gpu_cmd_unref_resource(vgdev, bo);
diff --git a/drivers/gpu/drm/virtio/virtgpu_prime.c b/drivers/gpu/drm/virtio/virtgpu_prime.c
index 44425f20d91a..29570a41fe7f 100644
--- a/drivers/gpu/drm/virtio/virtgpu_prime.c
+++ b/drivers/gpu/drm/virtio/virtgpu_prime.c
@@ -32,7 +32,7 @@ static int virtgpu_virtio_get_uuid(struct dma_buf *buf,
 {
 	struct drm_gem_object *obj = buf->priv;
 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
-	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
+	struct virtio_gpu_device *vgdev = obj->drm->dev_private;
 
 	wait_event(vgdev->resp_wq, bo->uuid_state != STATE_INITIALIZING);
 	if (bo->uuid_state != STATE_OK)
@@ -105,7 +105,7 @@ struct dma_buf *virtgpu_gem_prime_export(struct drm_gem_object *obj,
 					 int flags)
 {
 	struct dma_buf *buf;
-	struct drm_device *dev = obj->dev;
+	struct drm_device *dev = obj->drm;
 	struct virtio_gpu_device *vgdev = dev->dev_private;
 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
 	int ret = 0;
@@ -149,7 +149,7 @@ struct drm_gem_object *virtgpu_gem_prime_import(struct drm_device *dev,
 
 	if (buf->ops == &virtgpu_dmabuf_ops.ops) {
 		obj = buf->priv;
-		if (obj->dev == dev) {
+		if (obj->drm == dev) {
 			/*
 			 * Importing dmabuf exported from our own gem increases
 			 * refcount on gem itself instead of f_count of dmabuf.
diff --git a/drivers/gpu/drm/virtio/virtgpu_vram.c b/drivers/gpu/drm/virtio/virtgpu_vram.c
index 25df81c02783..41b46d2f5bfd 100644
--- a/drivers/gpu/drm/virtio/virtgpu_vram.c
+++ b/drivers/gpu/drm/virtio/virtgpu_vram.c
@@ -6,7 +6,7 @@
 static void virtio_gpu_vram_free(struct drm_gem_object *obj)
 {
 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
-	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
+	struct virtio_gpu_device *vgdev = obj->drm->dev_private;
 	struct virtio_gpu_object_vram *vram = to_virtio_gpu_vram(bo);
 	bool unmap;
 
@@ -33,7 +33,7 @@ static int virtio_gpu_vram_mmap(struct drm_gem_object *obj,
 				struct vm_area_struct *vma)
 {
 	int ret;
-	struct virtio_gpu_device *vgdev = obj->dev->dev_private;
+	struct virtio_gpu_device *vgdev = obj->drm->dev_private;
 	struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj);
 	struct virtio_gpu_object_vram *vram = to_virtio_gpu_vram(bo);
 	unsigned long vm_size = vma->vm_end - vma->vm_start;
@@ -70,7 +70,7 @@ struct sg_table *virtio_gpu_vram_map_dma_buf(struct virtio_gpu_object *bo,
 					     struct device *dev,
 					     enum dma_data_direction dir)
 {
-	struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private;
+	struct virtio_gpu_device *vgdev = bo->base.base.drm->dev_private;
 	struct virtio_gpu_object_vram *vram = to_virtio_gpu_vram(bo);
 	struct sg_table *sgt;
 	dma_addr_t addr;
@@ -143,7 +143,7 @@ static int virtio_gpu_vram_map(struct virtio_gpu_object *bo)
 	int ret;
 	uint64_t offset;
 	struct virtio_gpu_object_array *objs;
-	struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private;
+	struct virtio_gpu_device *vgdev = bo->base.base.drm->dev_private;
 	struct virtio_gpu_object_vram *vram = to_virtio_gpu_vram(bo);
 
 	if (!vgdev->has_host_visible)
diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_gem.c b/drivers/gpu/drm/vmwgfx/vmwgfx_gem.c
index c0da89e16e6f..2270c0f33b74 100644
--- a/drivers/gpu/drm/vmwgfx/vmwgfx_gem.c
+++ b/drivers/gpu/drm/vmwgfx/vmwgfx_gem.c
@@ -86,7 +86,8 @@ static struct sg_table *vmw_gem_object_get_sg_table(struct drm_gem_object *obj)
 	if (vmw_tt->vsgt.sgt)
 		return vmw_tt->vsgt.sgt;
 
-	return drm_prime_pages_to_sg(obj->dev, vmw_tt->dma_ttm.pages, vmw_tt->dma_ttm.num_pages);
+	return drm_prime_pages_to_sg(obj->drm, vmw_tt->dma_ttm.pages,
+				     vmw_tt->dma_ttm.num_pages);
 }
 
 static const struct vm_operations_struct vmw_vm_ops = {
diff --git a/drivers/gpu/drm/xen/xen_drm_front.c b/drivers/gpu/drm/xen/xen_drm_front.c
index 7e9431c50c5a..2220000f1b61 100644
--- a/drivers/gpu/drm/xen/xen_drm_front.c
+++ b/drivers/gpu/drm/xen/xen_drm_front.c
@@ -383,10 +383,10 @@ void xen_drm_front_on_frame_done(struct xen_drm_front_info *front_info,
 
 void xen_drm_front_gem_object_free(struct drm_gem_object *obj)
 {
-	struct xen_drm_front_drm_info *drm_info = obj->dev->dev_private;
+	struct xen_drm_front_drm_info *drm_info = obj->drm->dev_private;
 	int idx;
 
-	if (drm_dev_enter(obj->dev, &idx)) {
+	if (drm_dev_enter(obj->drm, &idx)) {
 		xen_drm_front_dbuf_destroy(drm_info->front_info,
 					   xen_drm_front_dbuf_to_cookie(obj));
 		drm_dev_exit(idx);
diff --git a/drivers/gpu/drm/xen/xen_drm_front_gem.c b/drivers/gpu/drm/xen/xen_drm_front_gem.c
index 3ad2b4cfd1f0..2054641d2ec1 100644
--- a/drivers/gpu/drm/xen/xen_drm_front_gem.c
+++ b/drivers/gpu/drm/xen/xen_drm_front_gem.c
@@ -235,7 +235,7 @@ struct sg_table *xen_drm_front_gem_get_sg_table(struct drm_gem_object *gem_obj)
 	if (!xen_obj->pages)
 		return ERR_PTR(-ENOMEM);
 
-	return drm_prime_pages_to_sg(gem_obj->dev,
+	return drm_prime_pages_to_sg(gem_obj->drm,
 				     xen_obj->pages, xen_obj->num_pages);
 }
 
diff --git a/include/drm/drm_gem.h b/include/drm/drm_gem.h
index bbc721870c13..7afda00286f6 100644
--- a/include/drm/drm_gem.h
+++ b/include/drm/drm_gem.h
@@ -285,9 +285,9 @@ struct drm_gem_object {
 	unsigned handle_count;
 
 	/**
-	 * @dev: DRM dev this object belongs to.
+	 * @drm: DRM dev this object belongs to.
 	 */
-	struct drm_device *dev;
+	struct drm_device *drm;
 
 	/**
 	 * @filp:
-- 
2.39.2




[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux