Re: [PATCH v2] dma-buf: Rename struct fence to dma_fence

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Oct 25, 2016 at 10:25:49AM +0100, Chris Wilson wrote:
> I plan to usurp the short name of struct fence for a core kernel struct,
> and so I need to rename the specialised fence/timeline for DMA
> operations to make room.
> 
> A consensus was reached in
> https://lists.freedesktop.org/archives/dri-devel/2016-July/113083.html
> that making clear this fence applies to DMA operations was a good thing.
> Since then the patch has grown a bit as usage increases, so hopefully it
> remains a good thing!
> 
> (v2...: rebase, rerun spatch)
> 
> coccinelle script:
> @@
> 
> @@
> - struct fence
> + struct dma_fence
> @@
> 
> @@
> - struct fence_ops
> + struct dma_fence_ops
> @@
> 
> @@
> - struct fence_cb
> + struct dma_fence_cb
> @@
> 
> @@
> - struct fence_array
> + struct dma_fence_array
> @@
> 
> @@
> - enum fence_flag_bits
> + enum dma_fence_flag_bits
> @@
> 
> @@
> (
> - fence_init
> + dma_fence_init
> |
> - fence_release
> + dma_fence_release
> |
> - fence_free
> + dma_fence_free
> |
> - fence_get
> + dma_fence_get
> |
> - fence_get_rcu
> + dma_fence_get_rcu
> |
> - fence_put
> + dma_fence_put
> |
> - fence_signal
> + dma_fence_signal
> |
> - fence_signal_locked
> + dma_fence_signal_locked
> |
> - fence_default_wait
> + dma_fence_default_wait
> |
> - fence_add_callback
> + dma_fence_add_callback
> |
> - fence_remove_callback
> + dma_fence_remove_callback
> |
> - fence_enable_sw_signaling
> + dma_fence_enable_sw_signaling
> |
> - fence_is_signaled_locked
> + dma_fence_is_signaled_locked
> |
> - fence_is_signaled
> + dma_fence_is_signaled
> |
> - fence_is_later
> + dma_fence_is_later
> |
> - fence_later
> + dma_fence_later
> |
> - fence_wait_timeout
> + dma_fence_wait_timeout
> |
> - fence_wait_any_timeout
> + dma_fence_wait_any_timeout
> |
> - fence_wait
> + dma_fence_wait
> |
> - fence_context_alloc
> + dma_fence_context_alloc
> |
> - fence_array_create
> + dma_fence_array_create
> |
> - to_fence_array
> + to_dma_fence_array
> |
> - fence_is_array
> + dma_fence_is_array
> |
> - trace_fence_emit
> + trace_dma_fence_emit
> |
> - FENCE_TRACE
> + DMA_FENCE_TRACE
> |
> - FENCE_WARN
> + DMA_FENCE_WARN
> |
> - FENCE_ERR
> + DMA_FENCE_ERR
> )
>  (
>  ...
>  )
> 
> Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx>
> Reviewed-by: Gustavo Padovan <gustavo.padovan@xxxxxxxxxxxxxxx>
> Acked-by: Sumit Semwal <sumit.semwal@xxxxxxxxxx>
> Acked-by: Christian König <christian.koenig@xxxxxxx>

Let's see how bad of a conflict-ride this is going to be. Applied to
drm-misc, thanks.
-Daniel

> ---
>  Documentation/sync_file.txt                        |   8 +-
>  drivers/base/Kconfig                               |   6 +-
>  drivers/dma-buf/Makefile                           |   2 +-
>  drivers/dma-buf/dma-buf.c                          |  28 +--
>  .../dma-buf/{fence-array.c => dma-fence-array.c}   |  91 ++++----
>  drivers/dma-buf/{fence.c => dma-fence.c}           | 195 ++++++++---------
>  drivers/dma-buf/reservation.c                      |  94 ++++----
>  drivers/dma-buf/seqno-fence.c                      |  18 +-
>  drivers/dma-buf/sw_sync.c                          |  48 ++---
>  drivers/dma-buf/sync_debug.c                       |  13 +-
>  drivers/dma-buf/sync_debug.h                       |   9 +-
>  drivers/dma-buf/sync_file.c                        |  63 +++---
>  drivers/gpu/drm/amd/amdgpu/amdgpu.h                |  54 ++---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c      |   8 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c             |  16 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c            |  22 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_device.c         |  14 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_display.c        |  16 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c          |  58 ++---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c             |   6 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_job.c            |  22 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_object.c         |  14 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_object.h         |   8 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c             |  24 +--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c           |  48 +++--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_test.c           |  12 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h          |   4 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c            |  10 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h            |   4 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c            |  26 +--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h            |   4 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c            |  26 +--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h            |   4 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c             |  80 +++----
>  drivers/gpu/drm/amd/amdgpu/cik_sdma.c              |   6 +-
>  drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c              |   6 +-
>  drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c              |   6 +-
>  drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c              |  12 +-
>  drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c             |   6 +-
>  drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c             |   6 +-
>  drivers/gpu/drm/amd/amdgpu/si_dma.c                |   6 +-
>  drivers/gpu/drm/amd/scheduler/gpu_sched_trace.h    |   4 +-
>  drivers/gpu/drm/amd/scheduler/gpu_scheduler.c      |  67 +++---
>  drivers/gpu/drm/amd/scheduler/gpu_scheduler.h      |  26 +--
>  drivers/gpu/drm/amd/scheduler/sched_fence.c        |  48 +++--
>  drivers/gpu/drm/drm_atomic.c                       |   2 +-
>  drivers/gpu/drm/drm_atomic_helper.c                |   8 +-
>  drivers/gpu/drm/drm_fops.c                         |   6 +-
>  drivers/gpu/drm/etnaviv/etnaviv_gem.c              |   6 +-
>  drivers/gpu/drm/etnaviv/etnaviv_gpu.c              |  50 ++---
>  drivers/gpu/drm/etnaviv/etnaviv_gpu.h              |   4 +-
>  drivers/gpu/drm/i915/i915_gem_request.c            |  32 +--
>  drivers/gpu/drm/i915/i915_gem_request.h            |  18 +-
>  drivers/gpu/drm/i915/i915_sw_fence.c               |  41 ++--
>  drivers/gpu/drm/i915/i915_sw_fence.h               |   8 +-
>  drivers/gpu/drm/i915/i915_trace.h                  |   2 +-
>  drivers/gpu/drm/i915/intel_breadcrumbs.c           |   4 +-
>  drivers/gpu/drm/i915/intel_engine_cs.c             |   2 +-
>  drivers/gpu/drm/msm/msm_drv.h                      |   2 +-
>  drivers/gpu/drm/msm/msm_fence.c                    |  34 +--
>  drivers/gpu/drm/msm/msm_fence.h                    |   2 +-
>  drivers/gpu/drm/msm/msm_gem.c                      |  14 +-
>  drivers/gpu/drm/msm/msm_gem.h                      |   2 +-
>  drivers/gpu/drm/msm/msm_gem_submit.c               |   8 +-
>  drivers/gpu/drm/msm/msm_gpu.c                      |   2 +-
>  drivers/gpu/drm/nouveau/nouveau_bo.c               |   6 +-
>  drivers/gpu/drm/nouveau/nouveau_fence.c            |  80 +++----
>  drivers/gpu/drm/nouveau/nouveau_fence.h            |   6 +-
>  drivers/gpu/drm/nouveau/nouveau_gem.c              |   2 +-
>  drivers/gpu/drm/nouveau/nv04_fence.c               |   2 +-
>  drivers/gpu/drm/nouveau/nv10_fence.c               |   2 +-
>  drivers/gpu/drm/nouveau/nv17_fence.c               |   2 +-
>  drivers/gpu/drm/nouveau/nv50_fence.c               |   2 +-
>  drivers/gpu/drm/nouveau/nv84_fence.c               |   2 +-
>  drivers/gpu/drm/qxl/qxl_drv.h                      |   4 +-
>  drivers/gpu/drm/qxl/qxl_release.c                  |  35 +--
>  drivers/gpu/drm/radeon/radeon.h                    |  10 +-
>  drivers/gpu/drm/radeon/radeon_device.c             |   2 +-
>  drivers/gpu/drm/radeon/radeon_display.c            |   8 +-
>  drivers/gpu/drm/radeon/radeon_fence.c              |  56 ++---
>  drivers/gpu/drm/radeon/radeon_sync.c               |   6 +-
>  drivers/gpu/drm/radeon/radeon_uvd.c                |   2 +-
>  drivers/gpu/drm/ttm/ttm_bo.c                       |  24 +--
>  drivers/gpu/drm/ttm/ttm_bo_util.c                  |  22 +-
>  drivers/gpu/drm/ttm/ttm_bo_vm.c                    |   8 +-
>  drivers/gpu/drm/ttm/ttm_execbuf_util.c             |   3 +-
>  drivers/gpu/drm/vgem/vgem_fence.c                  |  53 ++---
>  drivers/gpu/drm/virtio/virtgpu_drv.h               |   2 +-
>  drivers/gpu/drm/virtio/virtgpu_fence.c             |  26 +--
>  drivers/gpu/drm/virtio/virtgpu_ioctl.c             |  12 +-
>  drivers/gpu/drm/virtio/virtgpu_kms.c               |   2 +-
>  drivers/gpu/drm/virtio/virtgpu_plane.c             |   2 +-
>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.c              |  44 ++--
>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.h              |   8 +-
>  drivers/gpu/drm/vmwgfx/vmwgfx_resource.c           |   2 +-
>  include/drm/drmP.h                                 |   4 +-
>  include/drm/drm_crtc.h                             |   2 +-
>  include/drm/drm_plane.h                            |   2 +-
>  include/drm/ttm/ttm_bo_api.h                       |   2 +-
>  include/drm/ttm/ttm_bo_driver.h                    |   6 +-
>  include/drm/ttm/ttm_execbuf_util.h                 |   2 +-
>  include/linux/dma-buf.h                            |   4 +-
>  include/linux/dma-fence-array.h                    |  86 ++++++++
>  include/linux/{fence.h => dma-fence.h}             | 237 +++++++++++----------
>  include/linux/fence-array.h                        |  84 --------
>  include/linux/reservation.h                        |  28 +--
>  include/linux/seqno-fence.h                        |  20 +-
>  include/linux/sync_file.h                          |  14 +-
>  include/trace/events/{fence.h => dma_fence.h}      |  44 ++--
>  109 files changed, 1251 insertions(+), 1214 deletions(-)
>  rename drivers/dma-buf/{fence-array.c => dma-fence-array.c} (51%)
>  rename drivers/dma-buf/{fence.c => dma-fence.c} (65%)
>  create mode 100644 include/linux/dma-fence-array.h
>  rename include/linux/{fence.h => dma-fence.h} (59%)
>  delete mode 100644 include/linux/fence-array.h
>  rename include/trace/events/{fence.h => dma_fence.h} (72%)
> 
> diff --git a/Documentation/sync_file.txt b/Documentation/sync_file.txt
> index b63a68531afd..57effb2f1b91 100644
> --- a/Documentation/sync_file.txt
> +++ b/Documentation/sync_file.txt
> @@ -32,9 +32,9 @@ in-fences and out-fences
>  Sync files can go either to or from userspace. When a sync_file is sent from
>  the driver to userspace we call the fences it contains 'out-fences'. They are
>  related to a buffer that the driver is processing or is going to process, so
> -the driver creates an out-fence to be able to notify, through fence_signal(),
> -when it has finished using (or processing) that buffer. Out-fences are fences
> -that the driver creates.
> +the driver creates an out-fence to be able to notify, through
> +dma_fence_signal(), when it has finished using (or processing) that buffer.
> +Out-fences are fences that the driver creates.
>  
>  On the other hand if the driver receives fence(s) through a sync_file from
>  userspace we call these fence(s) 'in-fences'. Receiveing in-fences means that
> @@ -76,7 +76,7 @@ Interface:
>  
>  
>  The returned reference is owned by the caller and must be disposed of
> -afterwards using fence_put(). In case of error, a NULL is returned instead.
> +afterwards using dma_fence_put(). In case of error, a NULL is returned instead.
>  
>  References:
>  [1] struct sync_file in include/linux/sync_file.h
> diff --git a/drivers/base/Kconfig b/drivers/base/Kconfig
> index fdf44cac08e6..37bf25c6b4a6 100644
> --- a/drivers/base/Kconfig
> +++ b/drivers/base/Kconfig
> @@ -248,11 +248,11 @@ config DMA_SHARED_BUFFER
>  	  APIs extension; the file's descriptor can then be passed on to other
>  	  driver.
>  
> -config FENCE_TRACE
> -	bool "Enable verbose FENCE_TRACE messages"
> +config DMA_FENCE_TRACE
> +	bool "Enable verbose DMA_FENCE_TRACE messages"
>  	depends on DMA_SHARED_BUFFER
>  	help
> -	  Enable the FENCE_TRACE printks. This will add extra
> +	  Enable the DMA_FENCE_TRACE printks. This will add extra
>  	  spam to the console log, but will make it easier to diagnose
>  	  lockup related problems for dma-buffers shared across multiple
>  	  devices.
> diff --git a/drivers/dma-buf/Makefile b/drivers/dma-buf/Makefile
> index 210a10bfad2b..c33bf8863147 100644
> --- a/drivers/dma-buf/Makefile
> +++ b/drivers/dma-buf/Makefile
> @@ -1,3 +1,3 @@
> -obj-y := dma-buf.o fence.o reservation.o seqno-fence.o fence-array.o
> +obj-y := dma-buf.o dma-fence.o dma-fence-array.o reservation.o seqno-fence.o
>  obj-$(CONFIG_SYNC_FILE)		+= sync_file.o
>  obj-$(CONFIG_SW_SYNC)		+= sw_sync.o sync_debug.o
> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> index cf04d249a6a4..e72e64484131 100644
> --- a/drivers/dma-buf/dma-buf.c
> +++ b/drivers/dma-buf/dma-buf.c
> @@ -25,7 +25,7 @@
>  #include <linux/fs.h>
>  #include <linux/slab.h>
>  #include <linux/dma-buf.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/anon_inodes.h>
>  #include <linux/export.h>
>  #include <linux/debugfs.h>
> @@ -124,7 +124,7 @@ static loff_t dma_buf_llseek(struct file *file, loff_t offset, int whence)
>  	return base + offset;
>  }
>  
> -static void dma_buf_poll_cb(struct fence *fence, struct fence_cb *cb)
> +static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	struct dma_buf_poll_cb_t *dcb = (struct dma_buf_poll_cb_t *)cb;
>  	unsigned long flags;
> @@ -140,7 +140,7 @@ static unsigned int dma_buf_poll(struct file *file, poll_table *poll)
>  	struct dma_buf *dmabuf;
>  	struct reservation_object *resv;
>  	struct reservation_object_list *fobj;
> -	struct fence *fence_excl;
> +	struct dma_fence *fence_excl;
>  	unsigned long events;
>  	unsigned shared_count, seq;
>  
> @@ -187,20 +187,20 @@ static unsigned int dma_buf_poll(struct file *file, poll_table *poll)
>  		spin_unlock_irq(&dmabuf->poll.lock);
>  
>  		if (events & pevents) {
> -			if (!fence_get_rcu(fence_excl)) {
> +			if (!dma_fence_get_rcu(fence_excl)) {
>  				/* force a recheck */
>  				events &= ~pevents;
>  				dma_buf_poll_cb(NULL, &dcb->cb);
> -			} else if (!fence_add_callback(fence_excl, &dcb->cb,
> -						       dma_buf_poll_cb)) {
> +			} else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
> +							   dma_buf_poll_cb)) {
>  				events &= ~pevents;
> -				fence_put(fence_excl);
> +				dma_fence_put(fence_excl);
>  			} else {
>  				/*
>  				 * No callback queued, wake up any additional
>  				 * waiters.
>  				 */
> -				fence_put(fence_excl);
> +				dma_fence_put(fence_excl);
>  				dma_buf_poll_cb(NULL, &dcb->cb);
>  			}
>  		}
> @@ -222,9 +222,9 @@ static unsigned int dma_buf_poll(struct file *file, poll_table *poll)
>  			goto out;
>  
>  		for (i = 0; i < shared_count; ++i) {
> -			struct fence *fence = rcu_dereference(fobj->shared[i]);
> +			struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
>  
> -			if (!fence_get_rcu(fence)) {
> +			if (!dma_fence_get_rcu(fence)) {
>  				/*
>  				 * fence refcount dropped to zero, this means
>  				 * that fobj has been freed
> @@ -235,13 +235,13 @@ static unsigned int dma_buf_poll(struct file *file, poll_table *poll)
>  				dma_buf_poll_cb(NULL, &dcb->cb);
>  				break;
>  			}
> -			if (!fence_add_callback(fence, &dcb->cb,
> -						dma_buf_poll_cb)) {
> -				fence_put(fence);
> +			if (!dma_fence_add_callback(fence, &dcb->cb,
> +						    dma_buf_poll_cb)) {
> +				dma_fence_put(fence);
>  				events &= ~POLLOUT;
>  				break;
>  			}
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  		}
>  
>  		/* No callback queued, wake up any additional waiters. */
> diff --git a/drivers/dma-buf/fence-array.c b/drivers/dma-buf/dma-fence-array.c
> similarity index 51%
> rename from drivers/dma-buf/fence-array.c
> rename to drivers/dma-buf/dma-fence-array.c
> index f1989fcaf354..d9aa75dff089 100644
> --- a/drivers/dma-buf/fence-array.c
> +++ b/drivers/dma-buf/dma-fence-array.c
> @@ -1,5 +1,5 @@
>  /*
> - * fence-array: aggregate fences to be waited together
> + * dma-fence-array: aggregate fences to be waited together
>   *
>   * Copyright (C) 2016 Collabora Ltd
>   * Copyright (C) 2016 Advanced Micro Devices, Inc.
> @@ -19,35 +19,34 @@
>  
>  #include <linux/export.h>
>  #include <linux/slab.h>
> -#include <linux/fence-array.h>
> +#include <linux/dma-fence-array.h>
>  
> -static void fence_array_cb_func(struct fence *f, struct fence_cb *cb);
> -
> -static const char *fence_array_get_driver_name(struct fence *fence)
> +static const char *dma_fence_array_get_driver_name(struct dma_fence *fence)
>  {
> -	return "fence_array";
> +	return "dma_fence_array";
>  }
>  
> -static const char *fence_array_get_timeline_name(struct fence *fence)
> +static const char *dma_fence_array_get_timeline_name(struct dma_fence *fence)
>  {
>  	return "unbound";
>  }
>  
> -static void fence_array_cb_func(struct fence *f, struct fence_cb *cb)
> +static void dma_fence_array_cb_func(struct dma_fence *f,
> +				    struct dma_fence_cb *cb)
>  {
> -	struct fence_array_cb *array_cb =
> -		container_of(cb, struct fence_array_cb, cb);
> -	struct fence_array *array = array_cb->array;
> +	struct dma_fence_array_cb *array_cb =
> +		container_of(cb, struct dma_fence_array_cb, cb);
> +	struct dma_fence_array *array = array_cb->array;
>  
>  	if (atomic_dec_and_test(&array->num_pending))
> -		fence_signal(&array->base);
> -	fence_put(&array->base);
> +		dma_fence_signal(&array->base);
> +	dma_fence_put(&array->base);
>  }
>  
> -static bool fence_array_enable_signaling(struct fence *fence)
> +static bool dma_fence_array_enable_signaling(struct dma_fence *fence)
>  {
> -	struct fence_array *array = to_fence_array(fence);
> -	struct fence_array_cb *cb = (void *)(&array[1]);
> +	struct dma_fence_array *array = to_dma_fence_array(fence);
> +	struct dma_fence_array_cb *cb = (void *)(&array[1]);
>  	unsigned i;
>  
>  	for (i = 0; i < array->num_fences; ++i) {
> @@ -60,10 +59,10 @@ static bool fence_array_enable_signaling(struct fence *fence)
>  		 * until we signal the array as complete (but that is now
>  		 * insufficient).
>  		 */
> -		fence_get(&array->base);
> -		if (fence_add_callback(array->fences[i], &cb[i].cb,
> -				       fence_array_cb_func)) {
> -			fence_put(&array->base);
> +		dma_fence_get(&array->base);
> +		if (dma_fence_add_callback(array->fences[i], &cb[i].cb,
> +				       dma_fence_array_cb_func)) {
> +			dma_fence_put(&array->base);
>  			if (atomic_dec_and_test(&array->num_pending))
>  				return false;
>  		}
> @@ -72,69 +71,71 @@ static bool fence_array_enable_signaling(struct fence *fence)
>  	return true;
>  }
>  
> -static bool fence_array_signaled(struct fence *fence)
> +static bool dma_fence_array_signaled(struct dma_fence *fence)
>  {
> -	struct fence_array *array = to_fence_array(fence);
> +	struct dma_fence_array *array = to_dma_fence_array(fence);
>  
>  	return atomic_read(&array->num_pending) <= 0;
>  }
>  
> -static void fence_array_release(struct fence *fence)
> +static void dma_fence_array_release(struct dma_fence *fence)
>  {
> -	struct fence_array *array = to_fence_array(fence);
> +	struct dma_fence_array *array = to_dma_fence_array(fence);
>  	unsigned i;
>  
>  	for (i = 0; i < array->num_fences; ++i)
> -		fence_put(array->fences[i]);
> +		dma_fence_put(array->fences[i]);
>  
>  	kfree(array->fences);
> -	fence_free(fence);
> +	dma_fence_free(fence);
>  }
>  
> -const struct fence_ops fence_array_ops = {
> -	.get_driver_name = fence_array_get_driver_name,
> -	.get_timeline_name = fence_array_get_timeline_name,
> -	.enable_signaling = fence_array_enable_signaling,
> -	.signaled = fence_array_signaled,
> -	.wait = fence_default_wait,
> -	.release = fence_array_release,
> +const struct dma_fence_ops dma_fence_array_ops = {
> +	.get_driver_name = dma_fence_array_get_driver_name,
> +	.get_timeline_name = dma_fence_array_get_timeline_name,
> +	.enable_signaling = dma_fence_array_enable_signaling,
> +	.signaled = dma_fence_array_signaled,
> +	.wait = dma_fence_default_wait,
> +	.release = dma_fence_array_release,
>  };
> -EXPORT_SYMBOL(fence_array_ops);
> +EXPORT_SYMBOL(dma_fence_array_ops);
>  
>  /**
> - * fence_array_create - Create a custom fence array
> + * dma_fence_array_create - Create a custom fence array
>   * @num_fences:		[in]	number of fences to add in the array
>   * @fences:		[in]	array containing the fences
>   * @context:		[in]	fence context to use
>   * @seqno:		[in]	sequence number to use
>   * @signal_on_any:	[in]	signal on any fence in the array
>   *
> - * Allocate a fence_array object and initialize the base fence with fence_init().
> + * Allocate a dma_fence_array object and initialize the base fence with
> + * dma_fence_init().
>   * In case of error it returns NULL.
>   *
>   * The caller should allocate the fences array with num_fences size
>   * and fill it with the fences it wants to add to the object. Ownership of this
> - * array is taken and fence_put() is used on each fence on release.
> + * array is taken and dma_fence_put() is used on each fence on release.
>   *
>   * If @signal_on_any is true the fence array signals if any fence in the array
>   * signals, otherwise it signals when all fences in the array signal.
>   */
> -struct fence_array *fence_array_create(int num_fences, struct fence **fences,
> -				       u64 context, unsigned seqno,
> -				       bool signal_on_any)
> +struct dma_fence_array *dma_fence_array_create(int num_fences,
> +					       struct dma_fence **fences,
> +					       u64 context, unsigned seqno,
> +					       bool signal_on_any)
>  {
> -	struct fence_array *array;
> +	struct dma_fence_array *array;
>  	size_t size = sizeof(*array);
>  
>  	/* Allocate the callback structures behind the array. */
> -	size += num_fences * sizeof(struct fence_array_cb);
> +	size += num_fences * sizeof(struct dma_fence_array_cb);
>  	array = kzalloc(size, GFP_KERNEL);
>  	if (!array)
>  		return NULL;
>  
>  	spin_lock_init(&array->lock);
> -	fence_init(&array->base, &fence_array_ops, &array->lock,
> -		   context, seqno);
> +	dma_fence_init(&array->base, &dma_fence_array_ops, &array->lock,
> +		       context, seqno);
>  
>  	array->num_fences = num_fences;
>  	atomic_set(&array->num_pending, signal_on_any ? 1 : num_fences);
> @@ -142,4 +143,4 @@ struct fence_array *fence_array_create(int num_fences, struct fence **fences,
>  
>  	return array;
>  }
> -EXPORT_SYMBOL(fence_array_create);
> +EXPORT_SYMBOL(dma_fence_array_create);
> diff --git a/drivers/dma-buf/fence.c b/drivers/dma-buf/dma-fence.c
> similarity index 65%
> rename from drivers/dma-buf/fence.c
> rename to drivers/dma-buf/dma-fence.c
> index cc05dddc77a6..bf73730cf2bb 100644
> --- a/drivers/dma-buf/fence.c
> +++ b/drivers/dma-buf/dma-fence.c
> @@ -21,13 +21,13 @@
>  #include <linux/slab.h>
>  #include <linux/export.h>
>  #include <linux/atomic.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #define CREATE_TRACE_POINTS
> -#include <trace/events/fence.h>
> +#include <trace/events/dma_fence.h>
>  
> -EXPORT_TRACEPOINT_SYMBOL(fence_annotate_wait_on);
> -EXPORT_TRACEPOINT_SYMBOL(fence_emit);
> +EXPORT_TRACEPOINT_SYMBOL(dma_fence_annotate_wait_on);
> +EXPORT_TRACEPOINT_SYMBOL(dma_fence_emit);
>  
>  /*
>   * fence context counter: each execution context should have its own
> @@ -35,37 +35,37 @@ EXPORT_TRACEPOINT_SYMBOL(fence_emit);
>   * context or not. One device can have multiple separate contexts,
>   * and they're used if some engine can run independently of another.
>   */
> -static atomic64_t fence_context_counter = ATOMIC64_INIT(0);
> +static atomic64_t dma_fence_context_counter = ATOMIC64_INIT(0);
>  
>  /**
> - * fence_context_alloc - allocate an array of fence contexts
> + * dma_fence_context_alloc - allocate an array of fence contexts
>   * @num:	[in]	amount of contexts to allocate
>   *
>   * This function will return the first index of the number of fences allocated.
>   * The fence context is used for setting fence->context to a unique number.
>   */
> -u64 fence_context_alloc(unsigned num)
> +u64 dma_fence_context_alloc(unsigned num)
>  {
>  	BUG_ON(!num);
> -	return atomic64_add_return(num, &fence_context_counter) - num;
> +	return atomic64_add_return(num, &dma_fence_context_counter) - num;
>  }
> -EXPORT_SYMBOL(fence_context_alloc);
> +EXPORT_SYMBOL(dma_fence_context_alloc);
>  
>  /**
> - * fence_signal_locked - signal completion of a fence
> + * dma_fence_signal_locked - signal completion of a fence
>   * @fence: the fence to signal
>   *
>   * Signal completion for software callbacks on a fence, this will unblock
> - * fence_wait() calls and run all the callbacks added with
> - * fence_add_callback(). Can be called multiple times, but since a fence
> + * dma_fence_wait() calls and run all the callbacks added with
> + * dma_fence_add_callback(). Can be called multiple times, but since a fence
>   * can only go from unsignaled to signaled state, it will only be effective
>   * the first time.
>   *
> - * Unlike fence_signal, this function must be called with fence->lock held.
> + * Unlike dma_fence_signal, this function must be called with fence->lock held.
>   */
> -int fence_signal_locked(struct fence *fence)
> +int dma_fence_signal_locked(struct dma_fence *fence)
>  {
> -	struct fence_cb *cur, *tmp;
> +	struct dma_fence_cb *cur, *tmp;
>  	int ret = 0;
>  
>  	lockdep_assert_held(fence->lock);
> @@ -78,15 +78,15 @@ int fence_signal_locked(struct fence *fence)
>  		smp_mb__before_atomic();
>  	}
>  
> -	if (test_and_set_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
> +	if (test_and_set_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
>  		ret = -EINVAL;
>  
>  		/*
> -		 * we might have raced with the unlocked fence_signal,
> +		 * we might have raced with the unlocked dma_fence_signal,
>  		 * still run through all callbacks
>  		 */
>  	} else
> -		trace_fence_signaled(fence);
> +		trace_dma_fence_signaled(fence);
>  
>  	list_for_each_entry_safe(cur, tmp, &fence->cb_list, node) {
>  		list_del_init(&cur->node);
> @@ -94,19 +94,19 @@ int fence_signal_locked(struct fence *fence)
>  	}
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_signal_locked);
> +EXPORT_SYMBOL(dma_fence_signal_locked);
>  
>  /**
> - * fence_signal - signal completion of a fence
> + * dma_fence_signal - signal completion of a fence
>   * @fence: the fence to signal
>   *
>   * Signal completion for software callbacks on a fence, this will unblock
> - * fence_wait() calls and run all the callbacks added with
> - * fence_add_callback(). Can be called multiple times, but since a fence
> + * dma_fence_wait() calls and run all the callbacks added with
> + * dma_fence_add_callback(). Can be called multiple times, but since a fence
>   * can only go from unsignaled to signaled state, it will only be effective
>   * the first time.
>   */
> -int fence_signal(struct fence *fence)
> +int dma_fence_signal(struct dma_fence *fence)
>  {
>  	unsigned long flags;
>  
> @@ -118,13 +118,13 @@ int fence_signal(struct fence *fence)
>  		smp_mb__before_atomic();
>  	}
>  
> -	if (test_and_set_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_and_set_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		return -EINVAL;
>  
> -	trace_fence_signaled(fence);
> +	trace_dma_fence_signaled(fence);
>  
> -	if (test_bit(FENCE_FLAG_ENABLE_SIGNAL_BIT, &fence->flags)) {
> -		struct fence_cb *cur, *tmp;
> +	if (test_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT, &fence->flags)) {
> +		struct dma_fence_cb *cur, *tmp;
>  
>  		spin_lock_irqsave(fence->lock, flags);
>  		list_for_each_entry_safe(cur, tmp, &fence->cb_list, node) {
> @@ -135,10 +135,10 @@ int fence_signal(struct fence *fence)
>  	}
>  	return 0;
>  }
> -EXPORT_SYMBOL(fence_signal);
> +EXPORT_SYMBOL(dma_fence_signal);
>  
>  /**
> - * fence_wait_timeout - sleep until the fence gets signaled
> + * dma_fence_wait_timeout - sleep until the fence gets signaled
>   * or until timeout elapses
>   * @fence:	[in]	the fence to wait on
>   * @intr:	[in]	if true, do an interruptible wait
> @@ -154,7 +154,7 @@ EXPORT_SYMBOL(fence_signal);
>   * freed before return, resulting in undefined behavior.
>   */
>  signed long
> -fence_wait_timeout(struct fence *fence, bool intr, signed long timeout)
> +dma_fence_wait_timeout(struct dma_fence *fence, bool intr, signed long timeout)
>  {
>  	signed long ret;
>  
> @@ -162,70 +162,71 @@ fence_wait_timeout(struct fence *fence, bool intr, signed long timeout)
>  		return -EINVAL;
>  
>  	if (timeout == 0)
> -		return fence_is_signaled(fence);
> +		return dma_fence_is_signaled(fence);
>  
> -	trace_fence_wait_start(fence);
> +	trace_dma_fence_wait_start(fence);
>  	ret = fence->ops->wait(fence, intr, timeout);
> -	trace_fence_wait_end(fence);
> +	trace_dma_fence_wait_end(fence);
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_wait_timeout);
> +EXPORT_SYMBOL(dma_fence_wait_timeout);
>  
> -void fence_release(struct kref *kref)
> +void dma_fence_release(struct kref *kref)
>  {
> -	struct fence *fence =
> -			container_of(kref, struct fence, refcount);
> +	struct dma_fence *fence =
> +			container_of(kref, struct dma_fence, refcount);
>  
> -	trace_fence_destroy(fence);
> +	trace_dma_fence_destroy(fence);
>  
>  	BUG_ON(!list_empty(&fence->cb_list));
>  
>  	if (fence->ops->release)
>  		fence->ops->release(fence);
>  	else
> -		fence_free(fence);
> +		dma_fence_free(fence);
>  }
> -EXPORT_SYMBOL(fence_release);
> +EXPORT_SYMBOL(dma_fence_release);
>  
> -void fence_free(struct fence *fence)
> +void dma_fence_free(struct dma_fence *fence)
>  {
>  	kfree_rcu(fence, rcu);
>  }
> -EXPORT_SYMBOL(fence_free);
> +EXPORT_SYMBOL(dma_fence_free);
>  
>  /**
> - * fence_enable_sw_signaling - enable signaling on fence
> + * dma_fence_enable_sw_signaling - enable signaling on fence
>   * @fence:	[in]	the fence to enable
>   *
>   * this will request for sw signaling to be enabled, to make the fence
>   * complete as soon as possible
>   */
> -void fence_enable_sw_signaling(struct fence *fence)
> +void dma_fence_enable_sw_signaling(struct dma_fence *fence)
>  {
>  	unsigned long flags;
>  
> -	if (!test_and_set_bit(FENCE_FLAG_ENABLE_SIGNAL_BIT, &fence->flags) &&
> -	    !test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
> -		trace_fence_enable_signal(fence);
> +	if (!test_and_set_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
> +			      &fence->flags) &&
> +	    !test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
> +		trace_dma_fence_enable_signal(fence);
>  
>  		spin_lock_irqsave(fence->lock, flags);
>  
>  		if (!fence->ops->enable_signaling(fence))
> -			fence_signal_locked(fence);
> +			dma_fence_signal_locked(fence);
>  
>  		spin_unlock_irqrestore(fence->lock, flags);
>  	}
>  }
> -EXPORT_SYMBOL(fence_enable_sw_signaling);
> +EXPORT_SYMBOL(dma_fence_enable_sw_signaling);
>  
>  /**
> - * fence_add_callback - add a callback to be called when the fence
> + * dma_fence_add_callback - add a callback to be called when the fence
>   * is signaled
>   * @fence:	[in]	the fence to wait on
>   * @cb:		[in]	the callback to register
>   * @func:	[in]	the function to call
>   *
> - * cb will be initialized by fence_add_callback, no initialization
> + * cb will be initialized by dma_fence_add_callback, no initialization
>   * by the caller is required. Any number of callbacks can be registered
>   * to a fence, but a callback can only be registered to one fence at a time.
>   *
> @@ -234,15 +235,15 @@ EXPORT_SYMBOL(fence_enable_sw_signaling);
>   * *not* call the callback)
>   *
>   * Add a software callback to the fence. Same restrictions apply to
> - * refcount as it does to fence_wait, however the caller doesn't need to
> + * refcount as it does to dma_fence_wait, however the caller doesn't need to
>   * keep a refcount to fence afterwards: when software access is enabled,
>   * the creator of the fence is required to keep the fence alive until
> - * after it signals with fence_signal. The callback itself can be called
> + * after it signals with dma_fence_signal. The callback itself can be called
>   * from irq context.
>   *
>   */
> -int fence_add_callback(struct fence *fence, struct fence_cb *cb,
> -		       fence_func_t func)
> +int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb,
> +		       dma_fence_func_t func)
>  {
>  	unsigned long flags;
>  	int ret = 0;
> @@ -251,22 +252,23 @@ int fence_add_callback(struct fence *fence, struct fence_cb *cb,
>  	if (WARN_ON(!fence || !func))
>  		return -EINVAL;
>  
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
>  		INIT_LIST_HEAD(&cb->node);
>  		return -ENOENT;
>  	}
>  
>  	spin_lock_irqsave(fence->lock, flags);
>  
> -	was_set = test_and_set_bit(FENCE_FLAG_ENABLE_SIGNAL_BIT, &fence->flags);
> +	was_set = test_and_set_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
> +				   &fence->flags);
>  
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		ret = -ENOENT;
>  	else if (!was_set) {
> -		trace_fence_enable_signal(fence);
> +		trace_dma_fence_enable_signal(fence);
>  
>  		if (!fence->ops->enable_signaling(fence)) {
> -			fence_signal_locked(fence);
> +			dma_fence_signal_locked(fence);
>  			ret = -ENOENT;
>  		}
>  	}
> @@ -280,10 +282,10 @@ int fence_add_callback(struct fence *fence, struct fence_cb *cb,
>  
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_add_callback);
> +EXPORT_SYMBOL(dma_fence_add_callback);
>  
>  /**
> - * fence_remove_callback - remove a callback from the signaling list
> + * dma_fence_remove_callback - remove a callback from the signaling list
>   * @fence:	[in]	the fence to wait on
>   * @cb:		[in]	the callback to remove
>   *
> @@ -298,7 +300,7 @@ EXPORT_SYMBOL(fence_add_callback);
>   * with a reference held to the fence.
>   */
>  bool
> -fence_remove_callback(struct fence *fence, struct fence_cb *cb)
> +dma_fence_remove_callback(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	unsigned long flags;
>  	bool ret;
> @@ -313,15 +315,15 @@ fence_remove_callback(struct fence *fence, struct fence_cb *cb)
>  
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_remove_callback);
> +EXPORT_SYMBOL(dma_fence_remove_callback);
>  
>  struct default_wait_cb {
> -	struct fence_cb base;
> +	struct dma_fence_cb base;
>  	struct task_struct *task;
>  };
>  
>  static void
> -fence_default_wait_cb(struct fence *fence, struct fence_cb *cb)
> +dma_fence_default_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	struct default_wait_cb *wait =
>  		container_of(cb, struct default_wait_cb, base);
> @@ -330,7 +332,7 @@ fence_default_wait_cb(struct fence *fence, struct fence_cb *cb)
>  }
>  
>  /**
> - * fence_default_wait - default sleep until the fence gets signaled
> + * dma_fence_default_wait - default sleep until the fence gets signaled
>   * or until timeout elapses
>   * @fence:	[in]	the fence to wait on
>   * @intr:	[in]	if true, do an interruptible wait
> @@ -340,14 +342,14 @@ fence_default_wait_cb(struct fence *fence, struct fence_cb *cb)
>   * remaining timeout in jiffies on success.
>   */
>  signed long
> -fence_default_wait(struct fence *fence, bool intr, signed long timeout)
> +dma_fence_default_wait(struct dma_fence *fence, bool intr, signed long timeout)
>  {
>  	struct default_wait_cb cb;
>  	unsigned long flags;
>  	signed long ret = timeout;
>  	bool was_set;
>  
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		return timeout;
>  
>  	spin_lock_irqsave(fence->lock, flags);
> @@ -357,25 +359,26 @@ fence_default_wait(struct fence *fence, bool intr, signed long timeout)
>  		goto out;
>  	}
>  
> -	was_set = test_and_set_bit(FENCE_FLAG_ENABLE_SIGNAL_BIT, &fence->flags);
> +	was_set = test_and_set_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
> +				   &fence->flags);
>  
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		goto out;
>  
>  	if (!was_set) {
> -		trace_fence_enable_signal(fence);
> +		trace_dma_fence_enable_signal(fence);
>  
>  		if (!fence->ops->enable_signaling(fence)) {
> -			fence_signal_locked(fence);
> +			dma_fence_signal_locked(fence);
>  			goto out;
>  		}
>  	}
>  
> -	cb.base.func = fence_default_wait_cb;
> +	cb.base.func = dma_fence_default_wait_cb;
>  	cb.task = current;
>  	list_add(&cb.base.node, &fence->cb_list);
>  
> -	while (!test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags) && ret > 0) {
> +	while (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags) && ret > 0) {
>  		if (intr)
>  			__set_current_state(TASK_INTERRUPTIBLE);
>  		else
> @@ -397,23 +400,23 @@ fence_default_wait(struct fence *fence, bool intr, signed long timeout)
>  	spin_unlock_irqrestore(fence->lock, flags);
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_default_wait);
> +EXPORT_SYMBOL(dma_fence_default_wait);
>  
>  static bool
> -fence_test_signaled_any(struct fence **fences, uint32_t count)
> +dma_fence_test_signaled_any(struct dma_fence **fences, uint32_t count)
>  {
>  	int i;
>  
>  	for (i = 0; i < count; ++i) {
> -		struct fence *fence = fences[i];
> -		if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +		struct dma_fence *fence = fences[i];
> +		if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  			return true;
>  	}
>  	return false;
>  }
>  
>  /**
> - * fence_wait_any_timeout - sleep until any fence gets signaled
> + * dma_fence_wait_any_timeout - sleep until any fence gets signaled
>   * or until timeout elapses
>   * @fences:	[in]	array of fences to wait on
>   * @count:	[in]	number of fences to wait on
> @@ -429,7 +432,7 @@ fence_test_signaled_any(struct fence **fences, uint32_t count)
>   * fence might be freed before return, resulting in undefined behavior.
>   */
>  signed long
> -fence_wait_any_timeout(struct fence **fences, uint32_t count,
> +dma_fence_wait_any_timeout(struct dma_fence **fences, uint32_t count,
>  		       bool intr, signed long timeout)
>  {
>  	struct default_wait_cb *cb;
> @@ -441,7 +444,7 @@ fence_wait_any_timeout(struct fence **fences, uint32_t count,
>  
>  	if (timeout == 0) {
>  		for (i = 0; i < count; ++i)
> -			if (fence_is_signaled(fences[i]))
> +			if (dma_fence_is_signaled(fences[i]))
>  				return 1;
>  
>  		return 0;
> @@ -454,16 +457,16 @@ fence_wait_any_timeout(struct fence **fences, uint32_t count,
>  	}
>  
>  	for (i = 0; i < count; ++i) {
> -		struct fence *fence = fences[i];
> +		struct dma_fence *fence = fences[i];
>  
> -		if (fence->ops->wait != fence_default_wait) {
> +		if (fence->ops->wait != dma_fence_default_wait) {
>  			ret = -EINVAL;
>  			goto fence_rm_cb;
>  		}
>  
>  		cb[i].task = current;
> -		if (fence_add_callback(fence, &cb[i].base,
> -				       fence_default_wait_cb)) {
> +		if (dma_fence_add_callback(fence, &cb[i].base,
> +				       dma_fence_default_wait_cb)) {
>  			/* This fence is already signaled */
>  			goto fence_rm_cb;
>  		}
> @@ -475,7 +478,7 @@ fence_wait_any_timeout(struct fence **fences, uint32_t count,
>  		else
>  			set_current_state(TASK_UNINTERRUPTIBLE);
>  
> -		if (fence_test_signaled_any(fences, count))
> +		if (dma_fence_test_signaled_any(fences, count))
>  			break;
>  
>  		ret = schedule_timeout(ret);
> @@ -488,33 +491,33 @@ fence_wait_any_timeout(struct fence **fences, uint32_t count,
>  
>  fence_rm_cb:
>  	while (i-- > 0)
> -		fence_remove_callback(fences[i], &cb[i].base);
> +		dma_fence_remove_callback(fences[i], &cb[i].base);
>  
>  err_free_cb:
>  	kfree(cb);
>  
>  	return ret;
>  }
> -EXPORT_SYMBOL(fence_wait_any_timeout);
> +EXPORT_SYMBOL(dma_fence_wait_any_timeout);
>  
>  /**
> - * fence_init - Initialize a custom fence.
> + * dma_fence_init - Initialize a custom fence.
>   * @fence:	[in]	the fence to initialize
> - * @ops:	[in]	the fence_ops for operations on this fence
> + * @ops:	[in]	the dma_fence_ops for operations on this fence
>   * @lock:	[in]	the irqsafe spinlock to use for locking this fence
>   * @context:	[in]	the execution context this fence is run on
>   * @seqno:	[in]	a linear increasing sequence number for this context
>   *
>   * Initializes an allocated fence, the caller doesn't have to keep its
>   * refcount after committing with this fence, but it will need to hold a
> - * refcount again if fence_ops.enable_signaling gets called. This can
> + * refcount again if dma_fence_ops.enable_signaling gets called. This can
>   * be used for other implementing other types of fence.
>   *
>   * context and seqno are used for easy comparison between fences, allowing
> - * to check which fence is later by simply using fence_later.
> + * to check which fence is later by simply using dma_fence_later.
>   */
>  void
> -fence_init(struct fence *fence, const struct fence_ops *ops,
> +dma_fence_init(struct dma_fence *fence, const struct dma_fence_ops *ops,
>  	     spinlock_t *lock, u64 context, unsigned seqno)
>  {
>  	BUG_ON(!lock);
> @@ -529,6 +532,6 @@ fence_init(struct fence *fence, const struct fence_ops *ops,
>  	fence->seqno = seqno;
>  	fence->flags = 0UL;
>  
> -	trace_fence_init(fence);
> +	trace_dma_fence_init(fence);
>  }
> -EXPORT_SYMBOL(fence_init);
> +EXPORT_SYMBOL(dma_fence_init);
> diff --git a/drivers/dma-buf/reservation.c b/drivers/dma-buf/reservation.c
> index 82de59f7cbbd..7ed56f3edfb7 100644
> --- a/drivers/dma-buf/reservation.c
> +++ b/drivers/dma-buf/reservation.c
> @@ -102,17 +102,17 @@ EXPORT_SYMBOL(reservation_object_reserve_shared);
>  static void
>  reservation_object_add_shared_inplace(struct reservation_object *obj,
>  				      struct reservation_object_list *fobj,
> -				      struct fence *fence)
> +				      struct dma_fence *fence)
>  {
>  	u32 i;
>  
> -	fence_get(fence);
> +	dma_fence_get(fence);
>  
>  	preempt_disable();
>  	write_seqcount_begin(&obj->seq);
>  
>  	for (i = 0; i < fobj->shared_count; ++i) {
> -		struct fence *old_fence;
> +		struct dma_fence *old_fence;
>  
>  		old_fence = rcu_dereference_protected(fobj->shared[i],
>  						reservation_object_held(obj));
> @@ -123,7 +123,7 @@ reservation_object_add_shared_inplace(struct reservation_object *obj,
>  			write_seqcount_end(&obj->seq);
>  			preempt_enable();
>  
> -			fence_put(old_fence);
> +			dma_fence_put(old_fence);
>  			return;
>  		}
>  	}
> @@ -143,12 +143,12 @@ static void
>  reservation_object_add_shared_replace(struct reservation_object *obj,
>  				      struct reservation_object_list *old,
>  				      struct reservation_object_list *fobj,
> -				      struct fence *fence)
> +				      struct dma_fence *fence)
>  {
>  	unsigned i;
> -	struct fence *old_fence = NULL;
> +	struct dma_fence *old_fence = NULL;
>  
> -	fence_get(fence);
> +	dma_fence_get(fence);
>  
>  	if (!old) {
>  		RCU_INIT_POINTER(fobj->shared[0], fence);
> @@ -165,7 +165,7 @@ reservation_object_add_shared_replace(struct reservation_object *obj,
>  	fobj->shared_count = old->shared_count;
>  
>  	for (i = 0; i < old->shared_count; ++i) {
> -		struct fence *check;
> +		struct dma_fence *check;
>  
>  		check = rcu_dereference_protected(old->shared[i],
>  						reservation_object_held(obj));
> @@ -196,7 +196,7 @@ reservation_object_add_shared_replace(struct reservation_object *obj,
>  		kfree_rcu(old, rcu);
>  
>  	if (old_fence)
> -		fence_put(old_fence);
> +		dma_fence_put(old_fence);
>  }
>  
>  /**
> @@ -208,7 +208,7 @@ reservation_object_add_shared_replace(struct reservation_object *obj,
>   * reservation_object_reserve_shared() has been called.
>   */
>  void reservation_object_add_shared_fence(struct reservation_object *obj,
> -					 struct fence *fence)
> +					 struct dma_fence *fence)
>  {
>  	struct reservation_object_list *old, *fobj = obj->staged;
>  
> @@ -231,9 +231,9 @@ EXPORT_SYMBOL(reservation_object_add_shared_fence);
>   * Add a fence to the exclusive slot.  The obj->lock must be held.
>   */
>  void reservation_object_add_excl_fence(struct reservation_object *obj,
> -				       struct fence *fence)
> +				       struct dma_fence *fence)
>  {
> -	struct fence *old_fence = reservation_object_get_excl(obj);
> +	struct dma_fence *old_fence = reservation_object_get_excl(obj);
>  	struct reservation_object_list *old;
>  	u32 i = 0;
>  
> @@ -242,7 +242,7 @@ void reservation_object_add_excl_fence(struct reservation_object *obj,
>  		i = old->shared_count;
>  
>  	if (fence)
> -		fence_get(fence);
> +		dma_fence_get(fence);
>  
>  	preempt_disable();
>  	write_seqcount_begin(&obj->seq);
> @@ -255,11 +255,11 @@ void reservation_object_add_excl_fence(struct reservation_object *obj,
>  
>  	/* inplace update, no shared fences */
>  	while (i--)
> -		fence_put(rcu_dereference_protected(old->shared[i],
> +		dma_fence_put(rcu_dereference_protected(old->shared[i],
>  						reservation_object_held(obj)));
>  
>  	if (old_fence)
> -		fence_put(old_fence);
> +		dma_fence_put(old_fence);
>  }
>  EXPORT_SYMBOL(reservation_object_add_excl_fence);
>  
> @@ -276,12 +276,12 @@ EXPORT_SYMBOL(reservation_object_add_excl_fence);
>   * Zero or -errno
>   */
>  int reservation_object_get_fences_rcu(struct reservation_object *obj,
> -				      struct fence **pfence_excl,
> +				      struct dma_fence **pfence_excl,
>  				      unsigned *pshared_count,
> -				      struct fence ***pshared)
> +				      struct dma_fence ***pshared)
>  {
> -	struct fence **shared = NULL;
> -	struct fence *fence_excl;
> +	struct dma_fence **shared = NULL;
> +	struct dma_fence *fence_excl;
>  	unsigned int shared_count;
>  	int ret = 1;
>  
> @@ -296,12 +296,12 @@ int reservation_object_get_fences_rcu(struct reservation_object *obj,
>  		seq = read_seqcount_begin(&obj->seq);
>  
>  		fence_excl = rcu_dereference(obj->fence_excl);
> -		if (fence_excl && !fence_get_rcu(fence_excl))
> +		if (fence_excl && !dma_fence_get_rcu(fence_excl))
>  			goto unlock;
>  
>  		fobj = rcu_dereference(obj->fence);
>  		if (fobj) {
> -			struct fence **nshared;
> +			struct dma_fence **nshared;
>  			size_t sz = sizeof(*shared) * fobj->shared_max;
>  
>  			nshared = krealloc(shared, sz,
> @@ -322,15 +322,15 @@ int reservation_object_get_fences_rcu(struct reservation_object *obj,
>  
>  			for (i = 0; i < shared_count; ++i) {
>  				shared[i] = rcu_dereference(fobj->shared[i]);
> -				if (!fence_get_rcu(shared[i]))
> +				if (!dma_fence_get_rcu(shared[i]))
>  					break;
>  			}
>  		}
>  
>  		if (i != shared_count || read_seqcount_retry(&obj->seq, seq)) {
>  			while (i--)
> -				fence_put(shared[i]);
> -			fence_put(fence_excl);
> +				dma_fence_put(shared[i]);
> +			dma_fence_put(fence_excl);
>  			goto unlock;
>  		}
>  
> @@ -368,7 +368,7 @@ long reservation_object_wait_timeout_rcu(struct reservation_object *obj,
>  					 bool wait_all, bool intr,
>  					 unsigned long timeout)
>  {
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	unsigned seq, shared_count, i = 0;
>  	long ret = timeout;
>  
> @@ -389,16 +389,17 @@ long reservation_object_wait_timeout_rcu(struct reservation_object *obj,
>  			shared_count = fobj->shared_count;
>  
>  		for (i = 0; i < shared_count; ++i) {
> -			struct fence *lfence = rcu_dereference(fobj->shared[i]);
> +			struct dma_fence *lfence = rcu_dereference(fobj->shared[i]);
>  
> -			if (test_bit(FENCE_FLAG_SIGNALED_BIT, &lfence->flags))
> +			if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
> +				     &lfence->flags))
>  				continue;
>  
> -			if (!fence_get_rcu(lfence))
> +			if (!dma_fence_get_rcu(lfence))
>  				goto unlock_retry;
>  
> -			if (fence_is_signaled(lfence)) {
> -				fence_put(lfence);
> +			if (dma_fence_is_signaled(lfence)) {
> +				dma_fence_put(lfence);
>  				continue;
>  			}
>  
> @@ -408,15 +409,16 @@ long reservation_object_wait_timeout_rcu(struct reservation_object *obj,
>  	}
>  
>  	if (!shared_count) {
> -		struct fence *fence_excl = rcu_dereference(obj->fence_excl);
> +		struct dma_fence *fence_excl = rcu_dereference(obj->fence_excl);
>  
>  		if (fence_excl &&
> -		    !test_bit(FENCE_FLAG_SIGNALED_BIT, &fence_excl->flags)) {
> -			if (!fence_get_rcu(fence_excl))
> +		    !test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
> +			      &fence_excl->flags)) {
> +			if (!dma_fence_get_rcu(fence_excl))
>  				goto unlock_retry;
>  
> -			if (fence_is_signaled(fence_excl))
> -				fence_put(fence_excl);
> +			if (dma_fence_is_signaled(fence_excl))
> +				dma_fence_put(fence_excl);
>  			else
>  				fence = fence_excl;
>  		}
> @@ -425,12 +427,12 @@ long reservation_object_wait_timeout_rcu(struct reservation_object *obj,
>  	rcu_read_unlock();
>  	if (fence) {
>  		if (read_seqcount_retry(&obj->seq, seq)) {
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  			goto retry;
>  		}
>  
> -		ret = fence_wait_timeout(fence, intr, ret);
> -		fence_put(fence);
> +		ret = dma_fence_wait_timeout(fence, intr, ret);
> +		dma_fence_put(fence);
>  		if (ret > 0 && wait_all && (i + 1 < shared_count))
>  			goto retry;
>  	}
> @@ -444,18 +446,18 @@ EXPORT_SYMBOL_GPL(reservation_object_wait_timeout_rcu);
>  
>  
>  static inline int
> -reservation_object_test_signaled_single(struct fence *passed_fence)
> +reservation_object_test_signaled_single(struct dma_fence *passed_fence)
>  {
> -	struct fence *fence, *lfence = passed_fence;
> +	struct dma_fence *fence, *lfence = passed_fence;
>  	int ret = 1;
>  
> -	if (!test_bit(FENCE_FLAG_SIGNALED_BIT, &lfence->flags)) {
> -		fence = fence_get_rcu(lfence);
> +	if (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &lfence->flags)) {
> +		fence = dma_fence_get_rcu(lfence);
>  		if (!fence)
>  			return -1;
>  
> -		ret = !!fence_is_signaled(fence);
> -		fence_put(fence);
> +		ret = !!dma_fence_is_signaled(fence);
> +		dma_fence_put(fence);
>  	}
>  	return ret;
>  }
> @@ -492,7 +494,7 @@ bool reservation_object_test_signaled_rcu(struct reservation_object *obj,
>  			shared_count = fobj->shared_count;
>  
>  		for (i = 0; i < shared_count; ++i) {
> -			struct fence *fence = rcu_dereference(fobj->shared[i]);
> +			struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
>  
>  			ret = reservation_object_test_signaled_single(fence);
>  			if (ret < 0)
> @@ -506,7 +508,7 @@ bool reservation_object_test_signaled_rcu(struct reservation_object *obj,
>  	}
>  
>  	if (!shared_count) {
> -		struct fence *fence_excl = rcu_dereference(obj->fence_excl);
> +		struct dma_fence *fence_excl = rcu_dereference(obj->fence_excl);
>  
>  		if (fence_excl) {
>  			ret = reservation_object_test_signaled_single(
> diff --git a/drivers/dma-buf/seqno-fence.c b/drivers/dma-buf/seqno-fence.c
> index 71127f8f1626..f47112a64763 100644
> --- a/drivers/dma-buf/seqno-fence.c
> +++ b/drivers/dma-buf/seqno-fence.c
> @@ -21,35 +21,35 @@
>  #include <linux/export.h>
>  #include <linux/seqno-fence.h>
>  
> -static const char *seqno_fence_get_driver_name(struct fence *fence)
> +static const char *seqno_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	struct seqno_fence *seqno_fence = to_seqno_fence(fence);
>  
>  	return seqno_fence->ops->get_driver_name(fence);
>  }
>  
> -static const char *seqno_fence_get_timeline_name(struct fence *fence)
> +static const char *seqno_fence_get_timeline_name(struct dma_fence *fence)
>  {
>  	struct seqno_fence *seqno_fence = to_seqno_fence(fence);
>  
>  	return seqno_fence->ops->get_timeline_name(fence);
>  }
>  
> -static bool seqno_enable_signaling(struct fence *fence)
> +static bool seqno_enable_signaling(struct dma_fence *fence)
>  {
>  	struct seqno_fence *seqno_fence = to_seqno_fence(fence);
>  
>  	return seqno_fence->ops->enable_signaling(fence);
>  }
>  
> -static bool seqno_signaled(struct fence *fence)
> +static bool seqno_signaled(struct dma_fence *fence)
>  {
>  	struct seqno_fence *seqno_fence = to_seqno_fence(fence);
>  
>  	return seqno_fence->ops->signaled && seqno_fence->ops->signaled(fence);
>  }
>  
> -static void seqno_release(struct fence *fence)
> +static void seqno_release(struct dma_fence *fence)
>  {
>  	struct seqno_fence *f = to_seqno_fence(fence);
>  
> @@ -57,18 +57,18 @@ static void seqno_release(struct fence *fence)
>  	if (f->ops->release)
>  		f->ops->release(fence);
>  	else
> -		fence_free(&f->base);
> +		dma_fence_free(&f->base);
>  }
>  
> -static signed long seqno_wait(struct fence *fence, bool intr,
> -				signed long timeout)
> +static signed long seqno_wait(struct dma_fence *fence, bool intr,
> +			      signed long timeout)
>  {
>  	struct seqno_fence *f = to_seqno_fence(fence);
>  
>  	return f->ops->wait(fence, intr, timeout);
>  }
>  
> -const struct fence_ops seqno_fence_ops = {
> +const struct dma_fence_ops seqno_fence_ops = {
>  	.get_driver_name = seqno_fence_get_driver_name,
>  	.get_timeline_name = seqno_fence_get_timeline_name,
>  	.enable_signaling = seqno_enable_signaling,
> diff --git a/drivers/dma-buf/sw_sync.c b/drivers/dma-buf/sw_sync.c
> index 62e8e6dc7953..82e0ca4dd0c1 100644
> --- a/drivers/dma-buf/sw_sync.c
> +++ b/drivers/dma-buf/sw_sync.c
> @@ -68,9 +68,9 @@ struct sw_sync_create_fence_data {
>  
>  #define SW_SYNC_IOC_INC			_IOW(SW_SYNC_IOC_MAGIC, 1, __u32)
>  
> -static const struct fence_ops timeline_fence_ops;
> +static const struct dma_fence_ops timeline_fence_ops;
>  
> -static inline struct sync_pt *fence_to_sync_pt(struct fence *fence)
> +static inline struct sync_pt *dma_fence_to_sync_pt(struct dma_fence *fence)
>  {
>  	if (fence->ops != &timeline_fence_ops)
>  		return NULL;
> @@ -93,7 +93,7 @@ struct sync_timeline *sync_timeline_create(const char *name)
>  		return NULL;
>  
>  	kref_init(&obj->kref);
> -	obj->context = fence_context_alloc(1);
> +	obj->context = dma_fence_context_alloc(1);
>  	strlcpy(obj->name, name, sizeof(obj->name));
>  
>  	INIT_LIST_HEAD(&obj->child_list_head);
> @@ -146,7 +146,7 @@ static void sync_timeline_signal(struct sync_timeline *obj, unsigned int inc)
>  
>  	list_for_each_entry_safe(pt, next, &obj->active_list_head,
>  				 active_list) {
> -		if (fence_is_signaled_locked(&pt->base))
> +		if (dma_fence_is_signaled_locked(&pt->base))
>  			list_del_init(&pt->active_list);
>  	}
>  
> @@ -179,30 +179,30 @@ static struct sync_pt *sync_pt_create(struct sync_timeline *obj, int size,
>  
>  	spin_lock_irqsave(&obj->child_list_lock, flags);
>  	sync_timeline_get(obj);
> -	fence_init(&pt->base, &timeline_fence_ops, &obj->child_list_lock,
> -		   obj->context, value);
> +	dma_fence_init(&pt->base, &timeline_fence_ops, &obj->child_list_lock,
> +		       obj->context, value);
>  	list_add_tail(&pt->child_list, &obj->child_list_head);
>  	INIT_LIST_HEAD(&pt->active_list);
>  	spin_unlock_irqrestore(&obj->child_list_lock, flags);
>  	return pt;
>  }
>  
> -static const char *timeline_fence_get_driver_name(struct fence *fence)
> +static const char *timeline_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "sw_sync";
>  }
>  
> -static const char *timeline_fence_get_timeline_name(struct fence *fence)
> +static const char *timeline_fence_get_timeline_name(struct dma_fence *fence)
>  {
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  
>  	return parent->name;
>  }
>  
> -static void timeline_fence_release(struct fence *fence)
> +static void timeline_fence_release(struct dma_fence *fence)
>  {
> -	struct sync_pt *pt = fence_to_sync_pt(fence);
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_pt *pt = dma_fence_to_sync_pt(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  	unsigned long flags;
>  
>  	spin_lock_irqsave(fence->lock, flags);
> @@ -212,20 +212,20 @@ static void timeline_fence_release(struct fence *fence)
>  	spin_unlock_irqrestore(fence->lock, flags);
>  
>  	sync_timeline_put(parent);
> -	fence_free(fence);
> +	dma_fence_free(fence);
>  }
>  
> -static bool timeline_fence_signaled(struct fence *fence)
> +static bool timeline_fence_signaled(struct dma_fence *fence)
>  {
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  
>  	return (fence->seqno > parent->value) ? false : true;
>  }
>  
> -static bool timeline_fence_enable_signaling(struct fence *fence)
> +static bool timeline_fence_enable_signaling(struct dma_fence *fence)
>  {
> -	struct sync_pt *pt = fence_to_sync_pt(fence);
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_pt *pt = dma_fence_to_sync_pt(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  
>  	if (timeline_fence_signaled(fence))
>  		return false;
> @@ -234,26 +234,26 @@ static bool timeline_fence_enable_signaling(struct fence *fence)
>  	return true;
>  }
>  
> -static void timeline_fence_value_str(struct fence *fence,
> +static void timeline_fence_value_str(struct dma_fence *fence,
>  				    char *str, int size)
>  {
>  	snprintf(str, size, "%d", fence->seqno);
>  }
>  
> -static void timeline_fence_timeline_value_str(struct fence *fence,
> +static void timeline_fence_timeline_value_str(struct dma_fence *fence,
>  					     char *str, int size)
>  {
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  
>  	snprintf(str, size, "%d", parent->value);
>  }
>  
> -static const struct fence_ops timeline_fence_ops = {
> +static const struct dma_fence_ops timeline_fence_ops = {
>  	.get_driver_name = timeline_fence_get_driver_name,
>  	.get_timeline_name = timeline_fence_get_timeline_name,
>  	.enable_signaling = timeline_fence_enable_signaling,
>  	.signaled = timeline_fence_signaled,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = timeline_fence_release,
>  	.fence_value_str = timeline_fence_value_str,
>  	.timeline_value_str = timeline_fence_timeline_value_str,
> @@ -317,7 +317,7 @@ static long sw_sync_ioctl_create_fence(struct sync_timeline *obj,
>  
>  	sync_file = sync_file_create(&pt->base);
>  	if (!sync_file) {
> -		fence_put(&pt->base);
> +		dma_fence_put(&pt->base);
>  		err = -ENOMEM;
>  		goto err;
>  	}
> diff --git a/drivers/dma-buf/sync_debug.c b/drivers/dma-buf/sync_debug.c
> index 2dd4c3db6caa..48b20e34fb6d 100644
> --- a/drivers/dma-buf/sync_debug.c
> +++ b/drivers/dma-buf/sync_debug.c
> @@ -71,12 +71,13 @@ static const char *sync_status_str(int status)
>  	return "error";
>  }
>  
> -static void sync_print_fence(struct seq_file *s, struct fence *fence, bool show)
> +static void sync_print_fence(struct seq_file *s,
> +			     struct dma_fence *fence, bool show)
>  {
>  	int status = 1;
> -	struct sync_timeline *parent = fence_parent(fence);
> +	struct sync_timeline *parent = dma_fence_parent(fence);
>  
> -	if (fence_is_signaled_locked(fence))
> +	if (dma_fence_is_signaled_locked(fence))
>  		status = fence->status;
>  
>  	seq_printf(s, "  %s%sfence %s",
> @@ -135,10 +136,10 @@ static void sync_print_sync_file(struct seq_file *s,
>  	int i;
>  
>  	seq_printf(s, "[%p] %s: %s\n", sync_file, sync_file->name,
> -		   sync_status_str(!fence_is_signaled(sync_file->fence)));
> +		   sync_status_str(!dma_fence_is_signaled(sync_file->fence)));
>  
> -	if (fence_is_array(sync_file->fence)) {
> -		struct fence_array *array = to_fence_array(sync_file->fence);
> +	if (dma_fence_is_array(sync_file->fence)) {
> +		struct dma_fence_array *array = to_dma_fence_array(sync_file->fence);
>  
>  		for (i = 0; i < array->num_fences; ++i)
>  			sync_print_fence(s, array->fences[i], true);
> diff --git a/drivers/dma-buf/sync_debug.h b/drivers/dma-buf/sync_debug.h
> index d269aa6783aa..26fe8b9907b3 100644
> --- a/drivers/dma-buf/sync_debug.h
> +++ b/drivers/dma-buf/sync_debug.h
> @@ -15,7 +15,7 @@
>  
>  #include <linux/list.h>
>  #include <linux/spinlock.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include <linux/sync_file.h>
>  #include <uapi/linux/sync_file.h>
> @@ -45,10 +45,9 @@ struct sync_timeline {
>  	struct list_head	sync_timeline_list;
>  };
>  
> -static inline struct sync_timeline *fence_parent(struct fence *fence)
> +static inline struct sync_timeline *dma_fence_parent(struct dma_fence *fence)
>  {
> -	return container_of(fence->lock, struct sync_timeline,
> -			    child_list_lock);
> +	return container_of(fence->lock, struct sync_timeline, child_list_lock);
>  }
>  
>  /**
> @@ -58,7 +57,7 @@ static inline struct sync_timeline *fence_parent(struct fence *fence)
>   * @active_list: sync timeline active child's list
>   */
>  struct sync_pt {
> -	struct fence base;
> +	struct dma_fence base;
>  	struct list_head child_list;
>  	struct list_head active_list;
>  };
> diff --git a/drivers/dma-buf/sync_file.c b/drivers/dma-buf/sync_file.c
> index 235f8ac113cc..69d8ef98d34c 100644
> --- a/drivers/dma-buf/sync_file.c
> +++ b/drivers/dma-buf/sync_file.c
> @@ -54,7 +54,7 @@ static struct sync_file *sync_file_alloc(void)
>  	return NULL;
>  }
>  
> -static void fence_check_cb_func(struct fence *f, struct fence_cb *cb)
> +static void fence_check_cb_func(struct dma_fence *f, struct dma_fence_cb *cb)
>  {
>  	struct sync_file *sync_file;
>  
> @@ -71,7 +71,7 @@ static void fence_check_cb_func(struct fence *f, struct fence_cb *cb)
>   * takes ownership of @fence. The sync_file can be released with
>   * fput(sync_file->file). Returns the sync_file or NULL in case of error.
>   */
> -struct sync_file *sync_file_create(struct fence *fence)
> +struct sync_file *sync_file_create(struct dma_fence *fence)
>  {
>  	struct sync_file *sync_file;
>  
> @@ -79,7 +79,7 @@ struct sync_file *sync_file_create(struct fence *fence)
>  	if (!sync_file)
>  		return NULL;
>  
> -	sync_file->fence = fence_get(fence);
> +	sync_file->fence = dma_fence_get(fence);
>  
>  	snprintf(sync_file->name, sizeof(sync_file->name), "%s-%s%llu-%d",
>  		 fence->ops->get_driver_name(fence),
> @@ -121,16 +121,16 @@ static struct sync_file *sync_file_fdget(int fd)
>   * Ensures @fd references a valid sync_file and returns a fence that
>   * represents all fence in the sync_file. On error NULL is returned.
>   */
> -struct fence *sync_file_get_fence(int fd)
> +struct dma_fence *sync_file_get_fence(int fd)
>  {
>  	struct sync_file *sync_file;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  
>  	sync_file = sync_file_fdget(fd);
>  	if (!sync_file)
>  		return NULL;
>  
> -	fence = fence_get(sync_file->fence);
> +	fence = dma_fence_get(sync_file->fence);
>  	fput(sync_file->file);
>  
>  	return fence;
> @@ -138,22 +138,23 @@ struct fence *sync_file_get_fence(int fd)
>  EXPORT_SYMBOL(sync_file_get_fence);
>  
>  static int sync_file_set_fence(struct sync_file *sync_file,
> -			       struct fence **fences, int num_fences)
> +			       struct dma_fence **fences, int num_fences)
>  {
> -	struct fence_array *array;
> +	struct dma_fence_array *array;
>  
>  	/*
>  	 * The reference for the fences in the new sync_file and held
>  	 * in add_fence() during the merge procedure, so for num_fences == 1
>  	 * we already own a new reference to the fence. For num_fence > 1
> -	 * we own the reference of the fence_array creation.
> +	 * we own the reference of the dma_fence_array creation.
>  	 */
>  	if (num_fences == 1) {
>  		sync_file->fence = fences[0];
>  		kfree(fences);
>  	} else {
> -		array = fence_array_create(num_fences, fences,
> -					   fence_context_alloc(1), 1, false);
> +		array = dma_fence_array_create(num_fences, fences,
> +					       dma_fence_context_alloc(1),
> +					       1, false);
>  		if (!array)
>  			return -ENOMEM;
>  
> @@ -163,10 +164,11 @@ static int sync_file_set_fence(struct sync_file *sync_file,
>  	return 0;
>  }
>  
> -static struct fence **get_fences(struct sync_file *sync_file, int *num_fences)
> +static struct dma_fence **get_fences(struct sync_file *sync_file,
> +				     int *num_fences)
>  {
> -	if (fence_is_array(sync_file->fence)) {
> -		struct fence_array *array = to_fence_array(sync_file->fence);
> +	if (dma_fence_is_array(sync_file->fence)) {
> +		struct dma_fence_array *array = to_dma_fence_array(sync_file->fence);
>  
>  		*num_fences = array->num_fences;
>  		return array->fences;
> @@ -176,12 +178,13 @@ static struct fence **get_fences(struct sync_file *sync_file, int *num_fences)
>  	return &sync_file->fence;
>  }
>  
> -static void add_fence(struct fence **fences, int *i, struct fence *fence)
> +static void add_fence(struct dma_fence **fences,
> +		      int *i, struct dma_fence *fence)
>  {
>  	fences[*i] = fence;
>  
> -	if (!fence_is_signaled(fence)) {
> -		fence_get(fence);
> +	if (!dma_fence_is_signaled(fence)) {
> +		dma_fence_get(fence);
>  		(*i)++;
>  	}
>  }
> @@ -200,7 +203,7 @@ static struct sync_file *sync_file_merge(const char *name, struct sync_file *a,
>  					 struct sync_file *b)
>  {
>  	struct sync_file *sync_file;
> -	struct fence **fences, **nfences, **a_fences, **b_fences;
> +	struct dma_fence **fences, **nfences, **a_fences, **b_fences;
>  	int i, i_a, i_b, num_fences, a_num_fences, b_num_fences;
>  
>  	sync_file = sync_file_alloc();
> @@ -226,8 +229,8 @@ static struct sync_file *sync_file_merge(const char *name, struct sync_file *a,
>  	 * and sync_file_create, this is a reasonable assumption.
>  	 */
>  	for (i = i_a = i_b = 0; i_a < a_num_fences && i_b < b_num_fences; ) {
> -		struct fence *pt_a = a_fences[i_a];
> -		struct fence *pt_b = b_fences[i_b];
> +		struct dma_fence *pt_a = a_fences[i_a];
> +		struct dma_fence *pt_b = b_fences[i_b];
>  
>  		if (pt_a->context < pt_b->context) {
>  			add_fence(fences, &i, pt_a);
> @@ -255,7 +258,7 @@ static struct sync_file *sync_file_merge(const char *name, struct sync_file *a,
>  		add_fence(fences, &i, b_fences[i_b]);
>  
>  	if (i == 0)
> -		fences[i++] = fence_get(a_fences[0]);
> +		fences[i++] = dma_fence_get(a_fences[0]);
>  
>  	if (num_fences > i) {
>  		nfences = krealloc(fences, i * sizeof(*fences),
> @@ -286,8 +289,8 @@ static void sync_file_free(struct kref *kref)
>  						     kref);
>  
>  	if (test_bit(POLL_ENABLED, &sync_file->fence->flags))
> -		fence_remove_callback(sync_file->fence, &sync_file->cb);
> -	fence_put(sync_file->fence);
> +		dma_fence_remove_callback(sync_file->fence, &sync_file->cb);
> +	dma_fence_put(sync_file->fence);
>  	kfree(sync_file);
>  }
>  
> @@ -307,12 +310,12 @@ static unsigned int sync_file_poll(struct file *file, poll_table *wait)
>  
>  	if (!poll_does_not_wait(wait) &&
>  	    !test_and_set_bit(POLL_ENABLED, &sync_file->fence->flags)) {
> -		if (fence_add_callback(sync_file->fence, &sync_file->cb,
> -				       fence_check_cb_func) < 0)
> +		if (dma_fence_add_callback(sync_file->fence, &sync_file->cb,
> +					   fence_check_cb_func) < 0)
>  			wake_up_all(&sync_file->wq);
>  	}
>  
> -	return fence_is_signaled(sync_file->fence) ? POLLIN : 0;
> +	return dma_fence_is_signaled(sync_file->fence) ? POLLIN : 0;
>  }
>  
>  static long sync_file_ioctl_merge(struct sync_file *sync_file,
> @@ -370,14 +373,14 @@ static long sync_file_ioctl_merge(struct sync_file *sync_file,
>  	return err;
>  }
>  
> -static void sync_fill_fence_info(struct fence *fence,
> +static void sync_fill_fence_info(struct dma_fence *fence,
>  				 struct sync_fence_info *info)
>  {
>  	strlcpy(info->obj_name, fence->ops->get_timeline_name(fence),
>  		sizeof(info->obj_name));
>  	strlcpy(info->driver_name, fence->ops->get_driver_name(fence),
>  		sizeof(info->driver_name));
> -	if (fence_is_signaled(fence))
> +	if (dma_fence_is_signaled(fence))
>  		info->status = fence->status >= 0 ? 1 : fence->status;
>  	else
>  		info->status = 0;
> @@ -389,7 +392,7 @@ static long sync_file_ioctl_fence_info(struct sync_file *sync_file,
>  {
>  	struct sync_file_info info;
>  	struct sync_fence_info *fence_info = NULL;
> -	struct fence **fences;
> +	struct dma_fence **fences;
>  	__u32 size;
>  	int num_fences, ret, i;
>  
> @@ -429,7 +432,7 @@ static long sync_file_ioctl_fence_info(struct sync_file *sync_file,
>  
>  no_fences:
>  	strlcpy(info.name, sync_file->name, sizeof(info.name));
> -	info.status = fence_is_signaled(sync_file->fence);
> +	info.status = dma_fence_is_signaled(sync_file->fence);
>  	info.num_fences = num_fences;
>  
>  	if (copy_to_user((void __user *)arg, &info, sizeof(info)))
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
> index 039b57e4644c..283d05927d15 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
> @@ -34,7 +34,7 @@
>  #include <linux/kref.h>
>  #include <linux/interval_tree.h>
>  #include <linux/hashtable.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include <ttm/ttm_bo_api.h>
>  #include <ttm/ttm_bo_driver.h>
> @@ -378,7 +378,7 @@ struct amdgpu_fence_driver {
>  	struct timer_list		fallback_timer;
>  	unsigned			num_fences_mask;
>  	spinlock_t			lock;
> -	struct fence			**fences;
> +	struct dma_fence		**fences;
>  };
>  
>  /* some special values for the owner field */
> @@ -399,7 +399,7 @@ int amdgpu_fence_driver_start_ring(struct amdgpu_ring *ring,
>  				   unsigned irq_type);
>  void amdgpu_fence_driver_suspend(struct amdgpu_device *adev);
>  void amdgpu_fence_driver_resume(struct amdgpu_device *adev);
> -int amdgpu_fence_emit(struct amdgpu_ring *ring, struct fence **fence);
> +int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **fence);
>  void amdgpu_fence_process(struct amdgpu_ring *ring);
>  int amdgpu_fence_wait_empty(struct amdgpu_ring *ring);
>  unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring);
> @@ -427,7 +427,7 @@ struct amdgpu_bo_va_mapping {
>  struct amdgpu_bo_va {
>  	/* protected by bo being reserved */
>  	struct list_head		bo_list;
> -	struct fence		        *last_pt_update;
> +	struct dma_fence	        *last_pt_update;
>  	unsigned			ref_count;
>  
>  	/* protected by vm mutex and spinlock */
> @@ -543,7 +543,7 @@ struct amdgpu_sa_bo {
>  	struct amdgpu_sa_manager	*manager;
>  	unsigned			soffset;
>  	unsigned			eoffset;
> -	struct fence		        *fence;
> +	struct dma_fence	        *fence;
>  };
>  
>  /*
> @@ -566,19 +566,19 @@ int amdgpu_mode_dumb_mmap(struct drm_file *filp,
>   */
>  struct amdgpu_sync {
>  	DECLARE_HASHTABLE(fences, 4);
> -	struct fence	        *last_vm_update;
> +	struct dma_fence        *last_vm_update;
>  };
>  
>  void amdgpu_sync_create(struct amdgpu_sync *sync);
>  int amdgpu_sync_fence(struct amdgpu_device *adev, struct amdgpu_sync *sync,
> -		      struct fence *f);
> +		      struct dma_fence *f);
>  int amdgpu_sync_resv(struct amdgpu_device *adev,
>  		     struct amdgpu_sync *sync,
>  		     struct reservation_object *resv,
>  		     void *owner);
> -struct fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
> +struct dma_fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
>  				     struct amdgpu_ring *ring);
> -struct fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync);
> +struct dma_fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync);
>  void amdgpu_sync_free(struct amdgpu_sync *sync);
>  int amdgpu_sync_init(void);
>  void amdgpu_sync_fini(void);
> @@ -703,10 +703,10 @@ struct amdgpu_flip_work {
>  	uint64_t			base;
>  	struct drm_pending_vblank_event *event;
>  	struct amdgpu_bo		*old_abo;
> -	struct fence			*excl;
> +	struct dma_fence		*excl;
>  	unsigned			shared_count;
> -	struct fence			**shared;
> -	struct fence_cb			cb;
> +	struct dma_fence		**shared;
> +	struct dma_fence_cb		cb;
>  	bool				async;
>  };
>  
> @@ -742,7 +742,7 @@ void amdgpu_job_free_resources(struct amdgpu_job *job);
>  void amdgpu_job_free(struct amdgpu_job *job);
>  int amdgpu_job_submit(struct amdgpu_job *job, struct amdgpu_ring *ring,
>  		      struct amd_sched_entity *entity, void *owner,
> -		      struct fence **f);
> +		      struct dma_fence **f);
>  
>  struct amdgpu_ring {
>  	struct amdgpu_device		*adev;
> @@ -844,7 +844,7 @@ struct amdgpu_vm {
>  	/* contains the page directory */
>  	struct amdgpu_bo	*page_directory;
>  	unsigned		max_pde_used;
> -	struct fence		*page_directory_fence;
> +	struct dma_fence	*page_directory_fence;
>  	uint64_t		last_eviction_counter;
>  
>  	/* array of page tables, one for each page directory entry */
> @@ -865,14 +865,14 @@ struct amdgpu_vm {
>  
>  struct amdgpu_vm_id {
>  	struct list_head	list;
> -	struct fence		*first;
> +	struct dma_fence	*first;
>  	struct amdgpu_sync	active;
> -	struct fence		*last_flush;
> +	struct dma_fence	*last_flush;
>  	atomic64_t		owner;
>  
>  	uint64_t		pd_gpu_addr;
>  	/* last flushed PD/PT update */
> -	struct fence		*flushed_updates;
> +	struct dma_fence	*flushed_updates;
>  
>  	uint32_t                current_gpu_reset_count;
>  
> @@ -921,7 +921,7 @@ void amdgpu_vm_get_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>  void amdgpu_vm_move_pt_bos_in_lru(struct amdgpu_device *adev,
>  				  struct amdgpu_vm *vm);
>  int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
> -		      struct amdgpu_sync *sync, struct fence *fence,
> +		      struct amdgpu_sync *sync, struct dma_fence *fence,
>  		      struct amdgpu_job *job);
>  int amdgpu_vm_flush(struct amdgpu_ring *ring, struct amdgpu_job *job);
>  void amdgpu_vm_reset_id(struct amdgpu_device *adev, unsigned vm_id);
> @@ -957,7 +957,7 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
>  
>  struct amdgpu_ctx_ring {
>  	uint64_t		sequence;
> -	struct fence		**fences;
> +	struct dma_fence	**fences;
>  	struct amd_sched_entity	entity;
>  };
>  
> @@ -966,7 +966,7 @@ struct amdgpu_ctx {
>  	struct amdgpu_device    *adev;
>  	unsigned		reset_counter;
>  	spinlock_t		ring_lock;
> -	struct fence            **fences;
> +	struct dma_fence	**fences;
>  	struct amdgpu_ctx_ring	rings[AMDGPU_MAX_RINGS];
>  	bool preamble_presented;
>  };
> @@ -982,8 +982,8 @@ struct amdgpu_ctx *amdgpu_ctx_get(struct amdgpu_fpriv *fpriv, uint32_t id);
>  int amdgpu_ctx_put(struct amdgpu_ctx *ctx);
>  
>  uint64_t amdgpu_ctx_add_fence(struct amdgpu_ctx *ctx, struct amdgpu_ring *ring,
> -			      struct fence *fence);
> -struct fence *amdgpu_ctx_get_fence(struct amdgpu_ctx *ctx,
> +			      struct dma_fence *fence);
> +struct dma_fence *amdgpu_ctx_get_fence(struct amdgpu_ctx *ctx,
>  				   struct amdgpu_ring *ring, uint64_t seq);
>  
>  int amdgpu_ctx_ioctl(struct drm_device *dev, void *data,
> @@ -1181,10 +1181,10 @@ struct amdgpu_gfx {
>  int amdgpu_ib_get(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>  		  unsigned size, struct amdgpu_ib *ib);
>  void amdgpu_ib_free(struct amdgpu_device *adev, struct amdgpu_ib *ib,
> -		    struct fence *f);
> +		    struct dma_fence *f);
>  int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned num_ibs,
> -		       struct amdgpu_ib *ib, struct fence *last_vm_update,
> -		       struct amdgpu_job *job, struct fence **f);
> +		       struct amdgpu_ib *ib, struct dma_fence *last_vm_update,
> +		       struct amdgpu_job *job, struct dma_fence **f);
>  int amdgpu_ib_pool_init(struct amdgpu_device *adev);
>  void amdgpu_ib_pool_fini(struct amdgpu_device *adev);
>  int amdgpu_ib_ring_tests(struct amdgpu_device *adev);
> @@ -1225,7 +1225,7 @@ struct amdgpu_cs_parser {
>  	struct amdgpu_bo_list		*bo_list;
>  	struct amdgpu_bo_list_entry	vm_pd;
>  	struct list_head		validated;
> -	struct fence			*fence;
> +	struct dma_fence		*fence;
>  	uint64_t			bytes_moved_threshold;
>  	uint64_t			bytes_moved;
>  	struct amdgpu_bo_list_entry	*evictable;
> @@ -1245,7 +1245,7 @@ struct amdgpu_job {
>  	struct amdgpu_ring	*ring;
>  	struct amdgpu_sync	sync;
>  	struct amdgpu_ib	*ibs;
> -	struct fence		*fence; /* the hw fence */
> +	struct dma_fence	*fence; /* the hw fence */
>  	uint32_t		preamble_status;
>  	uint32_t		num_ibs;
>  	void			*owner;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c
> index 345305235349..cc97eee93226 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c
> @@ -33,7 +33,7 @@ static int amdgpu_benchmark_do_move(struct amdgpu_device *adev, unsigned size,
>  {
>  	unsigned long start_jiffies;
>  	unsigned long end_jiffies;
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  	int i, r;
>  
>  	start_jiffies = jiffies;
> @@ -43,17 +43,17 @@ static int amdgpu_benchmark_do_move(struct amdgpu_device *adev, unsigned size,
>  				       false);
>  		if (r)
>  			goto exit_do_move;
> -		r = fence_wait(fence, false);
> +		r = dma_fence_wait(fence, false);
>  		if (r)
>  			goto exit_do_move;
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  	}
>  	end_jiffies = jiffies;
>  	r = jiffies_to_msecs(end_jiffies - start_jiffies);
>  
>  exit_do_move:
>  	if (fence)
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  	return r;
>  }
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> index b0f6e6957536..5d582265e929 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> @@ -719,7 +719,7 @@ static void amdgpu_cs_parser_fini(struct amdgpu_cs_parser *parser, int error, bo
>  		ttm_eu_backoff_reservation(&parser->ticket,
>  					   &parser->validated);
>  	}
> -	fence_put(parser->fence);
> +	dma_fence_put(parser->fence);
>  
>  	if (parser->ctx)
>  		amdgpu_ctx_put(parser->ctx);
> @@ -756,7 +756,7 @@ static int amdgpu_bo_vm_update_pte(struct amdgpu_cs_parser *p,
>  
>  	if (p->bo_list) {
>  		for (i = 0; i < p->bo_list->num_entries; i++) {
> -			struct fence *f;
> +			struct dma_fence *f;
>  
>  			/* ignore duplicates */
>  			bo = p->bo_list->array[i].robj;
> @@ -956,7 +956,7 @@ static int amdgpu_cs_dependencies(struct amdgpu_device *adev,
>  		for (j = 0; j < num_deps; ++j) {
>  			struct amdgpu_ring *ring;
>  			struct amdgpu_ctx *ctx;
> -			struct fence *fence;
> +			struct dma_fence *fence;
>  
>  			r = amdgpu_cs_get_ring(adev, deps[j].ip_type,
>  					       deps[j].ip_instance,
> @@ -978,7 +978,7 @@ static int amdgpu_cs_dependencies(struct amdgpu_device *adev,
>  			} else if (fence) {
>  				r = amdgpu_sync_fence(adev, &p->job->sync,
>  						      fence);
> -				fence_put(fence);
> +				dma_fence_put(fence);
>  				amdgpu_ctx_put(ctx);
>  				if (r)
>  					return r;
> @@ -1008,7 +1008,7 @@ static int amdgpu_cs_submit(struct amdgpu_cs_parser *p,
>  
>  	job->owner = p->filp;
>  	job->fence_ctx = entity->fence_context;
> -	p->fence = fence_get(&job->base.s_fence->finished);
> +	p->fence = dma_fence_get(&job->base.s_fence->finished);
>  	cs->out.handle = amdgpu_ctx_add_fence(p->ctx, ring, p->fence);
>  	job->uf_sequence = cs->out.handle;
>  	amdgpu_job_free_resources(job);
> @@ -1091,7 +1091,7 @@ int amdgpu_cs_wait_ioctl(struct drm_device *dev, void *data,
>  	unsigned long timeout = amdgpu_gem_timeout(wait->in.timeout);
>  	struct amdgpu_ring *ring = NULL;
>  	struct amdgpu_ctx *ctx;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	long r;
>  
>  	r = amdgpu_cs_get_ring(adev, wait->in.ip_type, wait->in.ip_instance,
> @@ -1107,8 +1107,8 @@ int amdgpu_cs_wait_ioctl(struct drm_device *dev, void *data,
>  	if (IS_ERR(fence))
>  		r = PTR_ERR(fence);
>  	else if (fence) {
> -		r = fence_wait_timeout(fence, true, timeout);
> -		fence_put(fence);
> +		r = dma_fence_wait_timeout(fence, true, timeout);
> +		dma_fence_put(fence);
>  	} else
>  		r = 1;
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> index a5e2fcbef0f0..99bbc860322f 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
> @@ -35,7 +35,7 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev, struct amdgpu_ctx *ctx)
>  	kref_init(&ctx->refcount);
>  	spin_lock_init(&ctx->ring_lock);
>  	ctx->fences = kcalloc(amdgpu_sched_jobs * AMDGPU_MAX_RINGS,
> -			      sizeof(struct fence*), GFP_KERNEL);
> +			      sizeof(struct dma_fence*), GFP_KERNEL);
>  	if (!ctx->fences)
>  		return -ENOMEM;
>  
> @@ -79,7 +79,7 @@ static void amdgpu_ctx_fini(struct amdgpu_ctx *ctx)
>  
>  	for (i = 0; i < AMDGPU_MAX_RINGS; ++i)
>  		for (j = 0; j < amdgpu_sched_jobs; ++j)
> -			fence_put(ctx->rings[i].fences[j]);
> +			dma_fence_put(ctx->rings[i].fences[j]);
>  	kfree(ctx->fences);
>  	ctx->fences = NULL;
>  
> @@ -241,39 +241,39 @@ int amdgpu_ctx_put(struct amdgpu_ctx *ctx)
>  }
>  
>  uint64_t amdgpu_ctx_add_fence(struct amdgpu_ctx *ctx, struct amdgpu_ring *ring,
> -			      struct fence *fence)
> +			      struct dma_fence *fence)
>  {
>  	struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx];
>  	uint64_t seq = cring->sequence;
>  	unsigned idx = 0;
> -	struct fence *other = NULL;
> +	struct dma_fence *other = NULL;
>  
>  	idx = seq & (amdgpu_sched_jobs - 1);
>  	other = cring->fences[idx];
>  	if (other) {
>  		signed long r;
> -		r = fence_wait_timeout(other, false, MAX_SCHEDULE_TIMEOUT);
> +		r = dma_fence_wait_timeout(other, false, MAX_SCHEDULE_TIMEOUT);
>  		if (r < 0)
>  			DRM_ERROR("Error (%ld) waiting for fence!\n", r);
>  	}
>  
> -	fence_get(fence);
> +	dma_fence_get(fence);
>  
>  	spin_lock(&ctx->ring_lock);
>  	cring->fences[idx] = fence;
>  	cring->sequence++;
>  	spin_unlock(&ctx->ring_lock);
>  
> -	fence_put(other);
> +	dma_fence_put(other);
>  
>  	return seq;
>  }
>  
> -struct fence *amdgpu_ctx_get_fence(struct amdgpu_ctx *ctx,
> -				   struct amdgpu_ring *ring, uint64_t seq)
> +struct dma_fence *amdgpu_ctx_get_fence(struct amdgpu_ctx *ctx,
> +				       struct amdgpu_ring *ring, uint64_t seq)
>  {
>  	struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx];
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  
>  	spin_lock(&ctx->ring_lock);
>  
> @@ -288,7 +288,7 @@ struct fence *amdgpu_ctx_get_fence(struct amdgpu_ctx *ctx,
>  		return NULL;
>  	}
>  
> -	fence = fence_get(cring->fences[seq & (amdgpu_sched_jobs - 1)]);
> +	fence = dma_fence_get(cring->fences[seq & (amdgpu_sched_jobs - 1)]);
>  	spin_unlock(&ctx->ring_lock);
>  
>  	return fence;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> index b4f4a9239069..0262b43c8f0e 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> @@ -1599,7 +1599,7 @@ int amdgpu_device_init(struct amdgpu_device *adev,
>  	adev->vm_manager.vm_pte_funcs = NULL;
>  	adev->vm_manager.vm_pte_num_rings = 0;
>  	adev->gart.gart_funcs = NULL;
> -	adev->fence_context = fence_context_alloc(AMDGPU_MAX_RINGS);
> +	adev->fence_context = dma_fence_context_alloc(AMDGPU_MAX_RINGS);
>  
>  	adev->smc_rreg = &amdgpu_invalid_rreg;
>  	adev->smc_wreg = &amdgpu_invalid_wreg;
> @@ -2193,7 +2193,7 @@ bool amdgpu_need_backup(struct amdgpu_device *adev)
>  static int amdgpu_recover_vram_from_shadow(struct amdgpu_device *adev,
>  					   struct amdgpu_ring *ring,
>  					   struct amdgpu_bo *bo,
> -					   struct fence **fence)
> +					   struct dma_fence **fence)
>  {
>  	uint32_t domain;
>  	int r;
> @@ -2312,30 +2312,30 @@ int amdgpu_gpu_reset(struct amdgpu_device *adev)
>  		if (need_full_reset && amdgpu_need_backup(adev)) {
>  			struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring;
>  			struct amdgpu_bo *bo, *tmp;
> -			struct fence *fence = NULL, *next = NULL;
> +			struct dma_fence *fence = NULL, *next = NULL;
>  
>  			DRM_INFO("recover vram bo from shadow\n");
>  			mutex_lock(&adev->shadow_list_lock);
>  			list_for_each_entry_safe(bo, tmp, &adev->shadow_list, shadow_list) {
>  				amdgpu_recover_vram_from_shadow(adev, ring, bo, &next);
>  				if (fence) {
> -					r = fence_wait(fence, false);
> +					r = dma_fence_wait(fence, false);
>  					if (r) {
>  						WARN(r, "recovery from shadow isn't comleted\n");
>  						break;
>  					}
>  				}
>  
> -				fence_put(fence);
> +				dma_fence_put(fence);
>  				fence = next;
>  			}
>  			mutex_unlock(&adev->shadow_list_lock);
>  			if (fence) {
> -				r = fence_wait(fence, false);
> +				r = dma_fence_wait(fence, false);
>  				if (r)
>  					WARN(r, "recovery from shadow isn't comleted\n");
>  			}
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  		}
>  		for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  			struct amdgpu_ring *ring = adev->rings[i];
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> index 083e2b429872..075c0d7db205 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
> @@ -35,29 +35,29 @@
>  #include <drm/drm_crtc_helper.h>
>  #include <drm/drm_edid.h>
>  
> -static void amdgpu_flip_callback(struct fence *f, struct fence_cb *cb)
> +static void amdgpu_flip_callback(struct dma_fence *f, struct dma_fence_cb *cb)
>  {
>  	struct amdgpu_flip_work *work =
>  		container_of(cb, struct amdgpu_flip_work, cb);
>  
> -	fence_put(f);
> +	dma_fence_put(f);
>  	schedule_work(&work->flip_work.work);
>  }
>  
>  static bool amdgpu_flip_handle_fence(struct amdgpu_flip_work *work,
> -				     struct fence **f)
> +				     struct dma_fence **f)
>  {
> -	struct fence *fence= *f;
> +	struct dma_fence *fence= *f;
>  
>  	if (fence == NULL)
>  		return false;
>  
>  	*f = NULL;
>  
> -	if (!fence_add_callback(fence, &work->cb, amdgpu_flip_callback))
> +	if (!dma_fence_add_callback(fence, &work->cb, amdgpu_flip_callback))
>  		return true;
>  
> -	fence_put(fence);
> +	dma_fence_put(fence);
>  	return false;
>  }
>  
> @@ -244,9 +244,9 @@ int amdgpu_crtc_page_flip_target(struct drm_crtc *crtc,
>  
>  cleanup:
>  	amdgpu_bo_unref(&work->old_abo);
> -	fence_put(work->excl);
> +	dma_fence_put(work->excl);
>  	for (i = 0; i < work->shared_count; ++i)
> -		fence_put(work->shared[i]);
> +		dma_fence_put(work->shared[i]);
>  	kfree(work->shared);
>  	kfree(work);
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> index 3a2e42f4b897..57552c79ec58 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> @@ -48,7 +48,7 @@
>   */
>  
>  struct amdgpu_fence {
> -	struct fence base;
> +	struct dma_fence base;
>  
>  	/* RB, DMA, etc. */
>  	struct amdgpu_ring		*ring;
> @@ -73,8 +73,8 @@ void amdgpu_fence_slab_fini(void)
>  /*
>   * Cast helper
>   */
> -static const struct fence_ops amdgpu_fence_ops;
> -static inline struct amdgpu_fence *to_amdgpu_fence(struct fence *f)
> +static const struct dma_fence_ops amdgpu_fence_ops;
> +static inline struct amdgpu_fence *to_amdgpu_fence(struct dma_fence *f)
>  {
>  	struct amdgpu_fence *__f = container_of(f, struct amdgpu_fence, base);
>  
> @@ -130,11 +130,11 @@ static u32 amdgpu_fence_read(struct amdgpu_ring *ring)
>   * Emits a fence command on the requested ring (all asics).
>   * Returns 0 on success, -ENOMEM on failure.
>   */
> -int amdgpu_fence_emit(struct amdgpu_ring *ring, struct fence **f)
> +int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_fence *fence;
> -	struct fence *old, **ptr;
> +	struct dma_fence *old, **ptr;
>  	uint32_t seq;
>  
>  	fence = kmem_cache_alloc(amdgpu_fence_slab, GFP_KERNEL);
> @@ -143,10 +143,10 @@ int amdgpu_fence_emit(struct amdgpu_ring *ring, struct fence **f)
>  
>  	seq = ++ring->fence_drv.sync_seq;
>  	fence->ring = ring;
> -	fence_init(&fence->base, &amdgpu_fence_ops,
> -		   &ring->fence_drv.lock,
> -		   adev->fence_context + ring->idx,
> -		   seq);
> +	dma_fence_init(&fence->base, &amdgpu_fence_ops,
> +		       &ring->fence_drv.lock,
> +		       adev->fence_context + ring->idx,
> +		       seq);
>  	amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr,
>  			       seq, AMDGPU_FENCE_FLAG_INT);
>  
> @@ -155,12 +155,12 @@ int amdgpu_fence_emit(struct amdgpu_ring *ring, struct fence **f)
>  	 * emitting the fence would mess up the hardware ring buffer.
>  	 */
>  	old = rcu_dereference_protected(*ptr, 1);
> -	if (old && !fence_is_signaled(old)) {
> +	if (old && !dma_fence_is_signaled(old)) {
>  		DRM_INFO("rcu slot is busy\n");
> -		fence_wait(old, false);
> +		dma_fence_wait(old, false);
>  	}
>  
> -	rcu_assign_pointer(*ptr, fence_get(&fence->base));
> +	rcu_assign_pointer(*ptr, dma_fence_get(&fence->base));
>  
>  	*f = &fence->base;
>  
> @@ -211,7 +211,7 @@ void amdgpu_fence_process(struct amdgpu_ring *ring)
>  	seq &= drv->num_fences_mask;
>  
>  	do {
> -		struct fence *fence, **ptr;
> +		struct dma_fence *fence, **ptr;
>  
>  		++last_seq;
>  		last_seq &= drv->num_fences_mask;
> @@ -224,13 +224,13 @@ void amdgpu_fence_process(struct amdgpu_ring *ring)
>  		if (!fence)
>  			continue;
>  
> -		r = fence_signal(fence);
> +		r = dma_fence_signal(fence);
>  		if (!r)
> -			FENCE_TRACE(fence, "signaled from irq context\n");
> +			DMA_FENCE_TRACE(fence, "signaled from irq context\n");
>  		else
>  			BUG();
>  
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  	} while (last_seq != seq);
>  }
>  
> @@ -260,7 +260,7 @@ static void amdgpu_fence_fallback(unsigned long arg)
>  int amdgpu_fence_wait_empty(struct amdgpu_ring *ring)
>  {
>  	uint64_t seq = ACCESS_ONCE(ring->fence_drv.sync_seq);
> -	struct fence *fence, **ptr;
> +	struct dma_fence *fence, **ptr;
>  	int r;
>  
>  	if (!seq)
> @@ -269,14 +269,14 @@ int amdgpu_fence_wait_empty(struct amdgpu_ring *ring)
>  	ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask];
>  	rcu_read_lock();
>  	fence = rcu_dereference(*ptr);
> -	if (!fence || !fence_get_rcu(fence)) {
> +	if (!fence || !dma_fence_get_rcu(fence)) {
>  		rcu_read_unlock();
>  		return 0;
>  	}
>  	rcu_read_unlock();
>  
> -	r = fence_wait(fence, false);
> -	fence_put(fence);
> +	r = dma_fence_wait(fence, false);
> +	dma_fence_put(fence);
>  	return r;
>  }
>  
> @@ -452,7 +452,7 @@ void amdgpu_fence_driver_fini(struct amdgpu_device *adev)
>  		amd_sched_fini(&ring->sched);
>  		del_timer_sync(&ring->fence_drv.fallback_timer);
>  		for (j = 0; j <= ring->fence_drv.num_fences_mask; ++j)
> -			fence_put(ring->fence_drv.fences[j]);
> +			dma_fence_put(ring->fence_drv.fences[j]);
>  		kfree(ring->fence_drv.fences);
>  		ring->fence_drv.fences = NULL;
>  		ring->fence_drv.initialized = false;
> @@ -541,12 +541,12 @@ void amdgpu_fence_driver_force_completion(struct amdgpu_device *adev)
>   * Common fence implementation
>   */
>  
> -static const char *amdgpu_fence_get_driver_name(struct fence *fence)
> +static const char *amdgpu_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "amdgpu";
>  }
>  
> -static const char *amdgpu_fence_get_timeline_name(struct fence *f)
> +static const char *amdgpu_fence_get_timeline_name(struct dma_fence *f)
>  {
>  	struct amdgpu_fence *fence = to_amdgpu_fence(f);
>  	return (const char *)fence->ring->name;
> @@ -560,7 +560,7 @@ static const char *amdgpu_fence_get_timeline_name(struct fence *f)
>   * to fence_queue that checks if this fence is signaled, and if so it
>   * signals the fence and removes itself.
>   */
> -static bool amdgpu_fence_enable_signaling(struct fence *f)
> +static bool amdgpu_fence_enable_signaling(struct dma_fence *f)
>  {
>  	struct amdgpu_fence *fence = to_amdgpu_fence(f);
>  	struct amdgpu_ring *ring = fence->ring;
> @@ -568,7 +568,7 @@ static bool amdgpu_fence_enable_signaling(struct fence *f)
>  	if (!timer_pending(&ring->fence_drv.fallback_timer))
>  		amdgpu_fence_schedule_fallback(ring);
>  
> -	FENCE_TRACE(&fence->base, "armed on ring %i!\n", ring->idx);
> +	DMA_FENCE_TRACE(&fence->base, "armed on ring %i!\n", ring->idx);
>  
>  	return true;
>  }
> @@ -582,7 +582,7 @@ static bool amdgpu_fence_enable_signaling(struct fence *f)
>   */
>  static void amdgpu_fence_free(struct rcu_head *rcu)
>  {
> -	struct fence *f = container_of(rcu, struct fence, rcu);
> +	struct dma_fence *f = container_of(rcu, struct dma_fence, rcu);
>  	struct amdgpu_fence *fence = to_amdgpu_fence(f);
>  	kmem_cache_free(amdgpu_fence_slab, fence);
>  }
> @@ -595,16 +595,16 @@ static void amdgpu_fence_free(struct rcu_head *rcu)
>   * This function is called when the reference count becomes zero.
>   * It just RCU schedules freeing up the fence.
>   */
> -static void amdgpu_fence_release(struct fence *f)
> +static void amdgpu_fence_release(struct dma_fence *f)
>  {
>  	call_rcu(&f->rcu, amdgpu_fence_free);
>  }
>  
> -static const struct fence_ops amdgpu_fence_ops = {
> +static const struct dma_fence_ops amdgpu_fence_ops = {
>  	.get_driver_name = amdgpu_fence_get_driver_name,
>  	.get_timeline_name = amdgpu_fence_get_timeline_name,
>  	.enable_signaling = amdgpu_fence_enable_signaling,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = amdgpu_fence_release,
>  };
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
> index 6a6c86c9c169..c3672dfcfd6a 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
> @@ -89,7 +89,7 @@ int amdgpu_ib_get(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>   * Free an IB (all asics).
>   */
>  void amdgpu_ib_free(struct amdgpu_device *adev, struct amdgpu_ib *ib,
> -		    struct fence *f)
> +		    struct dma_fence *f)
>  {
>  	amdgpu_sa_bo_free(adev, &ib->sa_bo, f);
>  }
> @@ -116,8 +116,8 @@ void amdgpu_ib_free(struct amdgpu_device *adev, struct amdgpu_ib *ib,
>   * to SI there was just a DE IB.
>   */
>  int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned num_ibs,
> -		       struct amdgpu_ib *ibs, struct fence *last_vm_update,
> -		       struct amdgpu_job *job, struct fence **f)
> +		       struct amdgpu_ib *ibs, struct dma_fence *last_vm_update,
> +		       struct amdgpu_job *job, struct dma_fence **f)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib *ib = &ibs[0];
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> index 8c5807994073..a0de6286c453 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> @@ -81,7 +81,7 @@ int amdgpu_job_alloc_with_ib(struct amdgpu_device *adev, unsigned size,
>  
>  void amdgpu_job_free_resources(struct amdgpu_job *job)
>  {
> -	struct fence *f;
> +	struct dma_fence *f;
>  	unsigned i;
>  
>  	/* use sched fence if available */
> @@ -95,7 +95,7 @@ static void amdgpu_job_free_cb(struct amd_sched_job *s_job)
>  {
>  	struct amdgpu_job *job = container_of(s_job, struct amdgpu_job, base);
>  
> -	fence_put(job->fence);
> +	dma_fence_put(job->fence);
>  	amdgpu_sync_free(&job->sync);
>  	kfree(job);
>  }
> @@ -104,14 +104,14 @@ void amdgpu_job_free(struct amdgpu_job *job)
>  {
>  	amdgpu_job_free_resources(job);
>  
> -	fence_put(job->fence);
> +	dma_fence_put(job->fence);
>  	amdgpu_sync_free(&job->sync);
>  	kfree(job);
>  }
>  
>  int amdgpu_job_submit(struct amdgpu_job *job, struct amdgpu_ring *ring,
>  		      struct amd_sched_entity *entity, void *owner,
> -		      struct fence **f)
> +		      struct dma_fence **f)
>  {
>  	int r;
>  	job->ring = ring;
> @@ -125,19 +125,19 @@ int amdgpu_job_submit(struct amdgpu_job *job, struct amdgpu_ring *ring,
>  
>  	job->owner = owner;
>  	job->fence_ctx = entity->fence_context;
> -	*f = fence_get(&job->base.s_fence->finished);
> +	*f = dma_fence_get(&job->base.s_fence->finished);
>  	amdgpu_job_free_resources(job);
>  	amd_sched_entity_push_job(&job->base);
>  
>  	return 0;
>  }
>  
> -static struct fence *amdgpu_job_dependency(struct amd_sched_job *sched_job)
> +static struct dma_fence *amdgpu_job_dependency(struct amd_sched_job *sched_job)
>  {
>  	struct amdgpu_job *job = to_amdgpu_job(sched_job);
>  	struct amdgpu_vm *vm = job->vm;
>  
> -	struct fence *fence = amdgpu_sync_get_fence(&job->sync);
> +	struct dma_fence *fence = amdgpu_sync_get_fence(&job->sync);
>  
>  	if (fence == NULL && vm && !job->vm_id) {
>  		struct amdgpu_ring *ring = job->ring;
> @@ -155,9 +155,9 @@ static struct fence *amdgpu_job_dependency(struct amd_sched_job *sched_job)
>  	return fence;
>  }
>  
> -static struct fence *amdgpu_job_run(struct amd_sched_job *sched_job)
> +static struct dma_fence *amdgpu_job_run(struct amd_sched_job *sched_job)
>  {
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  	struct amdgpu_job *job;
>  	int r;
>  
> @@ -176,8 +176,8 @@ static struct fence *amdgpu_job_run(struct amd_sched_job *sched_job)
>  		DRM_ERROR("Error scheduling IBs (%d)\n", r);
>  
>  	/* if gpu reset, hw fence will be replaced here */
> -	fence_put(job->fence);
> -	job->fence = fence_get(fence);
> +	dma_fence_put(job->fence);
> +	job->fence = dma_fence_get(fence);
>  	amdgpu_job_free_resources(job);
>  	return fence;
>  }
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> index aa074fac0c7f..55e142a5ff5f 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> @@ -383,7 +383,7 @@ int amdgpu_bo_create_restricted(struct amdgpu_device *adev,
>  
>  	if (flags & AMDGPU_GEM_CREATE_VRAM_CLEARED &&
>  	    bo->tbo.mem.placement & TTM_PL_FLAG_VRAM) {
> -		struct fence *fence;
> +		struct dma_fence *fence;
>  
>  		if (adev->mman.buffer_funcs_ring == NULL ||
>  		   !adev->mman.buffer_funcs_ring->ready) {
> @@ -403,9 +403,9 @@ int amdgpu_bo_create_restricted(struct amdgpu_device *adev,
>  		amdgpu_fill_buffer(bo, 0, bo->tbo.resv, &fence);
>  		amdgpu_bo_fence(bo, fence, false);
>  		amdgpu_bo_unreserve(bo);
> -		fence_put(bo->tbo.moving);
> -		bo->tbo.moving = fence_get(fence);
> -		fence_put(fence);
> +		dma_fence_put(bo->tbo.moving);
> +		bo->tbo.moving = dma_fence_get(fence);
> +		dma_fence_put(fence);
>  	}
>  	*bo_ptr = bo;
>  
> @@ -491,7 +491,7 @@ int amdgpu_bo_backup_to_shadow(struct amdgpu_device *adev,
>  			       struct amdgpu_ring *ring,
>  			       struct amdgpu_bo *bo,
>  			       struct reservation_object *resv,
> -			       struct fence **fence,
> +			       struct dma_fence **fence,
>  			       bool direct)
>  
>  {
> @@ -523,7 +523,7 @@ int amdgpu_bo_restore_from_shadow(struct amdgpu_device *adev,
>  				  struct amdgpu_ring *ring,
>  				  struct amdgpu_bo *bo,
>  				  struct reservation_object *resv,
> -				  struct fence **fence,
> +				  struct dma_fence **fence,
>  				  bool direct)
>  
>  {
> @@ -926,7 +926,7 @@ int amdgpu_bo_fault_reserve_notify(struct ttm_buffer_object *bo)
>   * @shared: true if fence should be added shared
>   *
>   */
> -void amdgpu_bo_fence(struct amdgpu_bo *bo, struct fence *fence,
> +void amdgpu_bo_fence(struct amdgpu_bo *bo, struct dma_fence *fence,
>  		     bool shared)
>  {
>  	struct reservation_object *resv = bo->tbo.resv;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> index 8255034d73eb..3e785ed3cb4b 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> @@ -156,19 +156,19 @@ int amdgpu_bo_get_metadata(struct amdgpu_bo *bo, void *buffer,
>  void amdgpu_bo_move_notify(struct ttm_buffer_object *bo,
>  				  struct ttm_mem_reg *new_mem);
>  int amdgpu_bo_fault_reserve_notify(struct ttm_buffer_object *bo);
> -void amdgpu_bo_fence(struct amdgpu_bo *bo, struct fence *fence,
> +void amdgpu_bo_fence(struct amdgpu_bo *bo, struct dma_fence *fence,
>  		     bool shared);
>  u64 amdgpu_bo_gpu_offset(struct amdgpu_bo *bo);
>  int amdgpu_bo_backup_to_shadow(struct amdgpu_device *adev,
>  			       struct amdgpu_ring *ring,
>  			       struct amdgpu_bo *bo,
>  			       struct reservation_object *resv,
> -			       struct fence **fence, bool direct);
> +			       struct dma_fence **fence, bool direct);
>  int amdgpu_bo_restore_from_shadow(struct amdgpu_device *adev,
>  				  struct amdgpu_ring *ring,
>  				  struct amdgpu_bo *bo,
>  				  struct reservation_object *resv,
> -				  struct fence **fence,
> +				  struct dma_fence **fence,
>  				  bool direct);
>  
>  
> @@ -200,7 +200,7 @@ int amdgpu_sa_bo_new(struct amdgpu_sa_manager *sa_manager,
>  		     unsigned size, unsigned align);
>  void amdgpu_sa_bo_free(struct amdgpu_device *adev,
>  			      struct amdgpu_sa_bo **sa_bo,
> -			      struct fence *fence);
> +			      struct dma_fence *fence);
>  #if defined(CONFIG_DEBUG_FS)
>  void amdgpu_sa_bo_dump_debug_info(struct amdgpu_sa_manager *sa_manager,
>  					 struct seq_file *m);
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c
> index d8af37a845f4..fd26c4b8d793 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c
> @@ -147,7 +147,7 @@ static void amdgpu_sa_bo_remove_locked(struct amdgpu_sa_bo *sa_bo)
>  	}
>  	list_del_init(&sa_bo->olist);
>  	list_del_init(&sa_bo->flist);
> -	fence_put(sa_bo->fence);
> +	dma_fence_put(sa_bo->fence);
>  	kfree(sa_bo);
>  }
>  
> @@ -161,7 +161,7 @@ static void amdgpu_sa_bo_try_free(struct amdgpu_sa_manager *sa_manager)
>  	sa_bo = list_entry(sa_manager->hole->next, struct amdgpu_sa_bo, olist);
>  	list_for_each_entry_safe_from(sa_bo, tmp, &sa_manager->olist, olist) {
>  		if (sa_bo->fence == NULL ||
> -		    !fence_is_signaled(sa_bo->fence)) {
> +		    !dma_fence_is_signaled(sa_bo->fence)) {
>  			return;
>  		}
>  		amdgpu_sa_bo_remove_locked(sa_bo);
> @@ -244,7 +244,7 @@ static bool amdgpu_sa_event(struct amdgpu_sa_manager *sa_manager,
>  }
>  
>  static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager,
> -				   struct fence **fences,
> +				   struct dma_fence **fences,
>  				   unsigned *tries)
>  {
>  	struct amdgpu_sa_bo *best_bo = NULL;
> @@ -272,7 +272,7 @@ static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager,
>  		sa_bo = list_first_entry(&sa_manager->flist[i],
>  					 struct amdgpu_sa_bo, flist);
>  
> -		if (!fence_is_signaled(sa_bo->fence)) {
> +		if (!dma_fence_is_signaled(sa_bo->fence)) {
>  			fences[i] = sa_bo->fence;
>  			continue;
>  		}
> @@ -314,7 +314,7 @@ int amdgpu_sa_bo_new(struct amdgpu_sa_manager *sa_manager,
>  		     struct amdgpu_sa_bo **sa_bo,
>  		     unsigned size, unsigned align)
>  {
> -	struct fence *fences[AMDGPU_SA_NUM_FENCE_LISTS];
> +	struct dma_fence *fences[AMDGPU_SA_NUM_FENCE_LISTS];
>  	unsigned tries[AMDGPU_SA_NUM_FENCE_LISTS];
>  	unsigned count;
>  	int i, r;
> @@ -356,14 +356,14 @@ int amdgpu_sa_bo_new(struct amdgpu_sa_manager *sa_manager,
>  
>  		for (i = 0, count = 0; i < AMDGPU_SA_NUM_FENCE_LISTS; ++i)
>  			if (fences[i])
> -				fences[count++] = fence_get(fences[i]);
> +				fences[count++] = dma_fence_get(fences[i]);
>  
>  		if (count) {
>  			spin_unlock(&sa_manager->wq.lock);
> -			t = fence_wait_any_timeout(fences, count, false,
> -						   MAX_SCHEDULE_TIMEOUT);
> +			t = dma_fence_wait_any_timeout(fences, count, false,
> +						       MAX_SCHEDULE_TIMEOUT);
>  			for (i = 0; i < count; ++i)
> -				fence_put(fences[i]);
> +				dma_fence_put(fences[i]);
>  
>  			r = (t > 0) ? 0 : t;
>  			spin_lock(&sa_manager->wq.lock);
> @@ -384,7 +384,7 @@ int amdgpu_sa_bo_new(struct amdgpu_sa_manager *sa_manager,
>  }
>  
>  void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo,
> -		       struct fence *fence)
> +		       struct dma_fence *fence)
>  {
>  	struct amdgpu_sa_manager *sa_manager;
>  
> @@ -394,10 +394,10 @@ void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo,
>  
>  	sa_manager = (*sa_bo)->manager;
>  	spin_lock(&sa_manager->wq.lock);
> -	if (fence && !fence_is_signaled(fence)) {
> +	if (fence && !dma_fence_is_signaled(fence)) {
>  		uint32_t idx;
>  
> -		(*sa_bo)->fence = fence_get(fence);
> +		(*sa_bo)->fence = dma_fence_get(fence);
>  		idx = fence->context % AMDGPU_SA_NUM_FENCE_LISTS;
>  		list_add_tail(&(*sa_bo)->flist, &sa_manager->flist[idx]);
>  	} else {
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
> index 5c8d3022fb87..ed814e6d0207 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
> @@ -34,7 +34,7 @@
>  
>  struct amdgpu_sync_entry {
>  	struct hlist_node	node;
> -	struct fence		*fence;
> +	struct dma_fence	*fence;
>  };
>  
>  static struct kmem_cache *amdgpu_sync_slab;
> @@ -60,7 +60,8 @@ void amdgpu_sync_create(struct amdgpu_sync *sync)
>   *
>   * Test if the fence was issued by us.
>   */
> -static bool amdgpu_sync_same_dev(struct amdgpu_device *adev, struct fence *f)
> +static bool amdgpu_sync_same_dev(struct amdgpu_device *adev,
> +				 struct dma_fence *f)
>  {
>  	struct amd_sched_fence *s_fence = to_amd_sched_fence(f);
>  
> @@ -81,7 +82,7 @@ static bool amdgpu_sync_same_dev(struct amdgpu_device *adev, struct fence *f)
>   *
>   * Extract who originally created the fence.
>   */
> -static void *amdgpu_sync_get_owner(struct fence *f)
> +static void *amdgpu_sync_get_owner(struct dma_fence *f)
>  {
>  	struct amd_sched_fence *s_fence = to_amd_sched_fence(f);
>  
> @@ -99,13 +100,14 @@ static void *amdgpu_sync_get_owner(struct fence *f)
>   *
>   * Either keep the existing fence or the new one, depending which one is later.
>   */
> -static void amdgpu_sync_keep_later(struct fence **keep, struct fence *fence)
> +static void amdgpu_sync_keep_later(struct dma_fence **keep,
> +				   struct dma_fence *fence)
>  {
> -	if (*keep && fence_is_later(*keep, fence))
> +	if (*keep && dma_fence_is_later(*keep, fence))
>  		return;
>  
> -	fence_put(*keep);
> -	*keep = fence_get(fence);
> +	dma_fence_put(*keep);
> +	*keep = dma_fence_get(fence);
>  }
>  
>  /**
> @@ -117,7 +119,7 @@ static void amdgpu_sync_keep_later(struct fence **keep, struct fence *fence)
>   * Tries to add the fence to an existing hash entry. Returns true when an entry
>   * was found, false otherwise.
>   */
> -static bool amdgpu_sync_add_later(struct amdgpu_sync *sync, struct fence *f)
> +static bool amdgpu_sync_add_later(struct amdgpu_sync *sync, struct dma_fence *f)
>  {
>  	struct amdgpu_sync_entry *e;
>  
> @@ -139,7 +141,7 @@ static bool amdgpu_sync_add_later(struct amdgpu_sync *sync, struct fence *f)
>   *
>   */
>  int amdgpu_sync_fence(struct amdgpu_device *adev, struct amdgpu_sync *sync,
> -		      struct fence *f)
> +		      struct dma_fence *f)
>  {
>  	struct amdgpu_sync_entry *e;
>  
> @@ -158,7 +160,7 @@ int amdgpu_sync_fence(struct amdgpu_device *adev, struct amdgpu_sync *sync,
>  		return -ENOMEM;
>  
>  	hash_add(sync->fences, &e->node, f->context);
> -	e->fence = fence_get(f);
> +	e->fence = dma_fence_get(f);
>  	return 0;
>  }
>  
> @@ -177,7 +179,7 @@ int amdgpu_sync_resv(struct amdgpu_device *adev,
>  		     void *owner)
>  {
>  	struct reservation_object_list *flist;
> -	struct fence *f;
> +	struct dma_fence *f;
>  	void *fence_owner;
>  	unsigned i;
>  	int r = 0;
> @@ -231,15 +233,15 @@ int amdgpu_sync_resv(struct amdgpu_device *adev,
>   * Returns the next fence not signaled yet without removing it from the sync
>   * object.
>   */
> -struct fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
> -				     struct amdgpu_ring *ring)
> +struct dma_fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
> +					 struct amdgpu_ring *ring)
>  {
>  	struct amdgpu_sync_entry *e;
>  	struct hlist_node *tmp;
>  	int i;
>  
>  	hash_for_each_safe(sync->fences, i, tmp, e, node) {
> -		struct fence *f = e->fence;
> +		struct dma_fence *f = e->fence;
>  		struct amd_sched_fence *s_fence = to_amd_sched_fence(f);
>  
>  		if (ring && s_fence) {
> @@ -247,16 +249,16 @@ struct fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
>  			 * when they are scheduled.
>  			 */
>  			if (s_fence->sched == &ring->sched) {
> -				if (fence_is_signaled(&s_fence->scheduled))
> +				if (dma_fence_is_signaled(&s_fence->scheduled))
>  					continue;
>  
>  				return &s_fence->scheduled;
>  			}
>  		}
>  
> -		if (fence_is_signaled(f)) {
> +		if (dma_fence_is_signaled(f)) {
>  			hash_del(&e->node);
> -			fence_put(f);
> +			dma_fence_put(f);
>  			kmem_cache_free(amdgpu_sync_slab, e);
>  			continue;
>  		}
> @@ -274,11 +276,11 @@ struct fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync,
>   *
>   * Get and removes the next fence from the sync object not signaled yet.
>   */
> -struct fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync)
> +struct dma_fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync)
>  {
>  	struct amdgpu_sync_entry *e;
>  	struct hlist_node *tmp;
> -	struct fence *f;
> +	struct dma_fence *f;
>  	int i;
>  
>  	hash_for_each_safe(sync->fences, i, tmp, e, node) {
> @@ -288,10 +290,10 @@ struct fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync)
>  		hash_del(&e->node);
>  		kmem_cache_free(amdgpu_sync_slab, e);
>  
> -		if (!fence_is_signaled(f))
> +		if (!dma_fence_is_signaled(f))
>  			return f;
>  
> -		fence_put(f);
> +		dma_fence_put(f);
>  	}
>  	return NULL;
>  }
> @@ -311,11 +313,11 @@ void amdgpu_sync_free(struct amdgpu_sync *sync)
>  
>  	hash_for_each_safe(sync->fences, i, tmp, e, node) {
>  		hash_del(&e->node);
> -		fence_put(e->fence);
> +		dma_fence_put(e->fence);
>  		kmem_cache_free(amdgpu_sync_slab, e);
>  	}
>  
> -	fence_put(sync->last_vm_update);
> +	dma_fence_put(sync->last_vm_update);
>  }
>  
>  /**
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_test.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_test.c
> index b827c75e95de..e05a24325eeb 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_test.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_test.c
> @@ -78,7 +78,7 @@ static void amdgpu_do_test_moves(struct amdgpu_device *adev)
>  		void *gtt_map, *vram_map;
>  		void **gtt_start, **gtt_end;
>  		void **vram_start, **vram_end;
> -		struct fence *fence = NULL;
> +		struct dma_fence *fence = NULL;
>  
>  		r = amdgpu_bo_create(adev, size, PAGE_SIZE, true,
>  				     AMDGPU_GEM_DOMAIN_GTT, 0, NULL,
> @@ -118,13 +118,13 @@ static void amdgpu_do_test_moves(struct amdgpu_device *adev)
>  			goto out_lclean_unpin;
>  		}
>  
> -		r = fence_wait(fence, false);
> +		r = dma_fence_wait(fence, false);
>  		if (r) {
>  			DRM_ERROR("Failed to wait for GTT->VRAM fence %d\n", i);
>  			goto out_lclean_unpin;
>  		}
>  
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  
>  		r = amdgpu_bo_kmap(vram_obj, &vram_map);
>  		if (r) {
> @@ -163,13 +163,13 @@ static void amdgpu_do_test_moves(struct amdgpu_device *adev)
>  			goto out_lclean_unpin;
>  		}
>  
> -		r = fence_wait(fence, false);
> +		r = dma_fence_wait(fence, false);
>  		if (r) {
>  			DRM_ERROR("Failed to wait for VRAM->GTT fence %d\n", i);
>  			goto out_lclean_unpin;
>  		}
>  
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  
>  		r = amdgpu_bo_kmap(gtt_obj[i], &gtt_map);
>  		if (r) {
> @@ -216,7 +216,7 @@ static void amdgpu_do_test_moves(struct amdgpu_device *adev)
>  			amdgpu_bo_unref(&gtt_obj[i]);
>  		}
>  		if (fence)
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  		break;
>  	}
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> index 067e5e683bb3..bb964a8ff938 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> @@ -104,7 +104,7 @@ TRACE_EVENT(amdgpu_cs_ioctl,
>  			     __field(struct amdgpu_device *, adev)
>  			     __field(struct amd_sched_job *, sched_job)
>  			     __field(struct amdgpu_ib *, ib)
> -			     __field(struct fence *, fence)
> +			     __field(struct dma_fence *, fence)
>  			     __field(char *, ring_name)
>  			     __field(u32, num_ibs)
>  			     ),
> @@ -129,7 +129,7 @@ TRACE_EVENT(amdgpu_sched_run_job,
>  			     __field(struct amdgpu_device *, adev)
>  			     __field(struct amd_sched_job *, sched_job)
>  			     __field(struct amdgpu_ib *, ib)
> -			     __field(struct fence *, fence)
> +			     __field(struct dma_fence *, fence)
>  			     __field(char *, ring_name)
>  			     __field(u32, num_ibs)
>  			     ),
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> index dcaf691f56b5..a743aeabc767 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> @@ -268,7 +268,7 @@ static int amdgpu_move_blit(struct ttm_buffer_object *bo,
>  	struct amdgpu_device *adev;
>  	struct amdgpu_ring *ring;
>  	uint64_t old_start, new_start;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int r;
>  
>  	adev = amdgpu_get_adev(bo->bdev);
> @@ -316,7 +316,7 @@ static int amdgpu_move_blit(struct ttm_buffer_object *bo,
>  		return r;
>  
>  	r = ttm_bo_pipeline_move(bo, fence, evict, new_mem);
> -	fence_put(fence);
> +	dma_fence_put(fence);
>  	return r;
>  }
>  
> @@ -1247,7 +1247,7 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring,
>  		       uint64_t dst_offset,
>  		       uint32_t byte_count,
>  		       struct reservation_object *resv,
> -		       struct fence **fence, bool direct_submit)
> +		       struct dma_fence **fence, bool direct_submit)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_job *job;
> @@ -1294,7 +1294,7 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring,
>  	if (direct_submit) {
>  		r = amdgpu_ib_schedule(ring, job->num_ibs, job->ibs,
>  				       NULL, NULL, fence);
> -		job->fence = fence_get(*fence);
> +		job->fence = dma_fence_get(*fence);
>  		if (r)
>  			DRM_ERROR("Error scheduling IBs (%d)\n", r);
>  		amdgpu_job_free(job);
> @@ -1315,7 +1315,7 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring,
>  int amdgpu_fill_buffer(struct amdgpu_bo *bo,
>  		uint32_t src_data,
>  		struct reservation_object *resv,
> -		struct fence **fence)
> +		struct dma_fence **fence)
>  {
>  	struct amdgpu_device *adev = bo->adev;
>  	struct amdgpu_job *job;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> index 9812c805326c..3f293e189378 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> @@ -77,11 +77,11 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring,
>  		       uint64_t dst_offset,
>  		       uint32_t byte_count,
>  		       struct reservation_object *resv,
> -		       struct fence **fence, bool direct_submit);
> +		       struct dma_fence **fence, bool direct_submit);
>  int amdgpu_fill_buffer(struct amdgpu_bo *bo,
>  			uint32_t src_data,
>  			struct reservation_object *resv,
> -			struct fence **fence);
> +			struct dma_fence **fence);
>  
>  int amdgpu_mmap(struct file *filp, struct vm_area_struct *vma);
>  bool amdgpu_ttm_is_bound(struct ttm_tt *ttm);
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> index e3281cacc586..0f6575e7ef8c 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
> @@ -333,7 +333,7 @@ void amdgpu_uvd_free_handles(struct amdgpu_device *adev, struct drm_file *filp)
>  	for (i = 0; i < adev->uvd.max_handles; ++i) {
>  		uint32_t handle = atomic_read(&adev->uvd.handles[i]);
>  		if (handle != 0 && adev->uvd.filp[i] == filp) {
> -			struct fence *fence;
> +			struct dma_fence *fence;
>  
>  			r = amdgpu_uvd_get_destroy_msg(ring, handle,
>  						       false, &fence);
> @@ -342,8 +342,8 @@ void amdgpu_uvd_free_handles(struct amdgpu_device *adev, struct drm_file *filp)
>  				continue;
>  			}
>  
> -			fence_wait(fence, false);
> -			fence_put(fence);
> +			dma_fence_wait(fence, false);
> +			dma_fence_put(fence);
>  
>  			adev->uvd.filp[i] = NULL;
>  			atomic_set(&adev->uvd.handles[i], 0);
> @@ -909,14 +909,14 @@ int amdgpu_uvd_ring_parse_cs(struct amdgpu_cs_parser *parser, uint32_t ib_idx)
>  }
>  
>  static int amdgpu_uvd_send_msg(struct amdgpu_ring *ring, struct amdgpu_bo *bo,
> -			       bool direct, struct fence **fence)
> +			       bool direct, struct dma_fence **fence)
>  {
>  	struct ttm_validate_buffer tv;
>  	struct ww_acquire_ctx ticket;
>  	struct list_head head;
>  	struct amdgpu_job *job;
>  	struct amdgpu_ib *ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	struct amdgpu_device *adev = ring->adev;
>  	uint64_t addr;
>  	int i, r;
> @@ -960,7 +960,7 @@ static int amdgpu_uvd_send_msg(struct amdgpu_ring *ring, struct amdgpu_bo *bo,
>  
>  	if (direct) {
>  		r = amdgpu_ib_schedule(ring, 1, ib, NULL, NULL, &f);
> -		job->fence = fence_get(f);
> +		job->fence = dma_fence_get(f);
>  		if (r)
>  			goto err_free;
>  
> @@ -975,9 +975,9 @@ static int amdgpu_uvd_send_msg(struct amdgpu_ring *ring, struct amdgpu_bo *bo,
>  	ttm_eu_fence_buffer_objects(&ticket, &head, f);
>  
>  	if (fence)
> -		*fence = fence_get(f);
> +		*fence = dma_fence_get(f);
>  	amdgpu_bo_unref(&bo);
> -	fence_put(f);
> +	dma_fence_put(f);
>  
>  	return 0;
>  
> @@ -993,7 +993,7 @@ static int amdgpu_uvd_send_msg(struct amdgpu_ring *ring, struct amdgpu_bo *bo,
>     crash the vcpu so just try to emmit a dummy create/destroy msg to
>     avoid this */
>  int amdgpu_uvd_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			      struct fence **fence)
> +			      struct dma_fence **fence)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_bo *bo;
> @@ -1042,7 +1042,7 @@ int amdgpu_uvd_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
>  }
>  
>  int amdgpu_uvd_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			       bool direct, struct fence **fence)
> +			       bool direct, struct dma_fence **fence)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_bo *bo;
> @@ -1128,7 +1128,7 @@ void amdgpu_uvd_ring_end_use(struct amdgpu_ring *ring)
>   */
>  int amdgpu_uvd_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	long r;
>  
>  	r = amdgpu_uvd_get_create_msg(ring, 1, NULL);
> @@ -1143,7 +1143,7 @@ int amdgpu_uvd_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  		goto error;
>  	}
>  
> -	r = fence_wait_timeout(fence, false, timeout);
> +	r = dma_fence_wait_timeout(fence, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out.\n");
>  		r = -ETIMEDOUT;
> @@ -1154,7 +1154,7 @@ int amdgpu_uvd_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  		r = 0;
>  	}
>  
> -	fence_put(fence);
> +	dma_fence_put(fence);
>  
>  error:
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h
> index c850009602d1..6249ba1bde2a 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h
> @@ -29,9 +29,9 @@ int amdgpu_uvd_sw_fini(struct amdgpu_device *adev);
>  int amdgpu_uvd_suspend(struct amdgpu_device *adev);
>  int amdgpu_uvd_resume(struct amdgpu_device *adev);
>  int amdgpu_uvd_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			      struct fence **fence);
> +			      struct dma_fence **fence);
>  int amdgpu_uvd_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			       bool direct, struct fence **fence);
> +			       bool direct, struct dma_fence **fence);
>  void amdgpu_uvd_free_handles(struct amdgpu_device *adev,
>  			     struct drm_file *filp);
>  int amdgpu_uvd_ring_parse_cs(struct amdgpu_cs_parser *parser, uint32_t ib_idx);
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
> index 7fe8fd884f06..f0f8afb85585 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
> @@ -395,12 +395,12 @@ void amdgpu_vce_free_handles(struct amdgpu_device *adev, struct drm_file *filp)
>   * Open up a stream for HW test
>   */
>  int amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			      struct fence **fence)
> +			      struct dma_fence **fence)
>  {
>  	const unsigned ib_size_dw = 1024;
>  	struct amdgpu_job *job;
>  	struct amdgpu_ib *ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	uint64_t dummy;
>  	int i, r;
>  
> @@ -450,14 +450,14 @@ int amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
>  		ib->ptr[i] = 0x0;
>  
>  	r = amdgpu_ib_schedule(ring, 1, ib, NULL, NULL, &f);
> -	job->fence = fence_get(f);
> +	job->fence = dma_fence_get(f);
>  	if (r)
>  		goto err;
>  
>  	amdgpu_job_free(job);
>  	if (fence)
> -		*fence = fence_get(f);
> -	fence_put(f);
> +		*fence = dma_fence_get(f);
> +	dma_fence_put(f);
>  	return 0;
>  
>  err:
> @@ -476,12 +476,12 @@ int amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
>   * Close up a stream for HW test or if userspace failed to do so
>   */
>  int amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			       bool direct, struct fence **fence)
> +			       bool direct, struct dma_fence **fence)
>  {
>  	const unsigned ib_size_dw = 1024;
>  	struct amdgpu_job *job;
>  	struct amdgpu_ib *ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	int i, r;
>  
>  	r = amdgpu_job_alloc_with_ib(ring->adev, ib_size_dw * 4, &job);
> @@ -513,7 +513,7 @@ int amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
>  
>  	if (direct) {
>  		r = amdgpu_ib_schedule(ring, 1, ib, NULL, NULL, &f);
> -		job->fence = fence_get(f);
> +		job->fence = dma_fence_get(f);
>  		if (r)
>  			goto err;
>  
> @@ -526,8 +526,8 @@ int amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
>  	}
>  
>  	if (fence)
> -		*fence = fence_get(f);
> -	fence_put(f);
> +		*fence = dma_fence_get(f);
> +	dma_fence_put(f);
>  	return 0;
>  
>  err:
> @@ -883,7 +883,7 @@ int amdgpu_vce_ring_test_ring(struct amdgpu_ring *ring)
>   */
>  int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  	long r;
>  
>  	/* skip vce ring1/2 ib test for now, since it's not reliable */
> @@ -902,7 +902,7 @@ int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  		goto error;
>  	}
>  
> -	r = fence_wait_timeout(fence, false, timeout);
> +	r = dma_fence_wait_timeout(fence, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out.\n");
>  		r = -ETIMEDOUT;
> @@ -913,6 +913,6 @@ int amdgpu_vce_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  		r = 0;
>  	}
>  error:
> -	fence_put(fence);
> +	dma_fence_put(fence);
>  	return r;
>  }
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h
> index 12729d2852df..566c29ddeeb6 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h
> @@ -29,9 +29,9 @@ int amdgpu_vce_sw_fini(struct amdgpu_device *adev);
>  int amdgpu_vce_suspend(struct amdgpu_device *adev);
>  int amdgpu_vce_resume(struct amdgpu_device *adev);
>  int amdgpu_vce_get_create_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			      struct fence **fence);
> +			      struct dma_fence **fence);
>  int amdgpu_vce_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle,
> -			       bool direct, struct fence **fence);
> +			       bool direct, struct dma_fence **fence);
>  void amdgpu_vce_free_handles(struct amdgpu_device *adev, struct drm_file *filp);
>  int amdgpu_vce_ring_parse_cs(struct amdgpu_cs_parser *p, uint32_t ib_idx);
>  void amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index 06f24322e7c3..31713f18208f 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -25,7 +25,7 @@
>   *          Alex Deucher
>   *          Jerome Glisse
>   */
> -#include <linux/fence-array.h>
> +#include <linux/dma-fence-array.h>
>  #include <drm/drmP.h>
>  #include <drm/amdgpu_drm.h>
>  #include "amdgpu.h"
> @@ -194,14 +194,14 @@ static bool amdgpu_vm_is_gpu_reset(struct amdgpu_device *adev,
>   * Allocate an id for the vm, adding fences to the sync obj as necessary.
>   */
>  int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
> -		      struct amdgpu_sync *sync, struct fence *fence,
> +		      struct amdgpu_sync *sync, struct dma_fence *fence,
>  		      struct amdgpu_job *job)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	uint64_t fence_context = adev->fence_context + ring->idx;
> -	struct fence *updates = sync->last_vm_update;
> +	struct dma_fence *updates = sync->last_vm_update;
>  	struct amdgpu_vm_id *id, *idle;
> -	struct fence **fences;
> +	struct dma_fence **fences;
>  	unsigned i;
>  	int r = 0;
>  
> @@ -225,17 +225,17 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
>  	if (&idle->list == &adev->vm_manager.ids_lru) {
>  		u64 fence_context = adev->vm_manager.fence_context + ring->idx;
>  		unsigned seqno = ++adev->vm_manager.seqno[ring->idx];
> -		struct fence_array *array;
> +		struct dma_fence_array *array;
>  		unsigned j;
>  
>  		for (j = 0; j < i; ++j)
> -			fence_get(fences[j]);
> +			dma_fence_get(fences[j]);
>  
> -		array = fence_array_create(i, fences, fence_context,
> +		array = dma_fence_array_create(i, fences, fence_context,
>  					   seqno, true);
>  		if (!array) {
>  			for (j = 0; j < i; ++j)
> -				fence_put(fences[j]);
> +				dma_fence_put(fences[j]);
>  			kfree(fences);
>  			r = -ENOMEM;
>  			goto error;
> @@ -243,7 +243,7 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
>  
>  
>  		r = amdgpu_sync_fence(ring->adev, sync, &array->base);
> -		fence_put(&array->base);
> +		dma_fence_put(&array->base);
>  		if (r)
>  			goto error;
>  
> @@ -257,7 +257,7 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
>  	/* Check if we can use a VMID already assigned to this VM */
>  	i = ring->idx;
>  	do {
> -		struct fence *flushed;
> +		struct dma_fence *flushed;
>  
>  		id = vm->ids[i++];
>  		if (i == AMDGPU_MAX_RINGS)
> @@ -279,12 +279,11 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
>  			continue;
>  
>  		if (id->last_flush->context != fence_context &&
> -		    !fence_is_signaled(id->last_flush))
> -			continue;
> +		    !dma_fence_is_signaled(id->last_flush))
>  
>  		flushed  = id->flushed_updates;
>  		if (updates &&
> -		    (!flushed || fence_is_later(updates, flushed)))
> +		    (!flushed || dma_fence_is_later(updates, flushed)))
>  			continue;
>  
>  		/* Good we can use this VMID. Remember this submission as
> @@ -315,14 +314,14 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
>  	if (r)
>  		goto error;
>  
> -	fence_put(id->first);
> -	id->first = fence_get(fence);
> +	dma_fence_put(id->first);
> +	id->first = dma_fence_get(fence);
>  
> -	fence_put(id->last_flush);
> +	dma_fence_put(id->last_flush);
>  	id->last_flush = NULL;
>  
> -	fence_put(id->flushed_updates);
> -	id->flushed_updates = fence_get(updates);
> +	dma_fence_put(id->flushed_updates);
> +	id->flushed_updates = dma_fence_get(updates);
>  
>  	id->pd_gpu_addr = job->vm_pd_addr;
>  	id->current_gpu_reset_count = atomic_read(&adev->gpu_reset_counter);
> @@ -393,7 +392,7 @@ int amdgpu_vm_flush(struct amdgpu_ring *ring, struct amdgpu_job *job)
>  
>  	if (ring->funcs->emit_vm_flush && (job->vm_needs_flush ||
>  	    amdgpu_vm_is_gpu_reset(adev, id))) {
> -		struct fence *fence;
> +		struct dma_fence *fence;
>  
>  		trace_amdgpu_vm_flush(job->vm_pd_addr, ring->idx, job->vm_id);
>  		amdgpu_ring_emit_vm_flush(ring, job->vm_id, job->vm_pd_addr);
> @@ -403,7 +402,7 @@ int amdgpu_vm_flush(struct amdgpu_ring *ring, struct amdgpu_job *job)
>  			return r;
>  
>  		mutex_lock(&adev->vm_manager.lock);
> -		fence_put(id->last_flush);
> +		dma_fence_put(id->last_flush);
>  		id->last_flush = fence;
>  		mutex_unlock(&adev->vm_manager.lock);
>  	}
> @@ -537,7 +536,7 @@ static int amdgpu_vm_clear_bo(struct amdgpu_device *adev,
>  			      struct amdgpu_bo *bo)
>  {
>  	struct amdgpu_ring *ring;
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  	struct amdgpu_job *job;
>  	struct amdgpu_pte_update_params params;
>  	unsigned entries;
> @@ -578,7 +577,7 @@ static int amdgpu_vm_clear_bo(struct amdgpu_device *adev,
>  		goto error_free;
>  
>  	amdgpu_bo_fence(bo, fence, true);
> -	fence_put(fence);
> +	dma_fence_put(fence);
>  	return 0;
>  
>  error_free:
> @@ -625,7 +624,7 @@ static int amdgpu_vm_update_pd_or_shadow(struct amdgpu_device *adev,
>  	unsigned count = 0, pt_idx, ndw;
>  	struct amdgpu_job *job;
>  	struct amdgpu_pte_update_params params;
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  
>  	int r;
>  
> @@ -714,9 +713,9 @@ static int amdgpu_vm_update_pd_or_shadow(struct amdgpu_device *adev,
>  			goto error_free;
>  
>  		amdgpu_bo_fence(pd, fence, true);
> -		fence_put(vm->page_directory_fence);
> -		vm->page_directory_fence = fence_get(fence);
> -		fence_put(fence);
> +		dma_fence_put(vm->page_directory_fence);
> +		vm->page_directory_fence = dma_fence_get(fence);
> +		dma_fence_put(fence);
>  
>  	} else {
>  		amdgpu_job_free(job);
> @@ -929,20 +928,20 @@ static void amdgpu_vm_frag_ptes(struct amdgpu_pte_update_params	*params,
>   * Returns 0 for success, -EINVAL for failure.
>   */
>  static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
> -				       struct fence *exclusive,
> +				       struct dma_fence *exclusive,
>  				       uint64_t src,
>  				       dma_addr_t *pages_addr,
>  				       struct amdgpu_vm *vm,
>  				       uint64_t start, uint64_t last,
>  				       uint32_t flags, uint64_t addr,
> -				       struct fence **fence)
> +				       struct dma_fence **fence)
>  {
>  	struct amdgpu_ring *ring;
>  	void *owner = AMDGPU_FENCE_OWNER_VM;
>  	unsigned nptes, ncmds, ndw;
>  	struct amdgpu_job *job;
>  	struct amdgpu_pte_update_params params;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	int r;
>  
>  	memset(&params, 0, sizeof(params));
> @@ -1045,10 +1044,10 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  
>  	amdgpu_bo_fence(vm->page_directory, f, true);
>  	if (fence) {
> -		fence_put(*fence);
> -		*fence = fence_get(f);
> +		dma_fence_put(*fence);
> +		*fence = dma_fence_get(f);
>  	}
> -	fence_put(f);
> +	dma_fence_put(f);
>  	return 0;
>  
>  error_free:
> @@ -1074,13 +1073,13 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   * Returns 0 for success, -EINVAL for failure.
>   */
>  static int amdgpu_vm_bo_split_mapping(struct amdgpu_device *adev,
> -				      struct fence *exclusive,
> +				      struct dma_fence *exclusive,
>  				      uint32_t gtt_flags,
>  				      dma_addr_t *pages_addr,
>  				      struct amdgpu_vm *vm,
>  				      struct amdgpu_bo_va_mapping *mapping,
>  				      uint32_t flags, uint64_t addr,
> -				      struct fence **fence)
> +				      struct dma_fence **fence)
>  {
>  	const uint64_t max_size = 64ULL * 1024ULL * 1024ULL / AMDGPU_GPU_PAGE_SIZE;
>  
> @@ -1147,7 +1146,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>  	dma_addr_t *pages_addr = NULL;
>  	uint32_t gtt_flags, flags;
>  	struct ttm_mem_reg *mem;
> -	struct fence *exclusive;
> +	struct dma_fence *exclusive;
>  	uint64_t addr;
>  	int r;
>  
> @@ -1547,7 +1546,7 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
>  		kfree(mapping);
>  	}
>  
> -	fence_put(bo_va->last_pt_update);
> +	dma_fence_put(bo_va->last_pt_update);
>  	kfree(bo_va);
>  }
>  
> @@ -1709,7 +1708,7 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
>  
>  	amdgpu_bo_unref(&vm->page_directory->shadow);
>  	amdgpu_bo_unref(&vm->page_directory);
> -	fence_put(vm->page_directory_fence);
> +	dma_fence_put(vm->page_directory_fence);
>  }
>  
>  /**
> @@ -1733,7 +1732,8 @@ void amdgpu_vm_manager_init(struct amdgpu_device *adev)
>  			      &adev->vm_manager.ids_lru);
>  	}
>  
> -	adev->vm_manager.fence_context = fence_context_alloc(AMDGPU_MAX_RINGS);
> +	adev->vm_manager.fence_context =
> +		dma_fence_context_alloc(AMDGPU_MAX_RINGS);
>  	for (i = 0; i < AMDGPU_MAX_RINGS; ++i)
>  		adev->vm_manager.seqno[i] = 0;
>  
> @@ -1755,8 +1755,8 @@ void amdgpu_vm_manager_fini(struct amdgpu_device *adev)
>  	for (i = 0; i < AMDGPU_NUM_VM; ++i) {
>  		struct amdgpu_vm_id *id = &adev->vm_manager.ids[i];
>  
> -		fence_put(adev->vm_manager.ids[i].first);
> +		dma_fence_put(adev->vm_manager.ids[i].first);
>  		amdgpu_sync_free(&adev->vm_manager.ids[i].active);
> -		fence_put(id->flushed_updates);
> +		dma_fence_put(id->flushed_updates);
>  	}
>  }
> diff --git a/drivers/gpu/drm/amd/amdgpu/cik_sdma.c b/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
> index cb952acc7133..321b9d5a4e6e 100644
> --- a/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
> +++ b/drivers/gpu/drm/amd/amdgpu/cik_sdma.c
> @@ -622,7 +622,7 @@ static int cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	unsigned index;
>  	u32 tmp = 0;
>  	u64 gpu_addr;
> @@ -655,7 +655,7 @@ static int cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err1;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -675,7 +675,7 @@ static int cik_sdma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  
>  err1:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err0:
>  	amdgpu_wb_free(adev, index);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c b/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
> index 40abb6b81c09..7dc11a19e49d 100644
> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
> @@ -1522,7 +1522,7 @@ static int gfx_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	uint32_t scratch;
>  	uint32_t tmp = 0;
>  	long r;
> @@ -1548,7 +1548,7 @@ static int gfx_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err2;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -1569,7 +1569,7 @@ static int gfx_v6_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  
>  err2:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err1:
>  	amdgpu_gfx_scratch_free(adev, scratch);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c b/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
> index 71116da9e782..3865ffe7de55 100644
> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
> @@ -2286,7 +2286,7 @@ static int gfx_v7_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	uint32_t scratch;
>  	uint32_t tmp = 0;
>  	long r;
> @@ -2312,7 +2312,7 @@ static int gfx_v7_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err2;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -2333,7 +2333,7 @@ static int gfx_v7_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  
>  err2:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err1:
>  	amdgpu_gfx_scratch_free(adev, scratch);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
> index ee6a48a09214..a9dd18847c40 100644
> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
> @@ -798,7 +798,7 @@ static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	uint32_t scratch;
>  	uint32_t tmp = 0;
>  	long r;
> @@ -824,7 +824,7 @@ static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err2;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out.\n");
>  		r = -ETIMEDOUT;
> @@ -844,7 +844,7 @@ static int gfx_v8_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	}
>  err2:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err1:
>  	amdgpu_gfx_scratch_free(adev, scratch);
>  	return r;
> @@ -1575,7 +1575,7 @@ static int gfx_v8_0_do_edc_gpr_workarounds(struct amdgpu_device *adev)
>  {
>  	struct amdgpu_ring *ring = &adev->gfx.compute_ring[0];
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	int r, i;
>  	u32 tmp;
>  	unsigned total_size, vgpr_offset, sgpr_offset;
> @@ -1708,7 +1708,7 @@ static int gfx_v8_0_do_edc_gpr_workarounds(struct amdgpu_device *adev)
>  	}
>  
>  	/* wait for the GPU to finish processing the IB */
> -	r = fence_wait(f, false);
> +	r = dma_fence_wait(f, false);
>  	if (r) {
>  		DRM_ERROR("amdgpu: fence wait failed (%d).\n", r);
>  		goto fail;
> @@ -1729,7 +1729,7 @@ static int gfx_v8_0_do_edc_gpr_workarounds(struct amdgpu_device *adev)
>  
>  fail:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  
>  	return r;
>  }
> diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c b/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
> index 565dab3c7218..7edf6e8c63dc 100644
> --- a/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
> +++ b/drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
> @@ -668,7 +668,7 @@ static int sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	unsigned index;
>  	u32 tmp = 0;
>  	u64 gpu_addr;
> @@ -705,7 +705,7 @@ static int sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err1;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -725,7 +725,7 @@ static int sdma_v2_4_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  
>  err1:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err0:
>  	amdgpu_wb_free(adev, index);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c b/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
> index a9d10941fb53..1932a67c62ef 100644
> --- a/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
> @@ -871,7 +871,7 @@ static int sdma_v3_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	unsigned index;
>  	u32 tmp = 0;
>  	u64 gpu_addr;
> @@ -908,7 +908,7 @@ static int sdma_v3_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err1;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -927,7 +927,7 @@ static int sdma_v3_0_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	}
>  err1:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err0:
>  	amdgpu_wb_free(adev, index);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/si_dma.c b/drivers/gpu/drm/amd/amdgpu/si_dma.c
> index de358193a8f9..b4cf4e25bf91 100644
> --- a/drivers/gpu/drm/amd/amdgpu/si_dma.c
> +++ b/drivers/gpu/drm/amd/amdgpu/si_dma.c
> @@ -274,7 +274,7 @@ static int si_dma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  {
>  	struct amdgpu_device *adev = ring->adev;
>  	struct amdgpu_ib ib;
> -	struct fence *f = NULL;
> +	struct dma_fence *f = NULL;
>  	unsigned index;
>  	u32 tmp = 0;
>  	u64 gpu_addr;
> @@ -305,7 +305,7 @@ static int si_dma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  	if (r)
>  		goto err1;
>  
> -	r = fence_wait_timeout(f, false, timeout);
> +	r = dma_fence_wait_timeout(f, false, timeout);
>  	if (r == 0) {
>  		DRM_ERROR("amdgpu: IB test timed out\n");
>  		r = -ETIMEDOUT;
> @@ -325,7 +325,7 @@ static int si_dma_ring_test_ib(struct amdgpu_ring *ring, long timeout)
>  
>  err1:
>  	amdgpu_ib_free(adev, &ib, NULL);
> -	fence_put(f);
> +	dma_fence_put(f);
>  err0:
>  	amdgpu_wb_free(adev, index);
>  	return r;
> diff --git a/drivers/gpu/drm/amd/scheduler/gpu_sched_trace.h b/drivers/gpu/drm/amd/scheduler/gpu_sched_trace.h
> index b961a1c6caf3..dbd4fd3a810b 100644
> --- a/drivers/gpu/drm/amd/scheduler/gpu_sched_trace.h
> +++ b/drivers/gpu/drm/amd/scheduler/gpu_sched_trace.h
> @@ -17,7 +17,7 @@ TRACE_EVENT(amd_sched_job,
>  	    TP_STRUCT__entry(
>  			     __field(struct amd_sched_entity *, entity)
>  			     __field(struct amd_sched_job *, sched_job)
> -			     __field(struct fence *, fence)
> +			     __field(struct dma_fence *, fence)
>  			     __field(const char *, name)
>  			     __field(u32, job_count)
>  			     __field(int, hw_job_count)
> @@ -42,7 +42,7 @@ TRACE_EVENT(amd_sched_process_job,
>  	    TP_PROTO(struct amd_sched_fence *fence),
>  	    TP_ARGS(fence),
>  	    TP_STRUCT__entry(
> -		    __field(struct fence *, fence)
> +		    __field(struct dma_fence *, fence)
>  		    ),
>  
>  	    TP_fast_assign(
> diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
> index 963a24d46a93..5364e6a7ec8f 100644
> --- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
> +++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
> @@ -32,7 +32,7 @@
>  
>  static bool amd_sched_entity_is_ready(struct amd_sched_entity *entity);
>  static void amd_sched_wakeup(struct amd_gpu_scheduler *sched);
> -static void amd_sched_process_job(struct fence *f, struct fence_cb *cb);
> +static void amd_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb);
>  
>  struct kmem_cache *sched_fence_slab;
>  atomic_t sched_fence_slab_ref = ATOMIC_INIT(0);
> @@ -141,7 +141,7 @@ int amd_sched_entity_init(struct amd_gpu_scheduler *sched,
>  		return r;
>  
>  	atomic_set(&entity->fence_seq, 0);
> -	entity->fence_context = fence_context_alloc(2);
> +	entity->fence_context = dma_fence_context_alloc(2);
>  
>  	return 0;
>  }
> @@ -221,32 +221,32 @@ void amd_sched_entity_fini(struct amd_gpu_scheduler *sched,
>  	kfifo_free(&entity->job_queue);
>  }
>  
> -static void amd_sched_entity_wakeup(struct fence *f, struct fence_cb *cb)
> +static void amd_sched_entity_wakeup(struct dma_fence *f, struct dma_fence_cb *cb)
>  {
>  	struct amd_sched_entity *entity =
>  		container_of(cb, struct amd_sched_entity, cb);
>  	entity->dependency = NULL;
> -	fence_put(f);
> +	dma_fence_put(f);
>  	amd_sched_wakeup(entity->sched);
>  }
>  
> -static void amd_sched_entity_clear_dep(struct fence *f, struct fence_cb *cb)
> +static void amd_sched_entity_clear_dep(struct dma_fence *f, struct dma_fence_cb *cb)
>  {
>  	struct amd_sched_entity *entity =
>  		container_of(cb, struct amd_sched_entity, cb);
>  	entity->dependency = NULL;
> -	fence_put(f);
> +	dma_fence_put(f);
>  }
>  
>  static bool amd_sched_entity_add_dependency_cb(struct amd_sched_entity *entity)
>  {
>  	struct amd_gpu_scheduler *sched = entity->sched;
> -	struct fence * fence = entity->dependency;
> +	struct dma_fence * fence = entity->dependency;
>  	struct amd_sched_fence *s_fence;
>  
>  	if (fence->context == entity->fence_context) {
>  		/* We can ignore fences from ourself */
> -		fence_put(entity->dependency);
> +		dma_fence_put(entity->dependency);
>  		return false;
>  	}
>  
> @@ -257,23 +257,23 @@ static bool amd_sched_entity_add_dependency_cb(struct amd_sched_entity *entity)
>  		 * Fence is from the same scheduler, only need to wait for
>  		 * it to be scheduled
>  		 */
> -		fence = fence_get(&s_fence->scheduled);
> -		fence_put(entity->dependency);
> +		fence = dma_fence_get(&s_fence->scheduled);
> +		dma_fence_put(entity->dependency);
>  		entity->dependency = fence;
> -		if (!fence_add_callback(fence, &entity->cb,
> -					amd_sched_entity_clear_dep))
> +		if (!dma_fence_add_callback(fence, &entity->cb,
> +					    amd_sched_entity_clear_dep))
>  			return true;
>  
>  		/* Ignore it when it is already scheduled */
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  		return false;
>  	}
>  
> -	if (!fence_add_callback(entity->dependency, &entity->cb,
> -				amd_sched_entity_wakeup))
> +	if (!dma_fence_add_callback(entity->dependency, &entity->cb,
> +				    amd_sched_entity_wakeup))
>  		return true;
>  
> -	fence_put(entity->dependency);
> +	dma_fence_put(entity->dependency);
>  	return false;
>  }
>  
> @@ -354,7 +354,8 @@ static void amd_sched_job_finish(struct work_struct *work)
>  	sched->ops->free_job(s_job);
>  }
>  
> -static void amd_sched_job_finish_cb(struct fence *f, struct fence_cb *cb)
> +static void amd_sched_job_finish_cb(struct dma_fence *f,
> +				    struct dma_fence_cb *cb)
>  {
>  	struct amd_sched_job *job = container_of(cb, struct amd_sched_job,
>  						 finish_cb);
> @@ -388,8 +389,8 @@ void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched)
>  
>  	spin_lock(&sched->job_list_lock);
>  	list_for_each_entry_reverse(s_job, &sched->ring_mirror_list, node) {
> -		if (fence_remove_callback(s_job->s_fence->parent, &s_job->s_fence->cb)) {
> -			fence_put(s_job->s_fence->parent);
> +		if (dma_fence_remove_callback(s_job->s_fence->parent, &s_job->s_fence->cb)) {
> +			dma_fence_put(s_job->s_fence->parent);
>  			s_job->s_fence->parent = NULL;
>  		}
>  	}
> @@ -410,21 +411,21 @@ void amd_sched_job_recovery(struct amd_gpu_scheduler *sched)
>  
>  	list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) {
>  		struct amd_sched_fence *s_fence = s_job->s_fence;
> -		struct fence *fence;
> +		struct dma_fence *fence;
>  
>  		spin_unlock(&sched->job_list_lock);
>  		fence = sched->ops->run_job(s_job);
>  		atomic_inc(&sched->hw_rq_count);
>  		if (fence) {
> -			s_fence->parent = fence_get(fence);
> -			r = fence_add_callback(fence, &s_fence->cb,
> -					       amd_sched_process_job);
> +			s_fence->parent = dma_fence_get(fence);
> +			r = dma_fence_add_callback(fence, &s_fence->cb,
> +						   amd_sched_process_job);
>  			if (r == -ENOENT)
>  				amd_sched_process_job(fence, &s_fence->cb);
>  			else if (r)
>  				DRM_ERROR("fence add callback failed (%d)\n",
>  					  r);
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  		} else {
>  			DRM_ERROR("Failed to run job!\n");
>  			amd_sched_process_job(NULL, &s_fence->cb);
> @@ -446,8 +447,8 @@ void amd_sched_entity_push_job(struct amd_sched_job *sched_job)
>  	struct amd_sched_entity *entity = sched_job->s_entity;
>  
>  	trace_amd_sched_job(sched_job);
> -	fence_add_callback(&sched_job->s_fence->finished, &sched_job->finish_cb,
> -			   amd_sched_job_finish_cb);
> +	dma_fence_add_callback(&sched_job->s_fence->finished, &sched_job->finish_cb,
> +			       amd_sched_job_finish_cb);
>  	wait_event(entity->sched->job_scheduled,
>  		   amd_sched_entity_in(sched_job));
>  }
> @@ -511,7 +512,7 @@ amd_sched_select_entity(struct amd_gpu_scheduler *sched)
>  	return entity;
>  }
>  
> -static void amd_sched_process_job(struct fence *f, struct fence_cb *cb)
> +static void amd_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb)
>  {
>  	struct amd_sched_fence *s_fence =
>  		container_of(cb, struct amd_sched_fence, cb);
> @@ -521,7 +522,7 @@ static void amd_sched_process_job(struct fence *f, struct fence_cb *cb)
>  	amd_sched_fence_finished(s_fence);
>  
>  	trace_amd_sched_process_job(s_fence);
> -	fence_put(&s_fence->finished);
> +	dma_fence_put(&s_fence->finished);
>  	wake_up_interruptible(&sched->wake_up_worker);
>  }
>  
> @@ -547,7 +548,7 @@ static int amd_sched_main(void *param)
>  		struct amd_sched_entity *entity = NULL;
>  		struct amd_sched_fence *s_fence;
>  		struct amd_sched_job *sched_job;
> -		struct fence *fence;
> +		struct dma_fence *fence;
>  
>  		wait_event_interruptible(sched->wake_up_worker,
>  					 (!amd_sched_blocked(sched) &&
> @@ -569,15 +570,15 @@ static int amd_sched_main(void *param)
>  		fence = sched->ops->run_job(sched_job);
>  		amd_sched_fence_scheduled(s_fence);
>  		if (fence) {
> -			s_fence->parent = fence_get(fence);
> -			r = fence_add_callback(fence, &s_fence->cb,
> -					       amd_sched_process_job);
> +			s_fence->parent = dma_fence_get(fence);
> +			r = dma_fence_add_callback(fence, &s_fence->cb,
> +						   amd_sched_process_job);
>  			if (r == -ENOENT)
>  				amd_sched_process_job(fence, &s_fence->cb);
>  			else if (r)
>  				DRM_ERROR("fence add callback failed (%d)\n",
>  					  r);
> -			fence_put(fence);
> +			dma_fence_put(fence);
>  		} else {
>  			DRM_ERROR("Failed to run job!\n");
>  			amd_sched_process_job(NULL, &s_fence->cb);
> diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
> index 7cbbbfb502ef..876aa43b57df 100644
> --- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
> +++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
> @@ -25,7 +25,7 @@
>  #define _GPU_SCHEDULER_H_
>  
>  #include <linux/kfifo.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  struct amd_gpu_scheduler;
>  struct amd_sched_rq;
> @@ -50,8 +50,8 @@ struct amd_sched_entity {
>  	atomic_t			fence_seq;
>  	uint64_t                        fence_context;
>  
> -	struct fence			*dependency;
> -	struct fence_cb			cb;
> +	struct dma_fence		*dependency;
> +	struct dma_fence_cb		cb;
>  };
>  
>  /**
> @@ -66,10 +66,10 @@ struct amd_sched_rq {
>  };
>  
>  struct amd_sched_fence {
> -	struct fence                    scheduled;
> -	struct fence                    finished;
> -	struct fence_cb                 cb;
> -	struct fence                    *parent;
> +	struct dma_fence                scheduled;
> +	struct dma_fence                finished;
> +	struct dma_fence_cb             cb;
> +	struct dma_fence                *parent;
>  	struct amd_gpu_scheduler	*sched;
>  	spinlock_t			lock;
>  	void                            *owner;
> @@ -79,15 +79,15 @@ struct amd_sched_job {
>  	struct amd_gpu_scheduler        *sched;
>  	struct amd_sched_entity         *s_entity;
>  	struct amd_sched_fence          *s_fence;
> -	struct fence_cb			finish_cb;
> +	struct dma_fence_cb		finish_cb;
>  	struct work_struct		finish_work;
>  	struct list_head		node;
>  	struct delayed_work		work_tdr;
>  };
>  
> -extern const struct fence_ops amd_sched_fence_ops_scheduled;
> -extern const struct fence_ops amd_sched_fence_ops_finished;
> -static inline struct amd_sched_fence *to_amd_sched_fence(struct fence *f)
> +extern const struct dma_fence_ops amd_sched_fence_ops_scheduled;
> +extern const struct dma_fence_ops amd_sched_fence_ops_finished;
> +static inline struct amd_sched_fence *to_amd_sched_fence(struct dma_fence *f)
>  {
>  	if (f->ops == &amd_sched_fence_ops_scheduled)
>  		return container_of(f, struct amd_sched_fence, scheduled);
> @@ -103,8 +103,8 @@ static inline struct amd_sched_fence *to_amd_sched_fence(struct fence *f)
>   * these functions should be implemented in driver side
>  */
>  struct amd_sched_backend_ops {
> -	struct fence *(*dependency)(struct amd_sched_job *sched_job);
> -	struct fence *(*run_job)(struct amd_sched_job *sched_job);
> +	struct dma_fence *(*dependency)(struct amd_sched_job *sched_job);
> +	struct dma_fence *(*run_job)(struct amd_sched_job *sched_job);
>  	void (*timedout_job)(struct amd_sched_job *sched_job);
>  	void (*free_job)(struct amd_sched_job *sched_job);
>  };
> diff --git a/drivers/gpu/drm/amd/scheduler/sched_fence.c b/drivers/gpu/drm/amd/scheduler/sched_fence.c
> index 6b63beaf7574..c26fa298fe9e 100644
> --- a/drivers/gpu/drm/amd/scheduler/sched_fence.c
> +++ b/drivers/gpu/drm/amd/scheduler/sched_fence.c
> @@ -42,46 +42,50 @@ struct amd_sched_fence *amd_sched_fence_create(struct amd_sched_entity *entity,
>  	spin_lock_init(&fence->lock);
>  
>  	seq = atomic_inc_return(&entity->fence_seq);
> -	fence_init(&fence->scheduled, &amd_sched_fence_ops_scheduled,
> -		   &fence->lock, entity->fence_context, seq);
> -	fence_init(&fence->finished, &amd_sched_fence_ops_finished,
> -		   &fence->lock, entity->fence_context + 1, seq);
> +	dma_fence_init(&fence->scheduled, &amd_sched_fence_ops_scheduled,
> +		       &fence->lock, entity->fence_context, seq);
> +	dma_fence_init(&fence->finished, &amd_sched_fence_ops_finished,
> +		       &fence->lock, entity->fence_context + 1, seq);
>  
>  	return fence;
>  }
>  
>  void amd_sched_fence_scheduled(struct amd_sched_fence *fence)
>  {
> -	int ret = fence_signal(&fence->scheduled);
> +	int ret = dma_fence_signal(&fence->scheduled);
>  
>  	if (!ret)
> -		FENCE_TRACE(&fence->scheduled, "signaled from irq context\n");
> +		DMA_FENCE_TRACE(&fence->scheduled,
> +				"signaled from irq context\n");
>  	else
> -		FENCE_TRACE(&fence->scheduled, "was already signaled\n");
> +		DMA_FENCE_TRACE(&fence->scheduled,
> +				"was already signaled\n");
>  }
>  
>  void amd_sched_fence_finished(struct amd_sched_fence *fence)
>  {
> -	int ret = fence_signal(&fence->finished);
> +	int ret = dma_fence_signal(&fence->finished);
>  
>  	if (!ret)
> -		FENCE_TRACE(&fence->finished, "signaled from irq context\n");
> +		DMA_FENCE_TRACE(&fence->finished,
> +				"signaled from irq context\n");
>  	else
> -		FENCE_TRACE(&fence->finished, "was already signaled\n");
> +		DMA_FENCE_TRACE(&fence->finished,
> +				"was already signaled\n");
>  }
>  
> -static const char *amd_sched_fence_get_driver_name(struct fence *fence)
> +static const char *amd_sched_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "amd_sched";
>  }
>  
> -static const char *amd_sched_fence_get_timeline_name(struct fence *f)
> +static const char *amd_sched_fence_get_timeline_name(struct dma_fence *f)
>  {
>  	struct amd_sched_fence *fence = to_amd_sched_fence(f);
>  	return (const char *)fence->sched->name;
>  }
>  
> -static bool amd_sched_fence_enable_signaling(struct fence *f)
> +static bool amd_sched_fence_enable_signaling(struct dma_fence *f)
>  {
>  	return true;
>  }
> @@ -95,10 +99,10 @@ static bool amd_sched_fence_enable_signaling(struct fence *f)
>   */
>  static void amd_sched_fence_free(struct rcu_head *rcu)
>  {
> -	struct fence *f = container_of(rcu, struct fence, rcu);
> +	struct dma_fence *f = container_of(rcu, struct dma_fence, rcu);
>  	struct amd_sched_fence *fence = to_amd_sched_fence(f);
>  
> -	fence_put(fence->parent);
> +	dma_fence_put(fence->parent);
>  	kmem_cache_free(sched_fence_slab, fence);
>  }
>  
> @@ -110,7 +114,7 @@ static void amd_sched_fence_free(struct rcu_head *rcu)
>   * This function is called when the reference count becomes zero.
>   * It just RCU schedules freeing up the fence.
>   */
> -static void amd_sched_fence_release_scheduled(struct fence *f)
> +static void amd_sched_fence_release_scheduled(struct dma_fence *f)
>  {
>  	struct amd_sched_fence *fence = to_amd_sched_fence(f);
>  
> @@ -124,27 +128,27 @@ static void amd_sched_fence_release_scheduled(struct fence *f)
>   *
>   * Drop the extra reference from the scheduled fence to the base fence.
>   */
> -static void amd_sched_fence_release_finished(struct fence *f)
> +static void amd_sched_fence_release_finished(struct dma_fence *f)
>  {
>  	struct amd_sched_fence *fence = to_amd_sched_fence(f);
>  
> -	fence_put(&fence->scheduled);
> +	dma_fence_put(&fence->scheduled);
>  }
>  
> -const struct fence_ops amd_sched_fence_ops_scheduled = {
> +const struct dma_fence_ops amd_sched_fence_ops_scheduled = {
>  	.get_driver_name = amd_sched_fence_get_driver_name,
>  	.get_timeline_name = amd_sched_fence_get_timeline_name,
>  	.enable_signaling = amd_sched_fence_enable_signaling,
>  	.signaled = NULL,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = amd_sched_fence_release_scheduled,
>  };
>  
> -const struct fence_ops amd_sched_fence_ops_finished = {
> +const struct dma_fence_ops amd_sched_fence_ops_finished = {
>  	.get_driver_name = amd_sched_fence_get_driver_name,
>  	.get_timeline_name = amd_sched_fence_get_timeline_name,
>  	.enable_signaling = amd_sched_fence_enable_signaling,
>  	.signaled = NULL,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = amd_sched_fence_release_finished,
>  };
> diff --git a/drivers/gpu/drm/drm_atomic.c b/drivers/gpu/drm/drm_atomic.c
> index 1b5a32df9a9a..c32fb3c1d6f0 100644
> --- a/drivers/gpu/drm/drm_atomic.c
> +++ b/drivers/gpu/drm/drm_atomic.c
> @@ -1463,7 +1463,7 @@ EXPORT_SYMBOL(drm_atomic_nonblocking_commit);
>  
>  static struct drm_pending_vblank_event *create_vblank_event(
>  		struct drm_device *dev, struct drm_file *file_priv,
> -		struct fence *fence, uint64_t user_data)
> +		struct dma_fence *fence, uint64_t user_data)
>  {
>  	struct drm_pending_vblank_event *e = NULL;
>  	int ret;
> diff --git a/drivers/gpu/drm/drm_atomic_helper.c b/drivers/gpu/drm/drm_atomic_helper.c
> index f9362760bfb2..75ad01d595fd 100644
> --- a/drivers/gpu/drm/drm_atomic_helper.c
> +++ b/drivers/gpu/drm/drm_atomic_helper.c
> @@ -30,7 +30,7 @@
>  #include <drm/drm_plane_helper.h>
>  #include <drm/drm_crtc_helper.h>
>  #include <drm/drm_atomic_helper.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include "drm_crtc_internal.h"
>  
> @@ -1017,7 +1017,7 @@ EXPORT_SYMBOL(drm_atomic_helper_commit_modeset_enables);
>   * drm_atomic_helper_swap_state() so it uses the current plane state (and
>   * just uses the atomic state to find the changed planes)
>   *
> - * Returns zero if success or < 0 if fence_wait() fails.
> + * Returns zero if success or < 0 if dma_fence_wait() fails.
>   */
>  int drm_atomic_helper_wait_for_fences(struct drm_device *dev,
>  				      struct drm_atomic_state *state,
> @@ -1041,11 +1041,11 @@ int drm_atomic_helper_wait_for_fences(struct drm_device *dev,
>  		 * still interrupt the operation. Instead of blocking until the
>  		 * timer expires, make the wait interruptible.
>  		 */
> -		ret = fence_wait(plane_state->fence, pre_swap);
> +		ret = dma_fence_wait(plane_state->fence, pre_swap);
>  		if (ret)
>  			return ret;
>  
> -		fence_put(plane_state->fence);
> +		dma_fence_put(plane_state->fence);
>  		plane_state->fence = NULL;
>  	}
>  
> diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c
> index 8bed5f459182..cf993dbf602e 100644
> --- a/drivers/gpu/drm/drm_fops.c
> +++ b/drivers/gpu/drm/drm_fops.c
> @@ -665,7 +665,7 @@ void drm_event_cancel_free(struct drm_device *dev,
>  	spin_unlock_irqrestore(&dev->event_lock, flags);
>  
>  	if (p->fence)
> -		fence_put(p->fence);
> +		dma_fence_put(p->fence);
>  
>  	kfree(p);
>  }
> @@ -696,8 +696,8 @@ void drm_send_event_locked(struct drm_device *dev, struct drm_pending_event *e)
>  	}
>  
>  	if (e->fence) {
> -		fence_signal(e->fence);
> -		fence_put(e->fence);
> +		dma_fence_signal(e->fence);
> +		dma_fence_put(e->fence);
>  	}
>  
>  	if (!e->file_priv) {
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem.c b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> index 3755ef935af4..7d066a91d778 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem.c
> @@ -466,10 +466,10 @@ int etnaviv_gem_wait_bo(struct etnaviv_gpu *gpu, struct drm_gem_object *obj,
>  }
>  
>  #ifdef CONFIG_DEBUG_FS
> -static void etnaviv_gem_describe_fence(struct fence *fence,
> +static void etnaviv_gem_describe_fence(struct dma_fence *fence,
>  	const char *type, struct seq_file *m)
>  {
> -	if (!test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		seq_printf(m, "\t%9s: %s %s seq %u\n",
>  			   type,
>  			   fence->ops->get_driver_name(fence),
> @@ -482,7 +482,7 @@ static void etnaviv_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  	struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
>  	struct reservation_object *robj = etnaviv_obj->resv;
>  	struct reservation_object_list *fobj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	unsigned long off = drm_vma_node_start(&obj->vma_node);
>  
>  	seq_printf(m, "%08x: %c %2d (%2d) %08lx %p %zd\n",
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
> index b1254f885fed..8eacc5f03d8b 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
> @@ -15,7 +15,7 @@
>   */
>  
>  #include <linux/component.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/moduleparam.h>
>  #include <linux/of_device.h>
>  #include "etnaviv_dump.h"
> @@ -882,7 +882,7 @@ static void recover_worker(struct work_struct *work)
>  	for (i = 0; i < ARRAY_SIZE(gpu->event); i++) {
>  		if (!gpu->event[i].used)
>  			continue;
> -		fence_signal(gpu->event[i].fence);
> +		dma_fence_signal(gpu->event[i].fence);
>  		gpu->event[i].fence = NULL;
>  		gpu->event[i].used = false;
>  		complete(&gpu->event_free);
> @@ -952,55 +952,55 @@ static void hangcheck_disable(struct etnaviv_gpu *gpu)
>  /* fence object management */
>  struct etnaviv_fence {
>  	struct etnaviv_gpu *gpu;
> -	struct fence base;
> +	struct dma_fence base;
>  };
>  
> -static inline struct etnaviv_fence *to_etnaviv_fence(struct fence *fence)
> +static inline struct etnaviv_fence *to_etnaviv_fence(struct dma_fence *fence)
>  {
>  	return container_of(fence, struct etnaviv_fence, base);
>  }
>  
> -static const char *etnaviv_fence_get_driver_name(struct fence *fence)
> +static const char *etnaviv_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "etnaviv";
>  }
>  
> -static const char *etnaviv_fence_get_timeline_name(struct fence *fence)
> +static const char *etnaviv_fence_get_timeline_name(struct dma_fence *fence)
>  {
>  	struct etnaviv_fence *f = to_etnaviv_fence(fence);
>  
>  	return dev_name(f->gpu->dev);
>  }
>  
> -static bool etnaviv_fence_enable_signaling(struct fence *fence)
> +static bool etnaviv_fence_enable_signaling(struct dma_fence *fence)
>  {
>  	return true;
>  }
>  
> -static bool etnaviv_fence_signaled(struct fence *fence)
> +static bool etnaviv_fence_signaled(struct dma_fence *fence)
>  {
>  	struct etnaviv_fence *f = to_etnaviv_fence(fence);
>  
> -	return fence_completed(f->gpu, f->base.seqno);
> +	return dma_fence_completed(f->gpu, f->base.seqno);
>  }
>  
> -static void etnaviv_fence_release(struct fence *fence)
> +static void etnaviv_fence_release(struct dma_fence *fence)
>  {
>  	struct etnaviv_fence *f = to_etnaviv_fence(fence);
>  
>  	kfree_rcu(f, base.rcu);
>  }
>  
> -static const struct fence_ops etnaviv_fence_ops = {
> +static const struct dma_fence_ops etnaviv_fence_ops = {
>  	.get_driver_name = etnaviv_fence_get_driver_name,
>  	.get_timeline_name = etnaviv_fence_get_timeline_name,
>  	.enable_signaling = etnaviv_fence_enable_signaling,
>  	.signaled = etnaviv_fence_signaled,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = etnaviv_fence_release,
>  };
>  
> -static struct fence *etnaviv_gpu_fence_alloc(struct etnaviv_gpu *gpu)
> +static struct dma_fence *etnaviv_gpu_fence_alloc(struct etnaviv_gpu *gpu)
>  {
>  	struct etnaviv_fence *f;
>  
> @@ -1010,8 +1010,8 @@ static struct fence *etnaviv_gpu_fence_alloc(struct etnaviv_gpu *gpu)
>  
>  	f->gpu = gpu;
>  
> -	fence_init(&f->base, &etnaviv_fence_ops, &gpu->fence_spinlock,
> -		   gpu->fence_context, ++gpu->next_fence);
> +	dma_fence_init(&f->base, &etnaviv_fence_ops, &gpu->fence_spinlock,
> +		       gpu->fence_context, ++gpu->next_fence);
>  
>  	return &f->base;
>  }
> @@ -1021,7 +1021,7 @@ int etnaviv_gpu_fence_sync_obj(struct etnaviv_gem_object *etnaviv_obj,
>  {
>  	struct reservation_object *robj = etnaviv_obj->resv;
>  	struct reservation_object_list *fobj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int i, ret;
>  
>  	if (!exclusive) {
> @@ -1039,7 +1039,7 @@ int etnaviv_gpu_fence_sync_obj(struct etnaviv_gem_object *etnaviv_obj,
>  		/* Wait on any existing exclusive fence which isn't our own */
>  		fence = reservation_object_get_excl(robj);
>  		if (fence && fence->context != context) {
> -			ret = fence_wait(fence, true);
> +			ret = dma_fence_wait(fence, true);
>  			if (ret)
>  				return ret;
>  		}
> @@ -1052,7 +1052,7 @@ int etnaviv_gpu_fence_sync_obj(struct etnaviv_gem_object *etnaviv_obj,
>  		fence = rcu_dereference_protected(fobj->shared[i],
>  						reservation_object_held(robj));
>  		if (fence->context != context) {
> -			ret = fence_wait(fence, true);
> +			ret = dma_fence_wait(fence, true);
>  			if (ret)
>  				return ret;
>  		}
> @@ -1158,11 +1158,11 @@ static void retire_worker(struct work_struct *work)
>  
>  	mutex_lock(&gpu->lock);
>  	list_for_each_entry_safe(cmdbuf, tmp, &gpu->active_cmd_list, node) {
> -		if (!fence_is_signaled(cmdbuf->fence))
> +		if (!dma_fence_is_signaled(cmdbuf->fence))
>  			break;
>  
>  		list_del(&cmdbuf->node);
> -		fence_put(cmdbuf->fence);
> +		dma_fence_put(cmdbuf->fence);
>  
>  		for (i = 0; i < cmdbuf->nr_bos; i++) {
>  			struct etnaviv_vram_mapping *mapping = cmdbuf->bo_map[i];
> @@ -1204,7 +1204,7 @@ int etnaviv_gpu_wait_fence_interruptible(struct etnaviv_gpu *gpu,
>  
>  	if (!timeout) {
>  		/* No timeout was requested: just test for completion */
> -		ret = fence_completed(gpu, fence) ? 0 : -EBUSY;
> +		ret = dma_fence_completed(gpu, fence) ? 0 : -EBUSY;
>  	} else {
>  		unsigned long remaining = etnaviv_timeout_to_jiffies(timeout);
>  
> @@ -1275,7 +1275,7 @@ void etnaviv_gpu_pm_put(struct etnaviv_gpu *gpu)
>  int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
>  	struct etnaviv_gem_submit *submit, struct etnaviv_cmdbuf *cmdbuf)
>  {
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	unsigned int event, i;
>  	int ret;
>  
> @@ -1391,7 +1391,7 @@ static irqreturn_t irq_handler(int irq, void *data)
>  		}
>  
>  		while ((event = ffs(intr)) != 0) {
> -			struct fence *fence;
> +			struct dma_fence *fence;
>  
>  			event -= 1;
>  
> @@ -1401,7 +1401,7 @@ static irqreturn_t irq_handler(int irq, void *data)
>  
>  			fence = gpu->event[event].fence;
>  			gpu->event[event].fence = NULL;
> -			fence_signal(fence);
> +			dma_fence_signal(fence);
>  
>  			/*
>  			 * Events can be processed out of order.  Eg,
> @@ -1553,7 +1553,7 @@ static int etnaviv_gpu_bind(struct device *dev, struct device *master,
>  		return ret;
>  
>  	gpu->drm = drm;
> -	gpu->fence_context = fence_context_alloc(1);
> +	gpu->fence_context = dma_fence_context_alloc(1);
>  	spin_lock_init(&gpu->fence_spinlock);
>  
>  	INIT_LIST_HEAD(&gpu->active_cmd_list);
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h
> index 73c278dc3706..8c6b824e9d0a 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h
> @@ -89,7 +89,7 @@ struct etnaviv_chip_identity {
>  
>  struct etnaviv_event {
>  	bool used;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  };
>  
>  struct etnaviv_cmdbuf;
> @@ -163,7 +163,7 @@ struct etnaviv_cmdbuf {
>  	/* vram node used if the cmdbuf is mapped through the MMUv2 */
>  	struct drm_mm_node vram_node;
>  	/* fence after which this buffer is to be disposed */
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	/* target exec state */
>  	u32 exec_state;
>  	/* per GPU in-flight list */
> diff --git a/drivers/gpu/drm/i915/i915_gem_request.c b/drivers/gpu/drm/i915/i915_gem_request.c
> index 74ede1f53372..f9af2a00625e 100644
> --- a/drivers/gpu/drm/i915/i915_gem_request.c
> +++ b/drivers/gpu/drm/i915/i915_gem_request.c
> @@ -26,12 +26,12 @@
>  
>  #include "i915_drv.h"
>  
> -static const char *i915_fence_get_driver_name(struct fence *fence)
> +static const char *i915_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "i915";
>  }
>  
> -static const char *i915_fence_get_timeline_name(struct fence *fence)
> +static const char *i915_fence_get_timeline_name(struct dma_fence *fence)
>  {
>  	/* Timelines are bound by eviction to a VM. However, since
>  	 * we only have a global seqno at the moment, we only have
> @@ -42,12 +42,12 @@ static const char *i915_fence_get_timeline_name(struct fence *fence)
>  	return "global";
>  }
>  
> -static bool i915_fence_signaled(struct fence *fence)
> +static bool i915_fence_signaled(struct dma_fence *fence)
>  {
>  	return i915_gem_request_completed(to_request(fence));
>  }
>  
> -static bool i915_fence_enable_signaling(struct fence *fence)
> +static bool i915_fence_enable_signaling(struct dma_fence *fence)
>  {
>  	if (i915_fence_signaled(fence))
>  		return false;
> @@ -56,7 +56,7 @@ static bool i915_fence_enable_signaling(struct fence *fence)
>  	return true;
>  }
>  
> -static signed long i915_fence_wait(struct fence *fence,
> +static signed long i915_fence_wait(struct dma_fence *fence,
>  				   bool interruptible,
>  				   signed long timeout_jiffies)
>  {
> @@ -85,26 +85,26 @@ static signed long i915_fence_wait(struct fence *fence,
>  	return timeout_jiffies;
>  }
>  
> -static void i915_fence_value_str(struct fence *fence, char *str, int size)
> +static void i915_fence_value_str(struct dma_fence *fence, char *str, int size)
>  {
>  	snprintf(str, size, "%u", fence->seqno);
>  }
>  
> -static void i915_fence_timeline_value_str(struct fence *fence, char *str,
> +static void i915_fence_timeline_value_str(struct dma_fence *fence, char *str,
>  					  int size)
>  {
>  	snprintf(str, size, "%u",
>  		 intel_engine_get_seqno(to_request(fence)->engine));
>  }
>  
> -static void i915_fence_release(struct fence *fence)
> +static void i915_fence_release(struct dma_fence *fence)
>  {
>  	struct drm_i915_gem_request *req = to_request(fence);
>  
>  	kmem_cache_free(req->i915->requests, req);
>  }
>  
> -const struct fence_ops i915_fence_ops = {
> +const struct dma_fence_ops i915_fence_ops = {
>  	.get_driver_name = i915_fence_get_driver_name,
>  	.get_timeline_name = i915_fence_get_timeline_name,
>  	.enable_signaling = i915_fence_enable_signaling,
> @@ -388,8 +388,8 @@ i915_gem_request_alloc(struct intel_engine_cs *engine,
>  	 * The reference count is incremented atomically. If it is zero,
>  	 * the lookup knows the request is unallocated and complete. Otherwise,
>  	 * it is either still in use, or has been reallocated and reset
> -	 * with fence_init(). This increment is safe for release as we check
> -	 * that the request we have a reference to and matches the active
> +	 * with dma_fence_init(). This increment is safe for release as we
> +	 * check that the request we have a reference to and matches the active
>  	 * request.
>  	 *
>  	 * Before we increment the refcount, we chase the request->engine
> @@ -412,11 +412,11 @@ i915_gem_request_alloc(struct intel_engine_cs *engine,
>  		goto err;
>  
>  	spin_lock_init(&req->lock);
> -	fence_init(&req->fence,
> -		   &i915_fence_ops,
> -		   &req->lock,
> -		   engine->fence_context,
> -		   seqno);
> +	dma_fence_init(&req->fence,
> +		       &i915_fence_ops,
> +		       &req->lock,
> +		       engine->fence_context,
> +		       seqno);
>  
>  	i915_sw_fence_init(&req->submit, submit_notify);
>  
> diff --git a/drivers/gpu/drm/i915/i915_gem_request.h b/drivers/gpu/drm/i915/i915_gem_request.h
> index 974bd7bcc801..bceeaa3a5193 100644
> --- a/drivers/gpu/drm/i915/i915_gem_request.h
> +++ b/drivers/gpu/drm/i915/i915_gem_request.h
> @@ -25,7 +25,7 @@
>  #ifndef I915_GEM_REQUEST_H
>  #define I915_GEM_REQUEST_H
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include "i915_gem.h"
>  #include "i915_sw_fence.h"
> @@ -62,7 +62,7 @@ struct intel_signal_node {
>   * The requests are reference counted.
>   */
>  struct drm_i915_gem_request {
> -	struct fence fence;
> +	struct dma_fence fence;
>  	spinlock_t lock;
>  
>  	/** On Which ring this request was generated */
> @@ -145,9 +145,9 @@ struct drm_i915_gem_request {
>  	struct list_head execlist_link;
>  };
>  
> -extern const struct fence_ops i915_fence_ops;
> +extern const struct dma_fence_ops i915_fence_ops;
>  
> -static inline bool fence_is_i915(struct fence *fence)
> +static inline bool fence_is_i915(struct dma_fence *fence)
>  {
>  	return fence->ops == &i915_fence_ops;
>  }
> @@ -172,7 +172,7 @@ i915_gem_request_get_engine(struct drm_i915_gem_request *req)
>  }
>  
>  static inline struct drm_i915_gem_request *
> -to_request(struct fence *fence)
> +to_request(struct dma_fence *fence)
>  {
>  	/* We assume that NULL fence/request are interoperable */
>  	BUILD_BUG_ON(offsetof(struct drm_i915_gem_request, fence) != 0);
> @@ -183,19 +183,19 @@ to_request(struct fence *fence)
>  static inline struct drm_i915_gem_request *
>  i915_gem_request_get(struct drm_i915_gem_request *req)
>  {
> -	return to_request(fence_get(&req->fence));
> +	return to_request(dma_fence_get(&req->fence));
>  }
>  
>  static inline struct drm_i915_gem_request *
>  i915_gem_request_get_rcu(struct drm_i915_gem_request *req)
>  {
> -	return to_request(fence_get_rcu(&req->fence));
> +	return to_request(dma_fence_get_rcu(&req->fence));
>  }
>  
>  static inline void
>  i915_gem_request_put(struct drm_i915_gem_request *req)
>  {
> -	fence_put(&req->fence);
> +	dma_fence_put(&req->fence);
>  }
>  
>  static inline void i915_gem_request_assign(struct drm_i915_gem_request **pdst,
> @@ -497,7 +497,7 @@ __i915_gem_active_get_rcu(const struct i915_gem_active *active)
>  		 * compiler.
>  		 *
>  		 * The atomic operation at the heart of
> -		 * i915_gem_request_get_rcu(), see fence_get_rcu(), is
> +		 * i915_gem_request_get_rcu(), see dma_fence_get_rcu(), is
>  		 * atomic_inc_not_zero() which is only a full memory barrier
>  		 * when successful. That is, if i915_gem_request_get_rcu()
>  		 * returns the request (and so with the reference counted
> diff --git a/drivers/gpu/drm/i915/i915_sw_fence.c b/drivers/gpu/drm/i915/i915_sw_fence.c
> index 1e5cbc585ca2..8185002d7ec8 100644
> --- a/drivers/gpu/drm/i915/i915_sw_fence.c
> +++ b/drivers/gpu/drm/i915/i915_sw_fence.c
> @@ -8,7 +8,7 @@
>   */
>  
>  #include <linux/slab.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/reservation.h>
>  
>  #include "i915_sw_fence.h"
> @@ -226,49 +226,50 @@ int i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
>  	return pending;
>  }
>  
> -struct dma_fence_cb {
> -	struct fence_cb base;
> +struct i915_sw_dma_fence_cb {
> +	struct dma_fence_cb base;
>  	struct i915_sw_fence *fence;
> -	struct fence *dma;
> +	struct dma_fence *dma;
>  	struct timer_list timer;
>  };
>  
>  static void timer_i915_sw_fence_wake(unsigned long data)
>  {
> -	struct dma_fence_cb *cb = (struct dma_fence_cb *)data;
> +	struct i915_sw_dma_fence_cb *cb = (struct i915_sw_dma_fence_cb *)data;
>  
>  	printk(KERN_WARNING "asynchronous wait on fence %s:%s:%x timed out\n",
>  	       cb->dma->ops->get_driver_name(cb->dma),
>  	       cb->dma->ops->get_timeline_name(cb->dma),
>  	       cb->dma->seqno);
> -	fence_put(cb->dma);
> +	dma_fence_put(cb->dma);
>  	cb->dma = NULL;
>  
>  	i915_sw_fence_commit(cb->fence);
>  	cb->timer.function = NULL;
>  }
>  
> -static void dma_i915_sw_fence_wake(struct fence *dma, struct fence_cb *data)
> +static void dma_i915_sw_fence_wake(struct dma_fence *dma,
> +				   struct dma_fence_cb *data)
>  {
> -	struct dma_fence_cb *cb = container_of(data, typeof(*cb), base);
> +	struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
>  
>  	del_timer_sync(&cb->timer);
>  	if (cb->timer.function)
>  		i915_sw_fence_commit(cb->fence);
> -	fence_put(cb->dma);
> +	dma_fence_put(cb->dma);
>  
>  	kfree(cb);
>  }
>  
>  int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
> -				  struct fence *dma,
> +				  struct dma_fence *dma,
>  				  unsigned long timeout,
>  				  gfp_t gfp)
>  {
> -	struct dma_fence_cb *cb;
> +	struct i915_sw_dma_fence_cb *cb;
>  	int ret;
>  
> -	if (fence_is_signaled(dma))
> +	if (dma_fence_is_signaled(dma))
>  		return 0;
>  
>  	cb = kmalloc(sizeof(*cb), gfp);
> @@ -276,7 +277,7 @@ int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
>  		if (!gfpflags_allow_blocking(gfp))
>  			return -ENOMEM;
>  
> -		return fence_wait(dma, false);
> +		return dma_fence_wait(dma, false);
>  	}
>  
>  	cb->fence = i915_sw_fence_get(fence);
> @@ -287,11 +288,11 @@ int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
>  		      timer_i915_sw_fence_wake, (unsigned long)cb,
>  		      TIMER_IRQSAFE);
>  	if (timeout) {
> -		cb->dma = fence_get(dma);
> +		cb->dma = dma_fence_get(dma);
>  		mod_timer(&cb->timer, round_jiffies_up(jiffies + timeout));
>  	}
>  
> -	ret = fence_add_callback(dma, &cb->base, dma_i915_sw_fence_wake);
> +	ret = dma_fence_add_callback(dma, &cb->base, dma_i915_sw_fence_wake);
>  	if (ret == 0) {
>  		ret = 1;
>  	} else {
> @@ -305,16 +306,16 @@ int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
>  
>  int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
>  				    struct reservation_object *resv,
> -				    const struct fence_ops *exclude,
> +				    const struct dma_fence_ops *exclude,
>  				    bool write,
>  				    unsigned long timeout,
>  				    gfp_t gfp)
>  {
> -	struct fence *excl;
> +	struct dma_fence *excl;
>  	int ret = 0, pending;
>  
>  	if (write) {
> -		struct fence **shared;
> +		struct dma_fence **shared;
>  		unsigned int count, i;
>  
>  		ret = reservation_object_get_fences_rcu(resv,
> @@ -339,7 +340,7 @@ int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
>  		}
>  
>  		for (i = 0; i < count; i++)
> -			fence_put(shared[i]);
> +			dma_fence_put(shared[i]);
>  		kfree(shared);
>  	} else {
>  		excl = reservation_object_get_excl_rcu(resv);
> @@ -356,7 +357,7 @@ int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
>  			ret |= pending;
>  	}
>  
> -	fence_put(excl);
> +	dma_fence_put(excl);
>  
>  	return ret;
>  }
> diff --git a/drivers/gpu/drm/i915/i915_sw_fence.h b/drivers/gpu/drm/i915/i915_sw_fence.h
> index 373141602ca4..cd239e92f67f 100644
> --- a/drivers/gpu/drm/i915/i915_sw_fence.h
> +++ b/drivers/gpu/drm/i915/i915_sw_fence.h
> @@ -16,8 +16,8 @@
>  #include <linux/wait.h>
>  
>  struct completion;
> -struct fence;
> -struct fence_ops;
> +struct dma_fence;
> +struct dma_fence_ops;
>  struct reservation_object;
>  
>  struct i915_sw_fence {
> @@ -47,12 +47,12 @@ int i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
>  				 struct i915_sw_fence *after,
>  				 wait_queue_t *wq);
>  int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
> -				  struct fence *dma,
> +				  struct dma_fence *dma,
>  				  unsigned long timeout,
>  				  gfp_t gfp);
>  int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
>  				    struct reservation_object *resv,
> -				    const struct fence_ops *exclude,
> +				    const struct dma_fence_ops *exclude,
>  				    bool write,
>  				    unsigned long timeout,
>  				    gfp_t gfp);
> diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
> index 178798002a73..5c912c25f7d3 100644
> --- a/drivers/gpu/drm/i915/i915_trace.h
> +++ b/drivers/gpu/drm/i915/i915_trace.h
> @@ -491,7 +491,7 @@ TRACE_EVENT(i915_gem_ring_dispatch,
>  			   __entry->ring = req->engine->id;
>  			   __entry->seqno = req->fence.seqno;
>  			   __entry->flags = flags;
> -			   fence_enable_sw_signaling(&req->fence);
> +			   dma_fence_enable_sw_signaling(&req->fence);
>  			   ),
>  
>  	    TP_printk("dev=%u, ring=%u, seqno=%u, flags=%x",
> diff --git a/drivers/gpu/drm/i915/intel_breadcrumbs.c b/drivers/gpu/drm/i915/intel_breadcrumbs.c
> index 23fc1042fed4..56efcc507ea2 100644
> --- a/drivers/gpu/drm/i915/intel_breadcrumbs.c
> +++ b/drivers/gpu/drm/i915/intel_breadcrumbs.c
> @@ -464,7 +464,7 @@ static int intel_breadcrumbs_signaler(void *arg)
>  						 &request->signaling.wait);
>  
>  			local_bh_disable();
> -			fence_signal(&request->fence);
> +			dma_fence_signal(&request->fence);
>  			local_bh_enable(); /* kick start the tasklets */
>  
>  			/* Find the next oldest signal. Note that as we have
> @@ -502,7 +502,7 @@ void intel_engine_enable_signaling(struct drm_i915_gem_request *request)
>  	struct rb_node *parent, **p;
>  	bool first, wakeup;
>  
> -	/* locked by fence_enable_sw_signaling() */
> +	/* locked by dma_fence_enable_sw_signaling() */
>  	assert_spin_locked(&request->lock);
>  
>  	request->signaling.wait.tsk = b->signaler;
> diff --git a/drivers/gpu/drm/i915/intel_engine_cs.c b/drivers/gpu/drm/i915/intel_engine_cs.c
> index 2dc94812bea5..8cceb345aa0f 100644
> --- a/drivers/gpu/drm/i915/intel_engine_cs.c
> +++ b/drivers/gpu/drm/i915/intel_engine_cs.c
> @@ -245,7 +245,7 @@ void intel_engine_setup_common(struct intel_engine_cs *engine)
>  	INIT_LIST_HEAD(&engine->execlist_queue);
>  	spin_lock_init(&engine->execlist_lock);
>  
> -	engine->fence_context = fence_context_alloc(1);
> +	engine->fence_context = dma_fence_context_alloc(1);
>  
>  	intel_engine_init_requests(engine);
>  	intel_engine_init_hangcheck(engine);
> diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h
> index d0da52f2a806..06b5317f267e 100644
> --- a/drivers/gpu/drm/msm/msm_drv.h
> +++ b/drivers/gpu/drm/msm/msm_drv.h
> @@ -217,7 +217,7 @@ void msm_gem_vunmap(struct drm_gem_object *obj);
>  int msm_gem_sync_object(struct drm_gem_object *obj,
>  		struct msm_fence_context *fctx, bool exclusive);
>  void msm_gem_move_to_active(struct drm_gem_object *obj,
> -		struct msm_gpu *gpu, bool exclusive, struct fence *fence);
> +		struct dma_msm_gpu *gpu, bool exclusive, struct dma_fence *fence);
>  void msm_gem_move_to_inactive(struct drm_gem_object *obj);
>  int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout);
>  int msm_gem_cpu_fini(struct drm_gem_object *obj);
> diff --git a/drivers/gpu/drm/msm/msm_fence.c b/drivers/gpu/drm/msm/msm_fence.c
> index a9b9b1c95a2e..427ac74742a4 100644
> --- a/drivers/gpu/drm/msm/msm_fence.c
> +++ b/drivers/gpu/drm/msm/msm_fence.c
> @@ -15,7 +15,7 @@
>   * this program.  If not, see <http://www.gnu.org/licenses/>.
>   */
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include "msm_drv.h"
>  #include "msm_fence.h"
> @@ -32,7 +32,7 @@ msm_fence_context_alloc(struct drm_device *dev, const char *name)
>  
>  	fctx->dev = dev;
>  	fctx->name = name;
> -	fctx->context = fence_context_alloc(1);
> +	fctx->context = dma_fence_context_alloc(1);
>  	init_waitqueue_head(&fctx->event);
>  	spin_lock_init(&fctx->spinlock);
>  
> @@ -44,7 +44,7 @@ void msm_fence_context_free(struct msm_fence_context *fctx)
>  	kfree(fctx);
>  }
>  
> -static inline bool fence_completed(struct msm_fence_context *fctx, uint32_t fence)
> +static inline bool dma_fence_completed(struct msm_fence_context *fctx, uint32_t fence)
>  {
>  	return (int32_t)(fctx->completed_fence - fence) >= 0;
>  }
> @@ -63,7 +63,7 @@ int msm_wait_fence(struct msm_fence_context *fctx, uint32_t fence,
>  
>  	if (!timeout) {
>  		/* no-wait: */
> -		ret = fence_completed(fctx, fence) ? 0 : -EBUSY;
> +		ret = dma_fence_completed(fctx, fence) ? 0 : -EBUSY;
>  	} else {
>  		unsigned long remaining_jiffies = timeout_to_jiffies(timeout);
>  
> @@ -100,52 +100,52 @@ void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence)
>  
>  struct msm_fence {
>  	struct msm_fence_context *fctx;
> -	struct fence base;
> +	struct dma_fence base;
>  };
>  
> -static inline struct msm_fence *to_msm_fence(struct fence *fence)
> +static inline struct msm_fence *to_msm_fence(struct dma_fence *fence)
>  {
>  	return container_of(fence, struct msm_fence, base);
>  }
>  
> -static const char *msm_fence_get_driver_name(struct fence *fence)
> +static const char *msm_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "msm";
>  }
>  
> -static const char *msm_fence_get_timeline_name(struct fence *fence)
> +static const char *msm_fence_get_timeline_name(struct dma_fence *fence)
>  {
>  	struct msm_fence *f = to_msm_fence(fence);
>  	return f->fctx->name;
>  }
>  
> -static bool msm_fence_enable_signaling(struct fence *fence)
> +static bool msm_fence_enable_signaling(struct dma_fence *fence)
>  {
>  	return true;
>  }
>  
> -static bool msm_fence_signaled(struct fence *fence)
> +static bool msm_fence_signaled(struct dma_fence *fence)
>  {
>  	struct msm_fence *f = to_msm_fence(fence);
> -	return fence_completed(f->fctx, f->base.seqno);
> +	return dma_fence_completed(f->fctx, f->base.seqno);
>  }
>  
> -static void msm_fence_release(struct fence *fence)
> +static void msm_fence_release(struct dma_fence *fence)
>  {
>  	struct msm_fence *f = to_msm_fence(fence);
>  	kfree_rcu(f, base.rcu);
>  }
>  
> -static const struct fence_ops msm_fence_ops = {
> +static const struct dma_fence_ops msm_fence_ops = {
>  	.get_driver_name = msm_fence_get_driver_name,
>  	.get_timeline_name = msm_fence_get_timeline_name,
>  	.enable_signaling = msm_fence_enable_signaling,
>  	.signaled = msm_fence_signaled,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = msm_fence_release,
>  };
>  
> -struct fence *
> +struct dma_fence *
>  msm_fence_alloc(struct msm_fence_context *fctx)
>  {
>  	struct msm_fence *f;
> @@ -156,8 +156,8 @@ msm_fence_alloc(struct msm_fence_context *fctx)
>  
>  	f->fctx = fctx;
>  
> -	fence_init(&f->base, &msm_fence_ops, &fctx->spinlock,
> -			fctx->context, ++fctx->last_fence);
> +	dma_fence_init(&f->base, &msm_fence_ops, &fctx->spinlock,
> +		       fctx->context, ++fctx->last_fence);
>  
>  	return &f->base;
>  }
> diff --git a/drivers/gpu/drm/msm/msm_fence.h b/drivers/gpu/drm/msm/msm_fence.h
> index ceb5b3d314b4..56061aa1959d 100644
> --- a/drivers/gpu/drm/msm/msm_fence.h
> +++ b/drivers/gpu/drm/msm/msm_fence.h
> @@ -41,6 +41,6 @@ int msm_queue_fence_cb(struct msm_fence_context *fctx,
>  		struct msm_fence_cb *cb, uint32_t fence);
>  void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence);
>  
> -struct fence * msm_fence_alloc(struct msm_fence_context *fctx);
> +struct dma_fence * msm_fence_alloc(struct msm_fence_context *fctx);
>  
>  #endif
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index b6ac27e31929..57db7dbbb618 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -521,7 +521,7 @@ int msm_gem_sync_object(struct drm_gem_object *obj,
>  {
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
>  	struct reservation_object_list *fobj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int i, ret;
>  
>  	if (!exclusive) {
> @@ -540,7 +540,7 @@ int msm_gem_sync_object(struct drm_gem_object *obj,
>  		fence = reservation_object_get_excl(msm_obj->resv);
>  		/* don't need to wait on our own fences, since ring is fifo */
>  		if (fence && (fence->context != fctx->context)) {
> -			ret = fence_wait(fence, true);
> +			ret = dma_fence_wait(fence, true);
>  			if (ret)
>  				return ret;
>  		}
> @@ -553,7 +553,7 @@ int msm_gem_sync_object(struct drm_gem_object *obj,
>  		fence = rcu_dereference_protected(fobj->shared[i],
>  						reservation_object_held(msm_obj->resv));
>  		if (fence->context != fctx->context) {
> -			ret = fence_wait(fence, true);
> +			ret = dma_fence_wait(fence, true);
>  			if (ret)
>  				return ret;
>  		}
> @@ -563,7 +563,7 @@ int msm_gem_sync_object(struct drm_gem_object *obj,
>  }
>  
>  void msm_gem_move_to_active(struct drm_gem_object *obj,
> -		struct msm_gpu *gpu, bool exclusive, struct fence *fence)
> +		struct msm_gpu *gpu, bool exclusive, struct dma_fence *fence)
>  {
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
>  	WARN_ON(msm_obj->madv != MSM_MADV_WILLNEED);
> @@ -616,10 +616,10 @@ int msm_gem_cpu_fini(struct drm_gem_object *obj)
>  }
>  
>  #ifdef CONFIG_DEBUG_FS
> -static void describe_fence(struct fence *fence, const char *type,
> +static void describe_fence(struct dma_fence *fence, const char *type,
>  		struct seq_file *m)
>  {
> -	if (!fence_is_signaled(fence))
> +	if (!dma_fence_is_signaled(fence))
>  		seq_printf(m, "\t%9s: %s %s seq %u\n", type,
>  				fence->ops->get_driver_name(fence),
>  				fence->ops->get_timeline_name(fence),
> @@ -631,7 +631,7 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
>  	struct reservation_object *robj = msm_obj->resv;
>  	struct reservation_object_list *fobj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	uint64_t off = drm_vma_node_start(&obj->vma_node);
>  	const char *madv;
>  
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index b2f13cfe945e..2cb8551fda70 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -104,7 +104,7 @@ struct msm_gem_submit {
>  	struct list_head node;   /* node in gpu submit_list */
>  	struct list_head bo_list;
>  	struct ww_acquire_ctx ticket;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	struct pid *pid;    /* submitting process */
>  	bool valid;         /* true if no cmdstream patching needed */
>  	unsigned int nr_cmds;
> diff --git a/drivers/gpu/drm/msm/msm_gem_submit.c b/drivers/gpu/drm/msm/msm_gem_submit.c
> index b6a0f37a65f3..25e8786fa4ca 100644
> --- a/drivers/gpu/drm/msm/msm_gem_submit.c
> +++ b/drivers/gpu/drm/msm/msm_gem_submit.c
> @@ -60,7 +60,7 @@ static struct msm_gem_submit *submit_create(struct drm_device *dev,
>  
>  void msm_gem_submit_free(struct msm_gem_submit *submit)
>  {
> -	fence_put(submit->fence);
> +	dma_fence_put(submit->fence);
>  	list_del(&submit->node);
>  	put_pid(submit->pid);
>  	kfree(submit);
> @@ -380,7 +380,7 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
>  	struct msm_file_private *ctx = file->driver_priv;
>  	struct msm_gem_submit *submit;
>  	struct msm_gpu *gpu = priv->gpu;
> -	struct fence *in_fence = NULL;
> +	struct dma_fence *in_fence = NULL;
>  	struct sync_file *sync_file = NULL;
>  	int out_fence_fd = -1;
>  	unsigned i;
> @@ -439,7 +439,7 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
>  		 */
>  
>  		if (in_fence->context != gpu->fctx->context) {
> -			ret = fence_wait(in_fence, true);
> +			ret = dma_fence_wait(in_fence, true);
>  			if (ret)
>  				goto out;
>  		}
> @@ -542,7 +542,7 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
>  
>  out:
>  	if (in_fence)
> -		fence_put(in_fence);
> +		dma_fence_put(in_fence);
>  	submit_cleanup(submit);
>  	if (ret)
>  		msm_gem_submit_free(submit);
> diff --git a/drivers/gpu/drm/msm/msm_gpu.c b/drivers/gpu/drm/msm/msm_gpu.c
> index 5bb09838b5ae..3249707e6834 100644
> --- a/drivers/gpu/drm/msm/msm_gpu.c
> +++ b/drivers/gpu/drm/msm/msm_gpu.c
> @@ -476,7 +476,7 @@ static void retire_submits(struct msm_gpu *gpu)
>  		submit = list_first_entry(&gpu->submit_list,
>  				struct msm_gem_submit, node);
>  
> -		if (fence_is_signaled(submit->fence)) {
> +		if (dma_fence_is_signaled(submit->fence)) {
>  			retire_submit(gpu, submit);
>  		} else {
>  			break;
> diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c
> index 343b8659472c..ec8ac756aab4 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_bo.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c
> @@ -83,13 +83,13 @@ nv10_bo_get_tile_region(struct drm_device *dev, int i)
>  
>  static void
>  nv10_bo_put_tile_region(struct drm_device *dev, struct nouveau_drm_tile *tile,
> -			struct fence *fence)
> +			struct dma_fence *fence)
>  {
>  	struct nouveau_drm *drm = nouveau_drm(dev);
>  
>  	if (tile) {
>  		spin_lock(&drm->tile.lock);
> -		tile->fence = (struct nouveau_fence *)fence_get(fence);
> +		tile->fence = (struct nouveau_fence *)dma_fence_get(fence);
>  		tile->used = false;
>  		spin_unlock(&drm->tile.lock);
>  	}
> @@ -1243,7 +1243,7 @@ nouveau_bo_vm_cleanup(struct ttm_buffer_object *bo,
>  {
>  	struct nouveau_drm *drm = nouveau_bdev(bo->bdev);
>  	struct drm_device *dev = drm->dev;
> -	struct fence *fence = reservation_object_get_excl(bo->resv);
> +	struct dma_fence *fence = reservation_object_get_excl(bo->resv);
>  
>  	nv10_bo_put_tile_region(dev, *old_tile, fence);
>  	*old_tile = new_tile;
> diff --git a/drivers/gpu/drm/nouveau/nouveau_fence.c b/drivers/gpu/drm/nouveau/nouveau_fence.c
> index 4bb9ab892ae1..e9529ee6bc23 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_fence.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_fence.c
> @@ -28,7 +28,7 @@
>  
>  #include <linux/ktime.h>
>  #include <linux/hrtimer.h>
> -#include <trace/events/fence.h>
> +#include <trace/events/dma_fence.h>
>  
>  #include <nvif/cl826e.h>
>  #include <nvif/notify.h>
> @@ -38,11 +38,11 @@
>  #include "nouveau_dma.h"
>  #include "nouveau_fence.h"
>  
> -static const struct fence_ops nouveau_fence_ops_uevent;
> -static const struct fence_ops nouveau_fence_ops_legacy;
> +static const struct dma_fence_ops nouveau_fence_ops_uevent;
> +static const struct dma_fence_ops nouveau_fence_ops_legacy;
>  
>  static inline struct nouveau_fence *
> -from_fence(struct fence *fence)
> +from_fence(struct dma_fence *fence)
>  {
>  	return container_of(fence, struct nouveau_fence, base);
>  }
> @@ -58,23 +58,23 @@ nouveau_fence_signal(struct nouveau_fence *fence)
>  {
>  	int drop = 0;
>  
> -	fence_signal_locked(&fence->base);
> +	dma_fence_signal_locked(&fence->base);
>  	list_del(&fence->head);
>  	rcu_assign_pointer(fence->channel, NULL);
>  
> -	if (test_bit(FENCE_FLAG_USER_BITS, &fence->base.flags)) {
> +	if (test_bit(DMA_FENCE_FLAG_USER_BITS, &fence->base.flags)) {
>  		struct nouveau_fence_chan *fctx = nouveau_fctx(fence);
>  
>  		if (!--fctx->notify_ref)
>  			drop = 1;
>  	}
>  
> -	fence_put(&fence->base);
> +	dma_fence_put(&fence->base);
>  	return drop;
>  }
>  
>  static struct nouveau_fence *
> -nouveau_local_fence(struct fence *fence, struct nouveau_drm *drm) {
> +nouveau_local_fence(struct dma_fence *fence, struct nouveau_drm *drm) {
>  	struct nouveau_fence_priv *priv = (void*)drm->fence;
>  
>  	if (fence->ops != &nouveau_fence_ops_legacy &&
> @@ -201,7 +201,7 @@ nouveau_fence_context_new(struct nouveau_channel *chan, struct nouveau_fence_cha
>  
>  struct nouveau_fence_work {
>  	struct work_struct work;
> -	struct fence_cb cb;
> +	struct dma_fence_cb cb;
>  	void (*func)(void *);
>  	void *data;
>  };
> @@ -214,7 +214,7 @@ nouveau_fence_work_handler(struct work_struct *kwork)
>  	kfree(work);
>  }
>  
> -static void nouveau_fence_work_cb(struct fence *fence, struct fence_cb *cb)
> +static void nouveau_fence_work_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	struct nouveau_fence_work *work = container_of(cb, typeof(*work), cb);
>  
> @@ -222,12 +222,12 @@ static void nouveau_fence_work_cb(struct fence *fence, struct fence_cb *cb)
>  }
>  
>  void
> -nouveau_fence_work(struct fence *fence,
> +nouveau_fence_work(struct dma_fence *fence,
>  		   void (*func)(void *), void *data)
>  {
>  	struct nouveau_fence_work *work;
>  
> -	if (fence_is_signaled(fence))
> +	if (dma_fence_is_signaled(fence))
>  		goto err;
>  
>  	work = kmalloc(sizeof(*work), GFP_KERNEL);
> @@ -245,7 +245,7 @@ nouveau_fence_work(struct fence *fence,
>  	work->func = func;
>  	work->data = data;
>  
> -	if (fence_add_callback(fence, &work->cb, nouveau_fence_work_cb) < 0)
> +	if (dma_fence_add_callback(fence, &work->cb, nouveau_fence_work_cb) < 0)
>  		goto err_free;
>  	return;
>  
> @@ -266,17 +266,17 @@ nouveau_fence_emit(struct nouveau_fence *fence, struct nouveau_channel *chan)
>  	fence->timeout  = jiffies + (15 * HZ);
>  
>  	if (priv->uevent)
> -		fence_init(&fence->base, &nouveau_fence_ops_uevent,
> -			   &fctx->lock, fctx->context, ++fctx->sequence);
> +		dma_fence_init(&fence->base, &nouveau_fence_ops_uevent,
> +			       &fctx->lock, fctx->context, ++fctx->sequence);
>  	else
> -		fence_init(&fence->base, &nouveau_fence_ops_legacy,
> -			   &fctx->lock, fctx->context, ++fctx->sequence);
> +		dma_fence_init(&fence->base, &nouveau_fence_ops_legacy,
> +			       &fctx->lock, fctx->context, ++fctx->sequence);
>  	kref_get(&fctx->fence_ref);
>  
> -	trace_fence_emit(&fence->base);
> +	trace_dma_fence_emit(&fence->base);
>  	ret = fctx->emit(fence);
>  	if (!ret) {
> -		fence_get(&fence->base);
> +		dma_fence_get(&fence->base);
>  		spin_lock_irq(&fctx->lock);
>  
>  		if (nouveau_fence_update(chan, fctx))
> @@ -298,7 +298,7 @@ nouveau_fence_done(struct nouveau_fence *fence)
>  		struct nouveau_channel *chan;
>  		unsigned long flags;
>  
> -		if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
> +		if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
>  			return true;
>  
>  		spin_lock_irqsave(&fctx->lock, flags);
> @@ -307,11 +307,11 @@ nouveau_fence_done(struct nouveau_fence *fence)
>  			nvif_notify_put(&fctx->notify);
>  		spin_unlock_irqrestore(&fctx->lock, flags);
>  	}
> -	return fence_is_signaled(&fence->base);
> +	return dma_fence_is_signaled(&fence->base);
>  }
>  
>  static long
> -nouveau_fence_wait_legacy(struct fence *f, bool intr, long wait)
> +nouveau_fence_wait_legacy(struct dma_fence *f, bool intr, long wait)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  	unsigned long sleep_time = NSEC_PER_MSEC / 1000;
> @@ -378,7 +378,7 @@ nouveau_fence_wait(struct nouveau_fence *fence, bool lazy, bool intr)
>  	if (!lazy)
>  		return nouveau_fence_wait_busy(fence, intr);
>  
> -	ret = fence_wait_timeout(&fence->base, intr, 15 * HZ);
> +	ret = dma_fence_wait_timeout(&fence->base, intr, 15 * HZ);
>  	if (ret < 0)
>  		return ret;
>  	else if (!ret)
> @@ -391,7 +391,7 @@ int
>  nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool exclusive, bool intr)
>  {
>  	struct nouveau_fence_chan *fctx = chan->fence;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	struct reservation_object *resv = nvbo->bo.resv;
>  	struct reservation_object_list *fobj;
>  	struct nouveau_fence *f;
> @@ -421,7 +421,7 @@ nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool e
>  		}
>  
>  		if (must_wait)
> -			ret = fence_wait(fence, intr);
> +			ret = dma_fence_wait(fence, intr);
>  
>  		return ret;
>  	}
> @@ -446,7 +446,7 @@ nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool e
>  		}
>  
>  		if (must_wait)
> -			ret = fence_wait(fence, intr);
> +			ret = dma_fence_wait(fence, intr);
>  	}
>  
>  	return ret;
> @@ -456,7 +456,7 @@ void
>  nouveau_fence_unref(struct nouveau_fence **pfence)
>  {
>  	if (*pfence)
> -		fence_put(&(*pfence)->base);
> +		dma_fence_put(&(*pfence)->base);
>  	*pfence = NULL;
>  }
>  
> @@ -484,12 +484,12 @@ nouveau_fence_new(struct nouveau_channel *chan, bool sysmem,
>  	return ret;
>  }
>  
> -static const char *nouveau_fence_get_get_driver_name(struct fence *fence)
> +static const char *nouveau_fence_get_get_driver_name(struct dma_fence *fence)
>  {
>  	return "nouveau";
>  }
>  
> -static const char *nouveau_fence_get_timeline_name(struct fence *f)
> +static const char *nouveau_fence_get_timeline_name(struct dma_fence *f)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  	struct nouveau_fence_chan *fctx = nouveau_fctx(fence);
> @@ -503,7 +503,7 @@ static const char *nouveau_fence_get_timeline_name(struct fence *f)
>   * result. The drm node should still be there, so we can derive the index from
>   * the fence context.
>   */
> -static bool nouveau_fence_is_signaled(struct fence *f)
> +static bool nouveau_fence_is_signaled(struct dma_fence *f)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  	struct nouveau_fence_chan *fctx = nouveau_fctx(fence);
> @@ -519,7 +519,7 @@ static bool nouveau_fence_is_signaled(struct fence *f)
>  	return ret;
>  }
>  
> -static bool nouveau_fence_no_signaling(struct fence *f)
> +static bool nouveau_fence_no_signaling(struct dma_fence *f)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  
> @@ -530,30 +530,30 @@ static bool nouveau_fence_no_signaling(struct fence *f)
>  	WARN_ON(atomic_read(&fence->base.refcount.refcount) <= 1);
>  
>  	/*
> -	 * This needs uevents to work correctly, but fence_add_callback relies on
> +	 * This needs uevents to work correctly, but dma_fence_add_callback relies on
>  	 * being able to enable signaling. It will still get signaled eventually,
>  	 * just not right away.
>  	 */
>  	if (nouveau_fence_is_signaled(f)) {
>  		list_del(&fence->head);
>  
> -		fence_put(&fence->base);
> +		dma_fence_put(&fence->base);
>  		return false;
>  	}
>  
>  	return true;
>  }
>  
> -static void nouveau_fence_release(struct fence *f)
> +static void nouveau_fence_release(struct dma_fence *f)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  	struct nouveau_fence_chan *fctx = nouveau_fctx(fence);
>  
>  	kref_put(&fctx->fence_ref, nouveau_fence_context_put);
> -	fence_free(&fence->base);
> +	dma_fence_free(&fence->base);
>  }
>  
> -static const struct fence_ops nouveau_fence_ops_legacy = {
> +static const struct dma_fence_ops nouveau_fence_ops_legacy = {
>  	.get_driver_name = nouveau_fence_get_get_driver_name,
>  	.get_timeline_name = nouveau_fence_get_timeline_name,
>  	.enable_signaling = nouveau_fence_no_signaling,
> @@ -562,7 +562,7 @@ static const struct fence_ops nouveau_fence_ops_legacy = {
>  	.release = nouveau_fence_release
>  };
>  
> -static bool nouveau_fence_enable_signaling(struct fence *f)
> +static bool nouveau_fence_enable_signaling(struct dma_fence *f)
>  {
>  	struct nouveau_fence *fence = from_fence(f);
>  	struct nouveau_fence_chan *fctx = nouveau_fctx(fence);
> @@ -573,18 +573,18 @@ static bool nouveau_fence_enable_signaling(struct fence *f)
>  
>  	ret = nouveau_fence_no_signaling(f);
>  	if (ret)
> -		set_bit(FENCE_FLAG_USER_BITS, &fence->base.flags);
> +		set_bit(DMA_FENCE_FLAG_USER_BITS, &fence->base.flags);
>  	else if (!--fctx->notify_ref)
>  		nvif_notify_put(&fctx->notify);
>  
>  	return ret;
>  }
>  
> -static const struct fence_ops nouveau_fence_ops_uevent = {
> +static const struct dma_fence_ops nouveau_fence_ops_uevent = {
>  	.get_driver_name = nouveau_fence_get_get_driver_name,
>  	.get_timeline_name = nouveau_fence_get_timeline_name,
>  	.enable_signaling = nouveau_fence_enable_signaling,
>  	.signaled = nouveau_fence_is_signaled,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = NULL
>  };
> diff --git a/drivers/gpu/drm/nouveau/nouveau_fence.h b/drivers/gpu/drm/nouveau/nouveau_fence.h
> index 64c4ce7115ad..41f3c019e534 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_fence.h
> +++ b/drivers/gpu/drm/nouveau/nouveau_fence.h
> @@ -1,14 +1,14 @@
>  #ifndef __NOUVEAU_FENCE_H__
>  #define __NOUVEAU_FENCE_H__
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <nvif/notify.h>
>  
>  struct nouveau_drm;
>  struct nouveau_bo;
>  
>  struct nouveau_fence {
> -	struct fence base;
> +	struct dma_fence base;
>  
>  	struct list_head head;
>  
> @@ -24,7 +24,7 @@ void nouveau_fence_unref(struct nouveau_fence **);
>  
>  int  nouveau_fence_emit(struct nouveau_fence *, struct nouveau_channel *);
>  bool nouveau_fence_done(struct nouveau_fence *);
> -void nouveau_fence_work(struct fence *, void (*)(void *), void *);
> +void nouveau_fence_work(struct dma_fence *, void (*)(void *), void *);
>  int  nouveau_fence_wait(struct nouveau_fence *, bool lazy, bool intr);
>  int  nouveau_fence_sync(struct nouveau_bo *, struct nouveau_channel *, bool exclusive, bool intr);
>  
> diff --git a/drivers/gpu/drm/nouveau/nouveau_gem.c b/drivers/gpu/drm/nouveau/nouveau_gem.c
> index 0bd7164bc817..7f083c95f422 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_gem.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_gem.c
> @@ -119,7 +119,7 @@ nouveau_gem_object_unmap(struct nouveau_bo *nvbo, struct nvkm_vma *vma)
>  	const bool mapped = nvbo->bo.mem.mem_type != TTM_PL_SYSTEM;
>  	struct reservation_object *resv = nvbo->bo.resv;
>  	struct reservation_object_list *fobj;
> -	struct fence *fence = NULL;
> +	struct dma_fence *fence = NULL;
>  
>  	fobj = reservation_object_get_list(resv);
>  
> diff --git a/drivers/gpu/drm/nouveau/nv04_fence.c b/drivers/gpu/drm/nouveau/nv04_fence.c
> index 1915b7b82a59..fa8f2375c398 100644
> --- a/drivers/gpu/drm/nouveau/nv04_fence.c
> +++ b/drivers/gpu/drm/nouveau/nv04_fence.c
> @@ -110,6 +110,6 @@ nv04_fence_create(struct nouveau_drm *drm)
>  	priv->base.context_new = nv04_fence_context_new;
>  	priv->base.context_del = nv04_fence_context_del;
>  	priv->base.contexts = 15;
> -	priv->base.context_base = fence_context_alloc(priv->base.contexts);
> +	priv->base.context_base = dma_fence_context_alloc(priv->base.contexts);
>  	return 0;
>  }
> diff --git a/drivers/gpu/drm/nouveau/nv10_fence.c b/drivers/gpu/drm/nouveau/nv10_fence.c
> index 4e3de34ff6f4..f99fcf56928a 100644
> --- a/drivers/gpu/drm/nouveau/nv10_fence.c
> +++ b/drivers/gpu/drm/nouveau/nv10_fence.c
> @@ -107,7 +107,7 @@ nv10_fence_create(struct nouveau_drm *drm)
>  	priv->base.context_new = nv10_fence_context_new;
>  	priv->base.context_del = nv10_fence_context_del;
>  	priv->base.contexts = 31;
> -	priv->base.context_base = fence_context_alloc(priv->base.contexts);
> +	priv->base.context_base = dma_fence_context_alloc(priv->base.contexts);
>  	spin_lock_init(&priv->lock);
>  	return 0;
>  }
> diff --git a/drivers/gpu/drm/nouveau/nv17_fence.c b/drivers/gpu/drm/nouveau/nv17_fence.c
> index 7d5e562a55c5..79bc01111351 100644
> --- a/drivers/gpu/drm/nouveau/nv17_fence.c
> +++ b/drivers/gpu/drm/nouveau/nv17_fence.c
> @@ -126,7 +126,7 @@ nv17_fence_create(struct nouveau_drm *drm)
>  	priv->base.context_new = nv17_fence_context_new;
>  	priv->base.context_del = nv10_fence_context_del;
>  	priv->base.contexts = 31;
> -	priv->base.context_base = fence_context_alloc(priv->base.contexts);
> +	priv->base.context_base = dma_fence_context_alloc(priv->base.contexts);
>  	spin_lock_init(&priv->lock);
>  
>  	ret = nouveau_bo_new(drm->dev, 4096, 0x1000, TTM_PL_FLAG_VRAM,
> diff --git a/drivers/gpu/drm/nouveau/nv50_fence.c b/drivers/gpu/drm/nouveau/nv50_fence.c
> index 4d6f202b7770..8c5295414578 100644
> --- a/drivers/gpu/drm/nouveau/nv50_fence.c
> +++ b/drivers/gpu/drm/nouveau/nv50_fence.c
> @@ -97,7 +97,7 @@ nv50_fence_create(struct nouveau_drm *drm)
>  	priv->base.context_new = nv50_fence_context_new;
>  	priv->base.context_del = nv10_fence_context_del;
>  	priv->base.contexts = 127;
> -	priv->base.context_base = fence_context_alloc(priv->base.contexts);
> +	priv->base.context_base = dma_fence_context_alloc(priv->base.contexts);
>  	spin_lock_init(&priv->lock);
>  
>  	ret = nouveau_bo_new(drm->dev, 4096, 0x1000, TTM_PL_FLAG_VRAM,
> diff --git a/drivers/gpu/drm/nouveau/nv84_fence.c b/drivers/gpu/drm/nouveau/nv84_fence.c
> index 18bde9d8e6d6..23ef04b4e0b2 100644
> --- a/drivers/gpu/drm/nouveau/nv84_fence.c
> +++ b/drivers/gpu/drm/nouveau/nv84_fence.c
> @@ -229,7 +229,7 @@ nv84_fence_create(struct nouveau_drm *drm)
>  	priv->base.context_del = nv84_fence_context_del;
>  
>  	priv->base.contexts = fifo->nr;
> -	priv->base.context_base = fence_context_alloc(priv->base.contexts);
> +	priv->base.context_base = dma_fence_context_alloc(priv->base.contexts);
>  	priv->base.uevent = true;
>  
>  	/* Use VRAM if there is any ; otherwise fallback to system memory */
> diff --git a/drivers/gpu/drm/qxl/qxl_drv.h b/drivers/gpu/drm/qxl/qxl_drv.h
> index 5f3e5ad99de7..84995ebc6ffc 100644
> --- a/drivers/gpu/drm/qxl/qxl_drv.h
> +++ b/drivers/gpu/drm/qxl/qxl_drv.h
> @@ -31,7 +31,7 @@
>   * Definitions taken from spice-protocol, plus kernel driver specific bits.
>   */
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/workqueue.h>
>  #include <linux/firmware.h>
>  #include <linux/platform_device.h>
> @@ -190,7 +190,7 @@ enum {
>   * spice-protocol/qxl_dev.h */
>  #define QXL_MAX_RES 96
>  struct qxl_release {
> -	struct fence base;
> +	struct dma_fence base;
>  
>  	int id;
>  	int type;
> diff --git a/drivers/gpu/drm/qxl/qxl_release.c b/drivers/gpu/drm/qxl/qxl_release.c
> index cd83f050cf3e..00dc43be1369 100644
> --- a/drivers/gpu/drm/qxl/qxl_release.c
> +++ b/drivers/gpu/drm/qxl/qxl_release.c
> @@ -21,7 +21,7 @@
>   */
>  #include "qxl_drv.h"
>  #include "qxl_object.h"
> -#include <trace/events/fence.h>
> +#include <trace/events/dma_fence.h>
>  
>  /*
>   * drawable cmd cache - allocate a bunch of VRAM pages, suballocate
> @@ -40,23 +40,23 @@
>  static const int release_size_per_bo[] = { RELEASE_SIZE, SURFACE_RELEASE_SIZE, RELEASE_SIZE };
>  static const int releases_per_bo[] = { RELEASES_PER_BO, SURFACE_RELEASES_PER_BO, RELEASES_PER_BO };
>  
> -static const char *qxl_get_driver_name(struct fence *fence)
> +static const char *qxl_get_driver_name(struct dma_fence *fence)
>  {
>  	return "qxl";
>  }
>  
> -static const char *qxl_get_timeline_name(struct fence *fence)
> +static const char *qxl_get_timeline_name(struct dma_fence *fence)
>  {
>  	return "release";
>  }
>  
> -static bool qxl_nop_signaling(struct fence *fence)
> +static bool qxl_nop_signaling(struct dma_fence *fence)
>  {
>  	/* fences are always automatically signaled, so just pretend we did this.. */
>  	return true;
>  }
>  
> -static long qxl_fence_wait(struct fence *fence, bool intr, signed long timeout)
> +static long qxl_fence_wait(struct dma_fence *fence, bool intr, signed long timeout)
>  {
>  	struct qxl_device *qdev;
>  	struct qxl_release *release;
> @@ -71,7 +71,7 @@ static long qxl_fence_wait(struct fence *fence, bool intr, signed long timeout)
>  retry:
>  	sc++;
>  
> -	if (fence_is_signaled(fence))
> +	if (dma_fence_is_signaled(fence))
>  		goto signaled;
>  
>  	qxl_io_notify_oom(qdev);
> @@ -80,11 +80,11 @@ static long qxl_fence_wait(struct fence *fence, bool intr, signed long timeout)
>  		if (!qxl_queue_garbage_collect(qdev, true))
>  			break;
>  
> -		if (fence_is_signaled(fence))
> +		if (dma_fence_is_signaled(fence))
>  			goto signaled;
>  	}
>  
> -	if (fence_is_signaled(fence))
> +	if (dma_fence_is_signaled(fence))
>  		goto signaled;
>  
>  	if (have_drawable_releases || sc < 4) {
> @@ -96,9 +96,10 @@ static long qxl_fence_wait(struct fence *fence, bool intr, signed long timeout)
>  			return 0;
>  
>  		if (have_drawable_releases && sc > 300) {
> -			FENCE_WARN(fence, "failed to wait on release %llu "
> -					  "after spincount %d\n",
> -					  fence->context & ~0xf0000000, sc);
> +			DMA_FENCE_WARN(fence,
> +				       "failed to wait on release %llu "
> +				       "after spincount %d\n",
> +				       fence->context & ~0xf0000000, sc);
>  			goto signaled;
>  		}
>  		goto retry;
> @@ -115,7 +116,7 @@ static long qxl_fence_wait(struct fence *fence, bool intr, signed long timeout)
>  	return end - cur;
>  }
>  
> -static const struct fence_ops qxl_fence_ops = {
> +static const struct dma_fence_ops qxl_fence_ops = {
>  	.get_driver_name = qxl_get_driver_name,
>  	.get_timeline_name = qxl_get_timeline_name,
>  	.enable_signaling = qxl_nop_signaling,
> @@ -192,8 +193,8 @@ qxl_release_free(struct qxl_device *qdev,
>  		WARN_ON(list_empty(&release->bos));
>  		qxl_release_free_list(release);
>  
> -		fence_signal(&release->base);
> -		fence_put(&release->base);
> +		dma_fence_signal(&release->base);
> +		dma_fence_put(&release->base);
>  	} else {
>  		qxl_release_free_list(release);
>  		kfree(release);
> @@ -453,9 +454,9 @@ void qxl_release_fence_buffer_objects(struct qxl_release *release)
>  	 * Since we never really allocated a context and we don't want to conflict,
>  	 * set the highest bits. This will break if we really allow exporting of dma-bufs.
>  	 */
> -	fence_init(&release->base, &qxl_fence_ops, &qdev->release_lock,
> -		   release->id | 0xf0000000, release->base.seqno);
> -	trace_fence_emit(&release->base);
> +	dma_fence_init(&release->base, &qxl_fence_ops, &qdev->release_lock,
> +		       release->id | 0xf0000000, release->base.seqno);
> +	trace_dma_fence_emit(&release->base);
>  
>  	driver = bdev->driver;
>  	glob = bo->glob;
> diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h
> index 1b0dcad916b0..44e0c5ed6418 100644
> --- a/drivers/gpu/drm/radeon/radeon.h
> +++ b/drivers/gpu/drm/radeon/radeon.h
> @@ -66,7 +66,7 @@
>  #include <linux/kref.h>
>  #include <linux/interval_tree.h>
>  #include <linux/hashtable.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include <ttm/ttm_bo_api.h>
>  #include <ttm/ttm_bo_driver.h>
> @@ -367,7 +367,7 @@ struct radeon_fence_driver {
>  };
>  
>  struct radeon_fence {
> -	struct fence		base;
> +	struct dma_fence		base;
>  
>  	struct radeon_device	*rdev;
>  	uint64_t		seq;
> @@ -746,7 +746,7 @@ struct radeon_flip_work {
>  	uint64_t			base;
>  	struct drm_pending_vblank_event *event;
>  	struct radeon_bo		*old_rbo;
> -	struct fence			*fence;
> +	struct dma_fence		*fence;
>  	bool				async;
>  };
>  
> @@ -2514,9 +2514,9 @@ void cik_mm_wdoorbell(struct radeon_device *rdev, u32 index, u32 v);
>  /*
>   * Cast helper
>   */
> -extern const struct fence_ops radeon_fence_ops;
> +extern const struct dma_fence_ops radeon_fence_ops;
>  
> -static inline struct radeon_fence *to_radeon_fence(struct fence *f)
> +static inline struct radeon_fence *to_radeon_fence(struct dma_fence *f)
>  {
>  	struct radeon_fence *__f = container_of(f, struct radeon_fence, base);
>  
> diff --git a/drivers/gpu/drm/radeon/radeon_device.c b/drivers/gpu/drm/radeon/radeon_device.c
> index eb92aef46e3c..36b7ac7e57e5 100644
> --- a/drivers/gpu/drm/radeon/radeon_device.c
> +++ b/drivers/gpu/drm/radeon/radeon_device.c
> @@ -1320,7 +1320,7 @@ int radeon_device_init(struct radeon_device *rdev,
>  	for (i = 0; i < RADEON_NUM_RINGS; i++) {
>  		rdev->ring[i].idx = i;
>  	}
> -	rdev->fence_context = fence_context_alloc(RADEON_NUM_RINGS);
> +	rdev->fence_context = dma_fence_context_alloc(RADEON_NUM_RINGS);
>  
>  	DRM_INFO("initializing kernel modesetting (%s 0x%04X:0x%04X 0x%04X:0x%04X 0x%02X).\n",
>  		 radeon_family_name[rdev->family], pdev->vendor, pdev->device,
> diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c
> index cdb8cb568c15..e7409e8a9f87 100644
> --- a/drivers/gpu/drm/radeon/radeon_display.c
> +++ b/drivers/gpu/drm/radeon/radeon_display.c
> @@ -437,7 +437,7 @@ static void radeon_flip_work_func(struct work_struct *__work)
>  				down_read(&rdev->exclusive_lock);
>  			}
>  		} else
> -			r = fence_wait(work->fence, false);
> +			r = dma_fence_wait(work->fence, false);
>  
>  		if (r)
>  			DRM_ERROR("failed to wait on page flip fence (%d)!\n", r);
> @@ -447,7 +447,7 @@ static void radeon_flip_work_func(struct work_struct *__work)
>  		 * confused about which BO the CRTC is scanning out
>  		 */
>  
> -		fence_put(work->fence);
> +		dma_fence_put(work->fence);
>  		work->fence = NULL;
>  	}
>  
> @@ -542,7 +542,7 @@ static int radeon_crtc_page_flip_target(struct drm_crtc *crtc,
>  		DRM_ERROR("failed to pin new rbo buffer before flip\n");
>  		goto cleanup;
>  	}
> -	work->fence = fence_get(reservation_object_get_excl(new_rbo->tbo.resv));
> +	work->fence = dma_fence_get(reservation_object_get_excl(new_rbo->tbo.resv));
>  	radeon_bo_get_tiling_flags(new_rbo, &tiling_flags, NULL);
>  	radeon_bo_unreserve(new_rbo);
>  
> @@ -617,7 +617,7 @@ static int radeon_crtc_page_flip_target(struct drm_crtc *crtc,
>  
>  cleanup:
>  	drm_gem_object_unreference_unlocked(&work->old_rbo->gem_base);
> -	fence_put(work->fence);
> +	dma_fence_put(work->fence);
>  	kfree(work);
>  	return r;
>  }
> diff --git a/drivers/gpu/drm/radeon/radeon_fence.c b/drivers/gpu/drm/radeon/radeon_fence.c
> index 7ef075acde9c..ef09f0a63754 100644
> --- a/drivers/gpu/drm/radeon/radeon_fence.c
> +++ b/drivers/gpu/drm/radeon/radeon_fence.c
> @@ -141,8 +141,10 @@ int radeon_fence_emit(struct radeon_device *rdev,
>  	(*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring];
>  	(*fence)->ring = ring;
>  	(*fence)->is_vm_update = false;
> -	fence_init(&(*fence)->base, &radeon_fence_ops,
> -		   &rdev->fence_queue.lock, rdev->fence_context + ring, seq);
> +	dma_fence_init(&(*fence)->base, &radeon_fence_ops,
> +		       &rdev->fence_queue.lock,
> +		       rdev->fence_context + ring,
> +		       seq);
>  	radeon_fence_ring_emit(rdev, ring, *fence);
>  	trace_radeon_fence_emit(rdev->ddev, ring, (*fence)->seq);
>  	radeon_fence_schedule_check(rdev, ring);
> @@ -169,18 +171,18 @@ static int radeon_fence_check_signaled(wait_queue_t *wait, unsigned mode, int fl
>  	 */
>  	seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq);
>  	if (seq >= fence->seq) {
> -		int ret = fence_signal_locked(&fence->base);
> +		int ret = dma_fence_signal_locked(&fence->base);
>  
>  		if (!ret)
> -			FENCE_TRACE(&fence->base, "signaled from irq context\n");
> +			DMA_FENCE_TRACE(&fence->base, "signaled from irq context\n");
>  		else
> -			FENCE_TRACE(&fence->base, "was already signaled\n");
> +			DMA_FENCE_TRACE(&fence->base, "was already signaled\n");
>  
>  		radeon_irq_kms_sw_irq_put(fence->rdev, fence->ring);
>  		__remove_wait_queue(&fence->rdev->fence_queue, &fence->fence_wake);
> -		fence_put(&fence->base);
> +		dma_fence_put(&fence->base);
>  	} else
> -		FENCE_TRACE(&fence->base, "pending\n");
> +		DMA_FENCE_TRACE(&fence->base, "pending\n");
>  	return 0;
>  }
>  
> @@ -351,7 +353,7 @@ static bool radeon_fence_seq_signaled(struct radeon_device *rdev,
>  	return false;
>  }
>  
> -static bool radeon_fence_is_signaled(struct fence *f)
> +static bool radeon_fence_is_signaled(struct dma_fence *f)
>  {
>  	struct radeon_fence *fence = to_radeon_fence(f);
>  	struct radeon_device *rdev = fence->rdev;
> @@ -381,7 +383,7 @@ static bool radeon_fence_is_signaled(struct fence *f)
>   * to fence_queue that checks if this fence is signaled, and if so it
>   * signals the fence and removes itself.
>   */
> -static bool radeon_fence_enable_signaling(struct fence *f)
> +static bool radeon_fence_enable_signaling(struct dma_fence *f)
>  {
>  	struct radeon_fence *fence = to_radeon_fence(f);
>  	struct radeon_device *rdev = fence->rdev;
> @@ -414,9 +416,9 @@ static bool radeon_fence_enable_signaling(struct fence *f)
>  	fence->fence_wake.private = NULL;
>  	fence->fence_wake.func = radeon_fence_check_signaled;
>  	__add_wait_queue(&rdev->fence_queue, &fence->fence_wake);
> -	fence_get(f);
> +	dma_fence_get(f);
>  
> -	FENCE_TRACE(&fence->base, "armed on ring %i!\n", fence->ring);
> +	DMA_FENCE_TRACE(&fence->base, "armed on ring %i!\n", fence->ring);
>  	return true;
>  }
>  
> @@ -436,9 +438,9 @@ bool radeon_fence_signaled(struct radeon_fence *fence)
>  	if (radeon_fence_seq_signaled(fence->rdev, fence->seq, fence->ring)) {
>  		int ret;
>  
> -		ret = fence_signal(&fence->base);
> +		ret = dma_fence_signal(&fence->base);
>  		if (!ret)
> -			FENCE_TRACE(&fence->base, "signaled from radeon_fence_signaled\n");
> +			DMA_FENCE_TRACE(&fence->base, "signaled from radeon_fence_signaled\n");
>  		return true;
>  	}
>  	return false;
> @@ -552,7 +554,7 @@ long radeon_fence_wait_timeout(struct radeon_fence *fence, bool intr, long timeo
>  	 * exclusive_lock is not held in that case.
>  	 */
>  	if (WARN_ON_ONCE(!to_radeon_fence(&fence->base)))
> -		return fence_wait(&fence->base, intr);
> +		return dma_fence_wait(&fence->base, intr);
>  
>  	seq[fence->ring] = fence->seq;
>  	r = radeon_fence_wait_seq_timeout(fence->rdev, seq, intr, timeout);
> @@ -560,9 +562,9 @@ long radeon_fence_wait_timeout(struct radeon_fence *fence, bool intr, long timeo
>  		return r;
>  	}
>  
> -	r_sig = fence_signal(&fence->base);
> +	r_sig = dma_fence_signal(&fence->base);
>  	if (!r_sig)
> -		FENCE_TRACE(&fence->base, "signaled from fence_wait\n");
> +		DMA_FENCE_TRACE(&fence->base, "signaled from fence_wait\n");
>  	return r;
>  }
>  
> @@ -697,7 +699,7 @@ int radeon_fence_wait_empty(struct radeon_device *rdev, int ring)
>   */
>  struct radeon_fence *radeon_fence_ref(struct radeon_fence *fence)
>  {
> -	fence_get(&fence->base);
> +	dma_fence_get(&fence->base);
>  	return fence;
>  }
>  
> @@ -714,7 +716,7 @@ void radeon_fence_unref(struct radeon_fence **fence)
>  
>  	*fence = NULL;
>  	if (tmp) {
> -		fence_put(&tmp->base);
> +		dma_fence_put(&tmp->base);
>  	}
>  }
>  
> @@ -1028,12 +1030,12 @@ int radeon_debugfs_fence_init(struct radeon_device *rdev)
>  #endif
>  }
>  
> -static const char *radeon_fence_get_driver_name(struct fence *fence)
> +static const char *radeon_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "radeon";
>  }
>  
> -static const char *radeon_fence_get_timeline_name(struct fence *f)
> +static const char *radeon_fence_get_timeline_name(struct dma_fence *f)
>  {
>  	struct radeon_fence *fence = to_radeon_fence(f);
>  	switch (fence->ring) {
> @@ -1051,16 +1053,16 @@ static const char *radeon_fence_get_timeline_name(struct fence *f)
>  
>  static inline bool radeon_test_signaled(struct radeon_fence *fence)
>  {
> -	return test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags);
> +	return test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags);
>  }
>  
>  struct radeon_wait_cb {
> -	struct fence_cb base;
> +	struct dma_fence_cb base;
>  	struct task_struct *task;
>  };
>  
>  static void
> -radeon_fence_wait_cb(struct fence *fence, struct fence_cb *cb)
> +radeon_fence_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	struct radeon_wait_cb *wait =
>  		container_of(cb, struct radeon_wait_cb, base);
> @@ -1068,7 +1070,7 @@ radeon_fence_wait_cb(struct fence *fence, struct fence_cb *cb)
>  	wake_up_process(wait->task);
>  }
>  
> -static signed long radeon_fence_default_wait(struct fence *f, bool intr,
> +static signed long radeon_fence_default_wait(struct dma_fence *f, bool intr,
>  					     signed long t)
>  {
>  	struct radeon_fence *fence = to_radeon_fence(f);
> @@ -1077,7 +1079,7 @@ static signed long radeon_fence_default_wait(struct fence *f, bool intr,
>  
>  	cb.task = current;
>  
> -	if (fence_add_callback(f, &cb.base, radeon_fence_wait_cb))
> +	if (dma_fence_add_callback(f, &cb.base, radeon_fence_wait_cb))
>  		return t;
>  
>  	while (t > 0) {
> @@ -1105,12 +1107,12 @@ static signed long radeon_fence_default_wait(struct fence *f, bool intr,
>  	}
>  
>  	__set_current_state(TASK_RUNNING);
> -	fence_remove_callback(f, &cb.base);
> +	dma_fence_remove_callback(f, &cb.base);
>  
>  	return t;
>  }
>  
> -const struct fence_ops radeon_fence_ops = {
> +const struct dma_fence_ops radeon_fence_ops = {
>  	.get_driver_name = radeon_fence_get_driver_name,
>  	.get_timeline_name = radeon_fence_get_timeline_name,
>  	.enable_signaling = radeon_fence_enable_signaling,
> diff --git a/drivers/gpu/drm/radeon/radeon_sync.c b/drivers/gpu/drm/radeon/radeon_sync.c
> index 02ac8a1de4ff..be5d7a38d3aa 100644
> --- a/drivers/gpu/drm/radeon/radeon_sync.c
> +++ b/drivers/gpu/drm/radeon/radeon_sync.c
> @@ -92,7 +92,7 @@ int radeon_sync_resv(struct radeon_device *rdev,
>  		     bool shared)
>  {
>  	struct reservation_object_list *flist;
> -	struct fence *f;
> +	struct dma_fence *f;
>  	struct radeon_fence *fence;
>  	unsigned i;
>  	int r = 0;
> @@ -103,7 +103,7 @@ int radeon_sync_resv(struct radeon_device *rdev,
>  	if (fence && fence->rdev == rdev)
>  		radeon_sync_fence(sync, fence);
>  	else if (f)
> -		r = fence_wait(f, true);
> +		r = dma_fence_wait(f, true);
>  
>  	flist = reservation_object_get_list(resv);
>  	if (shared || !flist || r)
> @@ -116,7 +116,7 @@ int radeon_sync_resv(struct radeon_device *rdev,
>  		if (fence && fence->rdev == rdev)
>  			radeon_sync_fence(sync, fence);
>  		else
> -			r = fence_wait(f, true);
> +			r = dma_fence_wait(f, true);
>  
>  		if (r)
>  			break;
> diff --git a/drivers/gpu/drm/radeon/radeon_uvd.c b/drivers/gpu/drm/radeon/radeon_uvd.c
> index 0cd0e7bdee55..d34d1cf33895 100644
> --- a/drivers/gpu/drm/radeon/radeon_uvd.c
> +++ b/drivers/gpu/drm/radeon/radeon_uvd.c
> @@ -467,7 +467,7 @@ static int radeon_uvd_cs_msg(struct radeon_cs_parser *p, struct radeon_bo *bo,
>  {
>  	int32_t *msg, msg_type, handle;
>  	unsigned img_size = 0;
> -	struct fence *f;
> +	struct dma_fence *f;
>  	void *ptr;
>  
>  	int i, r;
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index fc6217dfe401..915e0d1c316a 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -148,7 +148,7 @@ static void ttm_bo_release_list(struct kref *list_kref)
>  	BUG_ON(!list_empty(&bo->ddestroy));
>  	ttm_tt_destroy(bo->ttm);
>  	atomic_dec(&bo->glob->bo_count);
> -	fence_put(bo->moving);
> +	dma_fence_put(bo->moving);
>  	if (bo->resv == &bo->ttm_resv)
>  		reservation_object_fini(&bo->ttm_resv);
>  	mutex_destroy(&bo->wu_mutex);
> @@ -426,20 +426,20 @@ static void ttm_bo_cleanup_memtype_use(struct ttm_buffer_object *bo)
>  static void ttm_bo_flush_all_fences(struct ttm_buffer_object *bo)
>  {
>  	struct reservation_object_list *fobj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int i;
>  
>  	fobj = reservation_object_get_list(bo->resv);
>  	fence = reservation_object_get_excl(bo->resv);
>  	if (fence && !fence->ops->signaled)
> -		fence_enable_sw_signaling(fence);
> +		dma_fence_enable_sw_signaling(fence);
>  
>  	for (i = 0; fobj && i < fobj->shared_count; ++i) {
>  		fence = rcu_dereference_protected(fobj->shared[i],
>  					reservation_object_held(bo->resv));
>  
>  		if (!fence->ops->signaled)
> -			fence_enable_sw_signaling(fence);
> +			dma_fence_enable_sw_signaling(fence);
>  	}
>  }
>  
> @@ -792,11 +792,11 @@ static int ttm_bo_add_move_fence(struct ttm_buffer_object *bo,
>  				 struct ttm_mem_type_manager *man,
>  				 struct ttm_mem_reg *mem)
>  {
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int ret;
>  
>  	spin_lock(&man->move_lock);
> -	fence = fence_get(man->move);
> +	fence = dma_fence_get(man->move);
>  	spin_unlock(&man->move_lock);
>  
>  	if (fence) {
> @@ -806,7 +806,7 @@ static int ttm_bo_add_move_fence(struct ttm_buffer_object *bo,
>  		if (unlikely(ret))
>  			return ret;
>  
> -		fence_put(bo->moving);
> +		dma_fence_put(bo->moving);
>  		bo->moving = fence;
>  	}
>  
> @@ -1286,7 +1286,7 @@ static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev,
>  {
>  	struct ttm_mem_type_manager *man = &bdev->man[mem_type];
>  	struct ttm_bo_global *glob = bdev->glob;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int ret;
>  
>  	/*
> @@ -1309,12 +1309,12 @@ static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev,
>  	spin_unlock(&glob->lru_lock);
>  
>  	spin_lock(&man->move_lock);
> -	fence = fence_get(man->move);
> +	fence = dma_fence_get(man->move);
>  	spin_unlock(&man->move_lock);
>  
>  	if (fence) {
> -		ret = fence_wait(fence, false);
> -		fence_put(fence);
> +		ret = dma_fence_wait(fence, false);
> +		dma_fence_put(fence);
>  		if (ret) {
>  			if (allow_errors) {
>  				return ret;
> @@ -1343,7 +1343,7 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type)
>  		       mem_type);
>  		return ret;
>  	}
> -	fence_put(man->move);
> +	dma_fence_put(man->move);
>  
>  	man->use_type = false;
>  	man->has_type = false;
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
> index bf6e21655c57..d0459b392e5e 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
> @@ -644,7 +644,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map)
>  EXPORT_SYMBOL(ttm_bo_kunmap);
>  
>  int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
> -			      struct fence *fence,
> +			      struct dma_fence *fence,
>  			      bool evict,
>  			      struct ttm_mem_reg *new_mem)
>  {
> @@ -674,8 +674,8 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
>  		 * operation has completed.
>  		 */
>  
> -		fence_put(bo->moving);
> -		bo->moving = fence_get(fence);
> +		dma_fence_put(bo->moving);
> +		bo->moving = dma_fence_get(fence);
>  
>  		ret = ttm_buffer_object_transfer(bo, &ghost_obj);
>  		if (ret)
> @@ -706,7 +706,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
>  EXPORT_SYMBOL(ttm_bo_move_accel_cleanup);
>  
>  int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
> -			 struct fence *fence, bool evict,
> +			 struct dma_fence *fence, bool evict,
>  			 struct ttm_mem_reg *new_mem)
>  {
>  	struct ttm_bo_device *bdev = bo->bdev;
> @@ -730,8 +730,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
>  		 * operation has completed.
>  		 */
>  
> -		fence_put(bo->moving);
> -		bo->moving = fence_get(fence);
> +		dma_fence_put(bo->moving);
> +		bo->moving = dma_fence_get(fence);
>  
>  		ret = ttm_buffer_object_transfer(bo, &ghost_obj);
>  		if (ret)
> @@ -761,16 +761,16 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
>  		 */
>  
>  		spin_lock(&from->move_lock);
> -		if (!from->move || fence_is_later(fence, from->move)) {
> -			fence_put(from->move);
> -			from->move = fence_get(fence);
> +		if (!from->move || dma_fence_is_later(fence, from->move)) {
> +			dma_fence_put(from->move);
> +			from->move = dma_fence_get(fence);
>  		}
>  		spin_unlock(&from->move_lock);
>  
>  		ttm_bo_free_old_node(bo);
>  
> -		fence_put(bo->moving);
> -		bo->moving = fence_get(fence);
> +		dma_fence_put(bo->moving);
> +		bo->moving = dma_fence_get(fence);
>  
>  	} else {
>  		/**
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> index a6ed9d5e5167..4748aedc933a 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> @@ -54,7 +54,7 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>  	/*
>  	 * Quick non-stalling check for idle.
>  	 */
> -	if (fence_is_signaled(bo->moving))
> +	if (dma_fence_is_signaled(bo->moving))
>  		goto out_clear;
>  
>  	/*
> @@ -67,14 +67,14 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>  			goto out_unlock;
>  
>  		up_read(&vma->vm_mm->mmap_sem);
> -		(void) fence_wait(bo->moving, true);
> +		(void) dma_fence_wait(bo->moving, true);
>  		goto out_unlock;
>  	}
>  
>  	/*
>  	 * Ordinary wait.
>  	 */
> -	ret = fence_wait(bo->moving, true);
> +	ret = dma_fence_wait(bo->moving, true);
>  	if (unlikely(ret != 0)) {
>  		ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
>  			VM_FAULT_NOPAGE;
> @@ -82,7 +82,7 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>  	}
>  
>  out_clear:
> -	fence_put(bo->moving);
> +	dma_fence_put(bo->moving);
>  	bo->moving = NULL;
>  
>  out_unlock:
> diff --git a/drivers/gpu/drm/ttm/ttm_execbuf_util.c b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> index a80717b35dc6..d35bc491e8de 100644
> --- a/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> @@ -179,7 +179,8 @@ int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket,
>  EXPORT_SYMBOL(ttm_eu_reserve_buffers);
>  
>  void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket,
> -				 struct list_head *list, struct fence *fence)
> +				 struct list_head *list,
> +				 struct dma_fence *fence)
>  {
>  	struct ttm_validate_buffer *entry;
>  	struct ttm_buffer_object *bo;
> diff --git a/drivers/gpu/drm/vgem/vgem_fence.c b/drivers/gpu/drm/vgem/vgem_fence.c
> index 5c57c1ffa1f9..488909a21ed8 100644
> --- a/drivers/gpu/drm/vgem/vgem_fence.c
> +++ b/drivers/gpu/drm/vgem/vgem_fence.c
> @@ -28,56 +28,57 @@
>  #define VGEM_FENCE_TIMEOUT (10*HZ)
>  
>  struct vgem_fence {
> -	struct fence base;
> +	struct dma_fence base;
>  	struct spinlock lock;
>  	struct timer_list timer;
>  };
>  
> -static const char *vgem_fence_get_driver_name(struct fence *fence)
> +static const char *vgem_fence_get_driver_name(struct dma_fence *fence)
>  {
>  	return "vgem";
>  }
>  
> -static const char *vgem_fence_get_timeline_name(struct fence *fence)
> +static const char *vgem_fence_get_timeline_name(struct dma_fence *fence)
>  {
>  	return "unbound";
>  }
>  
> -static bool vgem_fence_signaled(struct fence *fence)
> +static bool vgem_fence_signaled(struct dma_fence *fence)
>  {
>  	return false;
>  }
>  
> -static bool vgem_fence_enable_signaling(struct fence *fence)
> +static bool vgem_fence_enable_signaling(struct dma_fence *fence)
>  {
>  	return true;
>  }
>  
> -static void vgem_fence_release(struct fence *base)
> +static void vgem_fence_release(struct dma_fence *base)
>  {
>  	struct vgem_fence *fence = container_of(base, typeof(*fence), base);
>  
>  	del_timer_sync(&fence->timer);
> -	fence_free(&fence->base);
> +	dma_fence_free(&fence->base);
>  }
>  
> -static void vgem_fence_value_str(struct fence *fence, char *str, int size)
> +static void vgem_fence_value_str(struct dma_fence *fence, char *str, int size)
>  {
>  	snprintf(str, size, "%u", fence->seqno);
>  }
>  
> -static void vgem_fence_timeline_value_str(struct fence *fence, char *str,
> +static void vgem_fence_timeline_value_str(struct dma_fence *fence, char *str,
>  					  int size)
>  {
> -	snprintf(str, size, "%u", fence_is_signaled(fence) ? fence->seqno : 0);
> +	snprintf(str, size, "%u",
> +		 dma_fence_is_signaled(fence) ? fence->seqno : 0);
>  }
>  
> -static const struct fence_ops vgem_fence_ops = {
> +static const struct dma_fence_ops vgem_fence_ops = {
>  	.get_driver_name = vgem_fence_get_driver_name,
>  	.get_timeline_name = vgem_fence_get_timeline_name,
>  	.enable_signaling = vgem_fence_enable_signaling,
>  	.signaled = vgem_fence_signaled,
> -	.wait = fence_default_wait,
> +	.wait = dma_fence_default_wait,
>  	.release = vgem_fence_release,
>  
>  	.fence_value_str = vgem_fence_value_str,
> @@ -88,11 +89,11 @@ static void vgem_fence_timeout(unsigned long data)
>  {
>  	struct vgem_fence *fence = (struct vgem_fence *)data;
>  
> -	fence_signal(&fence->base);
> +	dma_fence_signal(&fence->base);
>  }
>  
> -static struct fence *vgem_fence_create(struct vgem_file *vfile,
> -				       unsigned int flags)
> +static struct dma_fence *vgem_fence_create(struct vgem_file *vfile,
> +					   unsigned int flags)
>  {
>  	struct vgem_fence *fence;
>  
> @@ -101,8 +102,8 @@ static struct fence *vgem_fence_create(struct vgem_file *vfile,
>  		return NULL;
>  
>  	spin_lock_init(&fence->lock);
> -	fence_init(&fence->base, &vgem_fence_ops, &fence->lock,
> -		   fence_context_alloc(1), 1);
> +	dma_fence_init(&fence->base, &vgem_fence_ops, &fence->lock,
> +		       dma_fence_context_alloc(1), 1);
>  
>  	setup_timer(&fence->timer, vgem_fence_timeout, (unsigned long)fence);
>  
> @@ -157,7 +158,7 @@ int vgem_fence_attach_ioctl(struct drm_device *dev,
>  	struct vgem_file *vfile = file->driver_priv;
>  	struct reservation_object *resv;
>  	struct drm_gem_object *obj;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int ret;
>  
>  	if (arg->flags & ~VGEM_FENCE_WRITE)
> @@ -209,8 +210,8 @@ int vgem_fence_attach_ioctl(struct drm_device *dev,
>  	}
>  err_fence:
>  	if (ret) {
> -		fence_signal(fence);
> -		fence_put(fence);
> +		dma_fence_signal(fence);
> +		dma_fence_put(fence);
>  	}
>  err:
>  	drm_gem_object_unreference_unlocked(obj);
> @@ -239,7 +240,7 @@ int vgem_fence_signal_ioctl(struct drm_device *dev,
>  {
>  	struct vgem_file *vfile = file->driver_priv;
>  	struct drm_vgem_fence_signal *arg = data;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	int ret = 0;
>  
>  	if (arg->flags)
> @@ -253,11 +254,11 @@ int vgem_fence_signal_ioctl(struct drm_device *dev,
>  	if (IS_ERR(fence))
>  		return PTR_ERR(fence);
>  
> -	if (fence_is_signaled(fence))
> +	if (dma_fence_is_signaled(fence))
>  		ret = -ETIMEDOUT;
>  
> -	fence_signal(fence);
> -	fence_put(fence);
> +	dma_fence_signal(fence);
> +	dma_fence_put(fence);
>  	return ret;
>  }
>  
> @@ -271,8 +272,8 @@ int vgem_fence_open(struct vgem_file *vfile)
>  
>  static int __vgem_fence_idr_fini(int id, void *p, void *data)
>  {
> -	fence_signal(p);
> -	fence_put(p);
> +	dma_fence_signal(p);
> +	dma_fence_put(p);
>  	return 0;
>  }
>  
> diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h
> index ae59080d63d1..ec1ebdcfe80b 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_drv.h
> +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
> @@ -82,7 +82,7 @@ struct virtio_gpu_fence_driver {
>  };
>  
>  struct virtio_gpu_fence {
> -	struct fence f;
> +	struct dma_fence f;
>  	struct virtio_gpu_fence_driver *drv;
>  	struct list_head node;
>  	uint64_t seq;
> diff --git a/drivers/gpu/drm/virtio/virtgpu_fence.c b/drivers/gpu/drm/virtio/virtgpu_fence.c
> index f3f70fa8a4c7..23353521f903 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_fence.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_fence.c
> @@ -26,22 +26,22 @@
>  #include <drm/drmP.h>
>  #include "virtgpu_drv.h"
>  
> -static const char *virtio_get_driver_name(struct fence *f)
> +static const char *virtio_get_driver_name(struct dma_fence *f)
>  {
>  	return "virtio_gpu";
>  }
>  
> -static const char *virtio_get_timeline_name(struct fence *f)
> +static const char *virtio_get_timeline_name(struct dma_fence *f)
>  {
>  	return "controlq";
>  }
>  
> -static bool virtio_enable_signaling(struct fence *f)
> +static bool virtio_enable_signaling(struct dma_fence *f)
>  {
>  	return true;
>  }
>  
> -static bool virtio_signaled(struct fence *f)
> +static bool virtio_signaled(struct dma_fence *f)
>  {
>  	struct virtio_gpu_fence *fence = to_virtio_fence(f);
>  
> @@ -50,26 +50,26 @@ static bool virtio_signaled(struct fence *f)
>  	return false;
>  }
>  
> -static void virtio_fence_value_str(struct fence *f, char *str, int size)
> +static void virtio_fence_value_str(struct dma_fence *f, char *str, int size)
>  {
>  	struct virtio_gpu_fence *fence = to_virtio_fence(f);
>  
>  	snprintf(str, size, "%llu", fence->seq);
>  }
>  
> -static void virtio_timeline_value_str(struct fence *f, char *str, int size)
> +static void virtio_timeline_value_str(struct dma_fence *f, char *str, int size)
>  {
>  	struct virtio_gpu_fence *fence = to_virtio_fence(f);
>  
>  	snprintf(str, size, "%llu", (u64)atomic64_read(&fence->drv->last_seq));
>  }
>  
> -static const struct fence_ops virtio_fence_ops = {
> +static const struct dma_fence_ops virtio_fence_ops = {
>  	.get_driver_name     = virtio_get_driver_name,
>  	.get_timeline_name   = virtio_get_timeline_name,
>  	.enable_signaling    = virtio_enable_signaling,
>  	.signaled            = virtio_signaled,
> -	.wait                = fence_default_wait,
> +	.wait                = dma_fence_default_wait,
>  	.fence_value_str     = virtio_fence_value_str,
>  	.timeline_value_str  = virtio_timeline_value_str,
>  };
> @@ -88,9 +88,9 @@ int virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev,
>  	spin_lock_irqsave(&drv->lock, irq_flags);
>  	(*fence)->drv = drv;
>  	(*fence)->seq = ++drv->sync_seq;
> -	fence_init(&(*fence)->f, &virtio_fence_ops, &drv->lock,
> -		   drv->context, (*fence)->seq);
> -	fence_get(&(*fence)->f);
> +	dma_fence_init(&(*fence)->f, &virtio_fence_ops, &drv->lock,
> +		       drv->context, (*fence)->seq);
> +	dma_fence_get(&(*fence)->f);
>  	list_add_tail(&(*fence)->node, &drv->fences);
>  	spin_unlock_irqrestore(&drv->lock, irq_flags);
>  
> @@ -111,9 +111,9 @@ void virtio_gpu_fence_event_process(struct virtio_gpu_device *vgdev,
>  	list_for_each_entry_safe(fence, tmp, &drv->fences, node) {
>  		if (last_seq < fence->seq)
>  			continue;
> -		fence_signal_locked(&fence->f);
> +		dma_fence_signal_locked(&fence->f);
>  		list_del(&fence->node);
> -		fence_put(&fence->f);
> +		dma_fence_put(&fence->f);
>  	}
>  	spin_unlock_irqrestore(&drv->lock, irq_flags);
>  }
> diff --git a/drivers/gpu/drm/virtio/virtgpu_ioctl.c b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> index 818478b4c4f0..61f3a963af95 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> @@ -172,7 +172,7 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
>  	/* fence the command bo */
>  	virtio_gpu_unref_list(&validate_list);
>  	drm_free_large(buflist);
> -	fence_put(&fence->f);
> +	dma_fence_put(&fence->f);
>  	return 0;
>  
>  out_unresv:
> @@ -298,7 +298,7 @@ static int virtio_gpu_resource_create_ioctl(struct drm_device *dev, void *data,
>  		drm_gem_object_release(obj);
>  		if (vgdev->has_virgl_3d) {
>  			virtio_gpu_unref_list(&validate_list);
> -			fence_put(&fence->f);
> +			dma_fence_put(&fence->f);
>  		}
>  		return ret;
>  	}
> @@ -309,13 +309,13 @@ static int virtio_gpu_resource_create_ioctl(struct drm_device *dev, void *data,
>  
>  	if (vgdev->has_virgl_3d) {
>  		virtio_gpu_unref_list(&validate_list);
> -		fence_put(&fence->f);
> +		dma_fence_put(&fence->f);
>  	}
>  	return 0;
>  fail_unref:
>  	if (vgdev->has_virgl_3d) {
>  		virtio_gpu_unref_list(&validate_list);
> -		fence_put(&fence->f);
> +		dma_fence_put(&fence->f);
>  	}
>  //fail_obj:
>  //	drm_gem_object_handle_unreference_unlocked(obj);
> @@ -383,7 +383,7 @@ static int virtio_gpu_transfer_from_host_ioctl(struct drm_device *dev,
>  	reservation_object_add_excl_fence(qobj->tbo.resv,
>  					  &fence->f);
>  
> -	fence_put(&fence->f);
> +	dma_fence_put(&fence->f);
>  out_unres:
>  	virtio_gpu_object_unreserve(qobj);
>  out:
> @@ -431,7 +431,7 @@ static int virtio_gpu_transfer_to_host_ioctl(struct drm_device *dev, void *data,
>  			 args->level, &box, &fence);
>  		reservation_object_add_excl_fence(qobj->tbo.resv,
>  						  &fence->f);
> -		fence_put(&fence->f);
> +		dma_fence_put(&fence->f);
>  	}
>  
>  out_unres:
> diff --git a/drivers/gpu/drm/virtio/virtgpu_kms.c b/drivers/gpu/drm/virtio/virtgpu_kms.c
> index 036b0fbae0fb..1235519853f4 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_kms.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_kms.c
> @@ -159,7 +159,7 @@ int virtio_gpu_driver_load(struct drm_device *dev, unsigned long flags)
>  	virtio_gpu_init_vq(&vgdev->ctrlq, virtio_gpu_dequeue_ctrl_func);
>  	virtio_gpu_init_vq(&vgdev->cursorq, virtio_gpu_dequeue_cursor_func);
>  
> -	vgdev->fence_drv.context = fence_context_alloc(1);
> +	vgdev->fence_drv.context = dma_fence_context_alloc(1);
>  	spin_lock_init(&vgdev->fence_drv.lock);
>  	INIT_LIST_HEAD(&vgdev->fence_drv.fences);
>  	INIT_LIST_HEAD(&vgdev->cap_cache);
> diff --git a/drivers/gpu/drm/virtio/virtgpu_plane.c b/drivers/gpu/drm/virtio/virtgpu_plane.c
> index ba28c0f6f28a..cb75f0663ba0 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_plane.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_plane.c
> @@ -152,7 +152,7 @@ static void virtio_gpu_cursor_plane_update(struct drm_plane *plane,
>  		if (!ret) {
>  			reservation_object_add_excl_fence(bo->tbo.resv,
>  							  &fence->f);
> -			fence_put(&fence->f);
> +			dma_fence_put(&fence->f);
>  			fence = NULL;
>  			virtio_gpu_object_unreserve(bo);
>  			virtio_gpu_object_wait(bo, false);
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> index 26ac8e80a478..6541dd8b82dc 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> @@ -108,7 +108,7 @@ fman_from_fence(struct vmw_fence_obj *fence)
>   * objects with actions attached to them.
>   */
>  
> -static void vmw_fence_obj_destroy(struct fence *f)
> +static void vmw_fence_obj_destroy(struct dma_fence *f)
>  {
>  	struct vmw_fence_obj *fence =
>  		container_of(f, struct vmw_fence_obj, base);
> @@ -123,17 +123,17 @@ static void vmw_fence_obj_destroy(struct fence *f)
>  	fence->destroy(fence);
>  }
>  
> -static const char *vmw_fence_get_driver_name(struct fence *f)
> +static const char *vmw_fence_get_driver_name(struct dma_fence *f)
>  {
>  	return "vmwgfx";
>  }
>  
> -static const char *vmw_fence_get_timeline_name(struct fence *f)
> +static const char *vmw_fence_get_timeline_name(struct dma_fence *f)
>  {
>  	return "svga";
>  }
>  
> -static bool vmw_fence_enable_signaling(struct fence *f)
> +static bool vmw_fence_enable_signaling(struct dma_fence *f)
>  {
>  	struct vmw_fence_obj *fence =
>  		container_of(f, struct vmw_fence_obj, base);
> @@ -152,12 +152,12 @@ static bool vmw_fence_enable_signaling(struct fence *f)
>  }
>  
>  struct vmwgfx_wait_cb {
> -	struct fence_cb base;
> +	struct dma_fence_cb base;
>  	struct task_struct *task;
>  };
>  
>  static void
> -vmwgfx_wait_cb(struct fence *fence, struct fence_cb *cb)
> +vmwgfx_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>  {
>  	struct vmwgfx_wait_cb *wait =
>  		container_of(cb, struct vmwgfx_wait_cb, base);
> @@ -167,7 +167,7 @@ vmwgfx_wait_cb(struct fence *fence, struct fence_cb *cb)
>  
>  static void __vmw_fences_update(struct vmw_fence_manager *fman);
>  
> -static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
> +static long vmw_fence_wait(struct dma_fence *f, bool intr, signed long timeout)
>  {
>  	struct vmw_fence_obj *fence =
>  		container_of(f, struct vmw_fence_obj, base);
> @@ -197,7 +197,7 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  
>  	while (ret > 0) {
>  		__vmw_fences_update(fman);
> -		if (test_bit(FENCE_FLAG_SIGNALED_BIT, &f->flags))
> +		if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &f->flags))
>  			break;
>  
>  		if (intr)
> @@ -225,7 +225,7 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  	return ret;
>  }
>  
> -static struct fence_ops vmw_fence_ops = {
> +static struct dma_fence_ops vmw_fence_ops = {
>  	.get_driver_name = vmw_fence_get_driver_name,
>  	.get_timeline_name = vmw_fence_get_timeline_name,
>  	.enable_signaling = vmw_fence_enable_signaling,
> @@ -298,7 +298,7 @@ struct vmw_fence_manager *vmw_fence_manager_init(struct vmw_private *dev_priv)
>  	fman->event_fence_action_size =
>  		ttm_round_pot(sizeof(struct vmw_event_fence_action));
>  	mutex_init(&fman->goal_irq_mutex);
> -	fman->ctx = fence_context_alloc(1);
> +	fman->ctx = dma_fence_context_alloc(1);
>  
>  	return fman;
>  }
> @@ -326,8 +326,8 @@ static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
>  	unsigned long irq_flags;
>  	int ret = 0;
>  
> -	fence_init(&fence->base, &vmw_fence_ops, &fman->lock,
> -		   fman->ctx, seqno);
> +	dma_fence_init(&fence->base, &vmw_fence_ops, &fman->lock,
> +		       fman->ctx, seqno);
>  	INIT_LIST_HEAD(&fence->seq_passed_actions);
>  	fence->destroy = destroy;
>  
> @@ -431,7 +431,7 @@ static bool vmw_fence_goal_check_locked(struct vmw_fence_obj *fence)
>  	u32 goal_seqno;
>  	u32 *fifo_mem;
>  
> -	if (fence_is_signaled_locked(&fence->base))
> +	if (dma_fence_is_signaled_locked(&fence->base))
>  		return false;
>  
>  	fifo_mem = fman->dev_priv->mmio_virt;
> @@ -459,7 +459,7 @@ static void __vmw_fences_update(struct vmw_fence_manager *fman)
>  	list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) {
>  		if (seqno - fence->base.seqno < VMW_FENCE_WRAP) {
>  			list_del_init(&fence->head);
> -			fence_signal_locked(&fence->base);
> +			dma_fence_signal_locked(&fence->base);
>  			INIT_LIST_HEAD(&action_list);
>  			list_splice_init(&fence->seq_passed_actions,
>  					 &action_list);
> @@ -500,18 +500,18 @@ bool vmw_fence_obj_signaled(struct vmw_fence_obj *fence)
>  {
>  	struct vmw_fence_manager *fman = fman_from_fence(fence);
>  
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
>  		return 1;
>  
>  	vmw_fences_update(fman);
>  
> -	return fence_is_signaled(&fence->base);
> +	return dma_fence_is_signaled(&fence->base);
>  }
>  
>  int vmw_fence_obj_wait(struct vmw_fence_obj *fence, bool lazy,
>  		       bool interruptible, unsigned long timeout)
>  {
> -	long ret = fence_wait_timeout(&fence->base, interruptible, timeout);
> +	long ret = dma_fence_wait_timeout(&fence->base, interruptible, timeout);
>  
>  	if (likely(ret > 0))
>  		return 0;
> @@ -530,7 +530,7 @@ void vmw_fence_obj_flush(struct vmw_fence_obj *fence)
>  
>  static void vmw_fence_destroy(struct vmw_fence_obj *fence)
>  {
> -	fence_free(&fence->base);
> +	dma_fence_free(&fence->base);
>  }
>  
>  int vmw_fence_create(struct vmw_fence_manager *fman,
> @@ -669,7 +669,7 @@ void vmw_fence_fifo_down(struct vmw_fence_manager *fman)
>  		struct vmw_fence_obj *fence =
>  			list_entry(fman->fence_list.prev, struct vmw_fence_obj,
>  				   head);
> -		fence_get(&fence->base);
> +		dma_fence_get(&fence->base);
>  		spin_unlock_irq(&fman->lock);
>  
>  		ret = vmw_fence_obj_wait(fence, false, false,
> @@ -677,7 +677,7 @@ void vmw_fence_fifo_down(struct vmw_fence_manager *fman)
>  
>  		if (unlikely(ret != 0)) {
>  			list_del_init(&fence->head);
> -			fence_signal(&fence->base);
> +			dma_fence_signal(&fence->base);
>  			INIT_LIST_HEAD(&action_list);
>  			list_splice_init(&fence->seq_passed_actions,
>  					 &action_list);
> @@ -685,7 +685,7 @@ void vmw_fence_fifo_down(struct vmw_fence_manager *fman)
>  		}
>  
>  		BUG_ON(!list_empty(&fence->head));
> -		fence_put(&fence->base);
> +		dma_fence_put(&fence->base);
>  		spin_lock_irq(&fman->lock);
>  	}
>  	spin_unlock_irq(&fman->lock);
> @@ -884,7 +884,7 @@ static void vmw_fence_obj_add_action(struct vmw_fence_obj *fence,
>  	spin_lock_irqsave(&fman->lock, irq_flags);
>  
>  	fman->pending_actions[action->type]++;
> -	if (fence_is_signaled_locked(&fence->base)) {
> +	if (dma_fence_is_signaled_locked(&fence->base)) {
>  		struct list_head action_list;
>  
>  		INIT_LIST_HEAD(&action_list);
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> index 83ae301ee141..d9d85aa6ed20 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> @@ -27,7 +27,7 @@
>  
>  #ifndef _VMWGFX_FENCE_H_
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #define VMW_FENCE_WAIT_TIMEOUT (5*HZ)
>  
> @@ -52,7 +52,7 @@ struct vmw_fence_action {
>  };
>  
>  struct vmw_fence_obj {
> -	struct fence base;
> +	struct dma_fence base;
>  
>  	struct list_head head;
>  	struct list_head seq_passed_actions;
> @@ -71,14 +71,14 @@ vmw_fence_obj_unreference(struct vmw_fence_obj **fence_p)
>  
>  	*fence_p = NULL;
>  	if (fence)
> -		fence_put(&fence->base);
> +		dma_fence_put(&fence->base);
>  }
>  
>  static inline struct vmw_fence_obj *
>  vmw_fence_obj_reference(struct vmw_fence_obj *fence)
>  {
>  	if (fence)
> -		fence_get(&fence->base);
> +		dma_fence_get(&fence->base);
>  	return fence;
>  }
>  
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c b/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
> index 1a85fb2d4dc6..8e86d6d4141b 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
> @@ -1454,7 +1454,7 @@ void vmw_fence_single_bo(struct ttm_buffer_object *bo,
>  	if (fence == NULL) {
>  		vmw_execbuf_fence_commands(NULL, dev_priv, &fence, NULL);
>  		reservation_object_add_excl_fence(bo->resv, &fence->base);
> -		fence_put(&fence->base);
> +		dma_fence_put(&fence->base);
>  	} else
>  		reservation_object_add_excl_fence(bo->resv, &fence->base);
>  }
> diff --git a/include/drm/drmP.h b/include/drm/drmP.h
> index 672644031bd5..e336e3901876 100644
> --- a/include/drm/drmP.h
> +++ b/include/drm/drmP.h
> @@ -57,7 +57,7 @@
>  #include <linux/types.h>
>  #include <linux/vmalloc.h>
>  #include <linux/workqueue.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  
>  #include <asm/mman.h>
>  #include <asm/pgalloc.h>
> @@ -362,7 +362,7 @@ struct drm_ioctl_desc {
>  struct drm_pending_event {
>  	struct completion *completion;
>  	struct drm_event *event;
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	struct list_head link;
>  	struct list_head pending_link;
>  	struct drm_file *file_priv;
> diff --git a/include/drm/drm_crtc.h b/include/drm/drm_crtc.h
> index bc860cfc67ca..fa1aa214c8ea 100644
> --- a/include/drm/drm_crtc.h
> +++ b/include/drm/drm_crtc.h
> @@ -54,7 +54,7 @@ struct drm_mode_set;
>  struct drm_file;
>  struct drm_clip_rect;
>  struct device_node;
> -struct fence;
> +struct dma_fence;
>  struct edid;
>  
>  static inline int64_t U642I64(uint64_t val)
> diff --git a/include/drm/drm_plane.h b/include/drm/drm_plane.h
> index 98b39d66eb32..c5e8a0df1623 100644
> --- a/include/drm/drm_plane.h
> +++ b/include/drm/drm_plane.h
> @@ -59,7 +59,7 @@ struct drm_plane_state {
>  
>  	struct drm_crtc *crtc;   /* do not write directly, use drm_atomic_set_crtc_for_plane() */
>  	struct drm_framebuffer *fb;  /* do not write directly, use drm_atomic_set_fb_for_plane() */
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  
>  	/* Signed dest location allows it to be partially off screen */
>  	int32_t crtc_x, crtc_y;
> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h
> index 9eb940d6755f..5beae7969bf7 100644
> --- a/include/drm/ttm/ttm_bo_api.h
> +++ b/include/drm/ttm/ttm_bo_api.h
> @@ -209,7 +209,7 @@ struct ttm_buffer_object {
>  	 * Members protected by a bo reservation.
>  	 */
>  
> -	struct fence *moving;
> +	struct dma_fence *moving;
>  
>  	struct drm_vma_offset_node vma_node;
>  
> diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
> index 4f0a92185995..27e9c26c9150 100644
> --- a/include/drm/ttm/ttm_bo_driver.h
> +++ b/include/drm/ttm/ttm_bo_driver.h
> @@ -303,7 +303,7 @@ struct ttm_mem_type_manager {
>  	/*
>  	 * Protected by @move_lock.
>  	 */
> -	struct fence *move;
> +	struct dma_fence *move;
>  };
>  
>  /**
> @@ -1025,7 +1025,7 @@ extern void ttm_bo_free_old_node(struct ttm_buffer_object *bo);
>   */
>  
>  extern int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
> -				     struct fence *fence, bool evict,
> +				     struct dma_fence *fence, bool evict,
>  				     struct ttm_mem_reg *new_mem);
>  
>  /**
> @@ -1040,7 +1040,7 @@ extern int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
>   * immediately or hang it on a temporary buffer object.
>   */
>  int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
> -			 struct fence *fence, bool evict,
> +			 struct dma_fence *fence, bool evict,
>  			 struct ttm_mem_reg *new_mem);
>  
>  /**
> diff --git a/include/drm/ttm/ttm_execbuf_util.h b/include/drm/ttm/ttm_execbuf_util.h
> index b620c317c772..47f35b8e6d09 100644
> --- a/include/drm/ttm/ttm_execbuf_util.h
> +++ b/include/drm/ttm/ttm_execbuf_util.h
> @@ -114,6 +114,6 @@ extern int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket,
>  
>  extern void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket,
>  					struct list_head *list,
> -					struct fence *fence);
> +					struct dma_fence *fence);
>  
>  #endif
> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> index e0b0741ae671..8daeb3ce0016 100644
> --- a/include/linux/dma-buf.h
> +++ b/include/linux/dma-buf.h
> @@ -30,7 +30,7 @@
>  #include <linux/list.h>
>  #include <linux/dma-mapping.h>
>  #include <linux/fs.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/wait.h>
>  
>  struct device;
> @@ -143,7 +143,7 @@ struct dma_buf {
>  	wait_queue_head_t poll;
>  
>  	struct dma_buf_poll_cb_t {
> -		struct fence_cb cb;
> +		struct dma_fence_cb cb;
>  		wait_queue_head_t *poll;
>  
>  		unsigned long active;
> diff --git a/include/linux/dma-fence-array.h b/include/linux/dma-fence-array.h
> new file mode 100644
> index 000000000000..5900945f962d
> --- /dev/null
> +++ b/include/linux/dma-fence-array.h
> @@ -0,0 +1,86 @@
> +/*
> + * fence-array: aggregates fence to be waited together
> + *
> + * Copyright (C) 2016 Collabora Ltd
> + * Copyright (C) 2016 Advanced Micro Devices, Inc.
> + * Authors:
> + *	Gustavo Padovan <gustavo@xxxxxxxxxxx>
> + *	Christian König <christian.koenig@xxxxxxx>
> + *
> + * This program is free software; you can redistribute it and/or modify it
> + * under the terms of the GNU General Public License version 2 as published by
> + * the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful, but WITHOUT
> + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
> + * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
> + * more details.
> + */
> +
> +#ifndef __LINUX_DMA_FENCE_ARRAY_H
> +#define __LINUX_DMA_FENCE_ARRAY_H
> +
> +#include <linux/dma-fence.h>
> +
> +/**
> + * struct dma_fence_array_cb - callback helper for fence array
> + * @cb: fence callback structure for signaling
> + * @array: reference to the parent fence array object
> + */
> +struct dma_fence_array_cb {
> +	struct dma_fence_cb cb;
> +	struct dma_fence_array *array;
> +};
> +
> +/**
> + * struct dma_fence_array - fence to represent an array of fences
> + * @base: fence base class
> + * @lock: spinlock for fence handling
> + * @num_fences: number of fences in the array
> + * @num_pending: fences in the array still pending
> + * @fences: array of the fences
> + */
> +struct dma_fence_array {
> +	struct dma_fence base;
> +
> +	spinlock_t lock;
> +	unsigned num_fences;
> +	atomic_t num_pending;
> +	struct dma_fence **fences;
> +};
> +
> +extern const struct dma_fence_ops dma_fence_array_ops;
> +
> +/**
> + * dma_fence_is_array - check if a fence is from the array subsclass
> + * @fence: fence to test
> + *
> + * Return true if it is a dma_fence_array and false otherwise.
> + */
> +static inline bool dma_fence_is_array(struct dma_fence *fence)
> +{
> +	return fence->ops == &dma_fence_array_ops;
> +}
> +
> +/**
> + * to_dma_fence_array - cast a fence to a dma_fence_array
> + * @fence: fence to cast to a dma_fence_array
> + *
> + * Returns NULL if the fence is not a dma_fence_array,
> + * or the dma_fence_array otherwise.
> + */
> +static inline struct dma_fence_array *
> +to_dma_fence_array(struct dma_fence *fence)
> +{
> +	if (fence->ops != &dma_fence_array_ops)
> +		return NULL;
> +
> +	return container_of(fence, struct dma_fence_array, base);
> +}
> +
> +struct dma_fence_array *dma_fence_array_create(int num_fences,
> +					       struct dma_fence **fences,
> +					       u64 context, unsigned seqno,
> +					       bool signal_on_any);
> +
> +#endif /* __LINUX_DMA_FENCE_ARRAY_H */
> diff --git a/include/linux/fence.h b/include/linux/dma-fence.h
> similarity index 59%
> rename from include/linux/fence.h
> rename to include/linux/dma-fence.h
> index c9c5ba98c302..ba60c043a5d3 100644
> --- a/include/linux/fence.h
> +++ b/include/linux/dma-fence.h
> @@ -18,8 +18,8 @@
>   * more details.
>   */
>  
> -#ifndef __LINUX_FENCE_H
> -#define __LINUX_FENCE_H
> +#ifndef __LINUX_DMA_FENCE_H
> +#define __LINUX_DMA_FENCE_H
>  
>  #include <linux/err.h>
>  #include <linux/wait.h>
> @@ -30,48 +30,48 @@
>  #include <linux/printk.h>
>  #include <linux/rcupdate.h>
>  
> -struct fence;
> -struct fence_ops;
> -struct fence_cb;
> +struct dma_fence;
> +struct dma_fence_ops;
> +struct dma_fence_cb;
>  
>  /**
> - * struct fence - software synchronization primitive
> + * struct dma_fence - software synchronization primitive
>   * @refcount: refcount for this fence
> - * @ops: fence_ops associated with this fence
> + * @ops: dma_fence_ops associated with this fence
>   * @rcu: used for releasing fence with kfree_rcu
>   * @cb_list: list of all callbacks to call
>   * @lock: spin_lock_irqsave used for locking
>   * @context: execution context this fence belongs to, returned by
> - *           fence_context_alloc()
> + *           dma_fence_context_alloc()
>   * @seqno: the sequence number of this fence inside the execution context,
>   * can be compared to decide which fence would be signaled later.
> - * @flags: A mask of FENCE_FLAG_* defined below
> + * @flags: A mask of DMA_FENCE_FLAG_* defined below
>   * @timestamp: Timestamp when the fence was signaled.
>   * @status: Optional, only valid if < 0, must be set before calling
> - * fence_signal, indicates that the fence has completed with an error.
> + * dma_fence_signal, indicates that the fence has completed with an error.
>   *
>   * the flags member must be manipulated and read using the appropriate
>   * atomic ops (bit_*), so taking the spinlock will not be needed most
>   * of the time.
>   *
> - * FENCE_FLAG_SIGNALED_BIT - fence is already signaled
> - * FENCE_FLAG_ENABLE_SIGNAL_BIT - enable_signaling might have been called*
> - * FENCE_FLAG_USER_BITS - start of the unused bits, can be used by the
> + * DMA_FENCE_FLAG_SIGNALED_BIT - fence is already signaled
> + * DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT - enable_signaling might have been called
> + * DMA_FENCE_FLAG_USER_BITS - start of the unused bits, can be used by the
>   * implementer of the fence for its own purposes. Can be used in different
>   * ways by different fence implementers, so do not rely on this.
>   *
>   * Since atomic bitops are used, this is not guaranteed to be the case.
> - * Particularly, if the bit was set, but fence_signal was called right
> + * Particularly, if the bit was set, but dma_fence_signal was called right
>   * before this bit was set, it would have been able to set the
> - * FENCE_FLAG_SIGNALED_BIT, before enable_signaling was called.
> - * Adding a check for FENCE_FLAG_SIGNALED_BIT after setting
> - * FENCE_FLAG_ENABLE_SIGNAL_BIT closes this race, and makes sure that
> - * after fence_signal was called, any enable_signaling call will have either
> + * DMA_FENCE_FLAG_SIGNALED_BIT, before enable_signaling was called.
> + * Adding a check for DMA_FENCE_FLAG_SIGNALED_BIT after setting
> + * DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT closes this race, and makes sure that
> + * after dma_fence_signal was called, any enable_signaling call will have either
>   * been completed, or never called at all.
>   */
> -struct fence {
> +struct dma_fence {
>  	struct kref refcount;
> -	const struct fence_ops *ops;
> +	const struct dma_fence_ops *ops;
>  	struct rcu_head rcu;
>  	struct list_head cb_list;
>  	spinlock_t *lock;
> @@ -82,34 +82,35 @@ struct fence {
>  	int status;
>  };
>  
> -enum fence_flag_bits {
> -	FENCE_FLAG_SIGNALED_BIT,
> -	FENCE_FLAG_ENABLE_SIGNAL_BIT,
> -	FENCE_FLAG_USER_BITS, /* must always be last member */
> +enum dma_fence_flag_bits {
> +	DMA_FENCE_FLAG_SIGNALED_BIT,
> +	DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
> +	DMA_FENCE_FLAG_USER_BITS, /* must always be last member */
>  };
>  
> -typedef void (*fence_func_t)(struct fence *fence, struct fence_cb *cb);
> +typedef void (*dma_fence_func_t)(struct dma_fence *fence,
> +				 struct dma_fence_cb *cb);
>  
>  /**
> - * struct fence_cb - callback for fence_add_callback
> - * @node: used by fence_add_callback to append this struct to fence::cb_list
> - * @func: fence_func_t to call
> + * struct dma_fence_cb - callback for dma_fence_add_callback
> + * @node: used by dma_fence_add_callback to append this struct to fence::cb_list
> + * @func: dma_fence_func_t to call
>   *
> - * This struct will be initialized by fence_add_callback, additional
> - * data can be passed along by embedding fence_cb in another struct.
> + * This struct will be initialized by dma_fence_add_callback, additional
> + * data can be passed along by embedding dma_fence_cb in another struct.
>   */
> -struct fence_cb {
> +struct dma_fence_cb {
>  	struct list_head node;
> -	fence_func_t func;
> +	dma_fence_func_t func;
>  };
>  
>  /**
> - * struct fence_ops - operations implemented for fence
> + * struct dma_fence_ops - operations implemented for fence
>   * @get_driver_name: returns the driver name.
>   * @get_timeline_name: return the name of the context this fence belongs to.
>   * @enable_signaling: enable software signaling of fence.
>   * @signaled: [optional] peek whether the fence is signaled, can be null.
> - * @wait: custom wait implementation, or fence_default_wait.
> + * @wait: custom wait implementation, or dma_fence_default_wait.
>   * @release: [optional] called on destruction of fence, can be null
>   * @fill_driver_data: [optional] callback to fill in free-form debug info
>   * Returns amount of bytes filled, or -errno.
> @@ -135,20 +136,20 @@ struct fence_cb {
>   * fence->status may be set in enable_signaling, but only when false is
>   * returned.
>   *
> - * Calling fence_signal before enable_signaling is called allows
> + * Calling dma_fence_signal before enable_signaling is called allows
>   * for a tiny race window in which enable_signaling is called during,
> - * before, or after fence_signal. To fight this, it is recommended
> + * before, or after dma_fence_signal. To fight this, it is recommended
>   * that before enable_signaling returns true an extra reference is
>   * taken on the fence, to be released when the fence is signaled.
> - * This will mean fence_signal will still be called twice, but
> + * This will mean dma_fence_signal will still be called twice, but
>   * the second time will be a noop since it was already signaled.
>   *
>   * Notes on signaled:
>   * May set fence->status if returning true.
>   *
>   * Notes on wait:
> - * Must not be NULL, set to fence_default_wait for default implementation.
> - * the fence_default_wait implementation should work for any fence, as long
> + * Must not be NULL, set to dma_fence_default_wait for default implementation.
> + * the dma_fence_default_wait implementation should work for any fence, as long
>   * as enable_signaling works correctly.
>   *
>   * Must return -ERESTARTSYS if the wait is intr = true and the wait was
> @@ -163,42 +164,44 @@ struct fence_cb {
>   * If pointer is set to NULL, kfree will get called instead.
>   */
>  
> -struct fence_ops {
> -	const char * (*get_driver_name)(struct fence *fence);
> -	const char * (*get_timeline_name)(struct fence *fence);
> -	bool (*enable_signaling)(struct fence *fence);
> -	bool (*signaled)(struct fence *fence);
> -	signed long (*wait)(struct fence *fence, bool intr, signed long timeout);
> -	void (*release)(struct fence *fence);
> -
> -	int (*fill_driver_data)(struct fence *fence, void *data, int size);
> -	void (*fence_value_str)(struct fence *fence, char *str, int size);
> -	void (*timeline_value_str)(struct fence *fence, char *str, int size);
> +struct dma_fence_ops {
> +	const char * (*get_driver_name)(struct dma_fence *fence);
> +	const char * (*get_timeline_name)(struct dma_fence *fence);
> +	bool (*enable_signaling)(struct dma_fence *fence);
> +	bool (*signaled)(struct dma_fence *fence);
> +	signed long (*wait)(struct dma_fence *fence,
> +			    bool intr, signed long timeout);
> +	void (*release)(struct dma_fence *fence);
> +
> +	int (*fill_driver_data)(struct dma_fence *fence, void *data, int size);
> +	void (*fence_value_str)(struct dma_fence *fence, char *str, int size);
> +	void (*timeline_value_str)(struct dma_fence *fence,
> +				   char *str, int size);
>  };
>  
> -void fence_init(struct fence *fence, const struct fence_ops *ops,
> -		spinlock_t *lock, u64 context, unsigned seqno);
> +void dma_fence_init(struct dma_fence *fence, const struct dma_fence_ops *ops,
> +		    spinlock_t *lock, u64 context, unsigned seqno);
>  
> -void fence_release(struct kref *kref);
> -void fence_free(struct fence *fence);
> +void dma_fence_release(struct kref *kref);
> +void dma_fence_free(struct dma_fence *fence);
>  
>  /**
> - * fence_put - decreases refcount of the fence
> + * dma_fence_put - decreases refcount of the fence
>   * @fence:	[in]	fence to reduce refcount of
>   */
> -static inline void fence_put(struct fence *fence)
> +static inline void dma_fence_put(struct dma_fence *fence)
>  {
>  	if (fence)
> -		kref_put(&fence->refcount, fence_release);
> +		kref_put(&fence->refcount, dma_fence_release);
>  }
>  
>  /**
> - * fence_get - increases refcount of the fence
> + * dma_fence_get - increases refcount of the fence
>   * @fence:	[in]	fence to increase refcount of
>   *
>   * Returns the same fence, with refcount increased by 1.
>   */
> -static inline struct fence *fence_get(struct fence *fence)
> +static inline struct dma_fence *dma_fence_get(struct dma_fence *fence)
>  {
>  	if (fence)
>  		kref_get(&fence->refcount);
> @@ -206,12 +209,13 @@ static inline struct fence *fence_get(struct fence *fence)
>  }
>  
>  /**
> - * fence_get_rcu - get a fence from a reservation_object_list with rcu read lock
> + * dma_fence_get_rcu - get a fence from a reservation_object_list with
> + *                     rcu read lock
>   * @fence:	[in]	fence to increase refcount of
>   *
>   * Function returns NULL if no refcount could be obtained, or the fence.
>   */
> -static inline struct fence *fence_get_rcu(struct fence *fence)
> +static inline struct dma_fence *dma_fence_get_rcu(struct dma_fence *fence)
>  {
>  	if (kref_get_unless_zero(&fence->refcount))
>  		return fence;
> @@ -220,7 +224,7 @@ static inline struct fence *fence_get_rcu(struct fence *fence)
>  }
>  
>  /**
> - * fence_get_rcu_safe  - acquire a reference to an RCU tracked fence
> + * dma_fence_get_rcu_safe  - acquire a reference to an RCU tracked fence
>   * @fence:	[in]	pointer to fence to increase refcount of
>   *
>   * Function returns NULL if no refcount could be obtained, or the fence.
> @@ -235,16 +239,17 @@ static inline struct fence *fence_get_rcu(struct fence *fence)
>   *
>   * The caller is required to hold the RCU read lock.
>   */
> -static inline struct fence *fence_get_rcu_safe(struct fence * __rcu *fencep)
> +static inline struct dma_fence *
> +dma_fence_get_rcu_safe(struct dma_fence * __rcu *fencep)
>  {
>  	do {
> -		struct fence *fence;
> +		struct dma_fence *fence;
>  
>  		fence = rcu_dereference(*fencep);
> -		if (!fence || !fence_get_rcu(fence))
> +		if (!fence || !dma_fence_get_rcu(fence))
>  			return NULL;
>  
> -		/* The atomic_inc_not_zero() inside fence_get_rcu()
> +		/* The atomic_inc_not_zero() inside dma_fence_get_rcu()
>  		 * provides a full memory barrier upon success (such as now).
>  		 * This is paired with the write barrier from assigning
>  		 * to the __rcu protected fence pointer so that if that
> @@ -261,37 +266,41 @@ static inline struct fence *fence_get_rcu_safe(struct fence * __rcu *fencep)
>  		if (fence == rcu_access_pointer(*fencep))
>  			return rcu_pointer_handoff(fence);
>  
> -		fence_put(fence);
> +		dma_fence_put(fence);
>  	} while (1);
>  }
>  
> -int fence_signal(struct fence *fence);
> -int fence_signal_locked(struct fence *fence);
> -signed long fence_default_wait(struct fence *fence, bool intr, signed long timeout);
> -int fence_add_callback(struct fence *fence, struct fence_cb *cb,
> -		       fence_func_t func);
> -bool fence_remove_callback(struct fence *fence, struct fence_cb *cb);
> -void fence_enable_sw_signaling(struct fence *fence);
> +int dma_fence_signal(struct dma_fence *fence);
> +int dma_fence_signal_locked(struct dma_fence *fence);
> +signed long dma_fence_default_wait(struct dma_fence *fence,
> +				   bool intr, signed long timeout);
> +int dma_fence_add_callback(struct dma_fence *fence,
> +			   struct dma_fence_cb *cb,
> +			   dma_fence_func_t func);
> +bool dma_fence_remove_callback(struct dma_fence *fence,
> +			       struct dma_fence_cb *cb);
> +void dma_fence_enable_sw_signaling(struct dma_fence *fence);
>  
>  /**
> - * fence_is_signaled_locked - Return an indication if the fence is signaled yet.
> + * dma_fence_is_signaled_locked - Return an indication if the fence
> + *                                is signaled yet.
>   * @fence:	[in]	the fence to check
>   *
>   * Returns true if the fence was already signaled, false if not. Since this
>   * function doesn't enable signaling, it is not guaranteed to ever return
> - * true if fence_add_callback, fence_wait or fence_enable_sw_signaling
> - * haven't been called before.
> + * true if dma_fence_add_callback, dma_fence_wait or
> + * dma_fence_enable_sw_signaling haven't been called before.
>   *
>   * This function requires fence->lock to be held.
>   */
>  static inline bool
> -fence_is_signaled_locked(struct fence *fence)
> +dma_fence_is_signaled_locked(struct dma_fence *fence)
>  {
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		return true;
>  
>  	if (fence->ops->signaled && fence->ops->signaled(fence)) {
> -		fence_signal_locked(fence);
> +		dma_fence_signal_locked(fence);
>  		return true;
>  	}
>  
> @@ -299,27 +308,27 @@ fence_is_signaled_locked(struct fence *fence)
>  }
>  
>  /**
> - * fence_is_signaled - Return an indication if the fence is signaled yet.
> + * dma_fence_is_signaled - Return an indication if the fence is signaled yet.
>   * @fence:	[in]	the fence to check
>   *
>   * Returns true if the fence was already signaled, false if not. Since this
>   * function doesn't enable signaling, it is not guaranteed to ever return
> - * true if fence_add_callback, fence_wait or fence_enable_sw_signaling
> - * haven't been called before.
> + * true if dma_fence_add_callback, dma_fence_wait or
> + * dma_fence_enable_sw_signaling haven't been called before.
>   *
> - * It's recommended for seqno fences to call fence_signal when the
> + * It's recommended for seqno fences to call dma_fence_signal when the
>   * operation is complete, it makes it possible to prevent issues from
>   * wraparound between time of issue and time of use by checking the return
>   * value of this function before calling hardware-specific wait instructions.
>   */
>  static inline bool
> -fence_is_signaled(struct fence *fence)
> +dma_fence_is_signaled(struct dma_fence *fence)
>  {
> -	if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> +	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
>  		return true;
>  
>  	if (fence->ops->signaled && fence->ops->signaled(fence)) {
> -		fence_signal(fence);
> +		dma_fence_signal(fence);
>  		return true;
>  	}
>  
> @@ -327,14 +336,15 @@ fence_is_signaled(struct fence *fence)
>  }
>  
>  /**
> - * fence_is_later - return if f1 is chronologically later than f2
> + * dma_fence_is_later - return if f1 is chronologically later than f2
>   * @f1:	[in]	the first fence from the same context
>   * @f2:	[in]	the second fence from the same context
>   *
>   * Returns true if f1 is chronologically later than f2. Both fences must be
>   * from the same context, since a seqno is not re-used across contexts.
>   */
> -static inline bool fence_is_later(struct fence *f1, struct fence *f2)
> +static inline bool dma_fence_is_later(struct dma_fence *f1,
> +				      struct dma_fence *f2)
>  {
>  	if (WARN_ON(f1->context != f2->context))
>  		return false;
> @@ -343,7 +353,7 @@ static inline bool fence_is_later(struct fence *f1, struct fence *f2)
>  }
>  
>  /**
> - * fence_later - return the chronologically later fence
> + * dma_fence_later - return the chronologically later fence
>   * @f1:	[in]	the first fence from the same context
>   * @f2:	[in]	the second fence from the same context
>   *
> @@ -351,28 +361,31 @@ static inline bool fence_is_later(struct fence *f1, struct fence *f2)
>   * signaled last. Both fences must be from the same context, since a seqno is
>   * not re-used across contexts.
>   */
> -static inline struct fence *fence_later(struct fence *f1, struct fence *f2)
> +static inline struct dma_fence *dma_fence_later(struct dma_fence *f1,
> +						struct dma_fence *f2)
>  {
>  	if (WARN_ON(f1->context != f2->context))
>  		return NULL;
>  
>  	/*
> -	 * can't check just FENCE_FLAG_SIGNALED_BIT here, it may never have been
> -	 * set if enable_signaling wasn't called, and enabling that here is
> -	 * overkill.
> +	 * Can't check just DMA_FENCE_FLAG_SIGNALED_BIT here, it may never
> +	 * have been set if enable_signaling wasn't called, and enabling that
> +	 * here is overkill.
>  	 */
> -	if (fence_is_later(f1, f2))
> -		return fence_is_signaled(f1) ? NULL : f1;
> +	if (dma_fence_is_later(f1, f2))
> +		return dma_fence_is_signaled(f1) ? NULL : f1;
>  	else
> -		return fence_is_signaled(f2) ? NULL : f2;
> +		return dma_fence_is_signaled(f2) ? NULL : f2;
>  }
>  
> -signed long fence_wait_timeout(struct fence *, bool intr, signed long timeout);
> -signed long fence_wait_any_timeout(struct fence **fences, uint32_t count,
> +signed long dma_fence_wait_timeout(struct dma_fence *,
>  				   bool intr, signed long timeout);
> +signed long dma_fence_wait_any_timeout(struct dma_fence **fences,
> +				       uint32_t count,
> +				       bool intr, signed long timeout);
>  
>  /**
> - * fence_wait - sleep until the fence gets signaled
> + * dma_fence_wait - sleep until the fence gets signaled
>   * @fence:	[in]	the fence to wait on
>   * @intr:	[in]	if true, do an interruptible wait
>   *
> @@ -384,41 +397,41 @@ signed long fence_wait_any_timeout(struct fence **fences, uint32_t count,
>   * directly or indirectly holds a reference to the fence, otherwise the
>   * fence might be freed before return, resulting in undefined behavior.
>   */
> -static inline signed long fence_wait(struct fence *fence, bool intr)
> +static inline signed long dma_fence_wait(struct dma_fence *fence, bool intr)
>  {
>  	signed long ret;
>  
> -	/* Since fence_wait_timeout cannot timeout with
> +	/* Since dma_fence_wait_timeout cannot timeout with
>  	 * MAX_SCHEDULE_TIMEOUT, only valid return values are
>  	 * -ERESTARTSYS and MAX_SCHEDULE_TIMEOUT.
>  	 */
> -	ret = fence_wait_timeout(fence, intr, MAX_SCHEDULE_TIMEOUT);
> +	ret = dma_fence_wait_timeout(fence, intr, MAX_SCHEDULE_TIMEOUT);
>  
>  	return ret < 0 ? ret : 0;
>  }
>  
> -u64 fence_context_alloc(unsigned num);
> +u64 dma_fence_context_alloc(unsigned num);
>  
> -#define FENCE_TRACE(f, fmt, args...) \
> +#define DMA_FENCE_TRACE(f, fmt, args...) \
>  	do {								\
> -		struct fence *__ff = (f);				\
> -		if (IS_ENABLED(CONFIG_FENCE_TRACE))			\
> +		struct dma_fence *__ff = (f);				\
> +		if (IS_ENABLED(CONFIG_DMA_FENCE_TRACE))			\
>  			pr_info("f %llu#%u: " fmt,			\
>  				__ff->context, __ff->seqno, ##args);	\
>  	} while (0)
>  
> -#define FENCE_WARN(f, fmt, args...) \
> +#define DMA_FENCE_WARN(f, fmt, args...) \
>  	do {								\
> -		struct fence *__ff = (f);				\
> +		struct dma_fence *__ff = (f);				\
>  		pr_warn("f %llu#%u: " fmt, __ff->context, __ff->seqno,	\
>  			 ##args);					\
>  	} while (0)
>  
> -#define FENCE_ERR(f, fmt, args...) \
> +#define DMA_FENCE_ERR(f, fmt, args...) \
>  	do {								\
> -		struct fence *__ff = (f);				\
> +		struct dma_fence *__ff = (f);				\
>  		pr_err("f %llu#%u: " fmt, __ff->context, __ff->seqno,	\
>  			##args);					\
>  	} while (0)
>  
> -#endif /* __LINUX_FENCE_H */
> +#endif /* __LINUX_DMA_FENCE_H */
> diff --git a/include/linux/fence-array.h b/include/linux/fence-array.h
> deleted file mode 100644
> index 9ea2bde10ac1..000000000000
> --- a/include/linux/fence-array.h
> +++ /dev/null
> @@ -1,84 +0,0 @@
> -/*
> - * fence-array: aggregates fence to be waited together
> - *
> - * Copyright (C) 2016 Collabora Ltd
> - * Copyright (C) 2016 Advanced Micro Devices, Inc.
> - * Authors:
> - *	Gustavo Padovan <gustavo@xxxxxxxxxxx>
> - *	Christian König <christian.koenig@xxxxxxx>
> - *
> - * This program is free software; you can redistribute it and/or modify it
> - * under the terms of the GNU General Public License version 2 as published by
> - * the Free Software Foundation.
> - *
> - * This program is distributed in the hope that it will be useful, but WITHOUT
> - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
> - * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
> - * more details.
> - */
> -
> -#ifndef __LINUX_FENCE_ARRAY_H
> -#define __LINUX_FENCE_ARRAY_H
> -
> -#include <linux/fence.h>
> -
> -/**
> - * struct fence_array_cb - callback helper for fence array
> - * @cb: fence callback structure for signaling
> - * @array: reference to the parent fence array object
> - */
> -struct fence_array_cb {
> -	struct fence_cb cb;
> -	struct fence_array *array;
> -};
> -
> -/**
> - * struct fence_array - fence to represent an array of fences
> - * @base: fence base class
> - * @lock: spinlock for fence handling
> - * @num_fences: number of fences in the array
> - * @num_pending: fences in the array still pending
> - * @fences: array of the fences
> - */
> -struct fence_array {
> -	struct fence base;
> -
> -	spinlock_t lock;
> -	unsigned num_fences;
> -	atomic_t num_pending;
> -	struct fence **fences;
> -};
> -
> -extern const struct fence_ops fence_array_ops;
> -
> -/**
> - * fence_is_array - check if a fence is from the array subsclass
> - * @fence: fence to test
> - *
> - * Return true if it is a fence_array and false otherwise.
> - */
> -static inline bool fence_is_array(struct fence *fence)
> -{
> -	return fence->ops == &fence_array_ops;
> -}
> -
> -/**
> - * to_fence_array - cast a fence to a fence_array
> - * @fence: fence to cast to a fence_array
> - *
> - * Returns NULL if the fence is not a fence_array,
> - * or the fence_array otherwise.
> - */
> -static inline struct fence_array *to_fence_array(struct fence *fence)
> -{
> -	if (fence->ops != &fence_array_ops)
> -		return NULL;
> -
> -	return container_of(fence, struct fence_array, base);
> -}
> -
> -struct fence_array *fence_array_create(int num_fences, struct fence **fences,
> -				       u64 context, unsigned seqno,
> -				       bool signal_on_any);
> -
> -#endif /* __LINUX_FENCE_ARRAY_H */
> diff --git a/include/linux/reservation.h b/include/linux/reservation.h
> index b0f305e77b7f..2e313cca08f0 100644
> --- a/include/linux/reservation.h
> +++ b/include/linux/reservation.h
> @@ -40,7 +40,7 @@
>  #define _LINUX_RESERVATION_H
>  
>  #include <linux/ww_mutex.h>
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/slab.h>
>  #include <linux/seqlock.h>
>  #include <linux/rcupdate.h>
> @@ -59,7 +59,7 @@ extern const char reservation_seqcount_string[];
>  struct reservation_object_list {
>  	struct rcu_head rcu;
>  	u32 shared_count, shared_max;
> -	struct fence __rcu *shared[];
> +	struct dma_fence __rcu *shared[];
>  };
>  
>  /**
> @@ -74,7 +74,7 @@ struct reservation_object {
>  	struct ww_mutex lock;
>  	seqcount_t seq;
>  
> -	struct fence __rcu *fence_excl;
> +	struct dma_fence __rcu *fence_excl;
>  	struct reservation_object_list __rcu *fence;
>  	struct reservation_object_list *staged;
>  };
> @@ -107,7 +107,7 @@ reservation_object_fini(struct reservation_object *obj)
>  {
>  	int i;
>  	struct reservation_object_list *fobj;
> -	struct fence *excl;
> +	struct dma_fence *excl;
>  
>  	/*
>  	 * This object should be dead and all references must have
> @@ -115,12 +115,12 @@ reservation_object_fini(struct reservation_object *obj)
>  	 */
>  	excl = rcu_dereference_protected(obj->fence_excl, 1);
>  	if (excl)
> -		fence_put(excl);
> +		dma_fence_put(excl);
>  
>  	fobj = rcu_dereference_protected(obj->fence, 1);
>  	if (fobj) {
>  		for (i = 0; i < fobj->shared_count; ++i)
> -			fence_put(rcu_dereference_protected(fobj->shared[i], 1));
> +			dma_fence_put(rcu_dereference_protected(fobj->shared[i], 1));
>  
>  		kfree(fobj);
>  	}
> @@ -155,7 +155,7 @@ reservation_object_get_list(struct reservation_object *obj)
>   * RETURNS
>   * The exclusive fence or NULL
>   */
> -static inline struct fence *
> +static inline struct dma_fence *
>  reservation_object_get_excl(struct reservation_object *obj)
>  {
>  	return rcu_dereference_protected(obj->fence_excl,
> @@ -173,10 +173,10 @@ reservation_object_get_excl(struct reservation_object *obj)
>   * RETURNS
>   * The exclusive fence or NULL if none
>   */
> -static inline struct fence *
> +static inline struct dma_fence *
>  reservation_object_get_excl_rcu(struct reservation_object *obj)
>  {
> -	struct fence *fence;
> +	struct dma_fence *fence;
>  	unsigned seq;
>  retry:
>  	seq = read_seqcount_begin(&obj->seq);
> @@ -186,22 +186,22 @@ reservation_object_get_excl_rcu(struct reservation_object *obj)
>  		rcu_read_unlock();
>  		goto retry;
>  	}
> -	fence = fence_get(fence);
> +	fence = dma_fence_get(fence);
>  	rcu_read_unlock();
>  	return fence;
>  }
>  
>  int reservation_object_reserve_shared(struct reservation_object *obj);
>  void reservation_object_add_shared_fence(struct reservation_object *obj,
> -					 struct fence *fence);
> +					 struct dma_fence *fence);
>  
>  void reservation_object_add_excl_fence(struct reservation_object *obj,
> -				       struct fence *fence);
> +				       struct dma_fence *fence);
>  
>  int reservation_object_get_fences_rcu(struct reservation_object *obj,
> -				      struct fence **pfence_excl,
> +				      struct dma_fence **pfence_excl,
>  				      unsigned *pshared_count,
> -				      struct fence ***pshared);
> +				      struct dma_fence ***pshared);
>  
>  long reservation_object_wait_timeout_rcu(struct reservation_object *obj,
>  					 bool wait_all, bool intr,
> diff --git a/include/linux/seqno-fence.h b/include/linux/seqno-fence.h
> index a1ba6a5ccdd6..c58c535d12a8 100644
> --- a/include/linux/seqno-fence.h
> +++ b/include/linux/seqno-fence.h
> @@ -20,7 +20,7 @@
>  #ifndef __LINUX_SEQNO_FENCE_H
>  #define __LINUX_SEQNO_FENCE_H
>  
> -#include <linux/fence.h>
> +#include <linux/dma-fence.h>
>  #include <linux/dma-buf.h>
>  
>  enum seqno_fence_condition {
> @@ -29,15 +29,15 @@ enum seqno_fence_condition {
>  };
>  
>  struct seqno_fence {
> -	struct fence base;
> +	struct dma_fence base;
>  
> -	const struct fence_ops *ops;
> +	const struct dma_fence_ops *ops;
>  	struct dma_buf *sync_buf;
>  	uint32_t seqno_ofs;
>  	enum seqno_fence_condition condition;
>  };
>  
> -extern const struct fence_ops seqno_fence_ops;
> +extern const struct dma_fence_ops seqno_fence_ops;
>  
>  /**
>   * to_seqno_fence - cast a fence to a seqno_fence
> @@ -47,7 +47,7 @@ extern const struct fence_ops seqno_fence_ops;
>   * or the seqno_fence otherwise.
>   */
>  static inline struct seqno_fence *
> -to_seqno_fence(struct fence *fence)
> +to_seqno_fence(struct dma_fence *fence)
>  {
>  	if (fence->ops != &seqno_fence_ops)
>  		return NULL;
> @@ -83,9 +83,9 @@ to_seqno_fence(struct fence *fence)
>   * dma-buf for sync_buf, since mapping or unmapping the sync_buf to the
>   * device's vm can be expensive.
>   *
> - * It is recommended for creators of seqno_fence to call fence_signal
> + * It is recommended for creators of seqno_fence to call dma_fence_signal()
>   * before destruction. This will prevent possible issues from wraparound at
> - * time of issue vs time of check, since users can check fence_is_signaled
> + * time of issue vs time of check, since users can check dma_fence_is_signaled()
>   * before submitting instructions for the hardware to wait on the fence.
>   * However, when ops.enable_signaling is not called, it doesn't have to be
>   * done as soon as possible, just before there's any real danger of seqno
> @@ -96,18 +96,18 @@ seqno_fence_init(struct seqno_fence *fence, spinlock_t *lock,
>  		 struct dma_buf *sync_buf,  uint32_t context,
>  		 uint32_t seqno_ofs, uint32_t seqno,
>  		 enum seqno_fence_condition cond,
> -		 const struct fence_ops *ops)
> +		 const struct dma_fence_ops *ops)
>  {
>  	BUG_ON(!fence || !sync_buf || !ops);
>  	BUG_ON(!ops->wait || !ops->enable_signaling ||
>  	       !ops->get_driver_name || !ops->get_timeline_name);
>  
>  	/*
> -	 * ops is used in fence_init for get_driver_name, so needs to be
> +	 * ops is used in dma_fence_init for get_driver_name, so needs to be
>  	 * initialized first
>  	 */
>  	fence->ops = ops;
> -	fence_init(&fence->base, &seqno_fence_ops, lock, context, seqno);
> +	dma_fence_init(&fence->base, &seqno_fence_ops, lock, context, seqno);
>  	get_dma_buf(sync_buf);
>  	fence->sync_buf = sync_buf;
>  	fence->seqno_ofs = seqno_ofs;
> diff --git a/include/linux/sync_file.h b/include/linux/sync_file.h
> index aa17ccfc2f57..3e3ab84fc4cd 100644
> --- a/include/linux/sync_file.h
> +++ b/include/linux/sync_file.h
> @@ -18,8 +18,8 @@
>  #include <linux/ktime.h>
>  #include <linux/list.h>
>  #include <linux/spinlock.h>
> -#include <linux/fence.h>
> -#include <linux/fence-array.h>
> +#include <linux/dma-fence.h>
> +#include <linux/dma-fence-array.h>
>  
>  /**
>   * struct sync_file - sync file to export to the userspace
> @@ -41,13 +41,13 @@ struct sync_file {
>  
>  	wait_queue_head_t	wq;
>  
> -	struct fence		*fence;
> -	struct fence_cb cb;
> +	struct dma_fence	*fence;
> +	struct dma_fence_cb cb;
>  };
>  
> -#define POLL_ENABLED FENCE_FLAG_USER_BITS
> +#define POLL_ENABLED DMA_FENCE_FLAG_USER_BITS
>  
> -struct sync_file *sync_file_create(struct fence *fence);
> -struct fence *sync_file_get_fence(int fd);
> +struct sync_file *sync_file_create(struct dma_fence *fence);
> +struct dma_fence *sync_file_get_fence(int fd);
>  
>  #endif /* _LINUX_SYNC_H */
> diff --git a/include/trace/events/fence.h b/include/trace/events/dma_fence.h
> similarity index 72%
> rename from include/trace/events/fence.h
> rename to include/trace/events/dma_fence.h
> index d6dfa05ba322..1157cb4c3c6f 100644
> --- a/include/trace/events/fence.h
> +++ b/include/trace/events/dma_fence.h
> @@ -1,17 +1,17 @@
>  #undef TRACE_SYSTEM
> -#define TRACE_SYSTEM fence
> +#define TRACE_SYSTEM dma_fence
>  
>  #if !defined(_TRACE_FENCE_H) || defined(TRACE_HEADER_MULTI_READ)
> -#define _TRACE_FENCE_H
> +#define _TRACE_DMA_FENCE_H
>  
>  #include <linux/tracepoint.h>
>  
> -struct fence;
> +struct dma_fence;
>  
> -TRACE_EVENT(fence_annotate_wait_on,
> +TRACE_EVENT(dma_fence_annotate_wait_on,
>  
>  	/* fence: the fence waiting on f1, f1: the fence to be waited on. */
> -	TP_PROTO(struct fence *fence, struct fence *f1),
> +	TP_PROTO(struct dma_fence *fence, struct dma_fence *f1),
>  
>  	TP_ARGS(fence, f1),
>  
> @@ -48,9 +48,9 @@ TRACE_EVENT(fence_annotate_wait_on,
>  		  __entry->waiting_context, __entry->waiting_seqno)
>  );
>  
> -DECLARE_EVENT_CLASS(fence,
> +DECLARE_EVENT_CLASS(dma_fence,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence),
>  
> @@ -73,56 +73,56 @@ DECLARE_EVENT_CLASS(fence,
>  		  __entry->seqno)
>  );
>  
> -DEFINE_EVENT(fence, fence_emit,
> +DEFINE_EVENT(dma_fence, dma_fence_emit,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_init,
> +DEFINE_EVENT(dma_fence, dma_fence_init,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_destroy,
> +DEFINE_EVENT(dma_fence, dma_fence_destroy,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_enable_signal,
> +DEFINE_EVENT(dma_fence, dma_fence_enable_signal,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_signaled,
> +DEFINE_EVENT(dma_fence, dma_fence_signaled,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_wait_start,
> +DEFINE_EVENT(dma_fence, dma_fence_wait_start,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -DEFINE_EVENT(fence, fence_wait_end,
> +DEFINE_EVENT(dma_fence, dma_fence_wait_end,
>  
> -	TP_PROTO(struct fence *fence),
> +	TP_PROTO(struct dma_fence *fence),
>  
>  	TP_ARGS(fence)
>  );
>  
> -#endif /*  _TRACE_FENCE_H */
> +#endif /*  _TRACE_DMA_FENCE_H */
>  
>  /* This part must be outside protection */
>  #include <trace/define_trace.h>
> -- 
> 2.10.1
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx

-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
_______________________________________________
Intel-gfx mailing list
Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/intel-gfx





[Index of Archives]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux