Re: [PATCH 20/22] drm/i915: Move i915_gem_object_wait_rendering()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On ke, 2016-07-27 at 12:14 +0100, Chris Wilson wrote:
> Just move it earlier so that we can use the companion nonblocking
> version in a couple of more callsites without having to add a forward
> declaration.
> 
> Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx>


Separate code motion patch, wow.

Reviewed-by: Joonas Lahtinen <joonas.lahtinen@xxxxxxxxxxxxxxx>

> ---
>  drivers/gpu/drm/i915/i915_gem.c | 202 ++++++++++++++++++++--------------------
>  1 file changed, 101 insertions(+), 101 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index ca9741525bf4..54d8a3863d11 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -301,6 +301,107 @@ i915_gem_object_unbind(struct drm_i915_gem_object *obj)
>  	return ret;
>  }
>  
> +/**
> + * Ensures that all rendering to the object has completed and the object is
> + * safe to unbind from the GTT or access from the CPU.
> + */
> +int
> +i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj,
> +			       bool readonly)
> +{
> +	struct reservation_object *resv;
> +	struct i915_gem_active *active;
> +	unsigned long active_mask;
> +	int idx;
> +
> +	lockdep_assert_held(&obj->base.dev->struct_mutex);
> +
> +	if (!readonly) {
> +		active = obj->last_read;
> +		active_mask = i915_gem_object_is_active(obj);
> +	} else {
> +		active_mask = 1;
> +		active = &obj->last_write;
> +	}
> +
> +	for_each_active(active_mask, idx) {
> +		int ret;
> +
> +		ret = i915_gem_active_wait(&active[idx],
> +					   &obj->base.dev->struct_mutex);
> +		if (ret)
> +			return ret;
> +	}
> +
> +	resv = i915_gem_object_get_dmabuf_resv(obj);
> +	if (resv) {
> +		long err;
> +
> +		err = reservation_object_wait_timeout_rcu(resv, !readonly, true,
> +							  MAX_SCHEDULE_TIMEOUT);
> +		if (err < 0)
> +			return err;
> +	}
> +
> +	return 0;
> +}
> +
> +/* A nonblocking variant of the above wait. This is a highly dangerous routine
> + * as the object state may change during this call.
> + */
> +static __must_check int
> +i915_gem_object_wait_rendering__nonblocking(struct drm_i915_gem_object *obj,
> +					    struct intel_rps_client *rps,
> +					    bool readonly)
> +{
> +	struct drm_device *dev = obj->base.dev;
> +	struct drm_i915_gem_request *requests[I915_NUM_ENGINES];
> +	struct i915_gem_active *active;
> +	unsigned long active_mask;
> +	int ret, i, n = 0;
> +
> +	lockdep_assert_held(&dev->struct_mutex);
> +	GEM_BUG_ON(!to_i915(dev)->mm.interruptible);
> +
> +	active_mask = i915_gem_object_is_active(obj);
> +	if (!active_mask)
> +		return 0;
> +
> +	if (!readonly) {
> +		active = obj->last_read;
> +	} else {
> +		active_mask = 1;
> +		active = &obj->last_write;
> +	}
> +
> +	for_each_active(active_mask, i) {
> +		struct drm_i915_gem_request *req;
> +
> +		req = i915_gem_active_get(&active[i],
> +					  &obj->base.dev->struct_mutex);
> +		if (req)
> +			requests[n++] = req;
> +	}
> +
> +	mutex_unlock(&dev->struct_mutex);
> +	ret = 0;
> +	for (i = 0; ret == 0 && i < n; i++)
> +		ret = i915_wait_request(requests[i], true, NULL, rps);
> +	mutex_lock(&dev->struct_mutex);
> +
> +	for (i = 0; i < n; i++)
> +		i915_gem_request_put(requests[i]);
> +
> +	return ret;
> +}
> +
> +static struct intel_rps_client *to_rps_client(struct drm_file *file)
> +{
> +	struct drm_i915_file_private *fpriv = file->driver_priv;
> +
> +	return &fpriv->rps;
> +}
> +
>  int
>  i915_gem_object_attach_phys(struct drm_i915_gem_object *obj,
>  			    int align)
> @@ -1335,107 +1436,6 @@ put_rpm:
>  	return ret;
>  }
>  
> -/**
> - * Ensures that all rendering to the object has completed and the object is
> - * safe to unbind from the GTT or access from the CPU.
> - * @obj: i915 gem object
> - * @readonly: waiting for read access or write
> - */
> -int
> -i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj,
> -			       bool readonly)
> -{
> -	struct reservation_object *resv;
> -	struct i915_gem_active *active;
> -	unsigned long active_mask;
> -	int idx, ret;
> -
> -	lockdep_assert_held(&obj->base.dev->struct_mutex);
> -
> -	if (!readonly) {
> -		active = obj->last_read;
> -		active_mask = i915_gem_object_is_active(obj);
> -	} else {
> -		active_mask = 1;
> -		active = &obj->last_write;
> -	}
> -
> -	for_each_active(active_mask, idx) {
> -		ret = i915_gem_active_wait(&active[idx],
> -					   &obj->base.dev->struct_mutex);
> -		if (ret)
> -			return ret;
> -	}
> -
> -	resv = i915_gem_object_get_dmabuf_resv(obj);
> -	if (resv) {
> -		long err;
> -
> -		err = reservation_object_wait_timeout_rcu(resv, !readonly, true,
> -							  MAX_SCHEDULE_TIMEOUT);
> -		if (err < 0)
> -			return err;
> -	}
> -
> -	return 0;
> -}
> -
> -/* A nonblocking variant of the above wait. This is a highly dangerous routine
> - * as the object state may change during this call.
> - */
> -static __must_check int
> -i915_gem_object_wait_rendering__nonblocking(struct drm_i915_gem_object *obj,
> -					    struct intel_rps_client *rps,
> -					    bool readonly)
> -{
> -	struct drm_device *dev = obj->base.dev;
> -	struct drm_i915_private *dev_priv = to_i915(dev);
> -	struct drm_i915_gem_request *requests[I915_NUM_ENGINES];
> -	struct i915_gem_active *active;
> -	unsigned long active_mask;
> -	int ret, i, n = 0;
> -
> -	BUG_ON(!mutex_is_locked(&dev->struct_mutex));
> -	BUG_ON(!dev_priv->mm.interruptible);
> -
> -	active_mask = i915_gem_object_is_active(obj);
> -	if (!active_mask)
> -		return 0;
> -
> -	if (!readonly) {
> -		active = obj->last_read;
> -	} else {
> -		active_mask = 1;
> -		active = &obj->last_write;
> -	}
> -
> -	for_each_active(active_mask, i) {
> -		struct drm_i915_gem_request *req;
> -
> -		req = i915_gem_active_get(&active[i],
> -					  &obj->base.dev->struct_mutex);
> -		if (req)
> -			requests[n++] = req;
> -	}
> -
> -	mutex_unlock(&dev->struct_mutex);
> -	ret = 0;
> -	for (i = 0; ret == 0 && i < n; i++)
> -		ret = i915_wait_request(requests[i], true, NULL, rps);
> -	mutex_lock(&dev->struct_mutex);
> -
> -	for (i = 0; i < n; i++)
> -		i915_gem_request_put(requests[i]);
> -
> -	return ret;
> -}
> -
> -static struct intel_rps_client *to_rps_client(struct drm_file *file)
> -{
> -	struct drm_i915_file_private *fpriv = file->driver_priv;
> -	return &fpriv->rps;
> -}
> -
>  static enum fb_op_origin
>  write_origin(struct drm_i915_gem_object *obj, unsigned domain)
>  {
-- 
Joonas Lahtinen
Open Source Technology Center
Intel Corporation
_______________________________________________
Intel-gfx mailing list
Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/intel-gfx




[Index of Archives]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux