Re: [PATCH 1/2] amdgpu: cleanup public interface

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Alex,

He sent them to me. I've applied them to my tree:

http://cgit.freedesktop.org/~mareko/mesa/log/?h=amdgpu

Marek

On Fri, Apr 24, 2015 at 5:38 PM, Alex Deucher <alexdeucher@xxxxxxxxx> wrote:
> On Thu, Apr 23, 2015 at 11:53 AM, Christian König
> <deathsimple@xxxxxxxxxxx> wrote:
>> From: Christian König <christian.koenig@xxxxxxx>
>>
>> Remove the mostly unused device parameter, for the few cases
>> where we really need it keep a copy in the context structure.
>
> Are you going to send out the mesa patches as well?
>
> Alex
>
>>
>> Signed-off-by: Christian König <christian.koenig@xxxxxxx>
>> Reviewed-by: Jammy Zhou <Jammy.Zhou@xxxxxxx>
>> Reviewed-by: Alex Deucher <alexander.deucher@xxxxxxx>
>> ---
>>  amdgpu/amdgpu.h            |  24 +++-------
>>  amdgpu/amdgpu_cs.c         | 115 +++++++++++++++++++--------------------------
>>  amdgpu/amdgpu_internal.h   |   2 +
>>  tests/amdgpu/basic_tests.c |  33 +++++++------
>>  tests/amdgpu/cs_tests.c    |  14 +++---
>>  5 files changed, 78 insertions(+), 110 deletions(-)
>>
>> diff --git a/amdgpu/amdgpu.h b/amdgpu/amdgpu.h
>> index 90dc33c..7a85982 100644
>> --- a/amdgpu/amdgpu.h
>> +++ b/amdgpu/amdgpu.h
>> @@ -882,7 +882,6 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
>>   *
>>   * Destroy GPU execution context when not needed any more
>>   *
>> - * \param   dev            - \c [in] Device handle. See #amdgpu_device_initialize()
>>   * \param   context - \c [in] GPU Context handle
>>   *
>>   * \return   0 on success\n
>> @@ -892,13 +891,11 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
>>   * \sa amdgpu_cs_ctx_create()
>>   *
>>  */
>> -int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
>> -                      amdgpu_context_handle context);
>> +int amdgpu_cs_ctx_free(amdgpu_context_handle context);
>>
>>  /**
>>   * Query reset state for the specific GPU Context
>>   *
>> - * \param   dev            - \c [in] Device handle. See #amdgpu_device_initialize()
>>   * \param   context - \c [in]  GPU Context handle
>>   * \param   state   - \c [out] Reset state status
>>   *
>> @@ -909,8 +906,7 @@ int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
>>   * \sa amdgpu_cs_ctx_create()
>>   *
>>  */
>> -int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
>> -                               amdgpu_context_handle context,
>> +int amdgpu_cs_query_reset_state(amdgpu_context_handle context,
>>                                 enum amdgpu_cs_ctx_reset_state *state);
>>
>>
>> @@ -924,7 +920,6 @@ int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
>>   * Allocate memory to be filled with PM4 packets and be served as the first
>>   * entry point of execution (a.k.a. Indirect Buffer)
>>   *
>> - * \param   dev            - \c [in] Device handle. See #amdgpu_device_initialize()
>>   * \param   context - \c [in]  GPU Context which will use IB
>>   * \param   ib_size - \c [in]  Size of allocation
>>   * \param   output  - \c [out] Pointer to structure to get information about
>> @@ -937,8 +932,7 @@ int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
>>   * \sa amdgpu_cs_free_ib()
>>   *
>>  */
>> -int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>> -                      amdgpu_context_handle context,
>> +int amdgpu_cs_alloc_ib(amdgpu_context_handle context,
>>                        enum amdgpu_cs_ib_size ib_size,
>>                        struct amdgpu_cs_ib_alloc_result *output);
>>
>> @@ -946,8 +940,6 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>>   * If UMD has allocates IBs which doesn’t need any more than those IBs must
>>   * be explicitly freed
>>   *
>> - * \param   dev            - \c [in] Device handle. See #amdgpu_device_initialize()
>> - * \param   context - \c [in] GPU Context containing IB
>>   * \param   handle  - \c [in] IB handle
>>   *
>>   * \return   0 on success\n
>> @@ -960,9 +952,7 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>>   * \sa amdgpu_cs_alloc_ib()
>>   *
>>  */
>> -int amdgpu_cs_free_ib(amdgpu_device_handle dev,
>> -                     amdgpu_context_handle context,
>> -                     amdgpu_ib_handle handle);
>> +int amdgpu_cs_free_ib(amdgpu_ib_handle handle);
>>
>>  /**
>>   * Send request to submit command buffers to hardware.
>> @@ -1011,8 +1001,7 @@ int amdgpu_cs_free_ib(amdgpu_device_handle dev,
>>   *     amdgpu_cs_query_fence_status()
>>   *
>>  */
>> -int amdgpu_cs_submit(amdgpu_device_handle  dev,
>> -                    amdgpu_context_handle context,
>> +int amdgpu_cs_submit(amdgpu_context_handle context,
>>                      uint64_t flags,
>>                      struct amdgpu_cs_request *ibs_request,
>>                      uint32_t number_of_requests,
>> @@ -1038,8 +1027,7 @@ int amdgpu_cs_submit(amdgpu_device_handle  dev,
>>   *
>>   * \sa amdgpu_cs_submit()
>>  */
>> -int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
>> -                                struct amdgpu_cs_query_fence *fence,
>> +int amdgpu_cs_query_fence_status(struct amdgpu_cs_query_fence *fence,
>>                                  uint32_t *expired);
>>
>>
>> diff --git a/amdgpu/amdgpu_cs.c b/amdgpu/amdgpu_cs.c
>> index 614904d..d6b4b2d 100644
>> --- a/amdgpu/amdgpu_cs.c
>> +++ b/amdgpu/amdgpu_cs.c
>> @@ -42,8 +42,7 @@
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
>> -                              amdgpu_context_handle context,
>> +static int amdgpu_cs_create_ib(amdgpu_context_handle context,
>>                                enum amdgpu_cs_ib_size ib_size,
>>                                amdgpu_ib_handle *ib)
>>  {
>> @@ -79,7 +78,7 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
>>
>>         alloc_buffer.preferred_heap = AMDGPU_GEM_DOMAIN_GTT;
>>
>> -       r = amdgpu_bo_alloc(dev,
>> +       r = amdgpu_bo_alloc(context->dev,
>>                             &alloc_buffer,
>>                             &info);
>>         if (r)
>> @@ -98,6 +97,7 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
>>                 return -ENOMEM;
>>         }
>>
>> +       new_ib->context = context;
>>         new_ib->buf_handle = info.buf_handle;
>>         new_ib->cpu = cpu;
>>         new_ib->virtual_mc_base_address = info.virtual_mc_base_address;
>> @@ -114,10 +114,10 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_destroy_ib(amdgpu_device_handle dev,
>> -                               amdgpu_ib_handle ib)
>> +static int amdgpu_cs_destroy_ib(amdgpu_ib_handle ib)
>>  {
>>         int r;
>> +
>>         r = amdgpu_bo_cpu_unmap(ib->buf_handle);
>>         if (r)
>>                 return r;
>> @@ -162,8 +162,7 @@ static int amdgpu_cs_init_ib_pool(amdgpu_context_handle context)
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_alloc_from_ib_pool(amdgpu_device_handle dev,
>> -                                       amdgpu_context_handle context,
>> +static int amdgpu_cs_alloc_from_ib_pool(amdgpu_context_handle context,
>>                                         enum amdgpu_cs_ib_size ib_size,
>>                                         amdgpu_ib_handle *ib)
>>  {
>> @@ -210,21 +209,19 @@ static void amdgpu_cs_free_to_ib_pool(amdgpu_context_handle context,
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_destroy_ib_pool(amdgpu_device_handle dev,
>> -                                    amdgpu_context_handle context)
>> +static int amdgpu_cs_destroy_ib_pool(amdgpu_context_handle context)
>>  {
>> -       int i;
>> -       int r;
>>         struct list_head *head;
>>         struct amdgpu_ib *next;
>>         struct amdgpu_ib *storage;
>> +       int i, r;
>>
>>         r = 0;
>>         pthread_mutex_lock(&context->pool_mutex);
>>         for (i = 0; i < AMDGPU_CS_IB_SIZE_NUM; i++) {
>>                 head = &context->ib_pools[i];
>>                 LIST_FOR_EACH_ENTRY_SAFE(next, storage, head, list_node) {
>> -                       r = amdgpu_cs_destroy_ib(dev, next);
>> +                       r = amdgpu_cs_destroy_ib(next);
>>                         if (r)
>>                                 break;
>>                 }
>> @@ -268,8 +265,7 @@ static int amdgpu_cs_init_pendings(amdgpu_context_handle context)
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_destroy_pendings(amdgpu_device_handle dev,
>> -                                     amdgpu_context_handle context)
>> +static int amdgpu_cs_destroy_pendings(amdgpu_context_handle context)
>>  {
>>         int ip, inst;
>>         uint32_t ring;
>> @@ -285,7 +281,7 @@ static int amdgpu_cs_destroy_pendings(amdgpu_device_handle dev,
>>                         for (ring = 0; ring < AMDGPU_CS_MAX_RINGS; ring++) {
>>                                 head = &context->pendings[ip][inst][ring];
>>                                 LIST_FOR_EACH_ENTRY_SAFE(next, s, head, list_node) {
>> -                                       r = amdgpu_cs_destroy_ib(dev, next);
>> +                                       r = amdgpu_cs_destroy_ib(next);
>>                                         if (r)
>>                                                 break;
>>                                 }
>> @@ -293,7 +289,7 @@ static int amdgpu_cs_destroy_pendings(amdgpu_device_handle dev,
>>
>>         head = &context->freed;
>>         LIST_FOR_EACH_ENTRY_SAFE(next, s, head, list_node) {
>> -               r = amdgpu_cs_destroy_ib(dev, next);
>> +               r = amdgpu_cs_destroy_ib(next);
>>                 if (r)
>>                         break;
>>         }
>> @@ -441,39 +437,35 @@ static void amdgpu_cs_all_pending_gc(amdgpu_context_handle context)
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -static int amdgpu_cs_alloc_ib_local(amdgpu_device_handle dev,
>> -                                   amdgpu_context_handle context,
>> +static int amdgpu_cs_alloc_ib_local(amdgpu_context_handle context,
>>                                     enum amdgpu_cs_ib_size ib_size,
>>                                     amdgpu_ib_handle *ib)
>>  {
>>         int r;
>>
>> -       r = amdgpu_cs_alloc_from_ib_pool(dev, context, ib_size, ib);
>> +       r = amdgpu_cs_alloc_from_ib_pool(context, ib_size, ib);
>>         if (!r)
>>                 return r;
>>
>>         amdgpu_cs_all_pending_gc(context);
>>
>>         /* Retry to allocate from free IB pools after garbage collector. */
>> -       r = amdgpu_cs_alloc_from_ib_pool(dev, context, ib_size, ib);
>> +       r = amdgpu_cs_alloc_from_ib_pool(context, ib_size, ib);
>>         if (!r)
>>                 return r;
>>
>>         /* There is no suitable IB in free pools. Create one. */
>> -       r = amdgpu_cs_create_ib(dev, context, ib_size, ib);
>> +       r = amdgpu_cs_create_ib(context, ib_size, ib);
>>         return r;
>>  }
>>
>> -int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>> -                      amdgpu_context_handle context,
>> +int amdgpu_cs_alloc_ib(amdgpu_context_handle context,
>>                        enum amdgpu_cs_ib_size ib_size,
>>                        struct amdgpu_cs_ib_alloc_result *output)
>>  {
>>         int r;
>>         amdgpu_ib_handle ib;
>>
>> -       if (NULL == dev)
>> -               return -EINVAL;
>>         if (NULL == context)
>>                 return -EINVAL;
>>         if (NULL == output)
>> @@ -481,7 +473,7 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>>         if (ib_size >= AMDGPU_CS_IB_SIZE_NUM)
>>                 return -EINVAL;
>>
>> -       r = amdgpu_cs_alloc_ib_local(dev, context, ib_size, &ib);
>> +       r = amdgpu_cs_alloc_ib_local(context, ib_size, &ib);
>>         if (!r) {
>>                 output->handle = ib;
>>                 output->cpu = ib->cpu;
>> @@ -491,17 +483,14 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
>>         return r;
>>  }
>>
>> -int amdgpu_cs_free_ib(amdgpu_device_handle dev,
>> -                     amdgpu_context_handle context,
>> -                     amdgpu_ib_handle handle)
>> +int amdgpu_cs_free_ib(amdgpu_ib_handle handle)
>>  {
>> -       if (NULL == dev)
>> -               return -EINVAL;
>> -       if (NULL == context)
>> -               return -EINVAL;
>> +       amdgpu_context_handle context;
>> +
>>         if (NULL == handle)
>>                 return -EINVAL;
>>
>> +       context = handle->context;
>>         pthread_mutex_lock(&context->pendings_mutex);
>>         LIST_ADD(&handle->list_node, &context->freed);
>>         pthread_mutex_unlock(&context->pendings_mutex);
>> @@ -532,6 +521,8 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
>>         if (NULL == gpu_context)
>>                 return -ENOMEM;
>>
>> +       gpu_context->dev = dev;
>> +
>>         r = pthread_mutex_init(&gpu_context->sequence_mutex, NULL);
>>         if (r)
>>                 goto error_mutex;
>> @@ -544,7 +535,7 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
>>         if (r)
>>                 goto error_pendings;
>>
>> -       r = amdgpu_cs_alloc_ib_local(dev, gpu_context, amdgpu_cs_ib_size_4K,
>> +       r = amdgpu_cs_alloc_ib_local(gpu_context, amdgpu_cs_ib_size_4K,
>>                                      &gpu_context->fence_ib);
>>         if (r)
>>                 goto error_fence_ib;
>> @@ -562,13 +553,13 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
>>         return 0;
>>
>>  error_kernel:
>> -       amdgpu_cs_free_ib(dev, gpu_context, gpu_context->fence_ib);
>> +       amdgpu_cs_free_ib(gpu_context->fence_ib);
>>
>>  error_fence_ib:
>> -       amdgpu_cs_destroy_pendings(dev, gpu_context);
>> +       amdgpu_cs_destroy_pendings(gpu_context);
>>
>>  error_pendings:
>> -       amdgpu_cs_destroy_ib_pool(dev, gpu_context);
>> +       amdgpu_cs_destroy_ib_pool(gpu_context);
>>
>>  error_pool:
>>         pthread_mutex_destroy(&gpu_context->sequence_mutex);
>> @@ -586,26 +577,23 @@ error_mutex:
>>   *
>>   * \return  0 on success otherwise POSIX Error code
>>  */
>> -int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
>> -                      amdgpu_context_handle context)
>> +int amdgpu_cs_ctx_free(amdgpu_context_handle context)
>>  {
>> -       int r;
>>         union drm_amdgpu_ctx args;
>> +       int r;
>>
>> -       if (NULL == dev)
>> -               return -EINVAL;
>>         if (NULL == context)
>>                 return -EINVAL;
>>
>> -       r = amdgpu_cs_free_ib(dev, context, context->fence_ib);
>> +       r = amdgpu_cs_free_ib(context->fence_ib);
>>         if (r)
>>                 return r;
>>
>> -       r = amdgpu_cs_destroy_pendings(dev, context);
>> +       r = amdgpu_cs_destroy_pendings(context);
>>         if (r)
>>                 return r;
>>
>> -       r = amdgpu_cs_destroy_ib_pool(dev, context);
>> +       r = amdgpu_cs_destroy_ib_pool(context);
>>         if (r)
>>                 return r;
>>
>> @@ -615,15 +603,15 @@ int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
>>         memset(&args, 0, sizeof(args));
>>         args.in.op = AMDGPU_CTX_OP_FREE_CTX;
>>         args.in.ctx_id = context->id;
>> -       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_CTX, &args, sizeof(args));
>> +       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_CTX,
>> +                               &args, sizeof(args));
>>
>>         free(context);
>>
>>         return r;
>>  }
>>
>> -static int amdgpu_cs_create_bo_list(amdgpu_device_handle dev,
>> -                                   amdgpu_context_handle context,
>> +static int amdgpu_cs_create_bo_list(amdgpu_context_handle context,
>>                                     struct amdgpu_cs_request *request,
>>                                     amdgpu_ib_handle fence_ib,
>>                                     uint32_t *handle)
>> @@ -657,7 +645,7 @@ static int amdgpu_cs_create_bo_list(amdgpu_device_handle dev,
>>         if (fence_ib)
>>                 list[i].bo_handle = fence_ib->buf_handle->handle;
>>
>> -       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
>> +       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_BO_LIST,
>>                                 &args, sizeof(args));
>>         if (r)
>>                 return r;
>> @@ -666,7 +654,7 @@ static int amdgpu_cs_create_bo_list(amdgpu_device_handle dev,
>>         return 0;
>>  }
>>
>> -static int amdgpu_cs_free_bo_list(amdgpu_device_handle dev, uint32_t handle)
>> +static int amdgpu_cs_free_bo_list(amdgpu_context_handle context, uint32_t handle)
>>  {
>>         union drm_amdgpu_bo_list args;
>>         int r;
>> @@ -675,7 +663,7 @@ static int amdgpu_cs_free_bo_list(amdgpu_device_handle dev, uint32_t handle)
>>         args.in.operation = AMDGPU_BO_LIST_OP_DESTROY;
>>         args.in.list_handle = handle;
>>
>> -       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
>> +       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_BO_LIST,
>>                                 &args, sizeof(args));
>>
>>         return r;
>> @@ -696,8 +684,7 @@ static uint32_t amdgpu_cs_fence_index(unsigned ip, unsigned ring)
>>   * \return  0 on success otherwise POSIX Error code
>>   * \sa amdgpu_cs_submit()
>>  */
>> -static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
>> -                               amdgpu_context_handle context,
>> +static int amdgpu_cs_submit_one(amdgpu_context_handle context,
>>                                 struct amdgpu_cs_request *ibs_request,
>>                                 uint64_t *fence)
>>  {
>> @@ -753,7 +740,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
>>                         chunk_data[i].ib_data.flags = AMDGPU_IB_FLAG_CE;
>>         }
>>
>> -       r = amdgpu_cs_create_bo_list(dev, context, ibs_request, NULL,
>> +       r = amdgpu_cs_create_bo_list(context, ibs_request, NULL,
>>                                      &cs.in.bo_list_handle);
>>         if (r)
>>                 goto error_unlock;
>> @@ -778,7 +765,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
>>                 chunk_data[i].fence_data.offset *= sizeof(uint64_t);
>>         }
>>
>> -       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_CS,
>> +       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_CS,
>>                                 &cs, sizeof(cs));
>>         if (r)
>>                 goto error_unlock;
>> @@ -805,7 +792,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
>>
>>         pthread_mutex_unlock(&context->sequence_mutex);
>>
>> -       r = amdgpu_cs_free_bo_list(dev, cs.in.bo_list_handle);
>> +       r = amdgpu_cs_free_bo_list(context, cs.in.bo_list_handle);
>>         if (r)
>>                 goto error_free;
>>
>> @@ -820,18 +807,15 @@ error_free:
>>         return r;
>>  }
>>
>> -int amdgpu_cs_submit(amdgpu_device_handle  dev,
>> -                    amdgpu_context_handle context,
>> +int amdgpu_cs_submit(amdgpu_context_handle context,
>>                      uint64_t flags,
>>                      struct amdgpu_cs_request *ibs_request,
>>                      uint32_t number_of_requests,
>>                      uint64_t *fences)
>>  {
>> -       int r;
>>         uint32_t i;
>> +       int r;
>>
>> -       if (NULL == dev)
>> -               return -EINVAL;
>>         if (NULL == context)
>>                 return -EINVAL;
>>         if (NULL == ibs_request)
>> @@ -841,7 +825,7 @@ int amdgpu_cs_submit(amdgpu_device_handle  dev,
>>
>>         r = 0;
>>         for (i = 0; i < number_of_requests; i++) {
>> -               r = amdgpu_cs_submit_one(dev, context, ibs_request, fences);
>> +               r = amdgpu_cs_submit_one(context, ibs_request, fences);
>>                 if (r)
>>                         break;
>>                 fences++;
>> @@ -904,8 +888,7 @@ static int amdgpu_ioctl_wait_cs(amdgpu_device_handle dev,
>>         return 0;
>>  }
>>
>> -int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
>> -                                struct amdgpu_cs_query_fence *fence,
>> +int amdgpu_cs_query_fence_status(struct amdgpu_cs_query_fence *fence,
>>                                  uint32_t *expired)
>>  {
>>         amdgpu_context_handle context;
>> @@ -916,8 +899,6 @@ int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
>>         bool busy = true;
>>         int r;
>>
>> -       if (NULL == dev)
>> -               return -EINVAL;
>>         if (NULL == fence)
>>                 return -EINVAL;
>>         if (NULL == expired)
>> @@ -958,7 +939,7 @@ int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
>>
>>         pthread_mutex_unlock(&context->sequence_mutex);
>>
>> -       r = amdgpu_ioctl_wait_cs(dev, ip_type, ip_instance, ring,
>> +       r = amdgpu_ioctl_wait_cs(context->dev, ip_type, ip_instance, ring,
>>                                  fence->fence, fence->timeout_ns, &busy);
>>         if (!r && !busy) {
>>                 *expired = true;
>> diff --git a/amdgpu/amdgpu_internal.h b/amdgpu/amdgpu_internal.h
>> index b27756d..55db645 100644
>> --- a/amdgpu/amdgpu_internal.h
>> +++ b/amdgpu/amdgpu_internal.h
>> @@ -93,6 +93,7 @@ struct amdgpu_bo {
>>   * sequence_mutex -> pendings_mutex -> pool_mutex.
>>  */
>>  struct amdgpu_context {
>> +       struct amdgpu_device *dev;
>>         /** Mutex for accessing fences and to maintain command submissions
>>             and pending lists in good sequence. */
>>         pthread_mutex_t sequence_mutex;
>> @@ -115,6 +116,7 @@ struct amdgpu_context {
>>  };
>>
>>  struct amdgpu_ib {
>> +       amdgpu_context_handle context;
>>         struct list_head list_node;
>>         amdgpu_bo_handle buf_handle;
>>         void *cpu;
>> diff --git a/tests/amdgpu/basic_tests.c b/tests/amdgpu/basic_tests.c
>> index c53f6a0..c01a7f4 100644
>> --- a/tests/amdgpu/basic_tests.c
>> +++ b/tests/amdgpu/basic_tests.c
>> @@ -168,11 +168,11 @@ static void amdgpu_command_submission_gfx(void)
>>         r = amdgpu_cs_ctx_create(device_handle, &context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> -       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> +       r = amdgpu_cs_alloc_ib(context_handle,
>>                                amdgpu_cs_ib_size_4K, &ib_result);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> -       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> +       r = amdgpu_cs_alloc_ib(context_handle,
>>                                amdgpu_cs_ib_size_4K, &ib_result_ce);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -199,7 +199,7 @@ static void amdgpu_command_submission_gfx(void)
>>         ibs_request.number_of_ibs = 2;
>>         ibs_request.ibs = ib_info;
>>
>> -       r = amdgpu_cs_submit(device_handle, context_handle, 0,
>> +       r = amdgpu_cs_submit(context_handle, 0,
>>                              &ibs_request, 1, &fence_status.fence);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -207,10 +207,10 @@ static void amdgpu_command_submission_gfx(void)
>>         fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;
>>         fence_status.ip_type = AMDGPU_HW_IP_GFX;
>>
>> -       r = amdgpu_cs_query_fence_status(device_handle, &fence_status, &expired);
>> +       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>>
>> @@ -230,7 +230,7 @@ static void amdgpu_command_submission_compute(void)
>>
>>         for (instance = 0; instance < 8; instance++) {
>>                 memset(&ib_result, 0, sizeof(struct amdgpu_cs_ib_alloc_result));
>> -               r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> +               r = amdgpu_cs_alloc_ib(context_handle,
>>                                        amdgpu_cs_ib_size_4K, &ib_result);
>>                 CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -249,7 +249,7 @@ static void amdgpu_command_submission_compute(void)
>>                 ibs_request.ibs = &ib_info;
>>
>>                 memset(&fence_status, 0, sizeof(struct amdgpu_cs_query_fence));
>> -               r = amdgpu_cs_submit(device_handle, context_handle, 0,
>> +               r = amdgpu_cs_submit(context_handle, 0,
>>                                      &ibs_request, 1, &fence_status.fence);
>>                 CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -258,11 +258,11 @@ static void amdgpu_command_submission_compute(void)
>>                 fence_status.ip_type = AMDGPU_HW_IP_COMPUTE;
>>                 fence_status.ring = instance;
>>
>> -               r = amdgpu_cs_query_fence_status(device_handle, &fence_status, &expired);
>> +               r = amdgpu_cs_query_fence_status(&fence_status, &expired);
>>                 CU_ASSERT_EQUAL(r, 0);
>>         }
>>
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>>
>> @@ -291,7 +291,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle context_handle,
>>         CU_ASSERT_TRUE(pm4_dw <= 1024);
>>
>>         /* allocate IB */
>> -       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> +       r = amdgpu_cs_alloc_ib(context_handle,
>>                            amdgpu_cs_ib_size_4K, &ib_result);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -313,7 +313,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle context_handle,
>>         CU_ASSERT_NOT_EQUAL(ibs_request, NULL);
>>
>>         /* submit CS */
>> -       r = amdgpu_cs_submit(device_handle, context_handle, 0,
>> +       r = amdgpu_cs_submit(context_handle, 0,
>>                          ibs_request, 1, &fence_status.fence);
>>         CU_ASSERT_EQUAL(r, 0);
>>
>> @@ -323,8 +323,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle context_handle,
>>         fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;
>>
>>         /* wait for IB accomplished */
>> -       r = amdgpu_cs_query_fence_status(device_handle, &fence_status,
>> -                                       &expired);
>> +       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
>>         CU_ASSERT_EQUAL(r, 0);
>>         CU_ASSERT_EQUAL(expired, true);
>>  }
>> @@ -410,7 +409,7 @@ static void amdgpu_command_submission_sdma_write_linear(void)
>>         free(pm4);
>>
>>         /* end of test */
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>>
>> @@ -493,7 +492,7 @@ static void amdgpu_command_submission_sdma_const_fill(void)
>>         free(pm4);
>>
>>         /* end of test */
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>>
>> @@ -596,7 +595,7 @@ static void amdgpu_command_submission_sdma_copy_linear(void)
>>         free(pm4);
>>
>>         /* end of test */
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>>
>> @@ -671,6 +670,6 @@ static void amdgpu_userptr_test(void)
>>         CU_ASSERT_EQUAL(r, 0);
>>         free(ptr);
>>
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         CU_ASSERT_EQUAL(r, 0);
>>  }
>> diff --git a/tests/amdgpu/cs_tests.c b/tests/amdgpu/cs_tests.c
>> index f30d923..c0d0c0d 100644
>> --- a/tests/amdgpu/cs_tests.c
>> +++ b/tests/amdgpu/cs_tests.c
>> @@ -73,8 +73,7 @@ int suite_cs_tests_init(void)
>>         if (r)
>>                 return CUE_SINIT_FAILED;
>>
>> -        r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> -                               IB_SIZE, &ib_result);
>> +        r = amdgpu_cs_alloc_ib(context_handle, IB_SIZE, &ib_result);
>>         if (r)
>>                 return CUE_SINIT_FAILED;
>>
>> @@ -88,11 +87,11 @@ int suite_cs_tests_clean(void)
>>  {
>>         int r;
>>
>> -       r = amdgpu_cs_free_ib(device_handle, context_handle, ib_handle);
>> +       r = amdgpu_cs_free_ib(ib_handle);
>>         if (r)
>>                 return CUE_SCLEAN_FAILED;
>>
>> -       r = amdgpu_cs_ctx_free(device_handle, context_handle);
>> +       r = amdgpu_cs_ctx_free(context_handle);
>>         if (r)
>>                 return CUE_SCLEAN_FAILED;
>>
>> @@ -121,13 +120,12 @@ static int submit(unsigned ndw, unsigned ip)
>>         ibs_request.number_of_ibs = 1;
>>         ibs_request.ibs = &ib_info;
>>
>> -       r = amdgpu_cs_submit(device_handle, context_handle, 0,
>> +       r = amdgpu_cs_submit(context_handle, 0,
>>                              &ibs_request, 1, &fence_status.fence);
>>         if (r)
>>                 return r;
>>
>> -       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
>> -                              IB_SIZE, &ib_result);
>> +       r = amdgpu_cs_alloc_ib(context_handle, IB_SIZE, &ib_result);
>>         if (r)
>>                 return r;
>>
>> @@ -138,7 +136,7 @@ static int submit(unsigned ndw, unsigned ip)
>>         fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;
>>         fence_status.ip_type = ip;
>>
>> -       r = amdgpu_cs_query_fence_status(device_handle, &fence_status, &expired);
>> +       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
>>         if (r)
>>                 return r;
>>
>> --
>> 1.9.1
>>
> _______________________________________________
> dri-devel mailing list
> dri-devel@xxxxxxxxxxxxxxxxxxxxx
> http://lists.freedesktop.org/mailman/listinfo/dri-devel
_______________________________________________
dri-devel mailing list
dri-devel@xxxxxxxxxxxxxxxxxxxxx
http://lists.freedesktop.org/mailman/listinfo/dri-devel





[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux