On Tue, 4 Aug 2020 at 21:25, Christian König <christian.koenig@xxxxxxx> wrote: > > Am 04.08.20 um 04:56 schrieb Dave Airlie: > > From: Dave Airlie <airlied@xxxxxxxxxx> > > > > This will allow different abstractions later. > > > > Signed-off-by: Dave Airlie <airlied@xxxxxxxxxx> > > Acked-by: Christian König <christian.koenig@xxxxxxx> Reviewed-by: Ben Skeggs <bskeggs@xxxxxxxxxx> > > > --- > > drivers/gpu/drm/ttm/ttm_bo.c | 34 +++++++++++++++---------------- > > drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++--------- > > drivers/gpu/drm/ttm/ttm_bo_vm.c | 2 +- > > include/drm/ttm/ttm_bo_driver.h | 6 ++++++ > > 4 files changed, 34 insertions(+), 28 deletions(-) > > > > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > > index ebecb796dd49..8777c323e7de 100644 > > --- a/drivers/gpu/drm/ttm/ttm_bo.c > > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > > @@ -108,7 +108,7 @@ static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, > > return; > > drm_printf(&p, " placement[%d]=0x%08X (%d)\n", > > i, placement->placement[i].flags, mem_type); > > - man = &bo->bdev->man[mem_type]; > > + man = ttm_manager_type(bo->bdev, mem_type); > > ttm_mem_type_manager_debug(man, &p); > > } > > } > > @@ -157,7 +157,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo, > > if (mem->placement & TTM_PL_FLAG_NO_EVICT) > > return; > > > > - man = &bdev->man[mem->mem_type]; > > + man = ttm_manager_type(bdev, mem->mem_type); > > list_add_tail(&bo->lru, &man->lru[bo->priority]); > > > > if (man->use_tt && bo->ttm && > > @@ -232,7 +232,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > > dma_resv_assert_held(pos->first->base.resv); > > dma_resv_assert_held(pos->last->base.resv); > > > > - man = &pos->first->bdev->man[TTM_PL_TT]; > > + man = ttm_manager_type(pos->first->bdev, TTM_PL_TT); > > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > > &pos->last->lru); > > } > > @@ -247,7 +247,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > > dma_resv_assert_held(pos->first->base.resv); > > dma_resv_assert_held(pos->last->base.resv); > > > > - man = &pos->first->bdev->man[TTM_PL_VRAM]; > > + man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM); > > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > > &pos->last->lru); > > } > > @@ -273,8 +273,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > > struct ttm_operation_ctx *ctx) > > { > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; > > - struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type); > > + struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type); > > int ret; > > > > ret = ttm_mem_io_lock(old_man, true); > > @@ -340,7 +340,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > > return 0; > > > > out_err: > > - new_man = &bdev->man[bo->mem.mem_type]; > > + new_man = ttm_manager_type(bdev, bo->mem.mem_type); > > if (!new_man->use_tt) { > > ttm_tt_destroy(bo->ttm); > > bo->ttm = NULL; > > @@ -552,7 +552,7 @@ static void ttm_bo_release(struct kref *kref) > > struct ttm_buffer_object *bo = > > container_of(kref, struct ttm_buffer_object, kref); > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > > size_t acc_size = bo->acc_size; > > int ret; > > > > @@ -844,7 +844,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > > const struct ttm_place *place, > > struct ttm_mem_reg *mem) > > { > > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > > > mem->mm_node = NULL; > > if (!man->func || !man->func->get_node) > > @@ -855,7 +855,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > > > > void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) > > { > > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > > > if (!man->func || !man->func->put_node) > > return; > > @@ -912,7 +912,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo, > > struct ttm_operation_ctx *ctx) > > { > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > > struct ww_acquire_ctx *ticket; > > int ret; > > > > @@ -1002,7 +1002,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo, > > if (ret) > > return ret; > > > > - man = &bdev->man[mem_type]; > > + man = ttm_manager_type(bdev, mem_type); > > if (!man->has_type || !man->use_type) > > return -EBUSY; > > > > @@ -1065,7 +1065,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo, > > if (unlikely(ret)) > > goto error; > > > > - man = &bdev->man[mem->mem_type]; > > + man = ttm_manager_type(bdev, mem->mem_type); > > ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu); > > if (unlikely(ret)) { > > ttm_bo_mem_put(bo, mem); > > @@ -1455,7 +1455,7 @@ EXPORT_SYMBOL(ttm_mem_type_manager_force_list_clean); > > > > int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) > > { > > - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); > > > > if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) { > > pr_err("Illegal memory manager memory type %u\n", mem_type); > > @@ -1558,7 +1558,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev) > > unsigned i; > > struct ttm_mem_type_manager *man; > > > > - man = &bdev->man[TTM_PL_SYSTEM]; > > + man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > > ttm_mem_type_manager_disable(man); > > > > mutex_lock(&ttm_global_mutex); > > @@ -1585,7 +1585,7 @@ EXPORT_SYMBOL(ttm_bo_device_release); > > > > static void ttm_bo_init_sysman(struct ttm_bo_device *bdev) > > { > > - struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > > > > /* > > * Initialize the system memory buffer type. > > @@ -1649,7 +1649,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) > > void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) > > { > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > > > > ttm_mem_io_lock(man, false); > > ttm_bo_unmap_virtual_locked(bo); > > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c > > index 1f502be0b646..879c8ded0cd8 100644 > > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > > @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man) > > int ttm_mem_io_reserve(struct ttm_bo_device *bdev, > > struct ttm_mem_reg *mem) > > { > > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > > int ret; > > > > if (mem->bus.io_reserved_count++) > > @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev, > > > > int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) > > { > > - struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type); > > struct ttm_mem_reg *mem = &bo->mem; > > int ret; > > > > @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, > > struct ttm_mem_reg *mem, > > void **virtual) > > { > > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > > int ret; > > void *addr; > > > > @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, > > { > > struct ttm_mem_type_manager *man; > > > > - man = &bdev->man[mem->mem_type]; > > + man = ttm_manager_type(bdev, mem->mem_type); > > > > if (virtual && mem->bus.addr == NULL) > > iounmap(virtual); > > @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, > > struct ttm_mem_reg *new_mem) > > { > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > > struct ttm_tt *ttm = bo->ttm; > > struct ttm_mem_reg *old_mem = &bo->mem; > > struct ttm_mem_reg old_copy = *old_mem; > > @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo, > > struct ttm_bo_kmap_obj *map) > > { > > struct ttm_mem_type_manager *man = > > - &bo->bdev->man[bo->mem.mem_type]; > > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > > unsigned long offset, size; > > int ret; > > > > @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map) > > { > > struct ttm_buffer_object *bo = map->bo; > > struct ttm_mem_type_manager *man = > > - &bo->bdev->man[bo->mem.mem_type]; > > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > > > > if (!map->virtual) > > return; > > @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, > > struct ttm_mem_reg *new_mem) > > { > > struct ttm_bo_device *bdev = bo->bdev; > > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > > struct ttm_mem_reg *old_mem = &bo->mem; > > int ret; > > struct ttm_buffer_object *ghost_obj; > > @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo, > > struct ttm_bo_device *bdev = bo->bdev; > > struct ttm_mem_reg *old_mem = &bo->mem; > > > > - struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type]; > > - struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type]; > > + struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type); > > + struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type); > > > > int ret; > > > > diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c > > index 468a0eb9e632..5ae679184eb5 100644 > > --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c > > +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c > > @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf, > > vm_fault_t ret = VM_FAULT_NOPAGE; > > unsigned long address = vmf->address; > > struct ttm_mem_type_manager *man = > > - &bdev->man[bo->mem.mem_type]; > > + ttm_manager_type(bdev, bo->mem.mem_type); > > > > /* > > * Refuse to fault imported pages. This should be handled > > diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h > > index 8cc39cd55a14..e80deee3ae99 100644 > > --- a/include/drm/ttm/ttm_bo_driver.h > > +++ b/include/drm/ttm/ttm_bo_driver.h > > @@ -444,6 +444,12 @@ struct ttm_bo_device { > > bool no_retry; > > }; > > > > +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev, > > + int mem_type) > > +{ > > + return &bdev->man[mem_type]; > > +} > > + > > /** > > * struct ttm_lru_bulk_move_pos > > * > > _______________________________________________ > dri-devel mailing list > dri-devel@xxxxxxxxxxxxxxxxxxxxx > https://lists.freedesktop.org/mailman/listinfo/dri-devel _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel