On Mon, Dec 7, 2020 at 5:33 PM Christian König <ckoenig.leichtzumerken@xxxxxxxxx> wrote: > > Based on an idea from Dave, but cleaned up a bit. > > We had multiple fields for essentially the same thing. > > Now bo->size is the original size of the BO in arbitrary > units, usually bytes. > > bo->mem.num_pages is the size in number of pages in the > resource domain of bo->mem.mem_type. > > Signed-off-by: Christian König <christian.koenig@xxxxxxx> We also have ttm_bo.base.size, do we want to reuse that one like we've done for the reservation object? I'd have said "follow up patch" but it's going to be exactly the same churn once more. -Daniel > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_object.h | 4 +-- > drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 6 ++-- > drivers/gpu/drm/amd/amdgpu/mes_v10_1.c | 2 +- > drivers/gpu/drm/nouveau/nouveau_bo.c | 10 +++--- > drivers/gpu/drm/nouveau/nouveau_display.c | 8 ++--- > drivers/gpu/drm/nouveau/nouveau_prime.c | 4 +-- > drivers/gpu/drm/nouveau/nv17_fence.c | 2 +- > drivers/gpu/drm/nouveau/nv50_fence.c | 2 +- > drivers/gpu/drm/qxl/qxl_object.h | 2 +- > drivers/gpu/drm/radeon/radeon_cs.c | 3 +- > drivers/gpu/drm/radeon/radeon_object.c | 13 ++++---- > drivers/gpu/drm/radeon/radeon_object.h | 4 +-- > drivers/gpu/drm/radeon/radeon_prime.c | 4 +-- > drivers/gpu/drm/radeon/radeon_trace.h | 2 +- > drivers/gpu/drm/radeon/radeon_ttm.c | 2 +- > drivers/gpu/drm/ttm/ttm_bo.c | 33 ++++++------------- > drivers/gpu/drm/ttm/ttm_bo_util.c | 11 +++---- > drivers/gpu/drm/ttm/ttm_bo_vm.c | 6 ++-- > drivers/gpu/drm/ttm/ttm_tt.c | 2 +- > drivers/gpu/drm/vmwgfx/vmwgfx_blit.c | 4 +-- > drivers/gpu/drm/vmwgfx/vmwgfx_bo.c | 6 ++-- > drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c | 2 +- > drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c | 4 +-- > drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c | 5 ++- > drivers/gpu/drm/vmwgfx/vmwgfx_kms.c | 2 +- > drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c | 8 ++--- > drivers/gpu/drm/vmwgfx/vmwgfx_resource.c | 2 +- > drivers/gpu/drm/vmwgfx/vmwgfx_shader.c | 3 +- > drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c | 4 +-- > drivers/gpu/drm/vmwgfx/vmwgfx_surface.c | 7 ++-- > include/drm/ttm/ttm_bo_api.h | 6 ++-- > include/drm/ttm/ttm_resource.h | 1 - > 36 files changed, 82 insertions(+), 100 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > index e5919efca870..c4c93f19d273 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > @@ -269,7 +269,7 @@ static struct sg_table *amdgpu_dma_buf_map(struct dma_buf_attachment *attach, > case TTM_PL_TT: > sgt = drm_prime_pages_to_sg(obj->dev, > bo->tbo.ttm->pages, > - bo->tbo.num_pages); > + bo->tbo.ttm->num_pages); > if (IS_ERR(sgt)) > return sgt; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c > index 056cb87d09ea..52bcd1b5582f 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c > @@ -121,7 +121,7 @@ uint64_t amdgpu_gmc_agp_addr(struct ttm_buffer_object *bo) > { > struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev); > > - if (bo->num_pages != 1 || bo->ttm->caching == ttm_cached) > + if (bo->ttm->num_pages != 1 || bo->ttm->caching == ttm_cached) > return AMDGPU_BO_INVALID_OFFSET; > > if (bo->ttm->dma_address[0] + PAGE_SIZE >= adev->gmc.agp_size) > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > index c6c9723d3d8a..381ecc4788d5 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > @@ -787,7 +787,7 @@ int amdgpu_bo_kmap(struct amdgpu_bo *bo, void **ptr) > if (r < 0) > return r; > > - r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); > + r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); > if (r) > return r; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > index ed47cbac4f75..176ed3615151 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > @@ -174,12 +174,12 @@ static inline void amdgpu_bo_unreserve(struct amdgpu_bo *bo) > > static inline unsigned long amdgpu_bo_size(struct amdgpu_bo *bo) > { > - return bo->tbo.num_pages << PAGE_SHIFT; > + return bo->tbo.size; > } > > static inline unsigned amdgpu_bo_ngpu_pages(struct amdgpu_bo *bo) > { > - return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE; > + return bo->tbo.size / AMDGPU_GPU_PAGE_SIZE; > } > > static inline unsigned amdgpu_bo_gpu_page_alignment(struct amdgpu_bo *bo) > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > index ee9480d14cbc..20715ddbb746 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > @@ -127,7 +127,7 @@ TRACE_EVENT(amdgpu_bo_create, > > TP_fast_assign( > __entry->bo = bo; > - __entry->pages = bo->tbo.num_pages; > + __entry->pages = bo->tbo.mem.num_pages; > __entry->type = bo->tbo.mem.mem_type; > __entry->prefer = bo->preferred_domains; > __entry->allow = bo->allowed_domains; > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c > index ec93d4fdabbd..dfb4122b4ff3 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c > @@ -636,7 +636,7 @@ static int amdgpu_bo_move(struct ttm_buffer_object *bo, bool evict, > > out: > /* update statistics */ > - atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &adev->num_bytes_moved); > + atomic64_add(bo->size, &adev->num_bytes_moved); > amdgpu_bo_move_notify(bo, evict, new_mem); > return 0; > } > @@ -2131,7 +2131,7 @@ int amdgpu_fill_buffer(struct amdgpu_bo *bo, > return r; > } > > - num_pages = bo->tbo.num_pages; > + num_pages = bo->tbo.mem.num_pages; > mm_node = bo->tbo.mem.mm_node; > num_loops = 0; > while (num_pages) { > @@ -2161,7 +2161,7 @@ int amdgpu_fill_buffer(struct amdgpu_bo *bo, > } > } > > - num_pages = bo->tbo.num_pages; > + num_pages = bo->tbo.mem.num_pages; > mm_node = bo->tbo.mem.mm_node; > > while (num_pages) { > diff --git a/drivers/gpu/drm/amd/amdgpu/mes_v10_1.c b/drivers/gpu/drm/amd/amdgpu/mes_v10_1.c > index 985e454463e1..9c59f3b229fa 100644 > --- a/drivers/gpu/drm/amd/amdgpu/mes_v10_1.c > +++ b/drivers/gpu/drm/amd/amdgpu/mes_v10_1.c > @@ -554,7 +554,7 @@ static int mes_v10_1_allocate_eop_buf(struct amdgpu_device *adev) > return r; > } > > - memset(eop, 0, adev->mes.eop_gpu_obj->tbo.mem.size); > + memset(eop, 0, adev->mes.eop_gpu_obj->tbo.size); > > amdgpu_bo_kunmap(adev->mes.eop_gpu_obj); > amdgpu_bo_unreserve(adev->mes.eop_gpu_obj); > diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c > index bc542ac4c4b6..f8c51291e9c4 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_bo.c > +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c > @@ -473,10 +473,10 @@ nouveau_bo_pin(struct nouveau_bo *nvbo, uint32_t domain, bool contig) > > switch (bo->mem.mem_type) { > case TTM_PL_VRAM: > - drm->gem.vram_available -= bo->mem.size; > + drm->gem.vram_available -= bo->size; > break; > case TTM_PL_TT: > - drm->gem.gart_available -= bo->mem.size; > + drm->gem.gart_available -= bo->size; > break; > default: > break; > @@ -504,10 +504,10 @@ nouveau_bo_unpin(struct nouveau_bo *nvbo) > if (!nvbo->bo.pin_count) { > switch (bo->mem.mem_type) { > case TTM_PL_VRAM: > - drm->gem.vram_available += bo->mem.size; > + drm->gem.vram_available += bo->size; > break; > case TTM_PL_TT: > - drm->gem.gart_available += bo->mem.size; > + drm->gem.gart_available += bo->size; > break; > default: > break; > @@ -913,7 +913,7 @@ nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct ttm_resource *new_reg, > return 0; > > if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_CELSIUS) { > - *new_tile = nv10_bo_set_tiling(dev, offset, new_reg->size, > + *new_tile = nv10_bo_set_tiling(dev, offset, bo->size, > nvbo->mode, nvbo->zeta); > } > > diff --git a/drivers/gpu/drm/nouveau/nouveau_display.c b/drivers/gpu/drm/nouveau/nouveau_display.c > index bceb48a2dfca..0f7804839b5f 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_display.c > +++ b/drivers/gpu/drm/nouveau/nouveau_display.c > @@ -286,11 +286,11 @@ nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo, > > bl_size = bw * bh * (1 << tile_mode) * gob_size; > > - DRM_DEBUG_KMS("offset=%u stride=%u h=%u tile_mode=0x%02x bw=%u bh=%u gob_size=%u bl_size=%llu size=%lu\n", > + DRM_DEBUG_KMS("offset=%u stride=%u h=%u tile_mode=0x%02x bw=%u bh=%u gob_size=%u bl_size=%llu size=%Lu\n", > offset, stride, h, tile_mode, bw, bh, gob_size, bl_size, > - nvbo->bo.mem.size); > + nvbo->bo.size); > > - if (bl_size + offset > nvbo->bo.mem.size) > + if (bl_size + offset > nvbo->bo.size) > return -ERANGE; > > return 0; > @@ -363,7 +363,7 @@ nouveau_framebuffer_new(struct drm_device *dev, > } else { > uint32_t size = mode_cmd->pitches[i] * height; > > - if (size + mode_cmd->offsets[i] > nvbo->bo.mem.size) > + if (size + mode_cmd->offsets[i] > nvbo->bo.size) > return -ERANGE; > } > } > diff --git a/drivers/gpu/drm/nouveau/nouveau_prime.c b/drivers/gpu/drm/nouveau/nouveau_prime.c > index 2f16b5249283..347488685f74 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_prime.c > +++ b/drivers/gpu/drm/nouveau/nouveau_prime.c > @@ -30,9 +30,9 @@ > struct sg_table *nouveau_gem_prime_get_sg_table(struct drm_gem_object *obj) > { > struct nouveau_bo *nvbo = nouveau_gem_object(obj); > - int npages = nvbo->bo.num_pages; > > - return drm_prime_pages_to_sg(obj->dev, nvbo->bo.ttm->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, nvbo->bo.ttm->pages, > + nvbo->bo.ttm->num_pages); > } > > struct drm_gem_object *nouveau_gem_prime_import_sg_table(struct drm_device *dev, > diff --git a/drivers/gpu/drm/nouveau/nv17_fence.c b/drivers/gpu/drm/nouveau/nv17_fence.c > index 1253fdec712d..bf3247d627b1 100644 > --- a/drivers/gpu/drm/nouveau/nv17_fence.c > +++ b/drivers/gpu/drm/nouveau/nv17_fence.c > @@ -80,7 +80,7 @@ nv17_fence_context_new(struct nouveau_channel *chan) > struct nv10_fence_chan *fctx; > struct ttm_resource *reg = &priv->bo->bo.mem; > u32 start = reg->start * PAGE_SIZE; > - u32 limit = start + reg->size - 1; > + u32 limit = start + priv->bo->bo.size - 1; > int ret = 0; > > fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); > diff --git a/drivers/gpu/drm/nouveau/nv50_fence.c b/drivers/gpu/drm/nouveau/nv50_fence.c > index 447238e3cbe7..1c89702fbba5 100644 > --- a/drivers/gpu/drm/nouveau/nv50_fence.c > +++ b/drivers/gpu/drm/nouveau/nv50_fence.c > @@ -39,7 +39,7 @@ nv50_fence_context_new(struct nouveau_channel *chan) > struct nv10_fence_chan *fctx; > struct ttm_resource *reg = &priv->bo->bo.mem; > u32 start = reg->start * PAGE_SIZE; > - u32 limit = start + reg->size - 1; > + u32 limit = start + priv->bo->bo.size - 1; > int ret; > > fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); > diff --git a/drivers/gpu/drm/qxl/qxl_object.h b/drivers/gpu/drm/qxl/qxl_object.h > index ebf24c9d2bf2..be4d95ed8298 100644 > --- a/drivers/gpu/drm/qxl/qxl_object.h > +++ b/drivers/gpu/drm/qxl/qxl_object.h > @@ -50,7 +50,7 @@ static inline void qxl_bo_unreserve(struct qxl_bo *bo) > > static inline unsigned long qxl_bo_size(struct qxl_bo *bo) > { > - return bo->tbo.num_pages << PAGE_SHIFT; > + return bo->tbo.size; > } > > static inline u64 qxl_bo_mmap_offset(struct qxl_bo *bo) > diff --git a/drivers/gpu/drm/radeon/radeon_cs.c b/drivers/gpu/drm/radeon/radeon_cs.c > index 21ce2f9502c0..4f35c32957be 100644 > --- a/drivers/gpu/drm/radeon/radeon_cs.c > +++ b/drivers/gpu/drm/radeon/radeon_cs.c > @@ -401,7 +401,8 @@ static int cmp_size_smaller_first(void *priv, struct list_head *a, > struct radeon_bo_list *lb = list_entry(b, struct radeon_bo_list, tv.head); > > /* Sort A before B if A is smaller. */ > - return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; > + return (int)la->robj->tbo.mem.num_pages - > + (int)lb->robj->tbo.mem.num_pages; > } > > /** > diff --git a/drivers/gpu/drm/radeon/radeon_object.c b/drivers/gpu/drm/radeon/radeon_object.c > index ab81e35cb060..785292fc8cff 100644 > --- a/drivers/gpu/drm/radeon/radeon_object.c > +++ b/drivers/gpu/drm/radeon/radeon_object.c > @@ -54,20 +54,19 @@ static void radeon_update_memory_usage(struct radeon_bo *bo, > unsigned mem_type, int sign) > { > struct radeon_device *rdev = bo->rdev; > - u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT; > > switch (mem_type) { > case TTM_PL_TT: > if (sign > 0) > - atomic64_add(size, &rdev->gtt_usage); > + atomic64_add(bo->tbo.size, &rdev->gtt_usage); > else > - atomic64_sub(size, &rdev->gtt_usage); > + atomic64_sub(bo->tbo.size, &rdev->gtt_usage); > break; > case TTM_PL_VRAM: > if (sign > 0) > - atomic64_add(size, &rdev->vram_usage); > + atomic64_add(bo->tbo.size, &rdev->vram_usage); > else > - atomic64_sub(size, &rdev->vram_usage); > + atomic64_sub(bo->tbo.size, &rdev->vram_usage); > break; > } > } > @@ -256,7 +255,7 @@ int radeon_bo_kmap(struct radeon_bo *bo, void **ptr) > } > return 0; > } > - r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); > + r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.mem.num_pages, &bo->kmap); > if (r) { > return r; > } > @@ -610,7 +609,7 @@ int radeon_bo_get_surface_reg(struct radeon_bo *bo) > out: > radeon_set_surface_reg(rdev, i, bo->tiling_flags, bo->pitch, > bo->tbo.mem.start << PAGE_SHIFT, > - bo->tbo.num_pages << PAGE_SHIFT); > + bo->tbo.size); > return 0; > } > > diff --git a/drivers/gpu/drm/radeon/radeon_object.h b/drivers/gpu/drm/radeon/radeon_object.h > index d606e9a935e3..63821d2d2c07 100644 > --- a/drivers/gpu/drm/radeon/radeon_object.h > +++ b/drivers/gpu/drm/radeon/radeon_object.h > @@ -109,12 +109,12 @@ static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) > > static inline unsigned long radeon_bo_size(struct radeon_bo *bo) > { > - return bo->tbo.num_pages << PAGE_SHIFT; > + return bo->tbo.size; > } > > static inline unsigned radeon_bo_ngpu_pages(struct radeon_bo *bo) > { > - return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; > + return bo->tbo.size / RADEON_GPU_PAGE_SIZE; > } > > static inline unsigned radeon_bo_gpu_page_alignment(struct radeon_bo *bo) > diff --git a/drivers/gpu/drm/radeon/radeon_prime.c b/drivers/gpu/drm/radeon/radeon_prime.c > index 088d39a51c0d..0a09dbaba289 100644 > --- a/drivers/gpu/drm/radeon/radeon_prime.c > +++ b/drivers/gpu/drm/radeon/radeon_prime.c > @@ -34,9 +34,9 @@ > struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj) > { > struct radeon_bo *bo = gem_to_radeon_bo(obj); > - int npages = bo->tbo.num_pages; > > - return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages, > + bo->tbo.ttm->num_pages); > } > > struct drm_gem_object *radeon_gem_prime_import_sg_table(struct drm_device *dev, > diff --git a/drivers/gpu/drm/radeon/radeon_trace.h b/drivers/gpu/drm/radeon/radeon_trace.h > index c93f3ab3c4e3..1729cb9a95c5 100644 > --- a/drivers/gpu/drm/radeon/radeon_trace.h > +++ b/drivers/gpu/drm/radeon/radeon_trace.h > @@ -22,7 +22,7 @@ TRACE_EVENT(radeon_bo_create, > > TP_fast_assign( > __entry->bo = bo; > - __entry->pages = bo->tbo.num_pages; > + __entry->pages = bo->tbo.mem.num_pages; > ), > TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages) > ); > diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c > index a3432c6343ba..51ad5eccd844 100644 > --- a/drivers/gpu/drm/radeon/radeon_ttm.c > +++ b/drivers/gpu/drm/radeon/radeon_ttm.c > @@ -274,7 +274,7 @@ static int radeon_bo_move(struct ttm_buffer_object *bo, bool evict, > > out: > /* update statistics */ > - atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &rdev->num_bytes_moved); > + atomic64_add(bo->size, &rdev->num_bytes_moved); > radeon_bo_move_notify(bo, evict, new_mem); > return 0; > } > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > index 02cc5d247c0d..1cb326056e92 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo.c > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > @@ -72,9 +72,9 @@ static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, > struct ttm_resource_manager *man; > int i, mem_type; > > - drm_printf(&p, "No space for %p (%lu pages, %luK, %luM)\n", > - bo, bo->mem.num_pages, bo->mem.size >> 10, > - bo->mem.size >> 20); > + drm_printf(&p, "No space for %p (%lu pages, %LuK, %LuM)\n", > + bo, bo->mem.num_pages, bo->size >> 10, > + bo->size >> 20); > for (i = 0; i < placement->num_placement; i++) { > mem_type = placement->placement[i].mem_type; > drm_printf(&p, " placement[%d]=0x%08X (%d)\n", > @@ -268,7 +268,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > goto out_err; > } > > - ctx->bytes_moved += bo->num_pages << PAGE_SHIFT; > + ctx->bytes_moved += bo->size; > return 0; > > out_err: > @@ -985,8 +985,7 @@ static int ttm_bo_move_buffer(struct ttm_buffer_object *bo, > > memset(&hop, 0, sizeof(hop)); > > - mem.num_pages = bo->num_pages; > - mem.size = mem.num_pages << PAGE_SHIFT; > + mem.num_pages = PAGE_ALIGN(bo->size) >> PAGE_SHIFT; > mem.page_alignment = bo->mem.page_alignment; > mem.bus.offset = 0; > mem.bus.addr = NULL; > @@ -1102,7 +1101,7 @@ EXPORT_SYMBOL(ttm_bo_validate); > > int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > struct ttm_buffer_object *bo, > - unsigned long size, > + uint64_t size, > enum ttm_bo_type type, > struct ttm_placement *placement, > uint32_t page_alignment, > @@ -1113,9 +1112,8 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > void (*destroy) (struct ttm_buffer_object *)) > { > struct ttm_mem_global *mem_glob = &ttm_mem_glob; > - int ret = 0; > - unsigned long num_pages; > bool locked; > + int ret = 0; > > ret = ttm_mem_global_alloc(mem_glob, acc_size, ctx); > if (ret) { > @@ -1127,16 +1125,6 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > return -ENOMEM; > } > > - num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; > - if (num_pages == 0) { > - pr_err("Illegal buffer object size\n"); > - if (destroy) > - (*destroy)(bo); > - else > - kfree(bo); > - ttm_mem_global_free(mem_glob, acc_size); > - return -EINVAL; > - } > bo->destroy = destroy ? destroy : ttm_bo_default_destroy; > > kref_init(&bo->kref); > @@ -1145,10 +1133,9 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > INIT_LIST_HEAD(&bo->swap); > bo->bdev = bdev; > bo->type = type; > - bo->num_pages = num_pages; > - bo->mem.size = num_pages << PAGE_SHIFT; > + bo->size = size; > bo->mem.mem_type = TTM_PL_SYSTEM; > - bo->mem.num_pages = bo->num_pages; > + bo->mem.num_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; > bo->mem.mm_node = NULL; > bo->mem.page_alignment = page_alignment; > bo->mem.bus.offset = 0; > @@ -1210,7 +1197,7 @@ EXPORT_SYMBOL(ttm_bo_init_reserved); > > int ttm_bo_init(struct ttm_bo_device *bdev, > struct ttm_buffer_object *bo, > - unsigned long size, > + uint64_t size, > enum ttm_bo_type type, > struct ttm_placement *placement, > uint32_t page_alignment, > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c > index 5bbc1339d28e..6758e5115ae5 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > @@ -431,9 +431,9 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo, > > map->virtual = NULL; > map->bo = bo; > - if (num_pages > bo->num_pages) > + if (num_pages > bo->mem.num_pages) > return -EINVAL; > - if (start_page > bo->num_pages) > + if ((start_page + num_pages) > bo->mem.num_pages) > return -EINVAL; > > ret = ttm_mem_io_reserve(bo->bdev, &bo->mem); > @@ -485,14 +485,13 @@ int ttm_bo_vmap(struct ttm_buffer_object *bo, struct dma_buf_map *map) > > if (mem->bus.is_iomem) { > void __iomem *vaddr_iomem; > - size_t size = bo->num_pages << PAGE_SHIFT; > > if (mem->bus.addr) > vaddr_iomem = (void __iomem *)mem->bus.addr; > else if (mem->bus.caching == ttm_write_combined) > - vaddr_iomem = ioremap_wc(mem->bus.offset, size); > + vaddr_iomem = ioremap_wc(mem->bus.offset, bo->size); > else > - vaddr_iomem = ioremap(mem->bus.offset, size); > + vaddr_iomem = ioremap(mem->bus.offset, bo->size); > > if (!vaddr_iomem) > return -ENOMEM; > @@ -517,7 +516,7 @@ int ttm_bo_vmap(struct ttm_buffer_object *bo, struct dma_buf_map *map) > * or to make the buffer object look contiguous. > */ > prot = ttm_io_prot(bo, mem, PAGE_KERNEL); > - vaddr = vmap(ttm->pages, bo->num_pages, 0, prot); > + vaddr = vmap(ttm->pages, ttm->num_pages, 0, prot); > if (!vaddr) > return -ENOMEM; > > diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c > index 144a4940b6b6..6dc96cf66744 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c > @@ -198,7 +198,7 @@ static vm_fault_t ttm_bo_vm_insert_huge(struct vm_fault *vmf, > > /* Fault should not cross bo boundary. */ > page_offset &= ~(fault_page_size - 1); > - if (page_offset + fault_page_size > bo->num_pages) > + if (page_offset + fault_page_size > bo->mem.num_pages) > goto out_fallback; > > if (bo->mem.bus.is_iomem) > @@ -306,7 +306,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf, > page_last = vma_pages(vma) + vma->vm_pgoff - > drm_vma_node_start(&bo->base.vma_node); > > - if (unlikely(page_offset >= bo->num_pages)) > + if (unlikely(page_offset >= bo->mem.num_pages)) > return VM_FAULT_SIGBUS; > > prot = ttm_io_prot(bo, &bo->mem, prot); > @@ -469,7 +469,7 @@ int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, > << PAGE_SHIFT); > int ret; > > - if (len < 1 || (offset + len) >> PAGE_SHIFT > bo->num_pages) > + if (len < 1 || (offset + len) >> PAGE_SHIFT > bo->mem.num_pages) > return -EIO; > > ret = ttm_bo_reserve(bo, true, false, NULL); > diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c > index 77ba784425dd..c5c973fa0736 100644 > --- a/drivers/gpu/drm/ttm/ttm_tt.c > +++ b/drivers/gpu/drm/ttm/ttm_tt.c > @@ -129,7 +129,7 @@ static void ttm_tt_init_fields(struct ttm_tt *ttm, > uint32_t page_flags, > enum ttm_caching caching) > { > - ttm->num_pages = bo->num_pages; > + ttm->num_pages = PAGE_ALIGN(bo->size) >> PAGE_SHIFT; > ttm->caching = ttm_cached; > ttm->page_flags = page_flags; > ttm->dma_address = NULL; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_blit.c b/drivers/gpu/drm/vmwgfx/vmwgfx_blit.c > index f21881e087db..9f2779ddcf08 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_blit.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_blit.c > @@ -482,8 +482,8 @@ int vmw_bo_cpu_blit(struct ttm_buffer_object *dst, > d.src_addr = NULL; > d.dst_pages = dst->ttm->pages; > d.src_pages = src->ttm->pages; > - d.dst_num_pages = dst->num_pages; > - d.src_num_pages = src->num_pages; > + d.dst_num_pages = dst->mem.num_pages; > + d.src_num_pages = src->mem.num_pages; > d.dst_prot = ttm_io_prot(dst, &dst->mem, PAGE_KERNEL); > d.src_prot = ttm_io_prot(src, &src->mem, PAGE_KERNEL); > d.diff = diff; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c b/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c > index 263d76ae43f0..63dbc44eebe0 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_bo.c > @@ -223,7 +223,7 @@ int vmw_bo_pin_in_start_of_vram(struct vmw_private *dev_priv, > uint32_t new_flags; > > place = vmw_vram_placement.placement[0]; > - place.lpfn = bo->num_pages; > + place.lpfn = bo->mem.num_pages; > placement.num_placement = 1; > placement.placement = &place; > placement.num_busy_placement = 1; > @@ -244,7 +244,7 @@ int vmw_bo_pin_in_start_of_vram(struct vmw_private *dev_priv, > * that situation. > */ > if (bo->mem.mem_type == TTM_PL_VRAM && > - bo->mem.start < bo->num_pages && > + bo->mem.start < bo->mem.num_pages && > bo->mem.start > 0 && > buf->base.pin_count == 0) { > ctx.interruptible = false; > @@ -391,7 +391,7 @@ void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) > if (virtual) > return virtual; > > - ret = ttm_bo_kmap(bo, 0, bo->num_pages, &vbo->map); > + ret = ttm_bo_kmap(bo, 0, bo->mem.num_pages, &vbo->map); > if (ret) > DRM_ERROR("Buffer object map failed: %d.\n", ret); > > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c b/drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c > index 984d8884357d..a077e420d2ff 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c > @@ -430,7 +430,7 @@ static int vmw_cotable_resize(struct vmw_resource *res, size_t new_size) > * Do a page by page copy of COTables. This eliminates slow vmap()s. > * This should really be a TTM utility. > */ > - for (i = 0; i < old_bo->num_pages; ++i) { > + for (i = 0; i < old_bo->mem.num_pages; ++i) { > bool dummy; > > ret = ttm_bo_kmap(old_bo, i, 1, &old_map); > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c b/drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c > index e67e2e8f6e6f..72bd6d941d24 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c > @@ -1042,7 +1042,7 @@ static int vmw_query_bo_switch_prepare(struct vmw_private *dev_priv, > > if (unlikely(new_query_bo != sw_context->cur_query_bo)) { > > - if (unlikely(new_query_bo->base.num_pages > 4)) { > + if (unlikely(new_query_bo->base.mem.num_pages > 4)) { > VMW_DEBUG_USER("Query buffer too large.\n"); > return -EINVAL; > } > @@ -1541,7 +1541,7 @@ static int vmw_cmd_dma(struct vmw_private *dev_priv, > return ret; > > /* Make sure DMA doesn't cross BO boundaries. */ > - bo_size = vmw_bo->base.num_pages * PAGE_SIZE; > + bo_size = vmw_bo->base.size; > if (unlikely(cmd->body.guest.ptr.offset > bo_size)) { > VMW_DEBUG_USER("Invalid DMA offset.\n"); > return -EINVAL; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c b/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c > index 8fe26e32f920..1774960d1b89 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c > @@ -64,20 +64,19 @@ static int vmw_gmrid_man_get_node(struct ttm_resource_manager *man, > spin_lock(&gman->lock); > > if (gman->max_gmr_pages > 0) { > - gman->used_gmr_pages += bo->num_pages; > + gman->used_gmr_pages += mem->num_pages; > if (unlikely(gman->used_gmr_pages > gman->max_gmr_pages)) > goto nospace; > } > > mem->mm_node = gman; > mem->start = id; > - mem->num_pages = bo->num_pages; > > spin_unlock(&gman->lock); > return 0; > > nospace: > - gman->used_gmr_pages -= bo->num_pages; > + gman->used_gmr_pages -= mem->num_pages; > spin_unlock(&gman->lock); > ida_free(&gman->gmr_ida, id); > return -ENOSPC; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c b/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c > index bc67f2b930e1..347aff6201f6 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c > @@ -1220,7 +1220,7 @@ static int vmw_kms_new_framebuffer_bo(struct vmw_private *dev_priv, > int ret; > > requested_size = mode_cmd->height * mode_cmd->pitches[0]; > - if (unlikely(requested_size > bo->base.num_pages * PAGE_SIZE)) { > + if (unlikely(requested_size > bo->base.size)) { > DRM_ERROR("Screen buffer object size is too small " > "for requested mode.\n"); > return -EINVAL; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c b/drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c > index 0b76b3d17d4c..0a900afc66ff 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c > @@ -232,7 +232,7 @@ void vmw_bo_dirty_unmap(struct vmw_buffer_object *vbo, > int vmw_bo_dirty_add(struct vmw_buffer_object *vbo) > { > struct vmw_bo_dirty *dirty = vbo->dirty; > - pgoff_t num_pages = vbo->base.num_pages; > + pgoff_t num_pages = vbo->base.mem.num_pages; > size_t size, acc_size; > int ret; > static struct ttm_operation_ctx ctx = { > @@ -413,7 +413,7 @@ vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) > return ret; > > page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); > - if (unlikely(page_offset >= bo->num_pages)) { > + if (unlikely(page_offset >= bo->mem.num_pages)) { > ret = VM_FAULT_SIGBUS; > goto out_unlock; > } > @@ -456,7 +456,7 @@ vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) > > page_offset = vmf->pgoff - > drm_vma_node_start(&bo->base.vma_node); > - if (page_offset >= bo->num_pages || > + if (page_offset >= bo->mem.num_pages || > vmw_resources_clean(vbo, page_offset, > page_offset + PAGE_SIZE, > &allowed_prefault)) { > @@ -531,7 +531,7 @@ vm_fault_t vmw_bo_vm_huge_fault(struct vm_fault *vmf, > > page_offset = vmf->pgoff - > drm_vma_node_start(&bo->base.vma_node); > - if (page_offset >= bo->num_pages || > + if (page_offset >= bo->mem.num_pages || > vmw_resources_clean(vbo, page_offset, > page_offset + PAGE_SIZE, > &allowed_prefault)) { > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c b/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c > index 00b535831a7a..84ed8376cd81 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_resource.c > @@ -360,7 +360,7 @@ static int vmw_resource_buf_alloc(struct vmw_resource *res, > int ret; > > if (likely(res->backup)) { > - BUG_ON(res->backup->base.num_pages * PAGE_SIZE < size); > + BUG_ON(res->backup->base.size < size); > return 0; > } > > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_shader.c b/drivers/gpu/drm/vmwgfx/vmwgfx_shader.c > index f328aa5839a2..c7220a770bc5 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_shader.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_shader.c > @@ -856,8 +856,7 @@ static int vmw_shader_define(struct drm_device *dev, struct drm_file *file_priv, > return ret; > } > > - if ((u64)buffer->base.num_pages * PAGE_SIZE < > - (u64)size + (u64)offset) { > + if ((u64)buffer->base.size < (u64)size + (u64)offset) { > VMW_DEBUG_USER("Illegal buffer- or shader size.\n"); > ret = -EINVAL; > goto out_bad_arg; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c b/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c > index 5b04ec047ef3..362a127b24bb 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c > @@ -512,7 +512,7 @@ static void vmw_stdu_bo_fifo_commit(struct vmw_kms_dirty *dirty) > cmd->body.host.mipmap = 0; > cmd->body.transfer = ddirty->transfer; > suffix->suffixSize = sizeof(*suffix); > - suffix->maximumOffset = ddirty->buf->base.num_pages * PAGE_SIZE; > + suffix->maximumOffset = ddirty->buf->base.size; > > if (ddirty->transfer == SVGA3D_WRITE_HOST_VRAM) { > blit_size += sizeof(struct vmw_stdu_update); > @@ -1238,7 +1238,7 @@ static uint32_t vmw_stdu_bo_populate_update(struct vmw_du_update_plane *update, > vfbbo = container_of(update->vfb, typeof(*vfbbo), base); > > suffix->suffixSize = sizeof(*suffix); > - suffix->maximumOffset = vfbbo->buffer->base.num_pages * PAGE_SIZE; > + suffix->maximumOffset = vfbbo->buffer->base.size; > > vmw_stdu_populate_update(&suffix[1], stdu->base.unit, bb->x1, bb->x2, > bb->y1, bb->y2); > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > index 3914bfee0533..6a02bad8416b 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > @@ -1550,8 +1550,7 @@ vmw_gb_surface_define_internal(struct drm_device *dev, > &res->backup, > &user_srf->backup_base); > if (ret == 0) { > - if (res->backup->base.num_pages * PAGE_SIZE < > - res->backup_size) { > + if (res->backup->base.size < res->backup_size) { > VMW_DEBUG_USER("Surface backup buffer too small.\n"); > vmw_bo_unreference(&res->backup); > ret = -EINVAL; > @@ -1614,7 +1613,7 @@ vmw_gb_surface_define_internal(struct drm_device *dev, > if (res->backup) { > rep->buffer_map_handle = > drm_vma_node_offset_addr(&res->backup->base.base.vma_node); > - rep->buffer_size = res->backup->base.num_pages * PAGE_SIZE; > + rep->buffer_size = res->backup->base.size; > rep->buffer_handle = backup_handle; > } else { > rep->buffer_map_handle = 0; > @@ -1692,7 +1691,7 @@ vmw_gb_surface_reference_internal(struct drm_device *dev, > rep->crep.buffer_handle = backup_handle; > rep->crep.buffer_map_handle = > drm_vma_node_offset_addr(&srf->res.backup->base.base.vma_node); > - rep->crep.buffer_size = srf->res.backup->base.num_pages * PAGE_SIZE; > + rep->crep.buffer_size = srf->res.backup->base.size; > > rep->creq.version = drm_vmw_gb_surface_v1; > rep->creq.svga3d_flags_upper_32_bits = > diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h > index 79b9367e0ffd..03675ce7ab3c 100644 > --- a/include/drm/ttm/ttm_bo_api.h > +++ b/include/drm/ttm/ttm_bo_api.h > @@ -125,7 +125,7 @@ struct ttm_buffer_object { > struct ttm_bo_device *bdev; > enum ttm_bo_type type; > void (*destroy) (struct ttm_buffer_object *); > - unsigned long num_pages; > + uint64_t size; > size_t acc_size; > > /** > @@ -397,7 +397,7 @@ size_t ttm_bo_dma_acc_size(struct ttm_bo_device *bdev, > > int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > struct ttm_buffer_object *bo, > - unsigned long size, > + uint64_t size, > enum ttm_bo_type type, > struct ttm_placement *placement, > uint32_t page_alignment, > @@ -445,7 +445,7 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev, > * -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources. > */ > int ttm_bo_init(struct ttm_bo_device *bdev, struct ttm_buffer_object *bo, > - unsigned long size, enum ttm_bo_type type, > + uint64_t size, enum ttm_bo_type type, > struct ttm_placement *placement, > uint32_t page_alignment, bool interrubtible, size_t acc_size, > struct sg_table *sg, struct dma_resv *resv, > diff --git a/include/drm/ttm/ttm_resource.h b/include/drm/ttm/ttm_resource.h > index f48a70d39ac5..ad6da99770e9 100644 > --- a/include/drm/ttm/ttm_resource.h > +++ b/include/drm/ttm/ttm_resource.h > @@ -171,7 +171,6 @@ struct ttm_bus_placement { > struct ttm_resource { > void *mm_node; > unsigned long start; > - unsigned long size; > unsigned long num_pages; > uint32_t page_alignment; > uint32_t mem_type; > -- > 2.25.1 > > _______________________________________________ > dri-devel mailing list > dri-devel@xxxxxxxxxxxxxxxxxxxxx > https://lists.freedesktop.org/mailman/listinfo/dri-devel -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel