I can't think of an use case when we don't want this to succeed. That is true. seems I can simplify more here. Thanks Roger(Hongbo.He) -----Original Message----- From: Koenig, Christian Sent: Thursday, February 08, 2018 8:58 PM To: He, Roger <Hongbo.He@xxxxxxx>; amd-gfx@xxxxxxxxxxxxxxxxxxxxx; dri-devel@xxxxxxxxxxxxxxxxxxxxx Subject: Re: [PATCH 3/4] drm/ttm: add input parameter force_alloc for ttm_bo_evict_mm Am 08.02.2018 um 10:06 schrieb Roger He: > if true, allocate TTM pages regardless of zone global memory account > limit. For suspend, We should avoid TTM memory allocate failure then > result in suspend failure. Why the extra parameter for amdgpu_bo_evict_vram ? I can't think of an use case when we don't want this to succeed. Christian. > > Signed-off-by: Roger He <Hongbo.He@xxxxxxx> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 4 ++-- > drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | 4 ++-- > drivers/gpu/drm/amd/amdgpu/amdgpu_object.h | 2 +- > drivers/gpu/drm/nouveau/nouveau_drm.c | 2 +- > drivers/gpu/drm/qxl/qxl_object.c | 4 ++-- > drivers/gpu/drm/radeon/radeon_device.c | 6 +++--- > drivers/gpu/drm/radeon/radeon_object.c | 5 +++-- > drivers/gpu/drm/radeon/radeon_object.h | 3 ++- > drivers/gpu/drm/ttm/ttm_bo.c | 16 ++++++++++------ > drivers/gpu/drm/vmwgfx/vmwgfx_drv.c | 6 +++--- > include/drm/ttm/ttm_bo_api.h | 5 ++++- > 12 files changed, 34 insertions(+), 25 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c > b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c > index ee76b46..59ee12c 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c > @@ -763,7 +763,7 @@ static int amdgpu_debugfs_evict_vram(struct seq_file *m, void *data) > struct drm_device *dev = node->minor->dev; > struct amdgpu_device *adev = dev->dev_private; > > - seq_printf(m, "(%d)\n", amdgpu_bo_evict_vram(adev)); > + seq_printf(m, "(%d)\n", amdgpu_bo_evict_vram(adev, true)); > return 0; > } > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > index e3fa3d7..3c5f9ca 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > @@ -2168,7 +2168,7 @@ int amdgpu_device_suspend(struct drm_device *dev, bool suspend, bool fbcon) > } > } > /* evict vram memory */ > - amdgpu_bo_evict_vram(adev); > + amdgpu_bo_evict_vram(adev, true); > > amdgpu_fence_driver_suspend(adev); > > @@ -2178,7 +2178,7 @@ int amdgpu_device_suspend(struct drm_device *dev, bool suspend, bool fbcon) > * This second call to evict vram is to evict the gart page table > * using the CPU. > */ > - amdgpu_bo_evict_vram(adev); > + amdgpu_bo_evict_vram(adev, true); > > pci_save_state(dev->pdev); > if (suspend) { > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > index 0338ef6..db813f9 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c > @@ -803,14 +803,14 @@ int amdgpu_bo_unpin(struct amdgpu_bo *bo) > return r; > } > > -int amdgpu_bo_evict_vram(struct amdgpu_device *adev) > +int amdgpu_bo_evict_vram(struct amdgpu_device *adev, bool > +force_alloc) > { > /* late 2.6.33 fix IGP hibernate - we need pm ops to do this correct */ > if (0 && (adev->flags & AMD_IS_APU)) { > /* Useless to evict on IGP chips */ > return 0; > } > - return ttm_bo_evict_mm(&adev->mman.bdev, TTM_PL_VRAM); > + return ttm_bo_evict_mm(&adev->mman.bdev, TTM_PL_VRAM, force_alloc); > } > > static const char *amdgpu_vram_names[] = { diff --git > a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > index c2b02f5..6724cdc 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h > @@ -227,7 +227,7 @@ int amdgpu_bo_pin_restricted(struct amdgpu_bo *bo, u32 domain, > u64 min_offset, u64 max_offset, > u64 *gpu_addr); > int amdgpu_bo_unpin(struct amdgpu_bo *bo); -int > amdgpu_bo_evict_vram(struct amdgpu_device *adev); > +int amdgpu_bo_evict_vram(struct amdgpu_device *adev, bool > +force_alloc); > int amdgpu_bo_init(struct amdgpu_device *adev); > void amdgpu_bo_fini(struct amdgpu_device *adev); > int amdgpu_bo_fbdev_mmap(struct amdgpu_bo *bo, diff --git > a/drivers/gpu/drm/nouveau/nouveau_drm.c > b/drivers/gpu/drm/nouveau/nouveau_drm.c > index 8d4a5be..c9627ef 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_drm.c > +++ b/drivers/gpu/drm/nouveau/nouveau_drm.c > @@ -702,7 +702,7 @@ nouveau_do_suspend(struct drm_device *dev, bool runtime) > } > > NV_DEBUG(drm, "evicting buffers...\n"); > - ttm_bo_evict_mm(&drm->ttm.bdev, TTM_PL_VRAM); > + ttm_bo_evict_mm(&drm->ttm.bdev, TTM_PL_VRAM, true); > > NV_DEBUG(drm, "waiting for kernel channels to go idle...\n"); > if (drm->cechan) { > diff --git a/drivers/gpu/drm/qxl/qxl_object.c > b/drivers/gpu/drm/qxl/qxl_object.c > index f6b80fe..d8d26c8 100644 > --- a/drivers/gpu/drm/qxl/qxl_object.c > +++ b/drivers/gpu/drm/qxl/qxl_object.c > @@ -350,10 +350,10 @@ int qxl_bo_check_id(struct qxl_device *qdev, > struct qxl_bo *bo) > > int qxl_surf_evict(struct qxl_device *qdev) > { > - return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_PRIV); > + return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_PRIV, true); > } > > int qxl_vram_evict(struct qxl_device *qdev) > { > - return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_VRAM); > + return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_VRAM, true); > } > diff --git a/drivers/gpu/drm/radeon/radeon_device.c > b/drivers/gpu/drm/radeon/radeon_device.c > index 8d3e3d2..c11ee06 100644 > --- a/drivers/gpu/drm/radeon/radeon_device.c > +++ b/drivers/gpu/drm/radeon/radeon_device.c > @@ -1522,7 +1522,7 @@ void radeon_device_fini(struct radeon_device *rdev) > DRM_INFO("radeon: finishing device.\n"); > rdev->shutdown = true; > /* evict vram memory */ > - radeon_bo_evict_vram(rdev); > + radeon_bo_evict_vram(rdev, true); > radeon_fini(rdev); > if (!pci_is_thunderbolt_attached(rdev->pdev)) > vga_switcheroo_unregister_client(rdev->pdev); > @@ -1607,7 +1607,7 @@ int radeon_suspend_kms(struct drm_device *dev, bool suspend, > } > } > /* evict vram memory */ > - radeon_bo_evict_vram(rdev); > + radeon_bo_evict_vram(rdev, true); > > /* wait for gpu to finish processing current batch */ > for (i = 0; i < RADEON_NUM_RINGS; i++) { @@ -1626,7 +1626,7 @@ int > radeon_suspend_kms(struct drm_device *dev, bool suspend, > * This second call to evict vram is to evict the gart page table > * using the CPU. > */ > - radeon_bo_evict_vram(rdev); > + radeon_bo_evict_vram(rdev, true); > > radeon_agp_suspend(rdev); > > diff --git a/drivers/gpu/drm/radeon/radeon_object.c > b/drivers/gpu/drm/radeon/radeon_object.c > index 15404af..99a9a45 100644 > --- a/drivers/gpu/drm/radeon/radeon_object.c > +++ b/drivers/gpu/drm/radeon/radeon_object.c > @@ -420,7 +420,8 @@ int radeon_bo_unpin(struct radeon_bo *bo) > return r; > } > > -int radeon_bo_evict_vram(struct radeon_device *rdev) > +int > +radeon_bo_evict_vram(struct radeon_device *rdev, bool force_alloc) > { > /* late 2.6.33 fix IGP hibernate - we need pm ops to do this correct */ > if (0 && (rdev->flags & RADEON_IS_IGP)) { @@ -428,7 +429,7 @@ int > radeon_bo_evict_vram(struct radeon_device *rdev) > /* Useless to evict on IGP chips */ > return 0; > } > - return ttm_bo_evict_mm(&rdev->mman.bdev, TTM_PL_VRAM); > + return ttm_bo_evict_mm(&rdev->mman.bdev, TTM_PL_VRAM, force_alloc); > } > > void radeon_bo_force_delete(struct radeon_device *rdev) diff --git > a/drivers/gpu/drm/radeon/radeon_object.h > b/drivers/gpu/drm/radeon/radeon_object.h > index 9ffd821..757ba88 100644 > --- a/drivers/gpu/drm/radeon/radeon_object.h > +++ b/drivers/gpu/drm/radeon/radeon_object.h > @@ -136,7 +136,8 @@ extern int radeon_bo_pin(struct radeon_bo *bo, u32 domain, u64 *gpu_addr); > extern int radeon_bo_pin_restricted(struct radeon_bo *bo, u32 domain, > u64 max_offset, u64 *gpu_addr); > extern int radeon_bo_unpin(struct radeon_bo *bo); -extern int > radeon_bo_evict_vram(struct radeon_device *rdev); > +extern int radeon_bo_evict_vram(struct radeon_device *rdev, > + bool force_alloc); > extern void radeon_bo_force_delete(struct radeon_device *rdev); > extern int radeon_bo_init(struct radeon_device *rdev); > extern void radeon_bo_fini(struct radeon_device *rdev); diff --git > a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c index > a907311..31d10f1 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo.c > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > @@ -1342,15 +1342,17 @@ int ttm_bo_create(struct ttm_bo_device *bdev, > EXPORT_SYMBOL(ttm_bo_create); > > static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev, > - unsigned mem_type) > + unsigned mem_type, bool force_alloc) > { > - struct ttm_operation_ctx ctx = { false, false }; > + struct ttm_operation_ctx ttm_opt_ctx = { false, false }; > struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > struct ttm_bo_global *glob = bdev->glob; > struct dma_fence *fence; > int ret; > unsigned i; > > + if (force_alloc) > + ttm_opt_ctx.flags = TTM_OPT_FLAG_FORCE_ALLOC; > /* > * Can't use standard list traversal since we're unlocking. > */ > @@ -1359,7 +1361,8 @@ static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev, > for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) { > while (!list_empty(&man->lru[i])) { > spin_unlock(&glob->lru_lock); > - ret = ttm_mem_evict_first(bdev, mem_type, NULL, &ctx); > + ret = ttm_mem_evict_first(bdev, mem_type, NULL, > + &ttm_opt_ctx); > if (ret) > return ret; > spin_lock(&glob->lru_lock); > @@ -1403,7 +1406,7 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, > unsigned mem_type) > > ret = 0; > if (mem_type > 0) { > - ret = ttm_bo_force_list_clean(bdev, mem_type); > + ret = ttm_bo_force_list_clean(bdev, mem_type, true); > if (ret) { > pr_err("Cleanup eviction failed\n"); > return ret; > @@ -1419,7 +1422,8 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type) > } > EXPORT_SYMBOL(ttm_bo_clean_mm); > > -int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) > +int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type, > + bool force_alloc) > { > struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > > @@ -1433,7 +1437,7 @@ int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) > return 0; > } > > - return ttm_bo_force_list_clean(bdev, mem_type); > + return ttm_bo_force_list_clean(bdev, mem_type, force_alloc); > } > EXPORT_SYMBOL(ttm_bo_evict_mm); > > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c > b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c > index 184340d..28f8e4f 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c > @@ -430,7 +430,7 @@ static int vmw_request_device(struct vmw_private *dev_priv) > if (dev_priv->cman) > vmw_cmdbuf_remove_pool(dev_priv->cman); > if (dev_priv->has_mob) { > - (void) ttm_bo_evict_mm(&dev_priv->bdev, VMW_PL_MOB); > + (void) ttm_bo_evict_mm(&dev_priv->bdev, VMW_PL_MOB, true); > vmw_otables_takedown(dev_priv); > } > if (dev_priv->cman) > @@ -463,7 +463,7 @@ static void vmw_release_device_early(struct vmw_private *dev_priv) > vmw_cmdbuf_remove_pool(dev_priv->cman); > > if (dev_priv->has_mob) { > - ttm_bo_evict_mm(&dev_priv->bdev, VMW_PL_MOB); > + ttm_bo_evict_mm(&dev_priv->bdev, VMW_PL_MOB, true); > vmw_otables_takedown(dev_priv); > } > } > @@ -1342,7 +1342,7 @@ void vmw_svga_disable(struct vmw_private *dev_priv) > if (dev_priv->bdev.man[TTM_PL_VRAM].use_type) { > dev_priv->bdev.man[TTM_PL_VRAM].use_type = false; > spin_unlock(&dev_priv->svga_lock); > - if (ttm_bo_evict_mm(&dev_priv->bdev, TTM_PL_VRAM)) > + if (ttm_bo_evict_mm(&dev_priv->bdev, TTM_PL_VRAM, true)) > DRM_ERROR("Failed evicting VRAM buffers.\n"); > vmw_write(dev_priv, SVGA_REG_ENABLE, > SVGA_REG_ENABLE_HIDE | > diff --git a/include/drm/ttm/ttm_bo_api.h > b/include/drm/ttm/ttm_bo_api.h index 2142639..6b5db9c 100644 > --- a/include/drm/ttm/ttm_bo_api.h > +++ b/include/drm/ttm/ttm_bo_api.h > @@ -636,6 +636,8 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type); > * > * @bdev: Pointer to a ttm_bo_device struct. > * @mem_type: The memory type. > + * @force_alloc: if true allow ttm pages allocation always > + * regardless of zone memory account limit > * > * Evicts all buffers on the lru list of the memory type. > * This is normally part of a VT switch or an @@ -649,7 +651,8 @@ > int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type); > * -ERESTARTSYS: The call was interrupted by a signal while waiting to > * evict a buffer. > */ > -int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type); > +int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type, > + bool force_alloc); > > /** > * ttm_kmap_obj_virtual _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel