Am 13.11.24 um 11:25 schrieb Tvrtko Ursulin:
On 13/11/2024 08:49, Christian König wrote:
Am 12.11.24 um 19:16 schrieb Li, Yunxiang (Teddy):
[SNIP]
+ size = sign * amdgpu_bo_size(bo);
+ res = bo->tbo.resource;
+ type = res ? res->mem_type :
amdgpu_bo_get_preferred_placement(bo);
Again, it's a clear NAK from my side to do stuff like that.
When there isn't any backing store the BO should *not* be accounted
to anything.
I don't have a preference either way, but I think it should be a
separate discussion to properly define what drm-total- means.
Total must show the total size of all BOs which exist even if they
don't currently have a backing store. That's how drm-usage-stats.rst
defines the field and that is how all the other drivers work.
In that case we should only look at the preferred placement and not the
backing store at all.
But that makes the total identical to the requested value, doesn't it?
+ type = res ? res->mem_type :
amdgpu_bo_get_preferred_placement(bo);
+ shared =
drm_gem_object_is_shared_for_memory_stats(&bo->tbo.base);
+
+ if (type >= __AMDGPU_PL_LAST)
+ return;
+
+ spin_lock(&vm->status_lock);
+
+ if (shared)
+ vm->stats[type].drm.shared += size;
+ else
+ vm->stats[type].drm.private += size;
+ if (res)
+ vm->stats[type].drm.resident += size;
+ if (bo->flags & AMDGPU_GEM_CREATE_DISCARDABLE)
+ vm->stats[type].drm.purgeable += size;
+
+ if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
+ vm->stats[TTM_PL_VRAM].requested += size;
+ if (type != TTM_PL_VRAM)
+ vm->stats[TTM_PL_VRAM].evicted += size;
Again that is incorrect. BOs can be created with VRAM|GTT as their
placement.
If such a BO is placed into GTT that doesn't mean it is evicted.
In that case, do we count BO with VRAM|GTT in both VRAM and GTT's
.requested field? and if they are not in either, they go in both
.evicted field?
Oh, good question depends on the definition of the requested field.
Accounting it to VRAM.evicted while GTT placement is desirable as
well is certainly not correct.
From my understanding they should go into the VRAM request, but not
account to evicted. But Tvrtko might see that differently.
Semantics of requested and evicted are kind of amdgpu 'legacy' thing.
So the question is whether or not they should keep matching.
Originally they were like this (I will edit out parts which deal with
CPU visible for ease of comparison, and which have since been removed
anyway):
if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
stats->requested_vram += size;
if (res->mem_type != TTM_PL_VRAM)
stats->evicted_vram += size;
} else if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_GTT) {
stats->requested_gtt += size;
}
So the part about accounting as evicted with dual preferred placement
was there from the start.
Then after my changes:
if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
stats[TTM_PL_VRAM].requested += size;
if (type != TTM_PL_VRAM) {
stats[TTM_PL_VRAM].evicted += size;
}
} else if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_GTT) {
stats[TTM_PL_TT].requested += size;
}
I mostly kept the same semantics.
Teddy's version keeps it the same:
if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
vm->stats[TTM_PL_VRAM].requested += size;
if (type != TTM_PL_VRAM)
vm->stats[TTM_PL_VRAM].evicted += size;
} else if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_GTT) {
vm->stats[TTM_PL_TT].requested += size;
}
If no AMD tools depend on the legacy semantics for evicted/requested
we can change them. There is some overlap with the standard keys
anyway and the fact preferred mask is unordered made the original
behaviour a bit presumptuous to begin with. In summary I think it
depends on whether we need to keep the legacy semantics, or even the
keys themselves.
As far as I know we don't have any dependency on the amdgpu specific keys.
But I need to double check what drm-usage-stats.rst actually defines.
Maybe that doesn't really match what we have in amdgpu and other TTM
drivers as information.
Thanks,
Christian.
Regards,
Tvrtko
@@ -2612,7 +2707,6 @@ void amdgpu_vm_fini(struct amdgpu_device *adev,
struct amdgpu_vm *vm)
root = amdgpu_bo_ref(vm->root.bo);
amdgpu_bo_reserve(root, true);
- amdgpu_vm_put_task_info(vm->task_info);
amdgpu_vm_set_pasid(adev, vm, 0);
dma_fence_wait(vm->last_unlocked, false);
dma_fence_put(vm->last_unlocked);
@@ -2660,6 +2754,15 @@ void amdgpu_vm_fini(struct amdgpu_device
*adev,
struct amdgpu_vm *vm)
}
}
+ if (!amdgpu_vm_stats_is_zero(vm)) {
+ struct amdgpu_task_info *ti = vm->task_info;
+
+ dev_warn(adev->dev,
+ "VM memory stats for proc %s(%d) task %s(%d)
is non-zero
when fini\n",
+ ti->process_name, ti->pid, ti->task_name, ti->tgid);
+ }
+
+ amdgpu_vm_put_task_info(vm->task_info);
Please don't move the call to amdgpu_vm_put_task_info().
Is keeping the task_info alive a hazard here? I could copy out the
info, it just seemed a bit wasteful.
Ah, now I see why you have moved that.
IIRC we need to free up the task info before releasing the PASID, but
that info might be outdated. Need to check the code.
Does it work if you move the message further up or does the root PD
then break your neck because it isn't released yet?
Thanks,
Christian.
Regards,
Teddy