Re: [PATCH] dma-buf: Eliminate all duplicate fences in dma_fence_unwrap_merge

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

On 18.10.24 10:56, Christian König wrote:
Am 18.10.24 um 07:46 schrieb Friedrich Vock:
When dma_fence_unwrap_merge is called on fence chains where the fences
aren't ordered by context, the merging logic breaks down and we end up
inserting fences twice. Doing this repeatedly leads to the number of
fences going up exponentially, and in some gaming workloads we'll end up
running out of memory to store the resulting array altogether, leading
to a warning such as:

Ah! I was searching for that one for quite a while now.

I own you a beer should you ever be near Cologne.

I visit the area somewhat regularly, I'll let you know the next time
around ;-)


Please also see my patch on the mailing list to use kvzalloc() to
mitigate this.


vkd3d_queue: page allocation failure: order:7,
mode:0x40dc0(GFP_KERNEL|__GFP_COMP|__GFP_ZERO),
nodemask=(null),cpuset=/,mems_allowed=0
CPU: 2 PID: 5287 Comm: vkd3d_queue Tainted: G S
6.10.7-200.fsync.fc40.x86_64 #1
Hardware name: Dell Inc. G5 5505/0NCW8W, BIOS 1.11.0 03/22/2022
Call Trace:
  <TASK>
  dump_stack_lvl+0x5d/0x80
  warn_alloc+0x164/0x190
  ? srso_return_thunk+0x5/0x5f
  ? __alloc_pages_direct_compact+0x1d9/0x220
  __alloc_pages_slowpath.constprop.2+0xd14/0xd80
  __alloc_pages_noprof+0x32b/0x350
  ? dma_fence_array_create+0x48/0x110
  __kmalloc_large_node+0x6f/0x130
  __kmalloc_noprof+0x2dd/0x4a0
  ? dma_fence_array_create+0x48/0x110
  dma_fence_array_create+0x48/0x110
  __dma_fence_unwrap_merge+0x481/0x5b0
  sync_file_merge.constprop.0+0xf8/0x180
  sync_file_ioctl+0x476/0x590
  ? srso_return_thunk+0x5/0x5f
  ? __seccomp_filter+0xe8/0x5a0
  __x64_sys_ioctl+0x97/0xd0
  do_syscall_64+0x82/0x160
  ? srso_return_thunk+0x5/0x5f
  ? drm_syncobj_destroy_ioctl+0x8b/0xb0
  ? srso_return_thunk+0x5/0x5f
  ? srso_return_thunk+0x5/0x5f
  ? __check_object_size+0x58/0x230
  ? srso_return_thunk+0x5/0x5f
  ? srso_return_thunk+0x5/0x5f
  ? drm_ioctl+0x2ba/0x530
  ? __pfx_drm_syncobj_destroy_ioctl+0x10/0x10
  ? srso_return_thunk+0x5/0x5f
  ? ktime_get_mono_fast_ns+0x3b/0xd0
  ? srso_return_thunk+0x5/0x5f
  ? amdgpu_drm_ioctl+0x71/0x90 [amdgpu]
  ? srso_return_thunk+0x5/0x5f
  ? syscall_exit_to_user_mode+0x72/0x200
  ? srso_return_thunk+0x5/0x5f
  ? do_syscall_64+0x8e/0x160
  ? syscall_exit_to_user_mode+0x72/0x200
  ? srso_return_thunk+0x5/0x5f
  ? do_syscall_64+0x8e/0x160
  ? srso_return_thunk+0x5/0x5f
  ? syscall_exit_to_user_mode+0x72/0x200
  ? srso_return_thunk+0x5/0x5f
  ? do_syscall_64+0x8e/0x160
  ? do_syscall_64+0x8e/0x160
  ? srso_return_thunk+0x5/0x5f
  entry_SYSCALL_64_after_hwframe+0x76/0x7e

It's a bit unfortunate that we end up with quadratic complexity w.r.t.
the number of merged fences in all cases, but I'd argue in practice
there shouldn't be more than a handful of in-flight fences to merge.

Link: https://gitlab.freedesktop.org/drm/amd/-/issues/3617
Signed-off-by: Friedrich Vock <friedrich.vock@xxxxxx>
---
  drivers/dma-buf/dma-fence-unwrap.c | 13 +++++++++++--
  1 file changed, 11 insertions(+), 2 deletions(-)

diff --git a/drivers/dma-buf/dma-fence-unwrap.c b/drivers/dma-buf/dma-
fence-unwrap.c
index 628af51c81af..46277cef0bc6 100644
--- a/drivers/dma-buf/dma-fence-unwrap.c
+++ b/drivers/dma-buf/dma-fence-unwrap.c
@@ -68,7 +68,7 @@ struct dma_fence *__dma_fence_unwrap_merge(unsigned
int num_fences,
      struct dma_fence *tmp, **array;
      ktime_t timestamp;
      unsigned int i;
-    size_t count;
+    size_t count, j;

      count = 0;
      timestamp = ns_to_ktime(0);
@@ -127,6 +127,10 @@ struct dma_fence
*__dma_fence_unwrap_merge(unsigned int num_fences,
               * function is used multiple times. So attempt to order
               * the fences by context as we pass over them and merge
               * fences with the same context.
+             *
+             * We will remove any remaining duplicate fences down
+             * below, but doing this here saves us from having to
+             * iterate over the array to detect the duplicate.
               */
              if (!tmp || tmp->context > next->context) {
                  tmp = next;
@@ -145,7 +149,12 @@ struct dma_fence
*__dma_fence_unwrap_merge(unsigned int num_fences,
          }

          if (tmp) {
-            array[count++] = dma_fence_get(tmp);
+            for (j = 0; j < count; ++j) {
+                if (array[count] == tmp)
+                    break;
+            }
+            if (j == count)
+                array[count++] = dma_fence_get(tmp);

That is clearly not the right solution. Since comparing the context
should have already removed all duplicates.

Sadly, not. This is true as long as the fences are ordered by context,
but this is not a given. The error manifests precisely when they are not
ordered.

Imagine we try to merge two chains/arrays that contain the same 4 fences
(I'll call them fences 1-4), but the second one has another fence with a
higher context (fence 5) in front of it.

I'll try to visualize the chains/arrays as ASCII art - we start out in a
state like this (the ^ marker referring to the current iterator position):

Result: <empty>

1. 1 -> 2 -> 3 -> 4
   ^
2. 5 -> 1 -> 2 -> 3 -> 4
   ^

Since fences 1-4 have a lower context than fence 5, we'll insert fences
1-4 first:

Result: 1 -> 2 -> 3 -> 4

1. 1 -> 2 -> 3 -> 4
                     ^
2. 5 -> 1 -> 2 -> 3 -> 4
   ^

Only then will we insert fence 5:

Result: 1 -> 2 -> 3 -> 4 -> 5

1. 1 -> 2 -> 3 -> 4
                     ^
2. 5 -> 1 -> 2 -> 3 -> 4
        ^

Now we're at fence 1 in the second chain/array, but we've already
inserted fences 1-4 from the first chain/array, so we fail to merge
them. Instead, they get reinserted:

Result: 1 -> 2 -> 3 -> 4 -> 5 -> 1 -> 2 -> 3 -> 4

1. 1 -> 2 -> 3 -> 4
                     ^
2. 5 -> 1 -> 2 -> 3 -> 4
                         ^

We can't assume the fences are in any sort of order w.r.t their context,
so if we want to check for duplicates exhaustively we'll always end up
with some kind of O(n^2) algorithm. I see only a handful of ways we can go:

a) Don't check exhaustively (current behavior). Obviously, this doesn't
work that well in practice.

b) Eat the O(n^2) cost (this patch). I've kept the current merging code
since it's an easy way to reduce the amount of times we have to do the
expensive duplicate check, but other than that I'm not sure we can do
much to reduce cost.

c) Enforce order w.r.t. context. I don't think we can require that fence
chains order their fences by context, they should be ordered by timeline
point (maybe it would work for arrays, but whatever). That leaves us
with having to sort the fences by context just before merging. That
would reduce complexity to some O(n log n) at worst, but in practice I
fear it might not be worth it compared to just iterating over the result
array a few times, especially given that once this bug is fixed, we
should be back to only a few fences to merge :)

Regards,
Friedrich


Going to double check the code.

Thanks,
Christian.

              fences[sel] = dma_fence_unwrap_next(&iter[sel]);
          }
      } while (tmp);
--
2.47.0







[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux