Re: [PATCH v6 2/2] drm/i915/gem: Don't try to map and fence large scanout buffers (v9)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 11/03/2022 09:39, Daniel Vetter wrote:
On Mon, 7 Mar 2022 at 21:38, Vivek Kasireddy <vivek.kasireddy@xxxxxxxxx> wrote:

On platforms capable of allowing 8K (7680 x 4320) modes, pinning 2 or
more framebuffers/scanout buffers results in only one that is mappable/
fenceable. Therefore, pageflipping between these 2 FBs where only one
is mappable/fenceable creates latencies large enough to miss alternate
vblanks thereby producing less optimal framerate.

This mainly happens because when i915_gem_object_pin_to_display_plane()
is called to pin one of the FB objs, the associated vma is identified
as misplaced and therefore i915_vma_unbind() is called which unbinds and
evicts it. This misplaced vma gets subseqently pinned only when
i915_gem_object_ggtt_pin_ww() is called without PIN_MAPPABLE. This
results in a latency of ~10ms and happens every other vblank/repaint cycle.
Therefore, to fix this issue, we try to see if there is space to map
at-least two objects of a given size and return early if there isn't. This
would ensure that we do not try with PIN_MAPPABLE for any objects that
are too big to map thereby preventing unncessary unbind.

Testcase:
Running Weston and weston-simple-egl on an Alderlake_S (ADLS) platform
with a 8K@60 mode results in only ~40 FPS. Since upstream Weston submits
a frame ~7ms before the next vblank, the latencies seen between atomic
commit and flip event are 7, 24 (7 + 16.66), 7, 24..... suggesting that
it misses the vblank every other frame.

Here is the ftrace snippet that shows the source of the ~10ms latency:
               i915_gem_object_pin_to_display_plane() {
0.102 us   |    i915_gem_object_set_cache_level();
                 i915_gem_object_ggtt_pin_ww() {
0.390 us   |      i915_vma_instance();
0.178 us   |      i915_vma_misplaced();
                   i915_vma_unbind() {
                   __i915_active_wait() {
0.082 us   |        i915_active_acquire_if_busy();
0.475 us   |      }
                   intel_runtime_pm_get() {
0.087 us   |        intel_runtime_pm_acquire();
0.259 us   |      }
                   __i915_active_wait() {
0.085 us   |        i915_active_acquire_if_busy();
0.240 us   |      }
                   __i915_vma_evict() {
                     ggtt_unbind_vma() {
                       gen8_ggtt_clear_range() {
10507.255 us |        }
10507.689 us |      }
10508.516 us |   }

v2: Instead of using bigjoiner checks, determine whether a scanout
     buffer is too big by checking to see if it is possible to map
     two of them into the ggtt.

v3 (Ville):
- Count how many fb objects can be fit into the available holes
   instead of checking for a hole twice the object size.
- Take alignment constraints into account.
- Limit this large scanout buffer check to >= Gen 11 platforms.

v4:
- Remove existing heuristic that checks just for size. (Ville)
- Return early if we find space to map at-least two objects. (Tvrtko)
- Slightly update the commit message.

v5: (Tvrtko)
- Rename the function to indicate that the object may be too big to
   map into the aperture.
- Account for guard pages while calculating the total size required
   for the object.
- Do not subject all objects to the heuristic check and instead
   consider objects only of a certain size.
- Do the hole walk using the rbtree.
- Preserve the existing PIN_NONBLOCK logic.
- Drop the PIN_MAPPABLE check while pinning the VMA.

v6: (Tvrtko)
- Return 0 on success and the specific error code on failure to
   preserve the existing behavior.

v7: (Ville)
- Drop the HAS_GMCH(i915), DISPLAY_VER(i915) < 11 and
   size < ggtt->mappable_end / 4 checks.
- Drop the redundant check that is based on previous heuristic.

v8:
- Make sure that we are holding the mutex associated with ggtt vm
   as we traverse the hole nodes.

v9: (Tvrtko)
- Use mutex_lock_interruptible_nested() instead of mutex_lock().

Cc: Ville Syrjälä <ville.syrjala@xxxxxxxxxxxxxxx>
Cc: Maarten Lankhorst <maarten.lankhorst@xxxxxxxxxxxxxxx>
Cc: Tvrtko Ursulin <tvrtko.ursulin@xxxxxxxxxxxxxxx>
Cc: Manasi Navare <manasi.d.navare@xxxxxxxxx>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@xxxxxxxxx>
Signed-off-by: Vivek Kasireddy <vivek.kasireddy@xxxxxxxxx>
---
  drivers/gpu/drm/i915/i915_gem.c | 128 +++++++++++++++++++++++---------
  1 file changed, 94 insertions(+), 34 deletions(-)

diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 9747924cc57b..e0d731b3f215 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -49,6 +49,7 @@
  #include "gem/i915_gem_pm.h"
  #include "gem/i915_gem_region.h"
  #include "gem/i915_gem_userptr.h"
+#include "gem/i915_gem_tiling.h"
  #include "gt/intel_engine_user.h"
  #include "gt/intel_gt.h"
  #include "gt/intel_gt_pm.h"
@@ -882,6 +883,96 @@ static void discard_ggtt_vma(struct i915_vma *vma)
         spin_unlock(&obj->vma.lock);
  }

+static int
+i915_gem_object_fits_in_aperture(struct drm_i915_gem_object *obj,
+                                u64 alignment, u64 flags)

Tvrtko asked me to ack the first patch, but then I looked at this and
started wondering.

Conceptually this doesn't pass the smell test. What if we have
multiple per-crtc buffers? Multiple planes on the same crtc? What if
the app does triple buffer? You'll be forever busy tuning this
heuristics, which can't fundamentally be fixed I think. The old "half
of mappable" heuristic isn't really better, but at least it was dead
simple.

Imo what we need here is a change in approach:
1. Check whether the useable view for scanout exists already. If yes,
use that. This should avoid the constant unbinding stalls.
2. Try to in buffer to mappabley, but without evicting anything (so
not the non-blocking thing)
3. Pin the buffer with the most lenient approach

Even the non-blocking interim stage is dangerous, since it'll just
result in other buffers (e.g. when triple-buffering) getting unbound
and we're back to the same stall. Note that this could have an impact
on cpu rendering compositors, where we might end up relying a lot more
partial views. But as long as we are a tad more aggressive (i.e. the
non-blocking binding) in the mmap path that should work out to keep
everything balanced, since usually you render first before you display
anything. And so the buffer should end up in the ideal place.

I'd try to first skip the 2. step since I think it'll require a bit of
work, and frankly I don't think we care about the potential fallout.

To be sure I understand, you propose to stop trying to pin mappable by default. Ie. stop respecting this comment from i915_gem_object_pin_to_display_plane:

	/*
	 * As the user may map the buffer once pinned in the display plane
	 * (e.g. libkms for the bootup splash), we have to ensure that we
	 * always use map_and_fenceable for all scanout buffers. However,
	 * it may simply be too big to fit into mappable, in which case
	 * put it anyway and hope that userspace can cope (but always first
	 * try to preserve the existing ABI).
	 */

By a quick look, for this case it appears we would end up creating partial views for CPU access (since the normal mapping would be busy/unpinnable). Worst case for this is to create a bunch of 1MiB VMAs so something to check would be how long those persist in memory before they get released. Or perhaps the bootup splash use case is not common these days?

Regards,

Tvrtko

-Daniel

+{
+       struct drm_i915_private *i915 = to_i915(obj->base.dev);
+       struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
+       struct drm_mm_node *hole;
+       u64 hole_start, hole_end, start, end;
+       u64 fence_size, fence_alignment;
+       unsigned int count = 0;
+       int err = 0;
+
+       /*
+        * If the required space is larger than the available
+        * aperture, we will not able to find a slot for the
+        * object and unbinding the object now will be in
+        * vain. Worse, doing so may cause us to ping-pong
+        * the object in and out of the Global GTT and
+        * waste a lot of cycles under the mutex.
+        */
+       if (obj->base.size > ggtt->mappable_end)
+               return -E2BIG;
+
+       /*
+        * If NONBLOCK is set the caller is optimistically
+        * trying to cache the full object within the mappable
+        * aperture, and *must* have a fallback in place for
+        * situations where we cannot bind the object. We
+        * can be a little more lax here and use the fallback
+        * more often to avoid costly migrations of ourselves
+        * and other objects within the aperture.
+        */
+       if (!(flags & PIN_NONBLOCK))
+               return 0;
+
+       /*
+        * Other objects such as batchbuffers are fairly small compared
+        * to FBs and are unlikely to exahust the aperture space.
+        * Therefore, return early if this obj is not an FB.
+        */
+       if (!i915_gem_object_is_framebuffer(obj))
+               return 0;
+
+       fence_size = i915_gem_fence_size(i915, obj->base.size,
+                                        i915_gem_object_get_tiling(obj),
+                                        i915_gem_object_get_stride(obj));
+
+       if (i915_vm_has_cache_coloring(&ggtt->vm))
+               fence_size += 2 * I915_GTT_PAGE_SIZE;
+
+       fence_alignment = i915_gem_fence_alignment(i915, obj->base.size,
+                                                  i915_gem_object_get_tiling(obj),
+                                                  i915_gem_object_get_stride(obj));
+       alignment = max_t(u64, alignment, fence_alignment);
+
+       err = mutex_lock_interruptible_nested(&ggtt->vm.mutex, 0);
+       if (err)
+               return err;
+
+       /*
+        * Assuming this object is a large scanout buffer, we try to find
+        * out if there is room to map at-least two of them. There could
+        * be space available to map one but to be consistent, we try to
+        * avoid mapping/fencing any of them.
+        */
+       drm_mm_for_each_suitable_hole(hole, &ggtt->vm.mm, 0, ggtt->mappable_end,
+                                     fence_size, DRM_MM_INSERT_LOW) {
+               hole_start = drm_mm_hole_node_start(hole);
+               hole_end = hole_start + hole->hole_size;
+
+               do {
+                       start = round_up(hole_start, alignment);
+                       end = min_t(u64, hole_end, ggtt->mappable_end);
+
+                       if (range_overflows(start, fence_size, end))
+                               break;
+
+                       if (++count >= 2) {
+                               mutex_unlock(&ggtt->vm.mutex);
+                               return 0;
+                       }
+
+                       hole_start = start + fence_size;
+               } while (1);
+       }
+
+       mutex_unlock(&ggtt->vm.mutex);
+       return -ENOSPC;
+}
+
  struct i915_vma *
  i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
                             struct i915_gem_ww_ctx *ww,
@@ -897,36 +988,9 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,

         if (flags & PIN_MAPPABLE &&
             (!view || view->type == I915_GGTT_VIEW_NORMAL)) {
-               /*
-                * If the required space is larger than the available
-                * aperture, we will not able to find a slot for the
-                * object and unbinding the object now will be in
-                * vain. Worse, doing so may cause us to ping-pong
-                * the object in and out of the Global GTT and
-                * waste a lot of cycles under the mutex.
-                */
-               if (obj->base.size > ggtt->mappable_end)
-                       return ERR_PTR(-E2BIG);
-
-               /*
-                * If NONBLOCK is set the caller is optimistically
-                * trying to cache the full object within the mappable
-                * aperture, and *must* have a fallback in place for
-                * situations where we cannot bind the object. We
-                * can be a little more lax here and use the fallback
-                * more often to avoid costly migrations of ourselves
-                * and other objects within the aperture.
-                *
-                * Half-the-aperture is used as a simple heuristic.
-                * More interesting would to do search for a free
-                * block prior to making the commitment to unbind.
-                * That caters for the self-harm case, and with a
-                * little more heuristics (e.g. NOFAULT, NOEVICT)
-                * we could try to minimise harm to others.
-                */
-               if (flags & PIN_NONBLOCK &&
-                   obj->base.size > ggtt->mappable_end / 2)
-                       return ERR_PTR(-ENOSPC);
+               ret = i915_gem_object_fits_in_aperture(obj, alignment, flags);
+               if (ret)
+                       return ERR_PTR(ret);
         }

  new_vma:
@@ -938,10 +1002,6 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
                 if (flags & PIN_NONBLOCK) {
                         if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma))
                                 return ERR_PTR(-ENOSPC);
-
-                       if (flags & PIN_MAPPABLE &&
-                           vma->fence_size > ggtt->mappable_end / 2)
-                               return ERR_PTR(-ENOSPC);
                 }

                 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) {
--
2.35.1






[Index of Archives]     [AMD Graphics]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux