On Fri, Aug 13, 2021 at 04:24:37PM -0700, Daniele Ceraolo Spurio wrote: > > > On 8/3/2021 6:23 PM, Matthew Brost wrote: > > The i915 currently has 2k visible priority levels which are currently > > unique. This is changing to statically map these 2k levels into 3 > > buckets: > > > > low: < 0 > > mid: 0 > > high: > 0 > > > > Update gem_exec_schedule to understand this. This entails updating > > promotion test to use 3 levels that will map into different buckets and > > also add bit of delay after releasing a cork beforing completing the > > spinners. > > This needs a line about why we add the delay, something like "to give time > to the i915 scheduler to process the fence release and queue the requests" > or something. Will reword, have typo here too. > BTW, any reason not to just add the delay unconditionally in > unplug_show_queue, instead of only in one test? Other tests might suffer > from the same problem even if they're not hitting it at the moment. > Yea, probably a better approach to future proof this as I could see other sections randomly failing in CI and wasting our time. Will fix this and subsequent patch too. Matt > Daniele > > > > > Also skip any tests that rely on having more than 3 priority levels. > > > > v2: Add a delay between starting releasing spinner and cork in > > promotion, add local define for static mapping engine info > > > > Signed-off-by: Matthew Brost <matthew.brost@xxxxxxxxx> > > --- > > lib/i915/gem_scheduler.c | 14 ++++++++ > > lib/i915/gem_scheduler.h | 1 + > > lib/i915/i915_drm_local.h | 10 ++++++ > > tests/i915/gem_exec_schedule.c | 62 +++++++++++++++++++++------------- > > 4 files changed, 63 insertions(+), 24 deletions(-) > > > > diff --git a/lib/i915/gem_scheduler.c b/lib/i915/gem_scheduler.c > > index cdddf42ad..d006b8676 100644 > > --- a/lib/i915/gem_scheduler.c > > +++ b/lib/i915/gem_scheduler.c > > @@ -28,6 +28,7 @@ > > #include "igt_core.h" > > #include "ioctl_wrappers.h" > > +#include "i915/i915_drm_local.h" > > #include "i915/gem_scheduler.h" > > #include "i915/gem_submission.h" > > @@ -90,6 +91,19 @@ bool gem_scheduler_has_ctx_priority(int fd) > > I915_SCHEDULER_CAP_PRIORITY; > > } > > +/** > > + * gem_scheduler_has_static_priority: > > + * @fd: open i915 drm file descriptor > > + * > > + * Feature test macro to query whether the driver supports priority assigned > > + * from user space are statically mapping into 3 buckets. > > + */ > > +bool gem_scheduler_has_static_priority(int fd) > > +{ > > + return gem_scheduler_capability(fd) & > > + I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP; > > +} > > + > > /** > > * gem_scheduler_has_preemption: > > * @fd: open i915 drm file descriptor > > diff --git a/lib/i915/gem_scheduler.h b/lib/i915/gem_scheduler.h > > index d43e84bd2..b00804f70 100644 > > --- a/lib/i915/gem_scheduler.h > > +++ b/lib/i915/gem_scheduler.h > > @@ -29,6 +29,7 @@ > > unsigned gem_scheduler_capability(int fd); > > bool gem_scheduler_enabled(int fd); > > bool gem_scheduler_has_ctx_priority(int fd); > > +bool gem_scheduler_has_static_priority(int fd); > > bool gem_scheduler_has_preemption(int fd); > > bool gem_scheduler_has_semaphores(int fd); > > bool gem_scheduler_has_engine_busy_stats(int fd); > > diff --git a/lib/i915/i915_drm_local.h b/lib/i915/i915_drm_local.h > > index dd646aedf..a1527ff21 100644 > > --- a/lib/i915/i915_drm_local.h > > +++ b/lib/i915/i915_drm_local.h > > @@ -20,6 +20,16 @@ extern "C" { > > * clean these up when kernel uapi headers are sync'd. > > */ > > +/* > > + * Indicates the 2k user priority levels are statically mapped into 3 buckets as > > + * follows: > > + * > > + * -1k to -1 Low priority > > + * 0 Normal priority > > + * 1 to 1k Highest priority > > + */ > > +#define I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP (1ul << 5) > > + > > #if defined(__cplusplus) > > } > > #endif > > diff --git a/tests/i915/gem_exec_schedule.c b/tests/i915/gem_exec_schedule.c > > index e5fb45982..bb9fb6c14 100644 > > --- a/tests/i915/gem_exec_schedule.c > > +++ b/tests/i915/gem_exec_schedule.c > > @@ -199,7 +199,8 @@ create_highest_priority(int fd, const intel_ctx_cfg_t *cfg) > > static void unplug_show_queue(int fd, struct igt_cork *c, > > const intel_ctx_cfg_t *cfg, > > - unsigned int engine) > > + unsigned int engine, > > + unsigned usec_delay) > > { > > igt_spin_t *spin[MAX_ELSP_QLEN]; > > int max = MAX_ELSP_QLEN; > > @@ -216,6 +217,7 @@ static void unplug_show_queue(int fd, struct igt_cork *c, > > igt_cork_unplug(c); /* batches will now be queued on the engine */ > > igt_debugfs_dump(fd, "i915_engine_info"); > > + usleep(usec_delay); > > for (int n = 0; n < max; n++) > > igt_spin_free(fd, spin[n]); > > @@ -237,7 +239,7 @@ static void fifo(int fd, const intel_ctx_t *ctx, unsigned ring) > > store_dword_fenced(fd, ctx, ring, scratch, 0, 1, fence, 0); > > store_dword_fenced(fd, ctx, ring, scratch, 0, 2, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, ring); > > + unplug_show_queue(fd, &cork, &ctx->cfg, ring, 0); > > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > > @@ -298,7 +300,7 @@ static void implicit_rw(int i915, const intel_ctx_t *ctx, unsigned int ring, > > ring, scratch, 0, ring, > > fence, I915_GEM_DOMAIN_RENDER); > > - unplug_show_queue(i915, &cork, &ctx->cfg, ring); > > + unplug_show_queue(i915, &cork, &ctx->cfg, ring, 0); > > close(fence); > > result = __sync_read_u32(i915, scratch, 0); > > @@ -355,7 +357,7 @@ static void independent(int fd, const intel_ctx_t *ctx, unsigned int engine, > > /* Same priority, but different timeline (as different engine) */ > > batch = __store_dword(fd, ctx, engine, scratch, 0, engine, 0, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, engine); > > + unplug_show_queue(fd, &cork, &ctx->cfg, engine, 0); > > close(fence); > > gem_sync(fd, batch); > > @@ -1326,7 +1328,7 @@ static void reorder(int fd, const intel_ctx_cfg_t *cfg, > > store_dword_fenced(fd, ctx[LO], ring, scratch, 0, ctx[LO]->id, fence, 0); > > store_dword_fenced(fd, ctx[HI], ring, scratch, 0, ctx[HI]->id, fence, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > > + unplug_show_queue(fd, &cork, cfg, ring, 0); > > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > > @@ -1353,10 +1355,10 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > gem_context_set_priority(fd, ctx[LO]->id, MIN_PRIO); > > ctx[HI] = intel_ctx_create(fd, cfg); > > - gem_context_set_priority(fd, ctx[HI]->id, 0); > > + gem_context_set_priority(fd, ctx[HI]->id, MAX_PRIO); > > ctx[NOISE] = intel_ctx_create(fd, cfg); > > - gem_context_set_priority(fd, ctx[NOISE]->id, MIN_PRIO/2); > > + gem_context_set_priority(fd, ctx[NOISE]->id, 0); > > result = gem_create(fd, 4096); > > dep = gem_create(fd, 4096); > > @@ -1377,7 +1379,7 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > store_dword(fd, ctx[HI], ring, result, 0, ctx[HI]->id, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > > + unplug_show_queue(fd, &cork, cfg, ring, 250000); > > close(fence); > > dep_read = __sync_read_u32(fd, dep, 0); > > @@ -1893,7 +1895,7 @@ static void deep(int fd, const intel_ctx_cfg_t *cfg, > > igt_info("Second deptree: %d requests [%.3fs]\n", > > n * XS, 1e-9*igt_nsec_elapsed(&tv)); > > - unplug_show_queue(fd, &cork, cfg, ring); > > + unplug_show_queue(fd, &cork, cfg, ring, 0); > > gem_close(fd, plug); > > igt_require(expected); /* too slow */ > > @@ -1962,7 +1964,7 @@ static void wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > igt_info("Submitted %d requests over %d contexts in %.1fms\n", > > count, MAX_CONTEXTS, igt_nsec_elapsed(&tv) * 1e-6); > > - unplug_show_queue(fd, &cork, cfg, ring); > > + unplug_show_queue(fd, &cork, cfg, ring, 0); > > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > > @@ -2067,7 +2069,7 @@ static void reorder_wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > intel_ctx_destroy(fd, tmp_ctx); > > } > > - unplug_show_queue(fd, &cork, cfg, ring); > > + unplug_show_queue(fd, &cork, cfg, ring, 0); > > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > > @@ -2963,19 +2965,25 @@ igt_main > > test_each_engine_store("preempt-other-chain", fd, ctx, e) > > preempt_other(fd, &ctx->cfg, e->flags, CHAIN); > > - test_each_engine_store("preempt-queue", fd, ctx, e) > > - preempt_queue(fd, &ctx->cfg, e->flags, 0); > > + test_each_engine_store("preempt-engines", fd, ctx, e) > > + preempt_engines(fd, e, 0); > > - test_each_engine_store("preempt-queue-chain", fd, ctx, e) > > - preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > > - test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > > + igt_subtest_group { > > + igt_fixture { > > + igt_require(!gem_scheduler_has_static_priority(fd)); > > + } > > - test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > > + test_each_engine_store("preempt-queue", fd, ctx, e) > > + preempt_queue(fd, &ctx->cfg, e->flags, 0); > > - test_each_engine_store("preempt-engines", fd, ctx, e) > > - preempt_engines(fd, e, 0); > > + test_each_engine_store("preempt-queue-chain", fd, ctx, e) > > + preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > > + test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > > + > > + test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > > + } > > igt_subtest_group { > > igt_hang_t hang; > > @@ -3017,11 +3025,17 @@ igt_main > > test_each_engine_store("wide", fd, ctx, e) > > wide(fd, &ctx->cfg, e->flags); > > - test_each_engine_store("reorder-wide", fd, ctx, e) > > - reorder_wide(fd, &ctx->cfg, e->flags); > > - > > test_each_engine_store("smoketest", fd, ctx, e) > > smoketest(fd, &ctx->cfg, e->flags, 5); > > + > > + igt_subtest_group { > > + igt_fixture { > > + igt_require(!gem_scheduler_has_static_priority(fd)); > > + } > > + > > + test_each_engine_store("reorder-wide", fd, ctx, e) > > + reorder_wide(fd, &ctx->cfg, e->flags); > > + } > > } > > igt_subtest_group { >