Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> writes: > Back in commit a4b2b01523a8 ("drm/i915: Don't mark an execlists > context-switch when idle") we noticed the presence of late > context-switch interrupts. We were able to filter those out by looking > at whether the ELSP remained active, but in commit beecec901790 > ("drm/i915/execlists: Preemption!") that became problematic as we now > anticipate receiving a context-switch event for preemption while ELSP > may be empty. To restore the spurious interrupt suppression, add a > counter for the expected number of pending context-switches and skip if > we do not need to handle this interrupt to make forward progress. > > v2: Don't forget to switch on for preempt. > v3: Reduce the counter to a on/off boolean tracker. Declare the HW as > active when we first submit, and idle after the final completion event > (with which we confirm the HW says it is idle), and track each source > of activity separately. With a finite number of sources, it should us > debug which gets stuck. > > Fixes: beecec901790 ("drm/i915/execlists: Preemption!") > Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> > Cc: Michal Winiarski <michal.winiarski@xxxxxxxxx> > Cc: Tvrtko Ursulin <tvrtko.ursulin@xxxxxxxxx> > Cc: Arkadiusz Hiler <arkadiusz.hiler@xxxxxxxxx> > Cc: Mika Kuoppala <mika.kuoppala@xxxxxxxxx> > Cc: Joonas Lahtinen <joonas.lahtinen@xxxxxxxxxxxxxxx> > --- > drivers/gpu/drm/i915/i915_guc_submission.c | 3 +++ > drivers/gpu/drm/i915/i915_irq.c | 6 ++++-- > drivers/gpu/drm/i915/intel_engine_cs.c | 5 +++-- > drivers/gpu/drm/i915/intel_lrc.c | 27 ++++++++++++++++++------ > drivers/gpu/drm/i915/intel_ringbuffer.h | 34 ++++++++++++++++++++++++++++-- > 5 files changed, 62 insertions(+), 13 deletions(-) > > diff --git a/drivers/gpu/drm/i915/i915_guc_submission.c b/drivers/gpu/drm/i915/i915_guc_submission.c > index a2e8114b739d..f84c267728fd 100644 > --- a/drivers/gpu/drm/i915/i915_guc_submission.c > +++ b/drivers/gpu/drm/i915/i915_guc_submission.c > @@ -610,6 +610,7 @@ static void i915_guc_dequeue(struct intel_engine_cs *engine) > execlists->first = rb; > if (submit) { > port_assign(port, last); > + execlists_set_active(execlists, EXECLISTS_ACTIVE_USER); > i915_guc_submit(engine); > } > spin_unlock_irq(&engine->timeline->lock); > @@ -633,6 +634,8 @@ static void i915_guc_irq_handler(unsigned long data) > > rq = port_request(&port[0]); > } > + if (!rq) > + execlists_clear_active(execlists, EXECLISTS_ACTIVE_USER); > > if (!port_isset(last_port)) > i915_guc_dequeue(engine); > diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c > index b1296a55c1e4..f8205841868b 100644 > --- a/drivers/gpu/drm/i915/i915_irq.c > +++ b/drivers/gpu/drm/i915/i915_irq.c > @@ -1388,8 +1388,10 @@ gen8_cs_irq_handler(struct intel_engine_cs *engine, u32 iir, int test_shift) > bool tasklet = false; > > if (iir & (GT_CONTEXT_SWITCH_INTERRUPT << test_shift)) { > - __set_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted); > - tasklet = true; > + if (READ_ONCE(engine->execlists.active)) { > + __set_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted); > + tasklet = true; These two in here feel naturally attracted to eachothers. Perhaps a future patch will meld them together. Reviewed-by: Mika Kuoppala <mika.kuoppala@xxxxxxxxxxxxxxx> > + } > } > > if (iir & (GT_RENDER_USER_INTERRUPT << test_shift)) { > diff --git a/drivers/gpu/drm/i915/intel_engine_cs.c b/drivers/gpu/drm/i915/intel_engine_cs.c > index a47a9c6bea52..ab5bf4e2e28e 100644 > --- a/drivers/gpu/drm/i915/intel_engine_cs.c > +++ b/drivers/gpu/drm/i915/intel_engine_cs.c > @@ -1548,8 +1548,8 @@ bool intel_engine_is_idle(struct intel_engine_cs *engine) > if (test_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted)) > return false; > > - /* Both ports drained, no more ELSP submission? */ > - if (port_request(&engine->execlists.port[0])) > + /* Waiting to drain ELSP? */ > + if (READ_ONCE(engine->execlists.active)) > return false; > > /* ELSP is empty, but there are ready requests? */ > @@ -1749,6 +1749,7 @@ void intel_engine_dump(struct intel_engine_cs *engine, struct drm_printer *m) > idx); > } > } > + drm_printf(m, "\t\tHW active? 0x%x\n", execlists->active); > rcu_read_unlock(); > } else if (INTEL_GEN(dev_priv) > 6) { > drm_printf(m, "\tPP_DIR_BASE: 0x%08x\n", > diff --git a/drivers/gpu/drm/i915/intel_lrc.c b/drivers/gpu/drm/i915/intel_lrc.c > index 7f45dd7dc3e5..233c992a886b 100644 > --- a/drivers/gpu/drm/i915/intel_lrc.c > +++ b/drivers/gpu/drm/i915/intel_lrc.c > @@ -575,7 +575,8 @@ static void execlists_dequeue(struct intel_engine_cs *engine) > * the state of the GPU is known (idle). > */ > inject_preempt_context(engine); > - execlists->preempt = true; > + execlists_set_active(execlists, > + EXECLISTS_ACTIVE_PREEMPT); > goto unlock; > } else { > /* > @@ -683,8 +684,10 @@ static void execlists_dequeue(struct intel_engine_cs *engine) > unlock: > spin_unlock_irq(&engine->timeline->lock); > > - if (submit) > + if (submit) { > + execlists_set_active(execlists, EXECLISTS_ACTIVE_USER); > execlists_submit_ports(engine); > + } > } > > static void > @@ -696,6 +699,7 @@ execlist_cancel_port_requests(struct intel_engine_execlists *execlists) > while (num_ports-- && port_isset(port)) { > struct drm_i915_gem_request *rq = port_request(port); > > + GEM_BUG_ON(!execlists->active); > execlists_context_status_change(rq, INTEL_CONTEXT_SCHEDULE_PREEMPTED); > i915_gem_request_put(rq); > > @@ -861,15 +865,21 @@ static void intel_lrc_irq_handler(unsigned long data) > unwind_incomplete_requests(engine); > spin_unlock_irq(&engine->timeline->lock); > > - GEM_BUG_ON(!execlists->preempt); > - execlists->preempt = false; > + GEM_BUG_ON(!execlists_is_active(execlists, > + EXECLISTS_ACTIVE_PREEMPT)); > + execlists_clear_active(execlists, > + EXECLISTS_ACTIVE_PREEMPT); > continue; > } > > if (status & GEN8_CTX_STATUS_PREEMPTED && > - execlists->preempt) > + execlists_is_active(execlists, > + EXECLISTS_ACTIVE_PREEMPT)) > continue; > > + GEM_BUG_ON(!execlists_is_active(execlists, > + EXECLISTS_ACTIVE_USER)); > + > /* Check the context/desc id for this event matches */ > GEM_DEBUG_BUG_ON(buf[2 * head + 1] != port->context_id); > > @@ -892,6 +902,9 @@ static void intel_lrc_irq_handler(unsigned long data) > /* After the final element, the hw should be idle */ > GEM_BUG_ON(port_count(port) == 0 && > !(status & GEN8_CTX_STATUS_ACTIVE_IDLE)); > + if (port_count(port) == 0) > + execlists_clear_active(execlists, > + EXECLISTS_ACTIVE_USER); > } > > if (head != execlists->csb_head) { > @@ -901,7 +914,7 @@ static void intel_lrc_irq_handler(unsigned long data) > } > } > > - if (!execlists->preempt) > + if (!execlists_is_active(execlists, EXECLISTS_ACTIVE_PREEMPT)) > execlists_dequeue(engine); > > intel_uncore_forcewake_put(dev_priv, execlists->fw_domains); > @@ -1460,7 +1473,7 @@ static int gen8_init_common_ring(struct intel_engine_cs *engine) > GT_CONTEXT_SWITCH_INTERRUPT << engine->irq_shift); > clear_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted); > execlists->csb_head = -1; > - execlists->preempt = false; > + execlists->active = 0; > > /* After a GPU reset, we may have requests to replay */ > if (!i915_modparams.enable_guc_submission && execlists->first) > diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h > index 17186f067408..51bc704bf413 100644 > --- a/drivers/gpu/drm/i915/intel_ringbuffer.h > +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h > @@ -241,9 +241,17 @@ struct intel_engine_execlists { > } port[EXECLIST_MAX_PORTS]; > > /** > - * @preempt: are we currently handling a preempting context switch? > + * @active: is the HW active? We consider the HW as active after > + * submitted any context for execution until we have seen the last > + * context completion event. After that, we do not expect any more > + * events until we submit, and so can park the HW. > + * > + * As we have a small number of different sources from which we feed > + * the HW, we track the state of each inside a single bitfield. > */ > - bool preempt; > + unsigned int active; > +#define EXECLISTS_ACTIVE_USER 0 > +#define EXECLISTS_ACTIVE_PREEMPT 1 > > /** > * @port_mask: number of execlist ports - 1 > @@ -525,6 +533,27 @@ struct intel_engine_cs { > u32 (*get_cmd_length_mask)(u32 cmd_header); > }; > > +static inline void > +execlists_set_active(struct intel_engine_execlists *execlists, > + unsigned int bit) > +{ > + __set_bit(bit, (unsigned long *)&execlists->active); > +} > + > +static inline void > +execlists_clear_active(struct intel_engine_execlists *execlists, > + unsigned int bit) > +{ > + __clear_bit(bit, (unsigned long *)&execlists->active); > +} > + > +static inline bool > +execlists_is_active(const struct intel_engine_execlists *execlists, > + unsigned int bit) > +{ > + return test_bit(bit, (unsigned long *)&execlists->active); > +} > + > static inline unsigned int > execlists_num_ports(const struct intel_engine_execlists * const execlists) > { > @@ -538,6 +567,7 @@ execlists_port_complete(struct intel_engine_execlists * const execlists, > const unsigned int m = execlists->port_mask; > > GEM_BUG_ON(port_index(port, execlists) != 0); > + GEM_BUG_ON(!execlists_is_active(execlists, EXECLISTS_ACTIVE_USER)); > > memmove(port, port + 1, m * sizeof(struct execlist_port)); > memset(port + m, 0, sizeof(struct execlist_port)); > -- > 2.15.0.rc1 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@xxxxxxxxxxxxxxxxxxxxx > https://lists.freedesktop.org/mailman/listinfo/intel-gfx _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx