On Mon, Mar 30, 2020 at 09:16:49PM +0300, Lisovskiy, Stanislav wrote: > On Mon, Mar 30, 2020 at 08:07:31PM +0300, Ville Syrjälä wrote: > > On Mon, Mar 30, 2020 at 03:23:53PM +0300, Stanislav Lisovskiy wrote: > > > According to BSpec max BW per slice is calculated using formula > > > Max BW = CDCLK * 64. Currently when calculating min CDCLK we > > > account only per plane requirements, however in order to avoid > > > FIFO underruns we need to estimate accumulated BW consumed by > > > all planes(ddb entries basically) residing on that particular > > > DBuf slice. This will allow us to put CDCLK lower and save power > > > when we don't need that much bandwidth or gain additional > > > performance once plane consumption grows. > > > > > > v2: - Fix long line warning > > > - Limited new DBuf bw checks to only gens >= 11 > > > > > > v3: - Lets track used Dbuf bw per slice and per crtc in bw state > > > (or may be in DBuf state in future), that way we don't need > > > to have all crtcs in state and those only if we detect if > > > are actually going to change cdclk, just same way as we > > > do with other stuff, i.e intel_atomic_serialize_global_state > > > and co. Just as per Ville's paradigm. > > > - Made dbuf bw calculation procedure look nicer by introducing > > > for_each_dbuf_slice_in_mask - we often will now need to iterate > > > slices using mask. > > > - According to experimental results CDCLK * 64 accounts for > > > overall bandwidth across all dbufs, not per dbuf. > > > > > > Signed-off-by: Stanislav Lisovskiy <stanislav.lisovskiy@xxxxxxxxx> > > > --- > > > drivers/gpu/drm/i915/display/intel_bw.c | 61 ++++++++++++++++++- > > > drivers/gpu/drm/i915/display/intel_bw.h | 8 +++ > > > drivers/gpu/drm/i915/display/intel_cdclk.c | 25 ++++++++ > > > drivers/gpu/drm/i915/display/intel_display.c | 8 +++ > > > .../drm/i915/display/intel_display_power.h | 2 + > > > drivers/gpu/drm/i915/intel_pm.c | 34 ++++++++++- > > > drivers/gpu/drm/i915/intel_pm.h | 3 + > > > 7 files changed, 138 insertions(+), 3 deletions(-) > > > > > > diff --git a/drivers/gpu/drm/i915/display/intel_bw.c b/drivers/gpu/drm/i915/display/intel_bw.c > > > index 573a1c206b60..e9d65820fb76 100644 > > > --- a/drivers/gpu/drm/i915/display/intel_bw.c > > > +++ b/drivers/gpu/drm/i915/display/intel_bw.c > > > @@ -6,6 +6,7 @@ > > > #include <drm/drm_atomic_state_helper.h> > > > > > > #include "intel_bw.h" > > > +#include "intel_pm.h" > > > #include "intel_display_types.h" > > > #include "intel_sideband.h" > > > #include "intel_atomic.h" > > > @@ -338,7 +339,6 @@ static unsigned int intel_bw_crtc_data_rate(const struct intel_crtc_state *crtc_ > > > > > > return data_rate; > > > } > > > - > > > void intel_bw_crtc_update(struct intel_bw_state *bw_state, > > > const struct intel_crtc_state *crtc_state) > > > { > > > @@ -419,6 +419,65 @@ intel_atomic_bw_get_state(struct intel_atomic_state *state) > > > return to_intel_bw_state(bw_state); > > > } > > > > > > +int intel_bw_calc_min_cdclk(struct intel_atomic_state *state) > > > +{ > > > + struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > > + int i = 0; > > > + enum plane_id plane_id; > > > + struct intel_crtc_state *crtc_state; > > > > const > > Why? There are lots of places where we use it that way. Always make it const if you don't have to change it. Makes life a lot easier when you immediately see that it can't mutate when you're not looking. Which is also why pure functions are preferrable to those with side effects. > Even in intel_compute_min_cdclk: > > struct intel_crtc_state *crtc_state; > int min_cdclk, i; > enum pipe pipe; > > for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) { > > Also I grepped - almost every place in intel_display.c > which has for_each_new_intel_crtc_in_state uses crtc_state as non-const. > > It is called that way intel_calc_active_pipes, intel_atomic_check_crtcs,... > almost everywhere in intel_display.c > > So can you please explain what do you mean?.. > > > > > > + struct intel_crtc *crtc; > > > + int max_bw = 0; > > > + int min_cdclk; > > > + enum pipe pipe; > > > + struct intel_bw_state *bw_state; > > > + int slice_id = 0; > > > > Bunch of needless intiialization, needlessly wide scope, etc. > > max_bw needs to be accumulated so I guess it should be init to 0. > > Agree regarding slice_id - just a leftover after I started to > use for_each_dbuf_slice macro. > > > > > > + > > > + bw_state = intel_atomic_bw_get_state(state); > > > + > > > > Spurious whitespace. > > > > > + if (IS_ERR(bw_state)) > > > + return PTR_ERR(bw_state); > > > + > > > + for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) { > > > + struct intel_crtc_bw *crtc_bw = &bw_state->dbuf_bw_used[crtc->pipe]; > > > + > > > + memset(&crtc_bw->dbuf_bw, 0, sizeof(crtc_bw->dbuf_bw)); > > > + > > > + for_each_plane_id_on_crtc(crtc, plane_id) { > > > + struct skl_ddb_entry *plane_alloc = > > > + &crtc_state->wm.skl.plane_ddb_y[plane_id]; > > > + struct skl_ddb_entry *uv_plane_alloc = > > > + &crtc_state->wm.skl.plane_ddb_uv[plane_id]; > > > > const > > > > > + unsigned int data_rate = crtc_state->data_rate[plane_id]; > > > + > > > > more strange whitespace > > > > > + unsigned int dbuf_mask = skl_ddb_dbuf_slice_mask(dev_priv, plane_alloc); > > > + > > > + dbuf_mask |= skl_ddb_dbuf_slice_mask(dev_priv, uv_plane_alloc); > > > > Looks bad when you initialize part of it in declaration and the rest > > later. > > Didn't want to have a long time here, will have to make a separate function then. Not sure why you want a separate function. foo = a | b ? > > > > > > + > > > + DRM_DEBUG_KMS("Got dbuf mask %x for pipe %c ddb %d-%d plane %d data rate %d\n", > > > + dbuf_mask, pipe_name(crtc->pipe), plane_alloc->start, > > > + plane_alloc->end, plane_id, data_rate); > > > + > > > + for_each_dbuf_slice_in_mask(slice_id, dbuf_mask) > > > + crtc_bw->dbuf_bw[slice_id] += data_rate; > > > > This doesn't feel quite right for planar formats. > > > > For pre-icl it works by accident since we only have the one slice so > > we don't end up accounting for the full bandwidth from both color planes > > to multiple slices. If we had multiple slices and chroma and luma had > > been allocated to different slices we'd count the same thing multiple > > times. > > > > For icl+ we seem to assign the full data rate to the UV plane's slice(s) > > since only the UV plane has data_rate[] != 0. > > This sounds interesting, so do you mean that we need something like data_rate > and data_rate_uv? Perhaps. Or maybe we just start tracking the data rate separately for the master and slave planes on icl+. No real issue on pre-icl so that would actually work AFAICS. > My logic here is that we just get a set of slices used by uv and y planes > in total and then account for data_rate for those. > We probably then need another patch to split data rate then. > > I was aware of that actually but anyway being pessimistic here is better. > Also same about if plane would be sitting on 2 slices at the same time. > > > > > > + } > > > + } > > > + > > > + for_each_dbuf_slice(slice_id) { > > > + int total_bw_per_slice = 0; > > > + > > > + for_each_pipe(dev_priv, pipe) { > > > + struct intel_crtc_bw *crtc_bw = &bw_state->dbuf_bw_used[pipe]; > > > + > > > + total_bw_per_slice += crtc_bw->dbuf_bw[slice_id]; > > > + } > > > + max_bw += total_bw_per_slice; > > > > So we're aggregating all the bw instead of per-slice? Is this based on > > the other mail you sent? Deserves a comment explaining why we do such > > odd things. > > Agree need to add a comment. Currently it seems the only way to make it > work without either having underruns or 8K not usable with bumped CDCLK. > > > > > > + } > > > + > > > + min_cdclk = max_bw / 64; > > > + > > > + return min_cdclk; > > > +} > > > + > > > int intel_bw_atomic_check(struct intel_atomic_state *state) > > > { > > > struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > > diff --git a/drivers/gpu/drm/i915/display/intel_bw.h b/drivers/gpu/drm/i915/display/intel_bw.h > > > index 9a5627be6876..d2b5f32b0791 100644 > > > --- a/drivers/gpu/drm/i915/display/intel_bw.h > > > +++ b/drivers/gpu/drm/i915/display/intel_bw.h > > > @@ -10,11 +10,16 @@ > > > > > > #include "intel_display.h" > > > #include "intel_global_state.h" > > > +#include "intel_display_power.h" > > > > > > struct drm_i915_private; > > > struct intel_atomic_state; > > > struct intel_crtc_state; > > > > > > +struct intel_crtc_bw { > > > + int dbuf_bw[I915_MAX_DBUF_SLICES]; > > > +}; > > > + > > > struct intel_bw_state { > > > struct intel_global_state base; > > > > > > @@ -31,6 +36,8 @@ struct intel_bw_state { > > > */ > > > u8 qgv_points_mask; > > > > > > + struct intel_crtc_bw dbuf_bw_used[I915_MAX_PIPES]; > > > > The name of the struct isn't very good if it just contains the > > dbuf bw numbers. > > > > > + > > > unsigned int data_rate[I915_MAX_PIPES]; > > > u8 num_active_planes[I915_MAX_PIPES]; > > > > Maybe collect all this to the per-crtc struct? > > > > > }; > > > @@ -53,5 +60,6 @@ void intel_bw_crtc_update(struct intel_bw_state *bw_state, > > > const struct intel_crtc_state *crtc_state); > > > int icl_pcode_restrict_qgv_points(struct drm_i915_private *dev_priv, > > > u32 points_mask); > > > +int intel_bw_calc_min_cdclk(struct intel_atomic_state *state); > > > > > > #endif /* __INTEL_BW_H__ */ > > > diff --git a/drivers/gpu/drm/i915/display/intel_cdclk.c b/drivers/gpu/drm/i915/display/intel_cdclk.c > > > index 979a0241fdcb..036774e7f3ec 100644 > > > --- a/drivers/gpu/drm/i915/display/intel_cdclk.c > > > +++ b/drivers/gpu/drm/i915/display/intel_cdclk.c > > > @@ -25,6 +25,7 @@ > > > #include "intel_cdclk.h" > > > #include "intel_display_types.h" > > > #include "intel_sideband.h" > > > +#include "intel_bw.h" > > > > > > /** > > > * DOC: CDCLK / RAWCLK > > > @@ -2001,11 +2002,19 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > > { > > > struct drm_i915_private *dev_priv = > > > to_i915(crtc_state->uapi.crtc->dev); > > > + struct intel_atomic_state *state = NULL; > > > int min_cdclk; > > > > > > if (!crtc_state->hw.enable) > > > return 0; > > > > > > + /* > > > + * FIXME: Unfortunately when this gets called from intel_modeset_setup_hw_state > > > + * there is no intel_atomic_state at all. So lets not then use it. > > > + */ > > > + if (crtc_state->uapi.state) > > > + state = to_intel_atomic_state(crtc_state->uapi.state); > > > > This still indicates that either this isn't the right place to call this > > or we have the state stored in the wrong place. > > > > I think I'd just move the thing into intel_compute_min_cdclk() as a start. > > Was thinking about this, why I didn't put it here is because just wanted > to keep all cdclk regulating logic in this function, because intel_compute_min_cdclk > is kind of higher level already which is supposed to call that one per crtc_state. > Otherwise CDCLK regulating logic will be kind of spread across those two. This is not a per-crtc limit so having it in the per-crtc function seems wrong anyway to me. > > > > > > + > > > min_cdclk = intel_pixel_rate_to_cdclk(crtc_state); > > > > > > /* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */ > > > @@ -2080,6 +2089,22 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > > if (IS_TIGERLAKE(dev_priv)) > > > min_cdclk = max(min_cdclk, (int)crtc_state->pixel_rate); > > > > > > + /* > > > + * Similar story as with skl_write_plane_wm and intel_enable_sagv > > > + * - in some certain driver parts, we don't have any guarantee that > > > + * parent exists. So we might be having a crtc_state without > > > + * parent state. > > > + */ > > > + if (INTEL_GEN(dev_priv) >= 11) { > > > + if (state) { > > > + int dbuf_bw_cdclk = intel_bw_calc_min_cdclk(state); > > > + > > > + DRM_DEBUG_KMS("DBuf bw min cdclk %d current min_cdclk %d\n", > > > + dbuf_bw_cdclk, min_cdclk); > > > + min_cdclk = max(min_cdclk, dbuf_bw_cdclk); > > > + } > > > + } > > > + > > > if (min_cdclk > dev_priv->max_cdclk_freq) { > > > drm_dbg_kms(&dev_priv->drm, > > > "required cdclk (%d kHz) exceeds max (%d kHz)\n", > > > diff --git a/drivers/gpu/drm/i915/display/intel_display.c b/drivers/gpu/drm/i915/display/intel_display.c > > > index 9fd32d61ebfe..fa2870c0d7fd 100644 > > > --- a/drivers/gpu/drm/i915/display/intel_display.c > > > +++ b/drivers/gpu/drm/i915/display/intel_display.c > > > @@ -14678,6 +14678,14 @@ static int intel_atomic_check_planes(struct intel_atomic_state *state, > > > if (old_active_planes == new_active_planes) > > > continue; > > > > > > + /* > > > + * active_planes bitmask has been updated, whenever amount > > > + * of active planes had changed we need to recalculate CDCLK > > > + * as it depends on total bandwidth now, not only min_cdclk > > > + * per plane. > > > + */ > > > + *need_cdclk_calc = true; > > > + > > > ret = intel_crtc_add_planes_to_state(state, crtc, new_active_planes); > > > if (ret) > > > return ret; > > > diff --git a/drivers/gpu/drm/i915/display/intel_display_power.h b/drivers/gpu/drm/i915/display/intel_display_power.h > > > index 468e8fb0203a..9e33fb90422f 100644 > > > --- a/drivers/gpu/drm/i915/display/intel_display_power.h > > > +++ b/drivers/gpu/drm/i915/display/intel_display_power.h > > > @@ -308,6 +308,8 @@ intel_display_power_put_async(struct drm_i915_private *i915, > > > } > > > #endif > > > > > > +#define I915_MAX_DBUF_SLICES 2 > > > + > > > enum dbuf_slice { > > > DBUF_S1, > > > DBUF_S2, > > > diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c > > > index 551933e3f7da..5dcd1cd09ad7 100644 > > > --- a/drivers/gpu/drm/i915/intel_pm.c > > > +++ b/drivers/gpu/drm/i915/intel_pm.c > > > @@ -4055,10 +4055,9 @@ icl_get_first_dbuf_slice_offset(u32 dbuf_slice_mask, > > > return offset; > > > } > > > > > > -static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv) > > > +u16 intel_get_ddb_size(struct drm_i915_private *dev_priv) > > > { > > > u16 ddb_size = INTEL_INFO(dev_priv)->ddb_size; > > > - > > > drm_WARN_ON(&dev_priv->drm, ddb_size == 0); > > > > > > if (INTEL_GEN(dev_priv) < 11) > > > @@ -4067,6 +4066,37 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv) > > > return ddb_size; > > > } > > > > > > +u32 skl_ddb_dbuf_slice_mask(struct drm_i915_private *dev_priv, > > > + const struct skl_ddb_entry *entry) > > > +{ > > > + u32 slice_mask = 0; > > > + u16 ddb_size = intel_get_ddb_size(dev_priv); > > > + u16 num_supported_slices = INTEL_INFO(dev_priv)->num_supported_dbuf_slices; > > > + u16 slice_size = ddb_size / num_supported_slices; > > > + u16 start_slice; > > > + u16 end_slice; > > > + > > > + if (!skl_ddb_entry_size(entry)) > > > + return 0; > > > + > > > + start_slice = entry->start / slice_size; > > > + end_slice = (entry->end - 1) / slice_size; > > > + > > > + DRM_DEBUG_KMS("ddb size %d slices %d slice size %d start slice %d end slice %d\n", > > > + ddb_size, num_supported_slices, slice_size, start_slice, end_slice); > > > + > > > + /* > > > + * Per plane DDB entry can in a really worst case be on multiple slices > > > + * but single entry is anyway contigious. > > > + */ > > > + while (start_slice <= end_slice) { > > > + slice_mask |= 1 << start_slice; > > > + start_slice++; > > > + } > > > > GENMASK() or somehting? > > Will fix, thanks for proposal. > > > > > > + > > > + return slice_mask; > > > +} > > > + > > > static u8 skl_compute_dbuf_slices(const struct intel_crtc_state *crtc_state, > > > u8 active_pipes); > > > > > > diff --git a/drivers/gpu/drm/i915/intel_pm.h b/drivers/gpu/drm/i915/intel_pm.h > > > index 069515f04170..41c61ad71ce6 100644 > > > --- a/drivers/gpu/drm/i915/intel_pm.h > > > +++ b/drivers/gpu/drm/i915/intel_pm.h > > > @@ -37,6 +37,9 @@ void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc, > > > struct skl_ddb_entry *ddb_y, > > > struct skl_ddb_entry *ddb_uv); > > > void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv); > > > +u16 intel_get_ddb_size(struct drm_i915_private *dev_priv); > > > +u32 skl_ddb_dbuf_slice_mask(struct drm_i915_private *dev_priv, > > > + const struct skl_ddb_entry *entry); > > > void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc, > > > struct skl_pipe_wm *out); > > > void g4x_wm_sanitize(struct drm_i915_private *dev_priv); > > > -- > > > 2.24.1.485.gad05a3d8e5 > > > > -- > > Ville Syrjälä > > Intel -- Ville Syrjälä Intel _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx