On Mon, May 15, 2017 at 02:04:36PM +0530, Mahesh Kumar wrote: > From: "Kumar, Mahesh" <mahesh1.kumar@xxxxxxxxx> > > This patch implements new DDB allocation algorithm as per HW team > recommendation. This algo takecare of scenario where we allocate less DDB > for the planes with lower relative pixel rate, but they require more DDB > to work. > It also takes care of enabling same watermark level for each > plane in crtc, for efficient power saving. > > Changes since v1: > - Rebase on top of Paulo's patch series > > Changes since v2: > - Fix the for loop condition to enable WM > > Changes since v3: > - Fix crash in cursor i-g-t reported by Maarten > - Rebase after addressing Paulo's comments > - Few other ULT fixes > Changes since v4: > - Rebase on drm-tip > - Added separate function to enable WM levels > Changes since v5: > - Fix a crash identified in skl-6770HQ system > > Signed-off-by: Mahesh Kumar <mahesh1.kumar@xxxxxxxxx> > --- > drivers/gpu/drm/i915/intel_pm.c | 253 ++++++++++++++++++++++++---------------- > 1 file changed, 152 insertions(+), 101 deletions(-) > > diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c > index d73369c2c2d9..d6b0ae0ef7a2 100644 > --- a/drivers/gpu/drm/i915/intel_pm.c > +++ b/drivers/gpu/drm/i915/intel_pm.c > @@ -4039,13 +4039,41 @@ skl_ddb_calc_min(const struct intel_crtc_state *cstate, int num_active, > minimum[PLANE_CURSOR] = skl_cursor_allocation(num_active); > } > > +static void > +skl_enable_plane_wm_levels(const struct drm_i915_private *dev_priv, > + uint16_t plane_ddb, > + uint16_t max_level, > + struct skl_plane_wm *wm) > +{ > + int level; > + /* > + * Now enable all levels in WM structure which can be enabled > + * using current DDB allocation > + */ > + for (level = ilk_wm_max_level(dev_priv); level >= 0; level--) { > + struct skl_wm_level *level_wm = &wm->wm[level]; > + > + if (level > max_level || level_wm->plane_res_b == 0 > + || level_wm->plane_res_l >= 31 > + || level_wm->plane_res_b >= plane_ddb) { Is it possible to hit level_wm->plane_res_b >= plane_ddb without hitting level > max_level given our new logic? > + level_wm->plane_en = false; > + level_wm->plane_res_b = 0; > + level_wm->plane_res_l = 0; > + } else { > + level_wm->plane_en = true; > + } > + } > +} > + > static int > skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > - struct skl_ddb_allocation *ddb /* out */) > + struct skl_ddb_allocation *ddb /* out */, > + struct skl_pipe_wm *pipe_wm) Not a huge deal, but it's a bit confusing to have the 'out' parameter in the middle of 'in' parameters. Maybe move the pipe_wm before the ddb just to make this a little cleaner? > { > struct drm_atomic_state *state = cstate->base.state; > struct drm_crtc *crtc = cstate->base.crtc; > struct drm_device *dev = crtc->dev; > + struct drm_i915_private *dev_priv = to_i915(dev); > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > enum pipe pipe = intel_crtc->pipe; > struct skl_ddb_entry *alloc = &cstate->wm.skl.ddb; > @@ -4058,6 +4086,8 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > unsigned plane_data_rate[I915_MAX_PLANES] = {}; > unsigned plane_y_data_rate[I915_MAX_PLANES] = {}; > uint16_t total_min_blocks = 0; > + uint16_t total_level_ddb; > + int max_level, level; > > /* Clear the partitioning for disabled planes. */ > memset(ddb->plane[pipe], 0, sizeof(ddb->plane[pipe])); > @@ -4096,10 +4126,43 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > return -EINVAL; > } > > - alloc_size -= total_min_blocks; > - ddb->plane[pipe][PLANE_CURSOR].start = alloc->end - minimum[PLANE_CURSOR]; > + alloc_size -= minimum[PLANE_CURSOR]; > + ddb->plane[pipe][PLANE_CURSOR].start = alloc->end - > + minimum[PLANE_CURSOR]; > ddb->plane[pipe][PLANE_CURSOR].end = alloc->end; > > + for (level = ilk_wm_max_level(dev_priv); level >= 0; level--) { > + total_level_ddb = 0; > + for_each_plane_id_on_crtc(intel_crtc, plane_id) { > + /* > + * TODO: We should calculate watermark values for Y/UV > + * plane both in case of NV12 format and use both values > + * for ddb calculation. NV12 is disabled as of now, So > + * using only single/UV plane value here. > + */ > + struct skl_plane_wm *wm = &pipe_wm->planes[plane_id]; > + uint16_t plane_res_b = wm->wm[level].plane_res_b; > + uint16_t min = minimum[plane_id] + y_minimum[plane_id]; > + > + if (plane_id == PLANE_CURSOR) > + continue; > + > + total_level_ddb += max(plane_res_b, min); > + } > + > + if (total_level_ddb <= alloc_size) > + break; Since I got confused on my first time through the review, I suspect other people might make the same mistake I did. You might want to put a comment above this like "If this level can successfully be enabled with the pipe's current DDB allocation, then all lower levels are guaranteed to succeed as well." > + } > + > + if ((level < 0) || (total_min_blocks > alloc_size)) { > + DRM_DEBUG_KMS("Requested display configuration exceeds system DDB limitations"); > + DRM_DEBUG_KMS("minimum required %d/%d\n", (level < 0) ? > + total_level_ddb : total_min_blocks, alloc_size); > + return -EINVAL; > + } > + max_level = level; > + alloc_size -= total_level_ddb; > + > /* > * 2. Distribute the remaining space in proportion to the amount of > * data each plane needs to fetch from memory. > @@ -4115,10 +4178,17 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > start = alloc->start; > for_each_plane_id_on_crtc(intel_crtc, plane_id) { > unsigned int data_rate, y_data_rate; > - uint16_t plane_blocks, y_plane_blocks = 0; > - > - if (plane_id == PLANE_CURSOR) > + uint16_t plane_blocks = 0, y_plane_blocks = 0; > + struct skl_plane_wm *wm = &pipe_wm->planes[plane_id]; > + uint16_t plane_res_b = wm->wm[max_level].plane_res_b; > + > + if (plane_id == PLANE_CURSOR) { > + plane_blocks = > + skl_ddb_entry_size(&ddb->plane[pipe][plane_id]); > + skl_enable_plane_wm_levels(dev_priv, plane_blocks, > + max_level, wm); > continue; > + } > > data_rate = plane_data_rate[plane_id]; > > @@ -4127,33 +4197,36 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > * promote the expression to 64 bits to avoid overflowing, the > * result is < available as data_rate / total_data_rate < 1 > */ > - plane_blocks = minimum[plane_id]; > - plane_blocks += div_u64((uint64_t)alloc_size * data_rate, > - total_data_rate); > > /* Leave disabled planes at (0,0) */ > if (data_rate) { > + plane_blocks = max(minimum[plane_id], plane_res_b); > + plane_blocks += div_u64((uint64_t)alloc_size * > + data_rate, total_data_rate); > ddb->plane[pipe][plane_id].start = start; > ddb->plane[pipe][plane_id].end = start + plane_blocks; > + start += plane_blocks; > } > > - start += plane_blocks; > - > /* > * allocation for y_plane part of planar format: > + * TODO: Once we start calculating watermark values for Y/UV > + * plane both consider it for initial allowed wm blocks. > */ > y_data_rate = plane_y_data_rate[plane_id]; > > - y_plane_blocks = y_minimum[plane_id]; > - y_plane_blocks += div_u64((uint64_t)alloc_size * y_data_rate, > - total_data_rate); > - > if (y_data_rate) { > + y_plane_blocks = y_minimum[plane_id]; > + y_plane_blocks += div_u64((uint64_t)alloc_size * > + y_data_rate, total_data_rate); > ddb->y_plane[pipe][plane_id].start = start; > ddb->y_plane[pipe][plane_id].end = start + y_plane_blocks; > + start += y_plane_blocks; > } > - > - start += y_plane_blocks; > + skl_enable_plane_wm_levels(dev_priv, > + plane_blocks, > + max_level, > + wm); > } > > return 0; > @@ -4243,11 +4316,9 @@ skl_adjusted_plane_pixel_rate(const struct intel_crtc_state *cstate, > static int skl_compute_plane_wm(const struct drm_i915_private *dev_priv, > struct intel_crtc_state *cstate, > const struct intel_plane_state *intel_pstate, > - uint16_t ddb_allocation, > int level, > uint16_t *out_blocks, /* out */ > - uint8_t *out_lines, /* out */ > - bool *enabled /* out */) > + uint8_t *out_lines /* out */) > { > struct intel_plane *plane = to_intel_plane(intel_pstate->base.plane); > const struct drm_plane_state *pstate = &intel_pstate->base; > @@ -4270,10 +4341,8 @@ static int skl_compute_plane_wm(const struct drm_i915_private *dev_priv, > bool y_tiled, x_tiled; > > if (latency == 0 || > - !intel_wm_plane_visible(cstate, intel_pstate)) { > - *enabled = false; > + !intel_wm_plane_visible(cstate, intel_pstate)) > return 0; > - } > > y_tiled = fb->modifier == I915_FORMAT_MOD_Y_TILED || > fb->modifier == I915_FORMAT_MOD_Yf_TILED; > @@ -4359,9 +4428,6 @@ static int skl_compute_plane_wm(const struct drm_i915_private *dev_priv, > if ((cpp * cstate->base.adjusted_mode.crtc_htotal / 512 < 1) && > (plane_bytes_per_line / 512 < 1)) > selected_result = method2; > - else if ((ddb_allocation && ddb_allocation / > - fixed_16_16_to_u32_round_up(plane_blocks_per_line)) >= 1) > - selected_result = min_fixed_16_16(method1, method2); > else if (latency >= linetime_us) > selected_result = method2; > else > @@ -4381,64 +4447,41 @@ static int skl_compute_plane_wm(const struct drm_i915_private *dev_priv, > } > } > > - if (res_blocks >= ddb_allocation || res_lines > 31) { > - *enabled = false; > - > - /* > - * If there are no valid level 0 watermarks, then we can't > - * support this display configuration. > - */ > - if (level) { > - return 0; > - } else { > - struct drm_plane *plane = pstate->plane; > + if (res_lines >= 31 && level == 0) { > + struct drm_plane *plane = pstate->plane; > > - DRM_DEBUG_KMS("Requested display configuration exceeds system watermark limitations\n"); > - DRM_DEBUG_KMS("[PLANE:%d:%s] blocks required = %u/%u, lines required = %u/31\n", > - plane->base.id, plane->name, > - res_blocks, ddb_allocation, res_lines); > - return -EINVAL; > - } > + DRM_DEBUG_KMS("Requested display configuration exceeds system watermark limitations\n"); > + DRM_DEBUG_KMS("[PLANE:%d:%s] lines required = %u/31\n", > + plane->base.id, plane->name, res_lines); Should this still be returning an EINVAL? It looks like we'll print the message here, but then continue on and disable watermark level 0 in skl_enable_plane_wm_levels() without failing the commit. > } > > *out_blocks = res_blocks; > *out_lines = res_lines; > - *enabled = true; > > return 0; > } > > static int > skl_compute_wm_levels(const struct drm_i915_private *dev_priv, > - struct skl_ddb_allocation *ddb, > struct intel_crtc_state *cstate, > const struct intel_plane_state *intel_pstate, > struct skl_plane_wm *wm) > { > - struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc); > - struct drm_plane *plane = intel_pstate->base.plane; > - struct intel_plane *intel_plane = to_intel_plane(plane); > - uint16_t ddb_blocks; > - enum pipe pipe = intel_crtc->pipe; > int level, max_level = ilk_wm_max_level(dev_priv); > int ret; > > if (WARN_ON(!intel_pstate->base.fb)) > return -EINVAL; > > - ddb_blocks = skl_ddb_entry_size(&ddb->plane[pipe][intel_plane->id]); > - > for (level = 0; level <= max_level; level++) { > struct skl_wm_level *result = &wm->wm[level]; > > ret = skl_compute_plane_wm(dev_priv, > cstate, > intel_pstate, > - ddb_blocks, > level, > &result->plane_res_b, > - &result->plane_res_l, > - &result->plane_en); > + &result->plane_res_l); > if (ret) > return ret; > } > @@ -4504,8 +4547,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate, > > wm = &pipe_wm->planes[plane_id]; > > - ret = skl_compute_wm_levels(dev_priv, ddb, cstate, > - intel_pstate, wm); > + ret = skl_compute_wm_levels(dev_priv, cstate, intel_pstate, wm); > if (ret) > return ret; > skl_compute_transition_wm(cstate, &wm->trans_wm); > @@ -4618,6 +4660,45 @@ bool skl_ddb_allocation_overlaps(const struct skl_ddb_entry **entries, > return false; > } > > +static int > +skl_ddb_add_affected_planes(struct intel_crtc_state *cstate, > + const struct skl_pipe_wm *old_pipe_wm, > + const struct skl_pipe_wm *pipe_wm) > +{ > + struct drm_atomic_state *state = cstate->base.state; > + struct drm_device *dev = state->dev; > + struct drm_crtc *crtc = cstate->base.crtc; > + struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > + struct drm_i915_private *dev_priv = to_i915(dev); > + struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > + struct skl_ddb_allocation *new_ddb = &intel_state->wm_results.ddb; > + struct skl_ddb_allocation *cur_ddb = &dev_priv->wm.skl_hw.ddb; > + struct drm_plane_state *plane_state; > + struct drm_plane *plane; > + enum pipe pipe = intel_crtc->pipe; > + > + WARN_ON(!drm_atomic_get_existing_crtc_state(state, crtc)); > + > + drm_for_each_plane_mask(plane, dev, cstate->base.plane_mask) { > + enum plane_id plane_id = to_intel_plane(plane)->id; > + const struct skl_plane_wm *wm = &pipe_wm->planes[plane_id]; > + const struct skl_plane_wm *old_wm = &old_pipe_wm->planes[plane_id]; > + > + if ((skl_ddb_entry_equal(&cur_ddb->plane[pipe][plane_id], > + &new_ddb->plane[pipe][plane_id]) && > + skl_ddb_entry_equal(&cur_ddb->y_plane[pipe][plane_id], > + &new_ddb->y_plane[pipe][plane_id])) && > + !memcmp(wm, old_wm, sizeof(struct skl_plane_wm))) > + continue; > + > + plane_state = drm_atomic_get_plane_state(state, plane); > + if (IS_ERR(plane_state)) > + return PTR_ERR(plane_state); > + } > + > + return 0; > +} > + > static int skl_update_pipe_wm(struct drm_crtc_state *cstate, > const struct skl_pipe_wm *old_pipe_wm, > struct skl_pipe_wm *pipe_wm, /* out */ > @@ -4631,6 +4712,18 @@ static int skl_update_pipe_wm(struct drm_crtc_state *cstate, > if (ret) > return ret; > > + ret = skl_allocate_pipe_ddb(intel_cstate, ddb, pipe_wm); > + if (ret) > + return ret; > + /* > + * TODO: Do we still need to add planes in state, As WM update is > + * not part of update_plane anymore, So wm for planes can be updated > + * irrespective of updade_plane call. > + */ I think the watermark registers get written by skl_write_plane_wm() regardless of whether we add the plane to the state, yes. However I think the watermark registers are only armed by writing to the actual plane registers, which I don't believe will happen anywhere unless you add the plane to the state. > + ret = skl_ddb_add_affected_planes(intel_cstate, old_pipe_wm, pipe_wm); > + if (ret) > + return ret; > + > if (!memcmp(old_pipe_wm, pipe_wm, sizeof(*pipe_wm))) > *changed = false; > else > @@ -4653,41 +4746,7 @@ pipes_modified(struct drm_atomic_state *state) > } > > static int > -skl_ddb_add_affected_planes(struct intel_crtc_state *cstate) > -{ > - struct drm_atomic_state *state = cstate->base.state; > - struct drm_device *dev = state->dev; > - struct drm_crtc *crtc = cstate->base.crtc; > - struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > - struct drm_i915_private *dev_priv = to_i915(dev); > - struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > - struct skl_ddb_allocation *new_ddb = &intel_state->wm_results.ddb; > - struct skl_ddb_allocation *cur_ddb = &dev_priv->wm.skl_hw.ddb; > - struct drm_plane_state *plane_state; > - struct drm_plane *plane; > - enum pipe pipe = intel_crtc->pipe; > - > - WARN_ON(!drm_atomic_get_existing_crtc_state(state, crtc)); > - > - drm_for_each_plane_mask(plane, dev, cstate->base.plane_mask) { > - enum plane_id plane_id = to_intel_plane(plane)->id; > - > - if (skl_ddb_entry_equal(&cur_ddb->plane[pipe][plane_id], > - &new_ddb->plane[pipe][plane_id]) && > - skl_ddb_entry_equal(&cur_ddb->y_plane[pipe][plane_id], > - &new_ddb->y_plane[pipe][plane_id])) > - continue; > - > - plane_state = drm_atomic_get_plane_state(state, plane); > - if (IS_ERR(plane_state)) > - return PTR_ERR(plane_state); > - } > - > - return 0; > -} > - > -static int > -skl_compute_ddb(struct drm_atomic_state *state) > +skl_include_affected_crtc(struct drm_atomic_state *state) Minor nitpick, but I'd make this plural as well. skl_include_affected_crtcs(). Matt > { > struct drm_device *dev = state->dev; > struct drm_i915_private *dev_priv = to_i915(dev); > @@ -4751,14 +4810,6 @@ skl_compute_ddb(struct drm_atomic_state *state) > cstate = intel_atomic_get_crtc_state(state, intel_crtc); > if (IS_ERR(cstate)) > return PTR_ERR(cstate); > - > - ret = skl_allocate_pipe_ddb(cstate, ddb); > - if (ret) > - return ret; > - > - ret = skl_ddb_add_affected_planes(cstate); > - if (ret) > - return ret; > } > > return 0; > @@ -4839,7 +4890,7 @@ skl_compute_wm(struct drm_atomic_state *state) > /* Clear all dirty flags */ > results->dirty_pipes = 0; > > - ret = skl_compute_ddb(state); > + ret = skl_include_affected_crtc(state); > if (ret) > return ret; > > -- > 2.11.0 > -- Matt Roper Graphics Software Engineer IoTG Platform Enabling & Development Intel Corporation (916) 356-2795 _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx