>-----Original Message----- >From: Dmitry Baryshkov <dmitry.baryshkov@xxxxxxxxxx> >Sent: Thursday, August 4, 2022 9:29 PM >To: Kalyan Thota (QUIC) <quic_kalyant@xxxxxxxxxxx> >Cc: dri-devel@xxxxxxxxxxxxxxxxxxxxx; linux-arm-msm@xxxxxxxxxxxxxxx; >freedreno@xxxxxxxxxxxxxxxxxxxxx; devicetree@xxxxxxxxxxxxxxx; linux- >kernel@xxxxxxxxxxxxxxx; robdclark@xxxxxxxxx; dianders@xxxxxxxxxxxx; >swboyd@xxxxxxxxxxxx; Vinod Polimera (QUIC) <quic_vpolimer@xxxxxxxxxxx>; >Abhinav Kumar (QUIC) <quic_abhinavk@xxxxxxxxxxx> >Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp >in sc7280 > >WARNING: This email originated from outside of Qualcomm. Please be wary of >any links or attachments, and do not enable macros. > >On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@xxxxxxxxxxx> wrote: >> >> Flush mechanism for DSPP blocks has changed in sc7280 family, it >> allows individual sub blocks to be flushed in coordination with master >> flush control. >> >> representation: master_flush && (PCC_flush | IGC_flush .. etc ) >> >> This change adds necessary support for the above design. >> >> Signed-off-by: Kalyan Thota <quic_kalyant@xxxxxxxxxxx> > >I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL >interface. Could you please rebase your patch on top of them? > Sure I'll wait for the series to rebase. @Doug can you comment if this is okay and this patch is not needed immediately ? >[1] https://patchwork.freedesktop.org/series/99909/ > >> --- >> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 >+++++++++++++++++++++++++- >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ >> 6 files changed, 59 insertions(+), 2 deletions(-) >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> index 7763558..4eca317 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct >drm_crtc *crtc) >> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, >> mixer[i].hw_dspp->idx); >> >> + if(ctl->ops.set_dspp_hierarchical_flush) >> + ctl->ops.set_dspp_hierarchical_flush(ctl, >> + mixer[i].hw_dspp->idx, >> + DSPP_SUB_PCC); >> + >> /* stage config flush mask */ >> ctl->ops.update_pending_flush(ctl, >> mixer[i].flush_mask); >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> index 021eb2f..3b27a87 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> @@ -58,7 +58,10 @@ >> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) >> >> #define CTL_SC7280_MASK \ >> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | >BIT(DPU_CTL_VM_CFG)) >> + (BIT(DPU_CTL_ACTIVE_CFG) | \ >> + BIT(DPU_CTL_FETCH_ACTIVE) | \ >> + BIT(DPU_CTL_VM_CFG) | \ >> + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) >> >> #define MERGE_3D_SM8150_MASK (0) >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> index b85b24b..7922f6c 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> @@ -185,6 +185,7 @@ enum { >> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display >> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) >> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs >> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical >> + flush >> * @DPU_CTL_MAX >> */ >> enum { >> @@ -192,6 +193,7 @@ enum { >> DPU_CTL_ACTIVE_CFG, >> DPU_CTL_FETCH_ACTIVE, >> DPU_CTL_VM_CFG, >> + DPU_CTL_HIERARCHICAL_FLUSH, >> DPU_CTL_MAX >> }; >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> index 3584f5e..b34fc30 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> @@ -28,6 +28,8 @@ >> #define CTL_INTF_FLUSH 0x110 >> #define CTL_INTF_MASTER 0x134 >> #define CTL_FETCH_PIPE_ACTIVE 0x0FC >> +#define CTL_DSPP_0_FLUSH 0x13C > >Please change to CTL_DSPP_n_FLUSH(n). > >> + >> >> #define CTL_MIXER_BORDER_OUT BIT(24) >> #define CTL_FLUSH_MASK_CTL BIT(17) >> @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct >dpu_hw_ctl *ctx, >> return flushbits; >> } >> >> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp dspp) >> +{ >> + return BIT(29); >> +} >> + >> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) { >> + uint32_t flushbits = 0, active = 0; >> + >> + switch (dspp_sub_blk) { >> + case DSPP_SUB_IGC: >> + flushbits = BIT(2); >> + break; >> + case DSPP_SUB_PCC: >> + flushbits = BIT(4); >> + break; >> + case DSPP_SUB_GC: >> + flushbits = BIT(5); >> + break; >> + default: >> + return; >> + } >> + >> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) >> + * 4)); > >So that this line will be simpler to read. > >> + >> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), >> +active | flushbits); } >> + >> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32 >> timeout_us) { >> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 @@ >> static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, >> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; >> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; >> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; >> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { >> + ops->get_bitmask_dspp = >> + dpu_hw_ctl_get_bitmask_dspp_v1; > >We have used _v1 for active CTLs. What is the relationship between >CTL_HIERARCHILCAL_FLUSH and active CTLs? Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers in grouping the resources such as WB, INTF, pingpong, DSC etc into the data path DSPP hierarchical flush will gives us a finer control on which post processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. etc ) These blocks are contained in DSPP package. > >> + ops->set_dspp_hierarchical_flush = >dpu_hw_ctl_set_dspp_hierarchical_flush; >> + } else { >> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >> + } >> + >> if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) >> ops->set_active_pipes = >> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git >> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> index ac15444..8ecab91 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { >> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, >> enum dpu_dspp blk); >> >> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp blk, enum dpu_dspp_sub_blk >> + dspp_sub_blk); > >The word "hierarchical" means particular (internal) implementation. >Please change to something like set_dspp_block_flush(). >Or with [2] in place, it can be hidden in the >update_pending_flush_dspp() function. Just pass the subblock to the function and >let the dpu_hw_ctl care about it. > >[2] https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 > > >> + >> /** >> * Set all blend stages to disabled >> * @ctx : ctl path ctx pointer >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> index bb9cead..561e2ab 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> @@ -166,6 +166,13 @@ enum dpu_dspp { >> DSPP_MAX >> }; >> >> +enum dpu_dspp_sub_blk{ >> + DSPP_SUB_PCC = 1, >> + DSPP_SUB_IGC, >> + DSPP_SUB_GC, >> + DSPP_SUB_MAX >> +}; > >I'd prefer if we can use DPU_DSPP_* definitions instead. > >> + >> enum dpu_ctl { >> CTL_0 = 1, >> CTL_1, > > > >-- >With best wishes >Dmitry