>-----Original Message----- >From: Dmitry Baryshkov <dmitry.baryshkov@xxxxxxxxxx> >Sent: Friday, August 26, 2022 5:02 PM >To: Kalyan Thota <kalyant@xxxxxxxxxxxxxxxx>; Kalyan Thota (QUIC) ><quic_kalyant@xxxxxxxxxxx>; dianders@xxxxxxxxxxxx >Cc: dri-devel@xxxxxxxxxxxxxxxxxxxxx; linux-arm-msm@xxxxxxxxxxxxxxx; >freedreno@xxxxxxxxxxxxxxxxxxxxx; devicetree@xxxxxxxxxxxxxxx; linux- >kernel@xxxxxxxxxxxxxxx; robdclark@xxxxxxxxx; swboyd@xxxxxxxxxxxx; Vinod >Polimera (QUIC) <quic_vpolimer@xxxxxxxxxxx>; Abhinav Kumar (QUIC) ><quic_abhinavk@xxxxxxxxxxx> >Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp >in sc7280 > >WARNING: This email originated from outside of Qualcomm. Please be wary of >any links or attachments, and do not enable macros. > >On 08/08/2022 13:44, Kalyan Thota wrote: >> >> >>> -----Original Message----- >>> From: Dmitry Baryshkov <dmitry.baryshkov@xxxxxxxxxx> >>> Sent: Thursday, August 4, 2022 9:29 PM >>> To: Kalyan Thota (QUIC) <quic_kalyant@xxxxxxxxxxx> >>> Cc: dri-devel@xxxxxxxxxxxxxxxxxxxxx; linux-arm-msm@xxxxxxxxxxxxxxx; >>> freedreno@xxxxxxxxxxxxxxxxxxxxx; devicetree@xxxxxxxxxxxxxxx; linux- >>> kernel@xxxxxxxxxxxxxxx; robdclark@xxxxxxxxx; dianders@xxxxxxxxxxxx; >>> swboyd@xxxxxxxxxxxx; Vinod Polimera (QUIC) >>> <quic_vpolimer@xxxxxxxxxxx>; Abhinav Kumar (QUIC) >>> <quic_abhinavk@xxxxxxxxxxx> >>> Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical >>> flush for dspp in sc7280 >>> >>> WARNING: This email originated from outside of Qualcomm. Please be >>> wary of any links or attachments, and do not enable macros. >>> >>> On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@xxxxxxxxxxx> >wrote: >>>> >>>> Flush mechanism for DSPP blocks has changed in sc7280 family, it >>>> allows individual sub blocks to be flushed in coordination with >>>> master flush control. >>>> >>>> representation: master_flush && (PCC_flush | IGC_flush .. etc ) >>>> >>>> This change adds necessary support for the above design. >>>> >>>> Signed-off-by: Kalyan Thota <quic_kalyant@xxxxxxxxxxx> >>> >>> I'd like to land at least patches 6-8 from [1] next cycle. They clean >>> up the CTL interface. Could you please rebase your patch on top of them? >>> >> >> Sure I'll wait for the series to rebase. @Doug can you comment if this is okay >and this patch is not needed immediately ? > >The respective patches have been picked up for 6.1 and were pushed to >https://gitlab.freedesktop.org/lumag/msm.git msm-next-lumag . Could you >please rebase your patch on top of them? > >All other comments also needs addressing. > >> >>> [1] https://patchwork.freedesktop.org/series/99909/ >>> >>>> --- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 >>> +++++++++++++++++++++++++- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ >>>> 6 files changed, 59 insertions(+), 2 deletions(-) >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> index 7763558..4eca317 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct >>> drm_crtc *crtc) >>>> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, >>>> mixer[i].hw_dspp->idx); >>>> >>>> + if(ctl->ops.set_dspp_hierarchical_flush) >>>> + ctl->ops.set_dspp_hierarchical_flush(ctl, >>>> + >>>> + mixer[i].hw_dspp->idx, DSPP_SUB_PCC); >>>> + >>>> /* stage config flush mask */ >>>> ctl->ops.update_pending_flush(ctl, >>>> mixer[i].flush_mask); >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> index 021eb2f..3b27a87 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> @@ -58,7 +58,10 @@ >>>> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) >>>> >>>> #define CTL_SC7280_MASK \ >>>> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | >>> BIT(DPU_CTL_VM_CFG)) >>>> + (BIT(DPU_CTL_ACTIVE_CFG) | \ >>>> + BIT(DPU_CTL_FETCH_ACTIVE) | \ >>>> + BIT(DPU_CTL_VM_CFG) | \ >>>> + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) >>>> >>>> #define MERGE_3D_SM8150_MASK (0) >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> index b85b24b..7922f6c 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> @@ -185,6 +185,7 @@ enum { >>>> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display >>>> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) >>>> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs >>>> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical >>>> + flush >>>> * @DPU_CTL_MAX >>>> */ >>>> enum { >>>> @@ -192,6 +193,7 @@ enum { >>>> DPU_CTL_ACTIVE_CFG, >>>> DPU_CTL_FETCH_ACTIVE, >>>> DPU_CTL_VM_CFG, >>>> + DPU_CTL_HIERARCHICAL_FLUSH, >>>> DPU_CTL_MAX >>>> }; >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> index 3584f5e..b34fc30 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> @@ -28,6 +28,8 @@ >>>> #define CTL_INTF_FLUSH 0x110 >>>> #define CTL_INTF_MASTER 0x134 >>>> #define CTL_FETCH_PIPE_ACTIVE 0x0FC >>>> +#define CTL_DSPP_0_FLUSH 0x13C >>> >>> Please change to CTL_DSPP_n_FLUSH(n). >>> >>>> + >>>> >>>> #define CTL_MIXER_BORDER_OUT BIT(24) >>>> #define CTL_FLUSH_MASK_CTL BIT(17) >>>> @@ -292,6 +294,36 @@ static uint32_t >>>> dpu_hw_ctl_get_bitmask_dspp(struct >>> dpu_hw_ctl *ctx, >>>> return flushbits; >>>> } >>>> >>>> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp dspp) >>>> +{ >>>> + return BIT(29); >>>> +} >>>> + >>>> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) { >>>> + uint32_t flushbits = 0, active = 0; >>>> + >>>> + switch (dspp_sub_blk) { >>>> + case DSPP_SUB_IGC: >>>> + flushbits = BIT(2); >>>> + break; >>>> + case DSPP_SUB_PCC: >>>> + flushbits = BIT(4); >>>> + break; >>>> + case DSPP_SUB_GC: >>>> + flushbits = BIT(5); >>>> + break; >>>> + default: >>>> + return; >>>> + } >>>> + >>>> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - >>>> + 1) >>>> + * 4)); >>> >>> So that this line will be simpler to read. >>> >>>> + >>>> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), >>>> +active | flushbits); } >>>> + >>>> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, >>>> u32 >>>> timeout_us) { >>>> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 >>>> @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, >>>> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; >>>> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; >>>> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; >>>> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>>> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { >>>> + ops->get_bitmask_dspp = >>>> + dpu_hw_ctl_get_bitmask_dspp_v1; >>> >>> We have used _v1 for active CTLs. What is the relationship between >>> CTL_HIERARCHILCAL_FLUSH and active CTLs? >> Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers >> in grouping the resources such as WB, INTF, pingpong, DSC etc into the >> data path DSPP hierarchical flush will gives us a finer control on which post >processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. >etc ) These blocks are contained in DSPP package. > >So, I assume that hierarchical DSPP flush does not exist on non-active CTL SoCs. >Which supported SoCs do support the hierarchichal DSPP flush? > Dspp Sub-block flush is supported from the DPU 7-series, the only target in the catalogue is sc7280. >>> >>>> + ops->set_dspp_hierarchical_flush = >>> dpu_hw_ctl_set_dspp_hierarchical_flush; >>>> + } else { >>>> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>>> + } >>>> + >>>> if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) >>>> ops->set_active_pipes = >>>> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git >>>> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> index ac15444..8ecab91 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { >>>> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, >>>> enum dpu_dspp blk); >>>> >>>> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp blk, enum dpu_dspp_sub_blk >>>> + dspp_sub_blk); >>> >>> The word "hierarchical" means particular (internal) implementation. >>> Please change to something like set_dspp_block_flush(). >>> Or with [2] in place, it can be hidden in the >>> update_pending_flush_dspp() function. Just pass the subblock to the >>> function and let the dpu_hw_ctl care about it. >>> >>> [2] >>> https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 >>> >>> >>>> + >>>> /** >>>> * Set all blend stages to disabled >>>> * @ctx : ctl path ctx pointer >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> index bb9cead..561e2ab 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> @@ -166,6 +166,13 @@ enum dpu_dspp { >>>> DSPP_MAX >>>> }; >>>> >>>> +enum dpu_dspp_sub_blk{ >>>> + DSPP_SUB_PCC = 1, >>>> + DSPP_SUB_IGC, >>>> + DSPP_SUB_GC, >>>> + DSPP_SUB_MAX >>>> +}; >>> >>> I'd prefer if we can use DPU_DSPP_* definitions instead. >>> >>>> + >>>> enum dpu_ctl { >>>> CTL_0 = 1, >>>> CTL_1, >>> >>> >>> >>> -- >>> With best wishes >>> Dmitry > >-- >With best wishes >Dmitry