[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <BN0PR02MB8142049D9867BFD4CBC9B285967A9@BN0PR02MB8142.namprd02.prod.outlook.com>
Date: Fri, 2 Sep 2022 09:35:24 +0000
From: Kalyan Thota <kalyant@....qualcomm.com>
To: "dmitry.baryshkov@...aro.org" <dmitry.baryshkov@...aro.org>,
"Kalyan Thota (QUIC)" <quic_kalyant@...cinc.com>,
"dianders@...omium.org" <dianders@...omium.org>
CC: "dri-devel@...ts.freedesktop.org" <dri-devel@...ts.freedesktop.org>,
"linux-arm-msm@...r.kernel.org" <linux-arm-msm@...r.kernel.org>,
"freedreno@...ts.freedesktop.org" <freedreno@...ts.freedesktop.org>,
"devicetree@...r.kernel.org" <devicetree@...r.kernel.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"robdclark@...il.com" <robdclark@...il.com>,
"swboyd@...omium.org" <swboyd@...omium.org>,
"Vinod Polimera (QUIC)" <quic_vpolimer@...cinc.com>,
"Abhinav Kumar (QUIC)" <quic_abhinavk@...cinc.com>
Subject: RE: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for
dspp in sc7280
>-----Original Message-----
>From: Dmitry Baryshkov <dmitry.baryshkov@...aro.org>
>Sent: Friday, August 26, 2022 5:02 PM
>To: Kalyan Thota <kalyant@....qualcomm.com>; Kalyan Thota (QUIC)
><quic_kalyant@...cinc.com>; dianders@...omium.org
>Cc: dri-devel@...ts.freedesktop.org; linux-arm-msm@...r.kernel.org;
>freedreno@...ts.freedesktop.org; devicetree@...r.kernel.org; linux-
>kernel@...r.kernel.org; robdclark@...il.com; swboyd@...omium.org; Vinod
>Polimera (QUIC) <quic_vpolimer@...cinc.com>; Abhinav Kumar (QUIC)
><quic_abhinavk@...cinc.com>
>Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp
>in sc7280
>
>WARNING: This email originated from outside of Qualcomm. Please be wary of
>any links or attachments, and do not enable macros.
>
>On 08/08/2022 13:44, Kalyan Thota wrote:
>>
>>
>>> -----Original Message-----
>>> From: Dmitry Baryshkov <dmitry.baryshkov@...aro.org>
>>> Sent: Thursday, August 4, 2022 9:29 PM
>>> To: Kalyan Thota (QUIC) <quic_kalyant@...cinc.com>
>>> Cc: dri-devel@...ts.freedesktop.org; linux-arm-msm@...r.kernel.org;
>>> freedreno@...ts.freedesktop.org; devicetree@...r.kernel.org; linux-
>>> kernel@...r.kernel.org; robdclark@...il.com; dianders@...omium.org;
>>> swboyd@...omium.org; Vinod Polimera (QUIC)
>>> <quic_vpolimer@...cinc.com>; Abhinav Kumar (QUIC)
>>> <quic_abhinavk@...cinc.com>
>>> Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical
>>> flush for dspp in sc7280
>>>
>>> WARNING: This email originated from outside of Qualcomm. Please be
>>> wary of any links or attachments, and do not enable macros.
>>>
>>> On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@...cinc.com>
>wrote:
>>>>
>>>> Flush mechanism for DSPP blocks has changed in sc7280 family, it
>>>> allows individual sub blocks to be flushed in coordination with
>>>> master flush control.
>>>>
>>>> representation: master_flush && (PCC_flush | IGC_flush .. etc )
>>>>
>>>> This change adds necessary support for the above design.
>>>>
>>>> Signed-off-by: Kalyan Thota <quic_kalyant@...cinc.com>
>>>
>>> I'd like to land at least patches 6-8 from [1] next cycle. They clean
>>> up the CTL interface. Could you please rebase your patch on top of them?
>>>
>>
>> Sure I'll wait for the series to rebase. @Doug can you comment if this is okay
>and this patch is not needed immediately ?
>
>The respective patches have been picked up for 6.1 and were pushed to
>https://gitlab.freedesktop.org/lumag/msm.git msm-next-lumag . Could you
>please rebase your patch on top of them?
>
>All other comments also needs addressing.
>
>>
>>> [1] https://patchwork.freedesktop.org/series/99909/
>>>
>>>> ---
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++-
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40
>>> +++++++++++++++++++++++++-
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++
>>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++
>>>> 6 files changed, 59 insertions(+), 2 deletions(-)
>>>>
>>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>>>> index 7763558..4eca317 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>>>> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct
>>> drm_crtc *crtc)
>>>> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl,
>>>> mixer[i].hw_dspp->idx);
>>>>
>>>> + if(ctl->ops.set_dspp_hierarchical_flush)
>>>> + ctl->ops.set_dspp_hierarchical_flush(ctl,
>>>> +
>>>> + mixer[i].hw_dspp->idx, DSPP_SUB_PCC);
>>>> +
>>>> /* stage config flush mask */
>>>> ctl->ops.update_pending_flush(ctl,
>>>> mixer[i].flush_mask);
>>>>
>>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>>>> index 021eb2f..3b27a87 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>>>> @@ -58,7 +58,10 @@
>>>> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2))
>>>>
>>>> #define CTL_SC7280_MASK \
>>>> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) |
>>> BIT(DPU_CTL_VM_CFG))
>>>> + (BIT(DPU_CTL_ACTIVE_CFG) | \
>>>> + BIT(DPU_CTL_FETCH_ACTIVE) | \
>>>> + BIT(DPU_CTL_VM_CFG) | \
>>>> + BIT(DPU_CTL_HIERARCHICAL_FLUSH))
>>>>
>>>> #define MERGE_3D_SM8150_MASK (0)
>>>>
>>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>>>> index b85b24b..7922f6c 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>>>> @@ -185,6 +185,7 @@ enum {
>>>> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display
>>>> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs)
>>>> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs
>>>> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical
>>>> + flush
>>>> * @DPU_CTL_MAX
>>>> */
>>>> enum {
>>>> @@ -192,6 +193,7 @@ enum {
>>>> DPU_CTL_ACTIVE_CFG,
>>>> DPU_CTL_FETCH_ACTIVE,
>>>> DPU_CTL_VM_CFG,
>>>> + DPU_CTL_HIERARCHICAL_FLUSH,
>>>> DPU_CTL_MAX
>>>> };
>>>>
>>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>>>> index 3584f5e..b34fc30 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>>>> @@ -28,6 +28,8 @@
>>>> #define CTL_INTF_FLUSH 0x110
>>>> #define CTL_INTF_MASTER 0x134
>>>> #define CTL_FETCH_PIPE_ACTIVE 0x0FC
>>>> +#define CTL_DSPP_0_FLUSH 0x13C
>>>
>>> Please change to CTL_DSPP_n_FLUSH(n).
>>>
>>>> +
>>>>
>>>> #define CTL_MIXER_BORDER_OUT BIT(24)
>>>> #define CTL_FLUSH_MASK_CTL BIT(17)
>>>> @@ -292,6 +294,36 @@ static uint32_t
>>>> dpu_hw_ctl_get_bitmask_dspp(struct
>>> dpu_hw_ctl *ctx,
>>>> return flushbits;
>>>> }
>>>>
>>>> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx,
>>>> + enum dpu_dspp dspp)
>>>> +{
>>>> + return BIT(29);
>>>> +}
>>>> +
>>>> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx,
>>>> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) {
>>>> + uint32_t flushbits = 0, active = 0;
>>>> +
>>>> + switch (dspp_sub_blk) {
>>>> + case DSPP_SUB_IGC:
>>>> + flushbits = BIT(2);
>>>> + break;
>>>> + case DSPP_SUB_PCC:
>>>> + flushbits = BIT(4);
>>>> + break;
>>>> + case DSPP_SUB_GC:
>>>> + flushbits = BIT(5);
>>>> + break;
>>>> + default:
>>>> + return;
>>>> + }
>>>> +
>>>> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp -
>>>> + 1)
>>>> + * 4));
>>>
>>> So that this line will be simpler to read.
>>>
>>>> +
>>>> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4),
>>>> +active | flushbits); }
>>>> +
>>>> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx,
>>>> u32
>>>> timeout_us) {
>>>> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13
>>>> @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops,
>>>> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage;
>>>> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp;
>>>> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer;
>>>> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp;
>>>> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) {
>>>> + ops->get_bitmask_dspp =
>>>> + dpu_hw_ctl_get_bitmask_dspp_v1;
>>>
>>> We have used _v1 for active CTLs. What is the relationship between
>>> CTL_HIERARCHILCAL_FLUSH and active CTLs?
>> Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers
>> in grouping the resources such as WB, INTF, pingpong, DSC etc into the
>> data path DSPP hierarchical flush will gives us a finer control on which post
>processing blocks to be flushed as part of the composition ( like IGC, PCC, GC ..
>etc ) These blocks are contained in DSPP package.
>
>So, I assume that hierarchical DSPP flush does not exist on non-active CTL SoCs.
>Which supported SoCs do support the hierarchichal DSPP flush?
>
Dspp Sub-block flush is supported from the DPU 7-series, the only target in the catalogue is sc7280.
>>>
>>>> + ops->set_dspp_hierarchical_flush =
>>> dpu_hw_ctl_set_dspp_hierarchical_flush;
>>>> + } else {
>>>> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp;
>>>> + }
>>>> +
>>>> if (cap & BIT(DPU_CTL_FETCH_ACTIVE))
>>>> ops->set_active_pipes =
>>>> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git
>>>> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>>>> index ac15444..8ecab91 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>>>> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops {
>>>> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx,
>>>> enum dpu_dspp blk);
>>>>
>>>> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx,
>>>> + enum dpu_dspp blk, enum dpu_dspp_sub_blk
>>>> + dspp_sub_blk);
>>>
>>> The word "hierarchical" means particular (internal) implementation.
>>> Please change to something like set_dspp_block_flush().
>>> Or with [2] in place, it can be hidden in the
>>> update_pending_flush_dspp() function. Just pass the subblock to the
>>> function and let the dpu_hw_ctl care about it.
>>>
>>> [2]
>>> https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1
>>>
>>>
>>>> +
>>>> /**
>>>> * Set all blend stages to disabled
>>>> * @ctx : ctl path ctx pointer
>>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>>>> index bb9cead..561e2ab 100644
>>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>>>> @@ -166,6 +166,13 @@ enum dpu_dspp {
>>>> DSPP_MAX
>>>> };
>>>>
>>>> +enum dpu_dspp_sub_blk{
>>>> + DSPP_SUB_PCC = 1,
>>>> + DSPP_SUB_IGC,
>>>> + DSPP_SUB_GC,
>>>> + DSPP_SUB_MAX
>>>> +};
>>>
>>> I'd prefer if we can use DPU_DSPP_* definitions instead.
>>>
>>>> +
>>>> enum dpu_ctl {
>>>> CTL_0 = 1,
>>>> CTL_1,
>>>
>>>
>>>
>>> --
>>> With best wishes
>>> Dmitry
>
>--
>With best wishes
>Dmitry
Powered by blists - more mailing lists