[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAE=gft7HObn4Q-6P_pd1QbpWJG-yvs4C87qpiQm6BbmHzwZW5w@mail.gmail.com>
Date: Fri, 9 Aug 2019 14:17:38 -0700
From: Evan Green <evgreen@...omium.org>
To: Georgi Djakov <georgi.djakov@...aro.org>
Cc: linux-pm@...r.kernel.org, David Dai <daidavid1@...eaurora.org>,
Vincent Guittot <vincent.guittot@...aro.org>,
Bjorn Andersson <bjorn.andersson@...aro.org>,
amit.kucheria@...aro.org, Doug Anderson <dianders@...omium.org>,
Sean Sweeney <seansw@....qualcomm.com>,
LKML <linux-kernel@...r.kernel.org>,
linux-arm Mailing List <linux-arm-kernel@...ts.infradead.org>,
linux-arm-msm <linux-arm-msm@...r.kernel.org>
Subject: Re: [PATCH v3 3/3] interconnect: qcom: Add tagging and wake/sleep
support for sdm845
On Fri, Aug 9, 2019 at 5:13 AM Georgi Djakov <georgi.djakov@...aro.org> wrote:
>
> From: David Dai <daidavid1@...eaurora.org>
>
> Add support for wake and sleep commands by using a tag to indicate
> whether or not the aggregate and set requests fall into execution
> state specific bucket.
>
> Signed-off-by: David Dai <daidavid1@...eaurora.org>
> Signed-off-by: Georgi Djakov <georgi.djakov@...aro.org>
> ---
> drivers/interconnect/qcom/sdm845.c | 134 ++++++++++++++++++++++-------
> 1 file changed, 103 insertions(+), 31 deletions(-)
>
> diff --git a/drivers/interconnect/qcom/sdm845.c b/drivers/interconnect/qcom/sdm845.c
> index fb526004c82e..b2047d1c6d84 100644
> --- a/drivers/interconnect/qcom/sdm845.c
> +++ b/drivers/interconnect/qcom/sdm845.c
> @@ -66,6 +66,17 @@ struct bcm_db {
> #define SDM845_MAX_BCM_PER_NODE 2
> #define SDM845_MAX_VCD 10
>
> +#define QCOM_ICC_BUCKET_AMC 0
> +#define QCOM_ICC_BUCKET_WAKE 1
> +#define QCOM_ICC_BUCKET_SLEEP 2
> +#define QCOM_ICC_NUM_BUCKETS 3
> +#define QCOM_ICC_TAG_AMC BIT(QCOM_ICC_BUCKET_AMC)
> +#define QCOM_ICC_TAG_WAKE BIT(QCOM_ICC_BUCKET_WAKE)
> +#define QCOM_ICC_TAG_SLEEP BIT(QCOM_ICC_BUCKET_SLEEP)
> +#define QCOM_ICC_TAG_ACTIVE_ONLY (QCOM_ICC_TAG_AMC | QCOM_ICC_TAG_WAKE)
> +#define QCOM_ICC_TAG_ALWAYS (QCOM_ICC_TAG_AMC | QCOM_ICC_TAG_WAKE |\
> + QCOM_ICC_TAG_SLEEP)
> +
> /**
> * struct qcom_icc_node - Qualcomm specific interconnect nodes
> * @name: the node name used in debugfs
> @@ -86,8 +97,8 @@ struct qcom_icc_node {
> u16 num_links;
> u16 channels;
> u16 buswidth;
> - u64 sum_avg;
> - u64 max_peak;
> + u64 sum_avg[QCOM_ICC_NUM_BUCKETS];
> + u64 max_peak[QCOM_ICC_NUM_BUCKETS];
> struct qcom_icc_bcm *bcms[SDM845_MAX_BCM_PER_NODE];
> size_t num_bcms;
> };
> @@ -112,8 +123,8 @@ struct qcom_icc_bcm {
> const char *name;
> u32 type;
> u32 addr;
> - u64 vote_x;
> - u64 vote_y;
> + u64 vote_x[QCOM_ICC_NUM_BUCKETS];
> + u64 vote_y[QCOM_ICC_NUM_BUCKETS];
> bool dirty;
> bool keepalive;
> struct bcm_db aux_data;
> @@ -555,7 +566,7 @@ inline void tcs_cmd_gen(struct tcs_cmd *cmd, u64 vote_x, u64 vote_y,
> cmd->wait = true;
> }
>
> -static void tcs_list_gen(struct list_head *bcm_list,
> +static void tcs_list_gen(struct list_head *bcm_list, int bucket,
> struct tcs_cmd tcs_list[SDM845_MAX_VCD],
> int n[SDM845_MAX_VCD])
> {
> @@ -573,8 +584,8 @@ static void tcs_list_gen(struct list_head *bcm_list,
> commit = true;
> cur_vcd_size = 0;
> }
> - tcs_cmd_gen(&tcs_list[idx], bcm->vote_x, bcm->vote_y,
> - bcm->addr, commit);
> + tcs_cmd_gen(&tcs_list[idx], bcm->vote_x[bucket],
> + bcm->vote_y[bucket], bcm->addr, commit);
> idx++;
> n[batch]++;
> /*
> @@ -595,51 +606,76 @@ static void tcs_list_gen(struct list_head *bcm_list,
>
> static void bcm_aggregate(struct qcom_icc_bcm *bcm)
> {
> - size_t i;
> - u64 agg_avg = 0;
> - u64 agg_peak = 0;
> + size_t i, bucket;
> + u64 agg_avg[QCOM_ICC_NUM_BUCKETS] = {0};
> + u64 agg_peak[QCOM_ICC_NUM_BUCKETS] = {0};
> u64 temp;
>
> - for (i = 0; i < bcm->num_nodes; i++) {
> - temp = bcm->nodes[i]->sum_avg * bcm->aux_data.width;
> - do_div(temp, bcm->nodes[i]->buswidth * bcm->nodes[i]->channels);
> - agg_avg = max(agg_avg, temp);
> + for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) {
> + for (i = 0; i < bcm->num_nodes; i++) {
> + temp = bcm->nodes[i]->sum_avg[bucket] * bcm->aux_data.width;
> + do_div(temp, bcm->nodes[i]->buswidth * bcm->nodes[i]->channels);
> + agg_avg[bucket] = max(agg_avg[bucket], temp);
>
> - temp = bcm->nodes[i]->max_peak * bcm->aux_data.width;
> - do_div(temp, bcm->nodes[i]->buswidth);
> - agg_peak = max(agg_peak, temp);
> - }
> + temp = bcm->nodes[i]->max_peak[bucket] * bcm->aux_data.width;
> + do_div(temp, bcm->nodes[i]->buswidth);
> + agg_peak[bucket] = max(agg_peak[bucket], temp);
> + }
>
> - temp = agg_avg * 1000ULL;
> - do_div(temp, bcm->aux_data.unit);
> - bcm->vote_x = temp;
> + temp = agg_avg[bucket] * 1000ULL;
> + do_div(temp, bcm->aux_data.unit);
> + bcm->vote_x[bucket] = temp;
>
> - temp = agg_peak * 1000ULL;
> - do_div(temp, bcm->aux_data.unit);
> - bcm->vote_y = temp;
> + temp = agg_peak[bucket] * 1000ULL;
> + do_div(temp, bcm->aux_data.unit);
> + bcm->vote_y[bucket] = temp;
> + }
>
> - if (bcm->keepalive && bcm->vote_x == 0 && bcm->vote_y == 0) {
> - bcm->vote_x = 1;
> - bcm->vote_y = 1;
> + if (bcm->keepalive && bcm->vote_x[0] == 0 && bcm->vote_y[0] == 0) {
Consider using QCOM_ICC_BUCKET_AMC instead of 0.
> + bcm->vote_x[QCOM_ICC_BUCKET_AMC] = 1;
> + bcm->vote_x[QCOM_ICC_BUCKET_WAKE] = 1;
> + bcm->vote_y[QCOM_ICC_BUCKET_AMC] = 1;
> + bcm->vote_y[QCOM_ICC_BUCKET_WAKE] = 1;
> }
>
> bcm->dirty = false;
> }
>
> +static void qcom_icc_pre_aggregate(struct icc_node *node)
> +{
> + size_t i;
> + struct qcom_icc_node *qn;
> +
> + qn = node->data;
> +
> + for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
> + qn->sum_avg[i] = 0;
> + qn->max_peak[i] = 0;
> + }
> +}
> +
> static int qcom_icc_aggregate(struct icc_node *node, u32 tag, u32 avg_bw,
> u32 peak_bw, u32 *agg_avg, u32 *agg_peak)
> {
> size_t i;
> struct qcom_icc_node *qn;
> + unsigned long tag_word = (unsigned long)tag;
Is this local still needed? Can't we just use tag directly?
>
> qn = node->data;
>
> + if (!tag)
> + tag_word = QCOM_ICC_TAG_ALWAYS;
> +
> + for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
> + if (test_bit(i, &tag_word)) {
> + qn->sum_avg[i] += avg_bw;
> + qn->max_peak[i] = max_t(u32, qn->max_peak[i], peak_bw);
> + }
> + }
> +
> *agg_avg += avg_bw;
> *agg_peak = max_t(u32, *agg_peak, peak_bw);
>
> - qn->sum_avg = *agg_avg;
> - qn->max_peak = *agg_peak;
> -
> for (i = 0; i < qn->num_bcms; i++)
> qn->bcms[i]->dirty = true;
>
> @@ -675,7 +711,7 @@ static int qcom_icc_set(struct icc_node *src, struct icc_node *dst)
> * Construct the command list based on a pre ordered list of BCMs
> * based on VCD.
> */
> - tcs_list_gen(&commit_list, cmds, commit_idx);
> + tcs_list_gen(&commit_list, QCOM_ICC_BUCKET_AMC, cmds, commit_idx);
>
> if (!commit_idx[0])
> return ret;
> @@ -693,6 +729,41 @@ static int qcom_icc_set(struct icc_node *src, struct icc_node *dst)
> return ret;
> }
>
> + INIT_LIST_HEAD(&commit_list);
> +
> + for (i = 0; i < qp->num_bcms; i++) {
> + /*
> + * Only generate WAKE and SLEEP commands if a resource's
> + * requirements change as the execution environment transitions
> + * between different power states.
> + */
> + if (qp->bcms[i]->vote_x[QCOM_ICC_BUCKET_WAKE] !=
> + qp->bcms[i]->vote_x[QCOM_ICC_BUCKET_SLEEP] ||
> + qp->bcms[i]->vote_y[QCOM_ICC_BUCKET_WAKE] !=
> + qp->bcms[i]->vote_y[QCOM_ICC_BUCKET_SLEEP]) {
> + list_add_tail(&qp->bcms[i]->list, &commit_list);
> + }
> + }
> +
> + if (list_empty(&commit_list))
> + return ret;
> +
> + tcs_list_gen(&commit_list, QCOM_ICC_BUCKET_WAKE, cmds, commit_idx);
> +
> + ret = rpmh_write_batch(qp->dev, RPMH_WAKE_ONLY_STATE, cmds, commit_idx);
> + if (ret) {
> + pr_err("Error sending WAKE RPMH requests (%d)\n", ret);
> + return ret;
> + }
> +
> + tcs_list_gen(&commit_list, QCOM_ICC_BUCKET_SLEEP, cmds, commit_idx);
> +
> + ret = rpmh_write_batch(qp->dev, RPMH_SLEEP_STATE, cmds, commit_idx);
> + if (ret) {
> + pr_err("Error sending SLEEP RPMH requests (%d)\n", ret);
> + return ret;
> + }
> +
> return ret;
> }
>
> @@ -738,6 +809,7 @@ static int qnoc_probe(struct platform_device *pdev)
> provider = &qp->provider;
> provider->dev = &pdev->dev;
> provider->set = qcom_icc_set;
> + provider->pre_aggregate = qcom_icc_pre_aggregate;
> provider->aggregate = qcom_icc_aggregate;
> provider->xlate = of_icc_xlate_onecell;
> INIT_LIST_HEAD(&provider->nodes);
Powered by blists - more mailing lists