[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZJL5OR4F_2-41Csw@FVFF77S0Q05N>
Date: Wed, 21 Jun 2023 14:20:57 +0100
From: Mark Rutland <mark.rutland@....com>
To: Anshuman Khandual <anshuman.khandual@....com>
Cc: linux-arm-kernel@...ts.infradead.org, linux-kernel@...r.kernel.org,
will@...nel.org, catalin.marinas@....com,
Mark Brown <broonie@...nel.org>,
James Clark <james.clark@....com>,
Rob Herring <robh@...nel.org>, Marc Zyngier <maz@...nel.org>,
Suzuki Poulose <suzuki.poulose@....com>,
Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...hat.com>,
Arnaldo Carvalho de Melo <acme@...nel.org>,
linux-perf-users@...r.kernel.org
Subject: Re: [PATCH V12 10/10] arm64/perf: Implement branch records save on
PMU IRQ
On Thu, Jun 15, 2023 at 07:02:39PM +0530, Anshuman Khandual wrote:
> This modifies armv8pmu_branch_read() to concatenate live entries along with
> task context stored entries and then process the resultant buffer to create
> perf branch entry array for perf_sample_data. It follows the same principle
> like task sched out.
>
> Cc: Catalin Marinas <catalin.marinas@....com>
> Cc: Will Deacon <will@...nel.org>
> Cc: Mark Rutland <mark.rutland@....com>
> Cc: linux-arm-kernel@...ts.infradead.org
> Cc: linux-kernel@...r.kernel.org
> Tested-by: James Clark <james.clark@....com>
> Signed-off-by: Anshuman Khandual <anshuman.khandual@....com>
The resulting logic looks fine here, but it would be nicer if we had the
resulting structure from the outset rather than having to rewrite it (i.e. if
when we introduced this we captured all the recores then processed them), as
that would keep the diff minimal and make it much clearer as to what wwas
happening here.
Either way:
Acked-by: Mark Rutland <mark.rutland@....com>
Mark.
> ---
> drivers/perf/arm_brbe.c | 69 +++++++++++++++++++----------------------
> 1 file changed, 32 insertions(+), 37 deletions(-)
>
> diff --git a/drivers/perf/arm_brbe.c b/drivers/perf/arm_brbe.c
> index 3bb17ced2b1d..d28067c896e2 100644
> --- a/drivers/perf/arm_brbe.c
> +++ b/drivers/perf/arm_brbe.c
> @@ -653,41 +653,44 @@ void armv8pmu_branch_reset(void)
> isb();
> }
>
> -static bool capture_branch_entry(struct pmu_hw_events *cpuc,
> - struct perf_event *event, int idx)
> +static void brbe_regset_branch_entries(struct pmu_hw_events *cpuc, struct perf_event *event,
> + struct brbe_regset *regset, int idx)
> {
> struct perf_branch_entry *entry = &cpuc->branches->branch_entries[idx];
> - u64 brbinf = get_brbinf_reg(idx);
> -
> - /*
> - * There are no valid entries anymore on the buffer.
> - * Abort the branch record processing to save some
> - * cycles and also reduce the capture/process load
> - * for the user space as well.
> - */
> - if (brbe_invalid(brbinf))
> - return false;
> + u64 brbinf = regset[idx].brbinf;
>
> perf_clear_branch_entry_bitfields(entry);
> if (brbe_record_is_complete(brbinf)) {
> - entry->from = get_brbsrc_reg(idx);
> - entry->to = get_brbtgt_reg(idx);
> + entry->from = regset[idx].brbsrc;
> + entry->to = regset[idx].brbtgt;
> } else if (brbe_record_is_source_only(brbinf)) {
> - entry->from = get_brbsrc_reg(idx);
> + entry->from = regset[idx].brbsrc;
> entry->to = 0;
> } else if (brbe_record_is_target_only(brbinf)) {
> entry->from = 0;
> - entry->to = get_brbtgt_reg(idx);
> + entry->to = regset[idx].brbtgt;
> }
> capture_brbe_flags(entry, event, brbinf);
> - return true;
> +}
> +
> +static void process_branch_entries(struct pmu_hw_events *cpuc, struct perf_event *event,
> + struct brbe_regset *regset, int nr_regset)
> +{
> + int idx;
> +
> + for (idx = 0; idx < nr_regset; idx++)
> + brbe_regset_branch_entries(cpuc, event, regset, idx);
> +
> + cpuc->branches->branch_stack.nr = nr_regset;
> + cpuc->branches->branch_stack.hw_idx = -1ULL;
> }
>
> void armv8pmu_branch_read(struct pmu_hw_events *cpuc, struct perf_event *event)
> {
> - int nr_hw_entries = brbe_get_numrec(cpuc->percpu_pmu->reg_brbidr);
> + struct arm64_perf_task_context *task_ctx = event->pmu_ctx->task_ctx_data;
> + struct brbe_regset live[BRBE_MAX_ENTRIES];
> + int nr_live, nr_store, nr_hw_entries;
> u64 brbfcr, brbcr;
> - int idx = 0;
>
> brbcr = read_sysreg_s(SYS_BRBCR_EL1);
> brbfcr = read_sysreg_s(SYS_BRBFCR_EL1);
> @@ -699,25 +702,17 @@ void armv8pmu_branch_read(struct pmu_hw_events *cpuc, struct perf_event *event)
> write_sysreg_s(brbfcr | BRBFCR_EL1_PAUSED, SYS_BRBFCR_EL1);
> isb();
>
> - /* Loop through bank 0 */
> - select_brbe_bank(BRBE_BANK_IDX_0);
> - while (idx < nr_hw_entries && idx < BRBE_BANK0_IDX_MAX) {
> - if (!capture_branch_entry(cpuc, event, idx))
> - goto skip_bank_1;
> - idx++;
> - }
> -
> - /* Loop through bank 1 */
> - select_brbe_bank(BRBE_BANK_IDX_1);
> - while (idx < nr_hw_entries && idx < BRBE_BANK1_IDX_MAX) {
> - if (!capture_branch_entry(cpuc, event, idx))
> - break;
> - idx++;
> + nr_hw_entries = brbe_get_numrec(cpuc->percpu_pmu->reg_brbidr);
> + nr_live = capture_brbe_regset(nr_hw_entries, live);
> + if (event->ctx->task) {
> + nr_store = task_ctx->nr_brbe_records;
> + nr_store = stitch_stored_live_entries(task_ctx->store, live, nr_store,
> + nr_live, nr_hw_entries);
> + process_branch_entries(cpuc, event, task_ctx->store, nr_store);
> + task_ctx->nr_brbe_records = 0;
> + } else {
> + process_branch_entries(cpuc, event, live, nr_live);
> }
> -
> -skip_bank_1:
> - cpuc->branches->branch_stack.nr = idx;
> - cpuc->branches->branch_stack.hw_idx = -1ULL;
> process_branch_aborts(cpuc);
>
> /* Unpause the buffer */
> --
> 2.25.1
>
Powered by blists - more mailing lists