lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <501e7c9a-be24-491-2abc-5895327d80e9@os.amperecomputing.com>
Date:   Fri, 20 Oct 2023 15:53:10 -0700 (PDT)
From:   Ilkka Koskinen <ilkka@...amperecomputing.com>
To:     Robin Murphy <robin.murphy@....com>
cc:     will@...nel.org, mark.rutland@....com,
        linux-arm-kernel@...ts.infradead.org, linux-kernel@...r.kernel.org,
        jeremy.linton@....com, ilkka@...amperecomputing.com,
        renyu.zj@...ux.alibaba.com
Subject: Re: [PATCH 2/3] perf/arm-cmn: Rework DTC counters (again)


On Fri, 20 Oct 2023, Robin Murphy wrote:
> The bitmap-based scheme for tracking DTC counter usage turns out to be a
> complete dead-end for its imagined purpose, since by the time we have to
> keep track of a per-DTC counter index anyway, we already have enough
> information to make the bitmap itself redundant. Revert the remains of
> it back to almost the original scheme, but now expanded to track per-DTC
> indices, in preparation for making use of them in anger.
>
> Note that since cycle count events always use a dedicated counter on a
> single DTC, we reuse the field to encode their DTC index directly.
>
> Signed-off-by: Robin Murphy <robin.murphy@....com>

Thanks! I had that on my task list but never had time to start working on 
it.

Reviewed-by: Ilkka Koskinen <ilkka@...amperecomputing.com>

Cheers, Ilkka

> ---
> drivers/perf/arm-cmn.c | 126 +++++++++++++++++++++--------------------
> 1 file changed, 64 insertions(+), 62 deletions(-)
>
> diff --git a/drivers/perf/arm-cmn.c b/drivers/perf/arm-cmn.c
> index f1ac8d0cdb3b..675f1638013e 100644
> --- a/drivers/perf/arm-cmn.c
> +++ b/drivers/perf/arm-cmn.c
> @@ -281,16 +281,13 @@ struct arm_cmn_node {
> 	u16 id, logid;
> 	enum cmn_node_type type;
>
> -	int dtm;
> -	union {
> -		/* DN/HN-F/CXHA */
> -		struct {
> -			u8 val : 4;
> -			u8 count : 4;
> -		} occupid[SEL_MAX];
> -		/* XP */
> -		u8 dtc;
> -	};
> +	u8 dtm;
> +	s8 dtc;
> +	/* DN/HN-F/CXHA */
> +	struct {
> +		u8 val : 4;
> +		u8 count : 4;
> +	} occupid[SEL_MAX];
> 	union {
> 		u8 event[4];
> 		__le32 event_sel;
> @@ -540,12 +537,12 @@ static int arm_cmn_map_show(struct seq_file *s, void *data)
>
> 		seq_puts(s, "\n     |");
> 		for (x = 0; x < cmn->mesh_x; x++) {
> -			u8 dtc = cmn->xps[xp_base + x].dtc;
> +			s8 dtc = cmn->xps[xp_base + x].dtc;
>
> -			if (dtc & (dtc - 1))
> +			if (dtc < 0)
> 				seq_puts(s, " DTC ?? |");
> 			else
> -				seq_printf(s, " DTC %ld  |", __ffs(dtc));
> +				seq_printf(s, " DTC %d  |", dtc);
> 		}
> 		seq_puts(s, "\n     |");
> 		for (x = 0; x < cmn->mesh_x; x++)
> @@ -589,8 +586,7 @@ static void arm_cmn_debugfs_init(struct arm_cmn *cmn, int id) {}
> struct arm_cmn_hw_event {
> 	struct arm_cmn_node *dn;
> 	u64 dtm_idx[4];
> -	unsigned int dtc_idx;
> -	u8 dtcs_used;
> +	s8 dtc_idx[CMN_MAX_DTCS];
> 	u8 num_dns;
> 	u8 dtm_offset;
> 	bool wide_sel;
> @@ -600,6 +596,10 @@ struct arm_cmn_hw_event {
> #define for_each_hw_dn(hw, dn, i) \
> 	for (i = 0, dn = hw->dn; i < hw->num_dns; i++, dn++)
>
> +/* @i is the DTC number, @idx is the counter index on that DTC */
> +#define for_each_hw_dtc_idx(hw, i, idx) \
> +	for (int i = 0, idx; i < CMN_MAX_DTCS; i++) if ((idx = hw->dtc_idx[i]) >= 0)
> +
> static struct arm_cmn_hw_event *to_cmn_hw(struct perf_event *event)
> {
> 	BUILD_BUG_ON(sizeof(struct arm_cmn_hw_event) > offsetof(struct hw_perf_event, target));
> @@ -1429,12 +1429,11 @@ static void arm_cmn_init_counter(struct perf_event *event)
> {
> 	struct arm_cmn *cmn = to_cmn(event->pmu);
> 	struct arm_cmn_hw_event *hw = to_cmn_hw(event);
> -	unsigned int i, pmevcnt = CMN_DT_PMEVCNT(hw->dtc_idx);
> 	u64 count;
>
> -	for (i = 0; hw->dtcs_used & (1U << i); i++) {
> -		writel_relaxed(CMN_COUNTER_INIT, cmn->dtc[i].base + pmevcnt);
> -		cmn->dtc[i].counters[hw->dtc_idx] = event;
> +	for_each_hw_dtc_idx(hw, i, idx) {
> +		writel_relaxed(CMN_COUNTER_INIT, cmn->dtc[i].base + CMN_DT_PMEVCNT(idx));
> +		cmn->dtc[i].counters[idx] = event;
> 	}
>
> 	count = arm_cmn_read_dtm(cmn, hw, false);
> @@ -1447,11 +1446,9 @@ static void arm_cmn_event_read(struct perf_event *event)
> 	struct arm_cmn_hw_event *hw = to_cmn_hw(event);
> 	u64 delta, new, prev;
> 	unsigned long flags;
> -	unsigned int i;
>
> -	if (hw->dtc_idx == CMN_DT_NUM_COUNTERS) {
> -		i = __ffs(hw->dtcs_used);
> -		delta = arm_cmn_read_cc(cmn->dtc + i);
> +	if (CMN_EVENT_TYPE(event) == CMN_TYPE_DTC) {
> +		delta = arm_cmn_read_cc(cmn->dtc + hw->dtc_idx[0]);
> 		local64_add(delta, &event->count);
> 		return;
> 	}
> @@ -1461,8 +1458,8 @@ static void arm_cmn_event_read(struct perf_event *event)
> 	delta = new - prev;
>
> 	local_irq_save(flags);
> -	for (i = 0; hw->dtcs_used & (1U << i); i++) {
> -		new = arm_cmn_read_counter(cmn->dtc + i, hw->dtc_idx);
> +	for_each_hw_dtc_idx(hw, i, idx) {
> +		new = arm_cmn_read_counter(cmn->dtc + i, idx);
> 		delta += new << 16;
> 	}
> 	local_irq_restore(flags);
> @@ -1518,7 +1515,7 @@ static void arm_cmn_event_start(struct perf_event *event, int flags)
> 	int i;
>
> 	if (type == CMN_TYPE_DTC) {
> -		i = __ffs(hw->dtcs_used);
> +		i = hw->dtc_idx[0];
> 		writeq_relaxed(CMN_CC_INIT, cmn->dtc[i].base + CMN_DT_PMCCNTR);
> 		cmn->dtc[i].cc_active = true;
> 	} else if (type == CMN_TYPE_WP) {
> @@ -1549,7 +1546,7 @@ static void arm_cmn_event_stop(struct perf_event *event, int flags)
> 	int i;
>
> 	if (type == CMN_TYPE_DTC) {
> -		i = __ffs(hw->dtcs_used);
> +		i = hw->dtc_idx[0];
> 		cmn->dtc[i].cc_active = false;
> 	} else if (type == CMN_TYPE_WP) {
> 		int wp_idx = arm_cmn_wp_idx(event);
> @@ -1735,12 +1732,19 @@ static int arm_cmn_event_init(struct perf_event *event)
> 	hw->dn = arm_cmn_node(cmn, type);
> 	if (!hw->dn)
> 		return -EINVAL;
> +
> +	memset(hw->dtc_idx, -1, sizeof(hw->dtc_idx));
> 	for (dn = hw->dn; dn->type == type; dn++) {
> 		if (bynodeid && dn->id != nodeid) {
> 			hw->dn++;
> 			continue;
> 		}
> 		hw->num_dns++;
> +		if (dn->dtc < 0)
> +			memset(hw->dtc_idx, 0, cmn->num_dtcs);
> +		else
> +			hw->dtc_idx[dn->dtc] = 0;
> +
> 		if (bynodeid)
> 			break;
> 	}
> @@ -1752,12 +1756,6 @@ static int arm_cmn_event_init(struct perf_event *event)
> 			nodeid, nid.x, nid.y, nid.port, nid.dev, type);
> 		return -EINVAL;
> 	}
> -	/*
> -	 * Keep assuming non-cycles events count in all DTC domains; turns out
> -	 * it's hard to make a worthwhile optimisation around this, short of
> -	 * going all-in with domain-local counter allocation as well.
> -	 */
> -	hw->dtcs_used = (1U << cmn->num_dtcs) - 1;
>
> 	return arm_cmn_validate_group(cmn, event);
> }
> @@ -1783,28 +1781,25 @@ static void arm_cmn_event_clear(struct arm_cmn *cmn, struct perf_event *event,
> 	}
> 	memset(hw->dtm_idx, 0, sizeof(hw->dtm_idx));
>
> -	for (i = 0; hw->dtcs_used & (1U << i); i++)
> -		cmn->dtc[i].counters[hw->dtc_idx] = NULL;
> +	for_each_hw_dtc_idx(hw, j, idx)
> +		cmn->dtc[j].counters[idx] = NULL;
> }
>
> static int arm_cmn_event_add(struct perf_event *event, int flags)
> {
> 	struct arm_cmn *cmn = to_cmn(event->pmu);
> 	struct arm_cmn_hw_event *hw = to_cmn_hw(event);
> -	struct arm_cmn_dtc *dtc = &cmn->dtc[0];
> 	struct arm_cmn_node *dn;
> 	enum cmn_node_type type = CMN_EVENT_TYPE(event);
> -	unsigned int i, dtc_idx, input_sel;
> +	unsigned int input_sel, i = 0;
>
> 	if (type == CMN_TYPE_DTC) {
> -		i = 0;
> 		while (cmn->dtc[i].cycles)
> 			if (++i == cmn->num_dtcs)
> 				return -ENOSPC;
>
> 		cmn->dtc[i].cycles = event;
> -		hw->dtc_idx = CMN_DT_NUM_COUNTERS;
> -		hw->dtcs_used = 1U << i;
> +		hw->dtc_idx[0] = i;
>
> 		if (flags & PERF_EF_START)
> 			arm_cmn_event_start(event, 0);
> @@ -1812,17 +1807,22 @@ static int arm_cmn_event_add(struct perf_event *event, int flags)
> 	}
>
> 	/* Grab a free global counter first... */
> -	dtc_idx = 0;
> -	while (dtc->counters[dtc_idx])
> -		if (++dtc_idx == CMN_DT_NUM_COUNTERS)
> -			return -ENOSPC;
> -
> -	hw->dtc_idx = dtc_idx;
> +	for_each_hw_dtc_idx(hw, j, idx) {
> +		if (j > 0) {
> +			idx = hw->dtc_idx[0];
> +		} else {
> +			idx = 0;
> +			while (cmn->dtc[j].counters[idx])
> +				if (++idx == CMN_DT_NUM_COUNTERS)
> +					goto free_dtms;
> +		}
> +		hw->dtc_idx[j] = idx;
> +	}
>
> 	/* ...then the local counters to feed it. */
> 	for_each_hw_dn(hw, dn, i) {
> 		struct arm_cmn_dtm *dtm = &cmn->dtms[dn->dtm] + hw->dtm_offset;
> -		unsigned int dtm_idx, shift;
> +		unsigned int dtm_idx, shift, d = 0;
> 		u64 reg;
>
> 		dtm_idx = 0;
> @@ -1841,11 +1841,11 @@ static int arm_cmn_event_add(struct perf_event *event, int flags)
>
> 			tmp = dtm->wp_event[wp_idx ^ 1];
> 			if (tmp >= 0 && CMN_EVENT_WP_COMBINE(event) !=
> -					CMN_EVENT_WP_COMBINE(dtc->counters[tmp]))
> +					CMN_EVENT_WP_COMBINE(cmn->dtc[d].counters[tmp]))
> 				goto free_dtms;
>
> 			input_sel = CMN__PMEVCNT0_INPUT_SEL_WP + wp_idx;
> -			dtm->wp_event[wp_idx] = dtc_idx;
> +			dtm->wp_event[wp_idx] = hw->dtc_idx[d];
> 			writel_relaxed(cfg, dtm->base + CMN_DTM_WPn_CONFIG(wp_idx));
> 		} else {
> 			struct arm_cmn_nodeid nid = arm_cmn_nid(cmn, dn->id);
> @@ -1865,7 +1865,7 @@ static int arm_cmn_event_add(struct perf_event *event, int flags)
> 		dtm->input_sel[dtm_idx] = input_sel;
> 		shift = CMN__PMEVCNTn_GLOBAL_NUM_SHIFT(dtm_idx);
> 		dtm->pmu_config_low &= ~(CMN__PMEVCNT0_GLOBAL_NUM << shift);
> -		dtm->pmu_config_low |= FIELD_PREP(CMN__PMEVCNT0_GLOBAL_NUM, dtc_idx) << shift;
> +		dtm->pmu_config_low |= FIELD_PREP(CMN__PMEVCNT0_GLOBAL_NUM, hw->dtc_idx[d]) << shift;
> 		dtm->pmu_config_low |= CMN__PMEVCNT_PAIRED(dtm_idx);
> 		reg = (u64)le32_to_cpu(dtm->pmu_config_high) << 32 | dtm->pmu_config_low;
> 		writeq_relaxed(reg, dtm->base + CMN_DTM_PMU_CONFIG);
> @@ -1893,7 +1893,7 @@ static void arm_cmn_event_del(struct perf_event *event, int flags)
> 	arm_cmn_event_stop(event, PERF_EF_UPDATE);
>
> 	if (type == CMN_TYPE_DTC)
> -		cmn->dtc[__ffs(hw->dtcs_used)].cycles = NULL;
> +		cmn->dtc[hw->dtc_idx[0]].cycles = NULL;
> 	else
> 		arm_cmn_event_clear(cmn, event, hw->num_dns);
> }
> @@ -2074,7 +2074,6 @@ static int arm_cmn_init_dtcs(struct arm_cmn *cmn)
> {
> 	struct arm_cmn_node *dn, *xp;
> 	int dtc_idx = 0;
> -	u8 dtcs_present = (1 << cmn->num_dtcs) - 1;
>
> 	cmn->dtc = devm_kcalloc(cmn->dev, cmn->num_dtcs, sizeof(cmn->dtc[0]), GFP_KERNEL);
> 	if (!cmn->dtc)
> @@ -2084,23 +2083,26 @@ static int arm_cmn_init_dtcs(struct arm_cmn *cmn)
>
> 	cmn->xps = arm_cmn_node(cmn, CMN_TYPE_XP);
>
> +	if (cmn->part == PART_CMN600 && cmn->num_dtcs > 1) {
> +		/* We do at least know that a DTC's XP must be in that DTC's domain */
> +		dn = arm_cmn_node(cmn, CMN_TYPE_DTC);
> +		for (int i = 0; i < cmn->num_dtcs; i++)
> +			arm_cmn_node_to_xp(cmn, dn + i)->dtc = i;
> +	}
> +
> 	for (dn = cmn->dns; dn->type; dn++) {
> -		if (dn->type == CMN_TYPE_XP) {
> -			dn->dtc &= dtcs_present;
> +		if (dn->type == CMN_TYPE_XP)
> 			continue;
> -		}
>
> 		xp = arm_cmn_node_to_xp(cmn, dn);
> +		dn->dtc = xp->dtc;
> 		dn->dtm = xp->dtm;
> 		if (cmn->multi_dtm)
> 			dn->dtm += arm_cmn_nid(cmn, dn->id).port / 2;
>
> 		if (dn->type == CMN_TYPE_DTC) {
> -			int err;
> -			/* We do at least know that a DTC's XP must be in that DTC's domain */
> -			if (xp->dtc == 0xf)
> -				xp->dtc = 1 << dtc_idx;
> -			err = arm_cmn_init_dtc(cmn, dn, dtc_idx++);
> +			int err = arm_cmn_init_dtc(cmn, dn, dtc_idx++);
> +
> 			if (err)
> 				return err;
> 		}
> @@ -2258,9 +2260,9 @@ static int arm_cmn_discover(struct arm_cmn *cmn, unsigned int rgn_offset)
> 			cmn->mesh_x = xp->logid;
>
> 		if (cmn->part == PART_CMN600)
> -			xp->dtc = 0xf;
> +			xp->dtc = -1;
> 		else
> -			xp->dtc = 1 << arm_cmn_dtc_domain(cmn, xp_region);
> +			xp->dtc = arm_cmn_dtc_domain(cmn, xp_region);
>
> 		xp->dtm = dtm - cmn->dtms;
> 		arm_cmn_init_dtm(dtm++, xp, 0);
> -- 
> 2.39.2.101.g768bb238c484.dirty
>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ