lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1249381489.7924.186.camel@twins>
Date:	Tue, 04 Aug 2009 12:24:49 +0200
From:	Peter Zijlstra <a.p.zijlstra@...llo.nl>
To:	Markus Metzger <markus.t.metzger@...el.com>
Cc:	mingo@...e.hu, tglx@...utronix.de, hpa@...or.com,
	markus.t.metzger@...il.com, linux-kernel@...r.kernel.org
Subject: Re: [patch] x86, perf_counter, bts: add bts to perf_counter

On Tue, 2009-07-21 at 15:56 +0200, Markus Metzger wrote:
> Implement a performance counter with:
>     attr.type           = PERF_TYPE_HARDWARE
>     attr.config         = PERF_COUNT_HW_BRANCH_INSTRUCTIONS
>     attr.sample_period  = 1
> 
> using branch trace store (BTS) on x86 hardware, if available.
> 
> The from and to address for each branch can be sampled using:
>     PERF_SAMPLE_IP      for the from address
>     PERF_SAMPLE_ADDR    for the to address
> 

Over all looks very nice, some comments below (could be addressed in a
delta patch).

Thanks Markus!

> +static int reserve_bts_hardware(void)
> +{
> +	int cpu, err = 0;
> +
> +	if (!bts_available())
> +		return -EOPNOTSUPP;
> +
> +	get_online_cpus();
> +
> +	for_each_possible_cpu(cpu) {
> +		struct debug_store *ds;
> +		void *buffer;
> +
> +		err = -ENOMEM;
> +		buffer = kzalloc(BTS_BUFFER_SIZE, GFP_KERNEL);
> +		if (unlikely(!buffer))
> +			break;
> +
> +		ds = kzalloc(sizeof(*ds), GFP_KERNEL);
> +		if (unlikely(!ds)) {
> +			kfree(buffer);
> +			break;
> +		}
> +
> +		ds->bts_buffer_base = (u64)(long)buffer;
> +		ds->bts_index = ds->bts_buffer_base;
> +		ds->bts_absolute_maximum =
> +			ds->bts_buffer_base + BTS_BUFFER_SIZE;
> +		ds->bts_interrupt_threshold =
> +			ds->bts_absolute_maximum - BTS_OVFL_TH;
> +
> +		per_cpu(cpu_hw_counters, cpu).ds = ds;
> +		err = 0;
> +	}
> +
> +	if (err)
> +		release_bts_hardware();
> +	else

 {

> +		for_each_online_cpu(cpu)
> +			init_debug_store_on_cpu(cpu);

 }

> +
> +	put_online_cpus();
> +
> +	return err;
> +}


> +static void intel_pmu_enable_bts(u64 config)
> +{
> +	unsigned long debugctlmsr;
> +
> +	debugctlmsr = get_debugctlmsr();
> +
> +	debugctlmsr |= (1 << 6);
> +	debugctlmsr |= (1 << 7);
> +
> +	if (!(config & ARCH_PERFMON_EVENTSEL_OS))
> +		debugctlmsr |= (1 << 9);
> +
> +	if (!(config & ARCH_PERFMON_EVENTSEL_USR))
> +		debugctlmsr |= (1 << 10);
> +
> +	update_debugctlmsr(debugctlmsr);
> +}
> +
> +static void intel_pmu_disable_bts(void)
> +{
> +	struct cpu_hw_counters *cpuc = &__get_cpu_var(cpu_hw_counters);
> +	unsigned long debugctlmsr;
> +
> +	if (!cpuc->ds)
> +		return;
> +
> +	debugctlmsr = get_debugctlmsr();
> +
> +	debugctlmsr &= ~(1 << 6);
> +	debugctlmsr &= ~(1 << 7);
> +	debugctlmsr &= ~(1 << 9);
> +	debugctlmsr &= ~(1 << 10);
> +
> +	update_debugctlmsr(debugctlmsr);
> +}

It would be good to not use these constants but instead use something
like:

#define X86_DEBUGCTL_TR		(1 << 6)
#define X86_DEBUGCTL_BTS	(1 << 7)
#define X86_DEBUGCTL_BTS_OS	(1 << 9)
#define X86_DEBUGCTL_BTS_USR	(1 << 10)

> @@ -1077,11 +1297,16 @@ fixed_mode_idx(struct perf_counter *coun
>  {
>  	unsigned int event;
>  
> +	event = hwc->config & ARCH_PERFMON_EVENT_MASK;
> +
> +	if (unlikely((event ==
> +		      x86_pmu.event_map(PERF_COUNT_HW_BRANCH_INSTRUCTIONS)) &&
> +		     (hwc->sample_period == 1)))
> +		return X86_PMC_IDX_FIXED_BTS;

I think we should validate this combination in hw_perf_counter_init()
and fail there if sample_period != 1.

> +static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
> +				       struct perf_sample_data *data)
> +{
> +	struct debug_store *ds = cpuc->ds;
> +	struct bts_record {
> +		u64	from;
> +		u64	to;
> +		u64	flags;
> +	};
> +	struct perf_counter *counter = cpuc->counters[X86_PMC_IDX_FIXED_BTS];
> +	unsigned long orig_ip = data->regs->ip;
> +	u64 at;
> +
> +	if (!counter)
> +		return;
> +
> +	if (!ds)
> +		return;
> +
> +	for (at = ds->bts_buffer_base;
> +	     at < ds->bts_index;
> +	     at += sizeof(struct bts_record)) {
> +		struct bts_record *rec = (struct bts_record *)(long)at;
> +
> +		data->regs->ip	= rec->from;
> +		data->addr	= rec->to;
> +
> +		perf_counter_output(counter, 1, data);
> +	}
> +
> +	ds->bts_index = ds->bts_buffer_base;
> +
> +	data->regs->ip	= orig_ip;
> +	data->addr	= 0;
> +}

You might want to set data->sample_period to 1 as well, just in case
someone is weird enough to request PERF_SAMPLE_PERIOD on a BTS
counter ;-)



--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ