lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Mon, 8 Aug 2016 15:52:39 -0300
From:	Arnaldo Carvalho de Melo <acme@...nel.org>
To:	Mark Rutland <mark.rutland@....com>
Cc:	linux-kernel@...r.kernel.org,
	Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
	Ingo Molnar <mingo@...hat.com>,
	Peter Zijlstra <peterz@...radead.org>
Subject: Re: [PATCH] perf-stat: avoid skew when reading events

Em Mon, Aug 08, 2016 at 12:16:23PM +0100, Mark Rutland escreveu:
> When we don't have a tracee (i.e. we're attaching to a task or CPU),
> counters can still be running after our workload finishes, and can still
> be running as we read their values. As we read events one-by-one, there
> can be arbitrary skew between values of events, even within a group.
> This means that ratios within an event group are not reliable.
> 
> This skew can be seen if measuring a group of identical events, e.g:
> 
>  # perf stat -a -C0 -e '{cycles,cycles}' sleep 1
> 
> To avoid this, we must stop groups from counting before we read the
> values of any constituent events. This patch adds and makes use of a new
> disable_counters() helper, which disables group leaders (and thus each
> group as a whole). This mirrors the use of enable_counters() for
> starting event groups in the absence of a tracee.
> 
> Closing a group leader splits the group, and without a disabled group
> leader the newly split events will begin counting. Thus to ensure counts
> are reliable we must defer closing group leaders until all counts have
> been read. To do so this patch splits the read_counters() helper into
> read_counters() and close_counters(), which also aids legibility.

Looks sane, just one nit, now that you close all counters in an evlist
in a loop, please just use perf_evlist__close(evsel_list), should be
equivalent to this new close_counters() routine.

- Arnaldo
 
> Signed-off-by: Mark Rutland <mark.rutland@....com>
> Cc: Alexander Shishkin <alexander.shishkin@...ux.intel.com>
> Cc: Arnaldo Carvalho de Melo <acme@...nel.org>
> Cc: Ingo Molnar <mingo@...hat.com>
> Cc: Peter Zijlstra <peterz@...radead.org>
> Cc: linux-kernel@...r.kernel.org
> ---
>  tools/perf/builtin-stat.c | 39 ++++++++++++++++++++++++++++++++-------
>  1 file changed, 32 insertions(+), 7 deletions(-)
> 
> Note: there is also a kernel-side issue which adds skew to group counts, which
> is addressed by [1]. In the absence of [1], this patch minimises skew, but does
> not remove it entirely.
> 
> Mark.
> 
> [1] http://lkml.kernel.org/r/1469553141-28314-1-git-send-email-mark.rutland@arm.com
> 
> diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c
> index 0c16d20..e8464b3 100644
> --- a/tools/perf/builtin-stat.c
> +++ b/tools/perf/builtin-stat.c
> @@ -331,7 +331,7 @@ static int read_counter(struct perf_evsel *counter)
>  	return 0;
>  }
>  
> -static void read_counters(bool close_counters)
> +static void read_counters(void)
>  {
>  	struct perf_evsel *counter;
>  
> @@ -341,11 +341,16 @@ static void read_counters(bool close_counters)
>  
>  		if (perf_stat_process_counter(&stat_config, counter))
>  			pr_warning("failed to process counter %s\n", counter->name);
> +	}
> +}
>  
> -		if (close_counters) {
> -			perf_evsel__close_fd(counter, perf_evsel__nr_cpus(counter),
> -					     thread_map__nr(evsel_list->threads));
> -		}
> +static void close_counters(void)
> +{
> +	struct perf_evsel *counter;
> +
> +	evlist__for_each(evsel_list, counter) {
> +		perf_evsel__close_fd(counter, perf_evsel__nr_cpus(counter),
> +				     thread_map__nr(evsel_list->threads));
>  	}
>  }
>  
> @@ -353,7 +358,7 @@ static void process_interval(void)
>  {
>  	struct timespec ts, rs;
>  
> -	read_counters(false);
> +	read_counters();
>  
>  	clock_gettime(CLOCK_MONOTONIC, &ts);
>  	diff_timespec(&rs, &ts, &ref_time);
> @@ -380,6 +385,17 @@ static void enable_counters(void)
>  		perf_evlist__enable(evsel_list);
>  }
>  
> +static void disable_counters(void)
> +{
> +	/*
> +	 * If we don't have tracee (attaching to task or cpu), counters may
> +	 * still be running. To get accurate group ratios, we must stop groups
> +	 * from counting before reading their constituent counters.
> +	 */
> +	if (!target__none(&target))
> +		perf_evlist__disable(evsel_list);
> +}
> +
>  static volatile int workload_exec_errno;
>  
>  /*
> @@ -657,11 +673,20 @@ try_again:
>  		}
>  	}
>  
> +	disable_counters();
> +
>  	t1 = rdclock();
>  
>  	update_stats(&walltime_nsecs_stats, t1 - t0);
>  
> -	read_counters(true);
> +	/*
> +	 * Closing a group leader splits the group, and as we only disable
> +	 * group leaders, results in remaining events becoming enabled. To
> +	 * avoid arbitrary skew, we must read all counters before closing any
> +	 * group leaders.
> +	 */
> +	read_counters();
> +	close_counters();
>  
>  	return WEXITSTATUS(status);
>  }
> -- 
> 1.9.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ