[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CO6PR11MB563588E1D93E96C5A09B54C2EEAC2@CO6PR11MB5635.namprd11.prod.outlook.com>
Date: Thu, 18 Jul 2024 23:55:16 +0000
From: "Wang, Weilin" <weilin.wang@...el.com>
To: Namhyung Kim <namhyung@...nel.org>
CC: Ian Rogers <irogers@...gle.com>, Arnaldo Carvalho de Melo
<acme@...nel.org>, Peter Zijlstra <peterz@...radead.org>, Ingo Molnar
<mingo@...hat.com>, Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
Jiri Olsa <jolsa@...nel.org>, "Hunter, Adrian" <adrian.hunter@...el.com>, Kan
Liang <kan.liang@...ux.intel.com>, "linux-perf-users@...r.kernel.org"
<linux-perf-users@...r.kernel.org>, "linux-kernel@...r.kernel.org"
<linux-kernel@...r.kernel.org>, "Taylor, Perry" <perry.taylor@...el.com>,
"Alt, Samantha" <samantha.alt@...el.com>, "Biggers, Caleb"
<caleb.biggers@...el.com>
Subject: RE: [RFC PATCH v17 3/8] perf stat: Fork and launch perf record when
perf stat needs to get retire latency value for a metric.
> -----Original Message-----
> From: Wang, Weilin
> Sent: Wednesday, July 17, 2024 11:28 PM
> To: Namhyung Kim <namhyung@...nel.org>
> Cc: Ian Rogers <irogers@...gle.com>; Arnaldo Carvalho de Melo
> <acme@...nel.org>; Peter Zijlstra <peterz@...radead.org>; Ingo Molnar
> <mingo@...hat.com>; Alexander Shishkin
> <alexander.shishkin@...ux.intel.com>; Jiri Olsa <jolsa@...nel.org>; Hunter,
> Adrian <adrian.hunter@...el.com>; Kan Liang <kan.liang@...ux.intel.com>;
> linux-perf-users@...r.kernel.org; linux-kernel@...r.kernel.org; Taylor, Perry
> <perry.taylor@...el.com>; Alt, Samantha <samantha.alt@...el.com>; Biggers,
> Caleb <caleb.biggers@...el.com>
> Subject: RE: [RFC PATCH v17 3/8] perf stat: Fork and launch perf record when
> perf stat needs to get retire latency value for a metric.
>
>
>
> > -----Original Message-----
> > From: Namhyung Kim <namhyung@...nel.org>
> > Sent: Wednesday, July 17, 2024 10:56 PM
> > To: Wang, Weilin <weilin.wang@...el.com>
> > Cc: Ian Rogers <irogers@...gle.com>; Arnaldo Carvalho de Melo
> > <acme@...nel.org>; Peter Zijlstra <peterz@...radead.org>; Ingo Molnar
> > <mingo@...hat.com>; Alexander Shishkin
> > <alexander.shishkin@...ux.intel.com>; Jiri Olsa <jolsa@...nel.org>; Hunter,
> > Adrian <adrian.hunter@...el.com>; Kan Liang <kan.liang@...ux.intel.com>;
> > linux-perf-users@...r.kernel.org; linux-kernel@...r.kernel.org; Taylor, Perry
> > <perry.taylor@...el.com>; Alt, Samantha <samantha.alt@...el.com>;
> Biggers,
> > Caleb <caleb.biggers@...el.com>
> > Subject: Re: [RFC PATCH v17 3/8] perf stat: Fork and launch perf record
> when
> > perf stat needs to get retire latency value for a metric.
> >
> > On Fri, Jul 12, 2024 at 03:09:25PM -0400, weilin.wang@...el.com wrote:
> > > From: Weilin Wang <weilin.wang@...el.com>
> > >
> > > When retire_latency value is used in a metric formula, evsel would fork a
> perf
> > > record process with "-e" and "-W" options. Perf record will collect required
> > > retire_latency values in parallel while perf stat is collecting counting values.
> > >
> > > At the point of time that perf stat stops counting, evsel would stop perf
> > record
> > > by sending sigterm signal to perf record process. Sampled data will be
> > process
> > > to get retire latency value. Another thread is required to synchronize
> > between
> > > perf stat and perf record when we pass data through pipe.
> > >
> > > Retire_latency evsel is not opened for perf stat so that there is no counter
> > > wasted on it. This commit includes code suggested by Namhyung to adjust
> > reading
> > > size for groups that include retire_latency evsels.
> > >
> > > Signed-off-by: Weilin Wang <weilin.wang@...el.com>
> > > ---
> > > tools/perf/builtin-stat.c | 4 +
> > > tools/perf/util/Build | 1 +
> > > tools/perf/util/evlist.c | 2 +
> > > tools/perf/util/evsel.c | 66 +++++-
> > > tools/perf/util/intel-tpebs.c | 430
> > ++++++++++++++++++++++++++++++++++
> > > tools/perf/util/intel-tpebs.h | 35 +++
> > > 6 files changed, 536 insertions(+), 2 deletions(-)
> > > create mode 100644 tools/perf/util/intel-tpebs.c
> > > create mode 100644 tools/perf/util/intel-tpebs.h
> > >
> > > diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c
> > > index 661832756a24..68125bd75b37 100644
> > > --- a/tools/perf/builtin-stat.c
> > > +++ b/tools/perf/builtin-stat.c
> > > @@ -70,6 +70,7 @@
> > > #include "util/bpf_counter.h"
> > > #include "util/iostat.h"
> > > #include "util/util.h"
> > > +#include "util/intel-tpebs.h"
> > > #include "asm/bug.h"
> > >
> > > #include <linux/time64.h>
> > > @@ -683,6 +684,9 @@ static enum counter_recovery
> > stat_handle_error(struct evsel *counter)
> > >
> > > if (child_pid != -1)
> > > kill(child_pid, SIGTERM);
> > > +
> > > + tpebs_delete();
> > > +
> > > return COUNTER_FATAL;
> > > }
> > >
> > > diff --git a/tools/perf/util/Build b/tools/perf/util/Build
> > > index 0f18fe81ef0b..25cae5c613ad 100644
> > > --- a/tools/perf/util/Build
> > > +++ b/tools/perf/util/Build
> > > @@ -154,6 +154,7 @@ perf-util-y += clockid.o
> > > perf-util-y += list_sort.o
> > > perf-util-y += mutex.o
> > > perf-util-y += sharded_mutex.o
> > > +perf-util-$(CONFIG_X86_64) += intel-tpebs.o
> > >
> > > perf-util-$(CONFIG_LIBBPF) += bpf_map.o
> > > perf-util-$(CONFIG_PERF_BPF_SKEL) += bpf_counter.o
> > > diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
> > > index 3a719edafc7a..78ce80f227aa 100644
> > > --- a/tools/perf/util/evlist.c
> > > +++ b/tools/perf/util/evlist.c
> > > @@ -33,6 +33,7 @@
> > > #include "util/bpf-filter.h"
> > > #include "util/stat.h"
> > > #include "util/util.h"
> > > +#include "util/intel-tpebs.h"
> > > #include <signal.h>
> > > #include <unistd.h>
> > > #include <sched.h>
> > > @@ -179,6 +180,7 @@ void evlist__delete(struct evlist *evlist)
> > > if (evlist == NULL)
> > > return;
> > >
> > > + tpebs_delete();
> > > evlist__free_stats(evlist);
> > > evlist__munmap(evlist);
> > > evlist__close(evlist);
> > > diff --git a/tools/perf/util/evsel.c b/tools/perf/util/evsel.c
> > > index bc603193c477..1d000fa2c6cf 100644
> > > --- a/tools/perf/util/evsel.c
> > > +++ b/tools/perf/util/evsel.c
> > > @@ -1554,6 +1554,60 @@ static void evsel__set_count(struct evsel
> > *counter, int cpu_map_idx, int thread,
> > > perf_counts__set_loaded(counter->counts, cpu_map_idx, thread,
> > true);
> > > }
> > >
> > > +static bool evsel__group_has_tpebs(struct evsel *leader)
> > > +{
> > > + struct evsel *evsel;
> > > +
> > > + for_each_group_evsel(evsel, leader) {
> > > + if (evsel__is_retire_lat(evsel))
> > > + return true;
> > > + }
> > > + return false;
> > > +}
> > > +
> > > +static u64 evsel__group_read_nr_members(struct evsel *leader)
> > > +{
> > > + u64 nr = leader->core.nr_members;
> > > + struct evsel *evsel;
> > > +
> > > + for_each_group_evsel(evsel, leader) {
> > > + if (evsel__is_retire_lat(evsel))
> > > + nr--;
> > > + }
> > > + return nr;
> > > +}
> > > +
> > > +static u64 evsel__group_read_size(struct evsel *leader)
> > > +{
> > > + u64 read_format = leader->core.attr.read_format;
> > > + int entry = sizeof(u64); /* value */
> > > + int size = 0;
> > > + int nr = 1;
> > > +
> > > + if (!evsel__group_has_tpebs(leader))
> > > + return perf_evsel__read_size(&leader->core);
> > > +
> > > + if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
> > > + size += sizeof(u64);
> > > +
> > > + if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING)
> > > + size += sizeof(u64);
> > > +
> > > + if (read_format & PERF_FORMAT_ID)
> > > + entry += sizeof(u64);
> > > +
> > > + if (read_format & PERF_FORMAT_LOST)
> > > + entry += sizeof(u64);
> > > +
> > > + if (read_format & PERF_FORMAT_GROUP) {
> > > + nr = evsel__group_read_nr_members(leader);
> > > + size += sizeof(u64);
> > > + }
> > > +
> > > + size += entry * nr;
> > > + return size;
> > > +}
> > > +
> > > static int evsel__process_group_data(struct evsel *leader, int
> cpu_map_idx,
> > int thread, u64 *data)
> > > {
> > > u64 read_format = leader->core.attr.read_format;
> > > @@ -1562,7 +1616,7 @@ static int evsel__process_group_data(struct
> evsel
> > *leader, int cpu_map_idx, int
> > >
> > > nr = *data++;
> > >
> > > - if (nr != (u64) leader->core.nr_members)
> > > + if (nr != evsel__group_read_nr_members(leader))
> > > return -EINVAL;
> > >
> > > if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
> > > @@ -1592,7 +1646,7 @@ static int evsel__read_group(struct evsel
> *leader,
> > int cpu_map_idx, int thread)
> > > {
> > > struct perf_stat_evsel *ps = leader->stats;
> > > u64 read_format = leader->core.attr.read_format;
> > > - int size = perf_evsel__read_size(&leader->core);
> > > + int size = evsel__group_read_size(leader);
> > > u64 *data = ps->group_data;
> > >
> > > if (!(read_format & PERF_FORMAT_ID))
> > > @@ -2200,6 +2254,9 @@ static int evsel__open_cpu(struct evsel *evsel,
> > struct perf_cpu_map *cpus,
> > > return 0;
> > > }
> > >
> > > + if (evsel__is_retire_lat(evsel))
> > > + return tpebs_start(evsel->evlist);
> > > +
> > > err = __evsel__prepare_open(evsel, cpus, threads);
> > > if (err)
> > > return err;
> > > @@ -2392,6 +2449,8 @@ int evsel__open(struct evsel *evsel, struct
> > perf_cpu_map *cpus,
> > >
> > > void evsel__close(struct evsel *evsel)
> > > {
> > > + if (evsel__is_retire_lat(evsel))
> > > + tpebs_delete();
> > > perf_evsel__close(&evsel->core);
> > > perf_evsel__free_id(&evsel->core);
> > > }
> > > @@ -3357,6 +3416,9 @@ static int store_evsel_ids(struct evsel *evsel,
> > struct evlist *evlist)
> > > {
> > > int cpu_map_idx, thread;
> > >
> > > + if (evsel__is_retire_lat(evsel))
> > > + return 0;
> > > +
> > > for (cpu_map_idx = 0; cpu_map_idx < xyarray__max_x(evsel->core.fd);
> > cpu_map_idx++) {
> > > for (thread = 0; thread < xyarray__max_y(evsel->core.fd);
> > > thread++) {
> > > diff --git a/tools/perf/util/intel-tpebs.c b/tools/perf/util/intel-tpebs.c
> > > new file mode 100644
> > > index 000000000000..6680ef887961
> > > --- /dev/null
> > > +++ b/tools/perf/util/intel-tpebs.c
> > > @@ -0,0 +1,430 @@
> > > +// SPDX-License-Identifier: GPL-2.0-only
> > > +/*
> > > + * intel_tpebs.c: Intel TPEBS support
> > > + */
> > > +
> > > +
> > > +#include <sys/param.h>
> > > +#include <subcmd/run-command.h>
> > > +#include <thread.h>
> > > +#include "intel-tpebs.h"
> > > +#include <linux/list.h>
> > > +#include <linux/zalloc.h>
> > > +#include <linux/err.h>
> > > +#include "sample.h"
> > > +#include "debug.h"
> > > +#include "evlist.h"
> > > +#include "evsel.h"
> > > +#include "session.h"
> > > +#include "tool.h"
> > > +#include "cpumap.h"
> > > +#include "metricgroup.h"
> > > +#include <sys/stat.h>
> > > +#include <sys/file.h>
> > > +#include <poll.h>
> > > +#include <math.h>
> > > +
> > > +#define PERF_DATA "-"
> > > +
> > > +bool tpebs_recording;
> > > +static pid_t tpebs_pid = -1;
> > > +static size_t tpebs_event_size;
> > > +static LIST_HEAD(tpebs_results);
> > > +static pthread_t tpebs_reader_thread;
> > > +static struct child_process *tpebs_cmd;
> > > +
> > > +struct tpebs_retire_lat {
> > > + struct list_head nd;
> > > + /* Event name */
> > > + const char *name;
> > > + /* Event name with the TPEBS modifier R */
> > > + const char *tpebs_name;
> > > + /* Count of retire_latency values found in sample data */
> > > + size_t count;
> > > + /* Sum of all the retire_latency values in sample data */
> > > + int sum;
> > > + /* Average of retire_latency, val = sum / count */
> > > + double val;
> > > +};
> > > +
> > > +static int get_perf_record_args(const char **record_argv, char buf[],
> > > + const char *cpumap_buf)
> > > +{
> > > + struct tpebs_retire_lat *e;
> > > + int i = 0;
> > > +
> > > + pr_debug("tpebs: Prepare perf record for retire_latency\n");
> > > +
> > > + record_argv[i++] = "perf";
> > > + record_argv[i++] = "record";
> > > + record_argv[i++] = "-W";
> > > + record_argv[i++] = "--synth=no";
> > > + record_argv[i++] = buf;
> > > +
> > > + if (!cpumap_buf) {
> > > + pr_err("tpebs: Require cpumap list to run sampling\n");
> > > + return -ECANCELED;
> > > + }
> > > + /* Use -C when cpumap_buf is not "-1" */
> > > + if (strcmp(cpumap_buf, "-1")) {
> > > + record_argv[i++] = "-C";
> > > + record_argv[i++] = cpumap_buf;
> > > + }
> > > +
> > > + list_for_each_entry(e, &tpebs_results, nd) {
> > > + record_argv[i++] = "-e";
> > > + record_argv[i++] = e->name;
> > > + }
> > > +
> > > + record_argv[i++] = "-o";
> > > + record_argv[i++] = PERF_DATA;
> > > +
> > > + return 0;
> > > +}
> > > +
> > > +static int prepare_run_command(const char **argv)
> > > +{
> > > + tpebs_cmd = zalloc(sizeof(struct child_process));
> > > + if (!tpebs_cmd)
> > > + return -ENOMEM;
> > > + tpebs_cmd->argv = argv;
> > > + tpebs_cmd->out = -1;
> > > + return 0;
> > > +}
> > > +
> > > +static int start_perf_record(int control_fd[], int ack_fd[],
> > > + const char *cpumap_buf)
> > > +{
> > > + const char **record_argv;
> > > + int ret;
> > > + char buf[32];
> > > +
> > > + scnprintf(buf, sizeof(buf), "--control=fd:%d,%d", control_fd[0],
> > ack_fd[1]);
> > > +
> > > + record_argv = calloc(12 + 2 * tpebs_event_size, sizeof(char *));
> > > + if (!record_argv)
> > > + return -ENOMEM;
> > > +
> > > + ret = get_perf_record_args(record_argv, buf, cpumap_buf);
> > > + if (ret)
> > > + goto out;
> > > +
> > > + ret = prepare_run_command(record_argv);
> > > + if (ret)
> > > + goto out;
> > > + ret = start_command(tpebs_cmd);
> > > +out:
> > > + free(record_argv);
> > > + return ret;
> > > +}
> > > +
> > > +static int process_sample_event(struct perf_tool *tool __maybe_unused,
> > > + union perf_event *event __maybe_unused,
> > > + struct perf_sample *sample,
> > > + struct evsel *evsel,
> > > + struct machine *machine __maybe_unused)
> > > +{
> > > + int ret = 0;
> > > + const char *evname;
> > > + struct tpebs_retire_lat *t;
> > > +
> > > + evname = evsel__name(evsel);
> > > +
> > > + /*
> > > + * Need to handle per core results? We are assuming average retire
> > > + * latency value will be used. Save the number of samples and the sum
> > of
> > > + * retire latency value for each event.
> > > + */
> > > + list_for_each_entry(t, &tpebs_results, nd) {
> > > + if (!strcmp(evname, t->name)) {
> > > + t->count += 1;
> > > + t->sum += sample->retire_lat;
> > > + t->val = (double) t->sum / t->count;
> > > + break;
> > > + }
> > > + }
> > > +
> > > + return ret;
> > > +}
> > > +
> > > +static int process_feature_event(struct perf_session *session,
> > > + union perf_event *event)
> > > +{
> > > + if (event->feat.feat_id < HEADER_LAST_FEATURE)
> > > + return perf_event__process_feature(session, event);
> > > + return 0;
> > > +}
> > > +
> > > +static void *__sample_reader(void *arg)
> > > +{
> > > + struct child_process *child = arg;
> > > + struct perf_session *session;
> > > + struct perf_data data = {
> > > + .mode = PERF_DATA_MODE_READ,
> > > + .path = PERF_DATA,
> > > + .file.fd = child->out,
> > > + };
> > > + struct perf_tool tool = {
> > > + .sample = process_sample_event,
> > > + .feature = process_feature_event,
> > > + .attr = perf_event__process_attr,
> > > + };
> > > +
> > > + session = perf_session__new(&data, &tool);
> > > + if (IS_ERR(session))
> > > + return NULL;
> > > + perf_session__process_events(session);
> > > + perf_session__delete(session);
> > > +
> > > + return NULL;
> > > +}
> > > +
> > > +/*
> > > + * tpebs_stop - stop the sample data read thread and the perf record
> > process.
> > > + */
> > > +static int tpebs_stop(void)
> > > +{
> > > + int ret = 0;
> > > +
> > > + /* Like tpebs_start, we should only run tpebs_end once. */
> > > + if (tpebs_pid != -1) {
> > > + kill(tpebs_cmd->pid, SIGTERM);
> > > + tpebs_pid = -1;
> > > + pthread_join(tpebs_reader_thread, NULL);
> > > + close(tpebs_cmd->out);
> > > + ret = finish_command(tpebs_cmd);
> > > + if (ret == -ERR_RUN_COMMAND_WAITPID_SIGNAL)
> > > + ret = 0;
> > > + }
> > > + return ret;
> > > +}
> > > +
> > > +/*
> > > + * tpebs_start - start tpebs execution.
> > > + * @evsel_list: retire_latency evsels in this list will be selected and sampled
> > > + * to get the average retire_latency value.
> > > + *
> > > + * This function will be called from evlist level later when evlist__open() is
> > > + * called consistently.
> > > + */
> > > +int tpebs_start(struct evlist *evsel_list)
> > > +{
> > > + int ret = 0;
> > > + struct evsel *evsel;
> > > + char cpumap_buf[50];
> > > +
> > > + /*
> > > + * We should only run tpebs_start when tpebs_recording is enabled.
> > > + * And we should only run it once with all the required events.
> > > + */
> > > + if (tpebs_pid != -1 || !tpebs_recording)
> > > + return 0;
> > > +
> > > + cpu_map__snprint(evsel_list->core.user_requested_cpus,
> > cpumap_buf, sizeof(cpumap_buf));
> > > + /*
> > > + * Prepare perf record for sampling event retire_latency before fork
> > and
> > > + * prepare workload
> > > + */
> > > + evlist__for_each_entry(evsel_list, evsel) {
> > > + int i;
> > > + char *name;
> > > + struct tpebs_retire_lat *new;
> > > +
> > > + if (!evsel->retire_lat)
> > > + continue;
> > > +
> > > + pr_debug("tpebs: Retire_latency of event %s is required\n",
> > evsel->name);
> > > + for (i = strlen(evsel->name) - 1; i > 0; i--) {
> > > + if (evsel->name[i] == 'R')
> > > + break;
> > > + }
> > > + if (i <= 0 || evsel->name[i] != 'R') {
> > > + ret = -1;
> > > + goto err;
> > > + }
> > > +
> > > + name = strdup(evsel->name);
> > > + if (!name) {
> > > + ret = -ENOMEM;
> > > + goto err;
> > > + }
> > > + name[i] = 'p';
> > > +
> > > + new = zalloc(sizeof(*new));
> > > + if (!new) {
> > > + ret = -1;
> > > + zfree(name);
> > > + goto err;
> > > + }
> > > + new->name = name;
> > > + new->tpebs_name = evsel->name;
> > > + list_add_tail(&new->nd, &tpebs_results);
> > > + tpebs_event_size += 1;
> > > + }
> > > +
> > > + if (tpebs_event_size > 0) {
> > > + struct pollfd pollfd = { .events = POLLIN, };
> > > + int control_fd[2], ack_fd[2], len;
> > > + char ack_buf[8];
> > > +
> > > + /*Create control and ack fd for --control*/
> > > + if (pipe(control_fd) < 0) {
> > > + pr_err("tpebs: Failed to create control fifo");
> > > + ret = -1;
> > > + goto out;
> > > + }
> > > + if (pipe(ack_fd) < 0) {
> > > + pr_err("tpebs: Failed to create control fifo");
> > > + ret = -1;
> > > + goto out;
> > > + }
> > > +
> > > + ret = start_perf_record(control_fd, ack_fd, cpumap_buf);
> > > + if (ret)
> > > + goto out;
> > > + tpebs_pid = tpebs_cmd->pid;
> > > + if (pthread_create(&tpebs_reader_thread, NULL,
> > __sample_reader, tpebs_cmd)) {
> > > + kill(tpebs_cmd->pid, SIGTERM);
> > > + close(tpebs_cmd->out);
> > > + pr_err("Could not create thread to process sample
> > data.\n");
> > > + ret = -1;
> > > + goto out;
> > > + }
> > > + /* Wait for perf record initialization.*/
> > > + len = strlen("enable");
> > > + ret = write(control_fd[1], "enable", len);
> >
> > Can we use EVLIST_CTL_CMD_ENABLE_TAG instead?
> >
> >
> > > + if (ret != len) {
> > > + pr_err("perf record control write control message
> > failed\n");
> > > + goto out;
> > > + }
> > > +
> > > + /* wait for an ack */
> > > + pollfd.fd = ack_fd[0];
> > > +
> > > + /*
> > > + * We need this poll to ensure the ack_fd PIPE will not hang
> > > + * when perf record failed for any reason. The timeout value
> > > + * 3000ms is an empirical selection.
> > > + */
> >
> > Oh, you changed it to 3 sec. But I think it's ok as we don't wait for
> > that long for the normal cases.
>
> Hi Namhyung,
>
> I found it's more reliable to use 3 secs because in some of my test cases 2 secs
> are not enough for perf record reach the point of sending ACK back.
Does this 3sec wait looks good to you? Please let me know if you have other suggestions.
Thanks,
Weilin
>
> Thanks,
> Weilin
>
> >
> >
> > > + if (!poll(&pollfd, 1, 3000)) {
> > > + pr_err("tpebs failed: perf record ack timeout\n");
> > > + ret = -1;
> > > + goto out;
> > > + }
> > > +
> > > + if (!(pollfd.revents & POLLIN)) {
> > > + pr_err("tpebs failed: did not received an ack\n");
> > > + ret = -1;
> > > + goto out;
> > > + }
> > > +
> > > + ret = read(ack_fd[0], ack_buf, sizeof(ack_buf));
> > > + if (ret > 0)
> > > + ret = strcmp(ack_buf, "ack\n");
> >
> > Same for EVLIST_CTL_CMD_ACK_TAG.
> >
> >
> > > + else {
> > > + pr_err("tpebs: perf record control ack failed\n");
> > > + goto out;
> > > + }
> > > +out:
> > > + close(control_fd[0]);
> > > + close(control_fd[1]);
> > > + close(ack_fd[0]);
> > > + close(ack_fd[1]);
> > > + }
> > > +err:
> > > + if (ret)
> > > + tpebs_delete();
> > > + return ret;
> > > +}
> > > +
> > > +
> > > +int tpebs_set_evsel(struct evsel *evsel, int cpu_map_idx, int thread)
> > > +{
> > > + __u64 val;
> > > + bool found = false;
> > > + struct tpebs_retire_lat *t;
> > > + struct perf_counts_values *count;
> > > +
> > > + /* Non reitre_latency evsel should never enter this function. */
> > > + if (!evsel__is_retire_lat(evsel))
> > > + return -1;
> > > +
> > > + /*
> > > + * Need to stop the forked record to ensure get sampled data from the
> > > + * PIPE to process and get non-zero retire_lat value for hybrid.
> > > + */
> > > + tpebs_stop();
> > > + count = perf_counts(evsel->counts, cpu_map_idx, thread);
> > > +
> > > + list_for_each_entry(t, &tpebs_results, nd) {
> > > + if (t->tpebs_name == evsel->name || (evsel->metric_id
> > && !strcmp(t->tpebs_name, evsel->metric_id))) {
> >
> > This line is too long, please break.
> >
> > Thanks,
> > Namhyung
> >
> >
> > > + found = true;
> > > + break;
> > > + }
> > > + }
> > > +
> > > + /* Set ena and run to non-zero */
> > > + count->ena = count->run = 1;
> > > + count->lost = 0;
> > > +
> > > + if (!found) {
> > > + /*
> > > + * Set default value or 0 when retire_latency for this event is
> > > + * not found from sampling data (record_tpebs not set or 0
> > > + * sample recorded).
> > > + */
> > > + count->val = 0;
> > > + return 0;
> > > + }
> > > +
> > > + /*
> > > + * Only set retire_latency value to the first CPU and thread.
> > > + */
> > > + if (cpu_map_idx == 0 && thread == 0)
> > > + val = rint(t->val);
> > > + else
> > > + val = 0;
> > > +
> > > + count->val = val;
> > > + return 0;
> > > +}
> > > +
> > > +static void tpebs_retire_lat__delete(struct tpebs_retire_lat *r)
> > > +{
> > > + zfree(&r->name);
> > > + free(r);
> > > +}
> > > +
> > > +
> > > +/*
> > > + * tpebs_delete - delete tpebs related data and stop the created thread
> and
> > > + * process by calling tpebs_stop().
> > > + *
> > > + * This function is called from evlist_delete() and also from builtin-stat
> > > + * stat_handle_error(). If tpebs_start() is called from places other then
> perf
> > > + * stat, need to ensure tpebs_delete() is also called to safely free mem and
> > > + * close the data read thread and the forked perf record process.
> > > + *
> > > + * This function is also called in evsel__close() to be symmetric with
> > > + * tpebs_start() being called in evsel__open(). We will update this call site
> > > + * when move tpebs_start() to evlist level.
> > > + */
> > > +void tpebs_delete(void)
> > > +{
> > > + struct tpebs_retire_lat *r, *rtmp;
> > > +
> > > + if (tpebs_pid == -1)
> > > + return;
> > > +
> > > + tpebs_stop();
> > > +
> > > + list_for_each_entry_safe(r, rtmp, &tpebs_results, nd) {
> > > + list_del_init(&r->nd);
> > > + tpebs_retire_lat__delete(r);
> > > + }
> > > +
> > > + if (tpebs_cmd) {
> > > + free(tpebs_cmd);
> > > + tpebs_cmd = NULL;
> > > + }
> > > +}
> > > diff --git a/tools/perf/util/intel-tpebs.h b/tools/perf/util/intel-tpebs.h
> > > new file mode 100644
> > > index 000000000000..766b3fbd79f1
> > > --- /dev/null
> > > +++ b/tools/perf/util/intel-tpebs.h
> > > @@ -0,0 +1,35 @@
> > > +/* SPDX-License-Identifier: GPL-2.0-only */
> > > +/*
> > > + * intel_tpebs.h: Intel TEPBS support
> > > + */
> > > +#ifndef INCLUDE__PERF_INTEL_TPEBS_H__
> > > +#define INCLUDE__PERF_INTEL_TPEBS_H__
> > > +
> > > +#include "stat.h"
> > > +#include "evsel.h"
> > > +
> > > +#ifdef HAVE_ARCH_X86_64_SUPPORT
> > > +
> > > +extern bool tpebs_recording;
> > > +int tpebs_start(struct evlist *evsel_list);
> > > +void tpebs_delete(void);
> > > +int tpebs_set_evsel(struct evsel *evsel, int cpu_map_idx, int thread);
> > > +
> > > +#else
> > > +
> > > +static inline int tpebs_start(struct evlist *evsel_list __maybe_unused)
> > > +{
> > > + return 0;
> > > +}
> > > +
> > > +static inline void tpebs_delete(void) {};
> > > +
> > > +static inline int tpebs_set_evsel(struct evsel *evsel __maybe_unused,
> > > + int cpu_map_idx __maybe_unused,
> > > + int thread __maybe_unused)
> > > +{
> > > + return 0;
> > > +}
> > > +
> > > +#endif
> > > +#endif
> > > --
> > > 2.43.0
> > >
Powered by blists - more mailing lists