lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAM9d7cgNK7c4Co6Sxar1XaW02nAg5F82S=AojuEsJzPGXsdfag@mail.gmail.com>
Date:	Wed, 24 Dec 2014 23:14:47 +0900
From:	Namhyung Kim <namhyung@...nel.org>
To:	Arnaldo Carvalho de Melo <acme@...nel.org>
Cc:	Ingo Molnar <mingo@...nel.org>,
	Peter Zijlstra <a.p.zijlstra@...llo.nl>,
	Jiri Olsa <jolsa@...hat.com>,
	LKML <linux-kernel@...r.kernel.org>,
	David Ahern <dsahern@...il.com>,
	Stephane Eranian <eranian@...gle.com>,
	Adrian Hunter <adrian.hunter@...el.com>,
	Andi Kleen <andi@...stfloor.org>,
	Frederic Weisbecker <fweisbec@...il.com>
Subject: Re: [PATCH 37/37] perf data: Implement 'split' subcommand

Hi Arnaldo,

On Wed, Dec 24, 2014 at 10:51 PM, Arnaldo Carvalho de Melo
<acme@...nel.org> wrote:
> Em Wed, Dec 24, 2014 at 04:15:33PM +0900, Namhyung Kim escreveu:
>> The perf data split command is for splitting a (large) single data
>> file into multiple files under a directory (perf.data.dir by default)
>> so that it can be processed and reported using multiple threads.
>
> How is it split? By CPU?
> Will the metadata stay in a different file?
> Please be as verbose on the description as possible :-)

It depends on the data file - if it's recorded system-wide, the split
will be done by cpu, otherwise by thread.  It's determined by checking
sample type has PERF_SAMPLE_CPU.  And metadata is saved in a separate
perf.header file like other multi-file data.  Will add it to the
change log.

Thanks,
Namhyung


>> Signed-off-by: Namhyung Kim <namhyung@...nel.org>
>> ---
>>  tools/perf/Documentation/perf-data.txt |  28 +++++
>>  tools/perf/builtin-data.c              | 223 +++++++++++++++++++++++++++++++++
>>  2 files changed, 251 insertions(+)
>>
>> diff --git a/tools/perf/Documentation/perf-data.txt b/tools/perf/Documentation/perf-data.txt
>> index b8c83947715c..42708702f10c 100644
>> --- a/tools/perf/Documentation/perf-data.txt
>> +++ b/tools/perf/Documentation/perf-data.txt
>> @@ -13,3 +13,31 @@ SYNOPSIS
>>  DESCRIPTION
>>  -----------
>>  Data file related processing.
>> +
>> +COMMANDS
>> +--------
>> +split::
>> +     Split single data file (perf.data) into multiple files under a directory
>> +     in order to be reported by multiple threads.
>> +
>> +OPTIONS for 'split'
>> +---------------------
>> +-i::
>> +--input::
>> +     Specify input perf data file path.
>> +
>> +-o::
>> +--output::
>> +     Specify output perf data directory path.
>> +
>> +-v::
>> +--verbose::
>> +        Be more verbose (show counter open errors, etc).
>> +
>> +-f::
>> +--force::
>> +        Don't complain, do it.
>> +
>> +SEE ALSO
>> +--------
>> +linkperf:perf[1], linkperf:perf-report[1]
>> diff --git a/tools/perf/builtin-data.c b/tools/perf/builtin-data.c
>> index 1eee97d020fa..5f3173826850 100644
>> --- a/tools/perf/builtin-data.c
>> +++ b/tools/perf/builtin-data.c
>> @@ -2,10 +2,14 @@
>>  #include "builtin.h"
>>  #include "perf.h"
>>  #include "debug.h"
>> +#include "session.h"
>> +#include "evlist.h"
>>  #include "parse-options.h"
>>
>>  typedef int (*data_cmd_fn_t)(int argc, const char **argv, const char *prefix);
>>
>> +static const char *output_name;
>> +
>>  struct data_cmd {
>>       const char      *name;
>>       const char      *summary;
>> @@ -41,10 +45,229 @@ static void print_usage(void)
>>       printf("\n");
>>  }
>>
>> +static int data_cmd_split(int argc, const char **argv, const char *prefix);
>> +
>>  static struct data_cmd data_cmds[] = {
>> +     { "split", "split single data file into multi-file", data_cmd_split },
>>       { NULL },
>>  };
>>
>> +#define FD_HASH_BITS  7
>> +#define FD_HASH_SIZE  (1 << FD_HASH_BITS)
>> +#define FD_HASH_MASK  (FD_HASH_SIZE - 1)
>> +
>> +struct data_split {
>> +     struct perf_tool        tool;
>> +     struct perf_session     *session;
>> +     enum {
>> +             PER_CPU,
>> +             PER_THREAD,
>> +     } mode;
>> +     int                     header_fd;
>> +     u64                     header_written;
>> +     struct hlist_head       fd_hash[FD_HASH_SIZE];
>> +     int                     fd_hash_nr;
>> +};
>> +
>> +struct fdhash_node {
>> +     int                     id;
>> +     int                     fd;
>> +     struct hlist_node       list;
>> +};
>> +
>> +static struct hlist_head *get_hash(struct data_split *split, int id)
>> +{
>> +     return &split->fd_hash[id % FD_HASH_MASK];
>> +}
>> +
>> +static int perf_event__rewrite_header(struct perf_tool *tool,
>> +                                   union perf_event *event)
>> +{
>> +     struct data_split *split = container_of(tool, struct data_split, tool);
>> +     ssize_t size;
>> +
>> +     size = writen(split->header_fd, event, event->header.size);
>> +     if (size < 0)
>> +             return -errno;
>> +
>> +     split->header_written += size;
>> +     return 0;
>> +}
>> +
>> +static int split_other_events(struct perf_tool *tool,
>> +                             union perf_event *event,
>> +                             struct perf_sample *sample __maybe_unused,
>> +                             struct machine *machine __maybe_unused)
>> +{
>> +     return perf_event__rewrite_header(tool, event);
>> +}
>> +
>> +static int split_sample_event(struct perf_tool *tool,
>> +                             union perf_event *event,
>> +                             struct perf_sample *sample,
>> +                             struct perf_evsel *evsel __maybe_unused,
>> +                             struct machine *machine __maybe_unused)
>> +{
>> +     struct data_split *split = container_of(tool, struct data_split, tool);
>> +     int id = split->mode == PER_CPU ? sample->cpu : sample->tid;
>> +     int fd = -1;
>> +     char buf[PATH_MAX];
>> +     struct hlist_head *head;
>> +     struct fdhash_node *node;
>> +
>> +     head = get_hash(split, id);
>> +     hlist_for_each_entry(node, head, list) {
>> +             if (node->id == id) {
>> +                     fd = node->fd;
>> +                     break;
>> +             }
>> +     }
>> +
>> +     if (fd == -1) {
>> +             scnprintf(buf, sizeof(buf), "%s/perf.data.%d",
>> +                       output_name, split->fd_hash_nr++);
>> +
>> +             fd = open(buf, O_RDWR|O_CREAT|O_TRUNC, 0600);
>> +             if (fd < 0) {
>> +                     pr_err("cannot open data file: %s: %m\n", buf);
>> +                     return -1;
>> +             }
>> +
>> +             node = malloc(sizeof(*node));
>> +             if (node == NULL) {
>> +                     pr_err("memory allocation failed\n");
>> +                     return -1;
>> +             }
>> +
>> +             node->id = id;
>> +             node->fd = fd;
>> +
>> +             hlist_add_head(&node->list, head);
>> +     }
>> +
>> +     return writen(fd, event, event->header.size) > 0 ? 0 : -errno;
>> +}
>> +
>> +static int __data_cmd_split(struct data_split *split)
>> +{
>> +     struct perf_session *session = split->session;
>> +     char *output = NULL;
>> +     char buf[PATH_MAX];
>> +     u64 sample_type;
>> +     int header_fd;
>> +     int ret = -1;
>> +     int i;
>> +
>> +     if (!output_name) {
>> +             if (asprintf(&output, "%s.dir", input_name) < 0) {
>> +                     pr_err("memory allocation failed\n");
>> +                     return -1;
>> +             }
>> +             output_name = output;
>> +     }
>> +
>> +     mkdir(output_name, 0700);
>> +
>> +     /*
>> +      * This is necessary to write (copy) build-id table.  After
>> +      * processing header, dsos list will contain dso which was on
>> +      * the original build-id table.
>> +      */
>> +     dsos__hit_all(session);
>> +
>> +     scnprintf(buf, sizeof(buf), "%s/perf.header", output_name);
>> +     header_fd = open(buf, O_RDWR|O_CREAT|O_TRUNC, 0600);
>> +     if (header_fd < 0) {
>> +             pr_err("cannot open header file: %s: %m\n", buf);
>> +             goto out;
>> +     }
>> +
>> +     lseek(header_fd, session->header.data_offset, SEEK_SET);
>> +
>> +     sample_type = perf_evlist__combined_sample_type(session->evlist);
>> +     if (sample_type & PERF_SAMPLE_CPU)
>> +             split->mode = PER_CPU;
>> +     else
>> +             split->mode = PER_THREAD;
>> +
>> +     pr_debug("splitting data file for %s\n",
>> +              split->mode == PER_CPU ? "CPUs" : "threads");
>> +
>> +     split->header_fd = header_fd;
>> +     perf_session__process_events(session, &split->tool);
>> +
>> +     for (i = 0; i < FD_HASH_SIZE; i++) {
>> +             struct fdhash_node *pos;
>> +             struct hlist_node *tmp;
>> +
>> +             hlist_for_each_entry_safe(pos, tmp, &split->fd_hash[i], list) {
>> +                     hlist_del(&pos->list);
>> +                     close(pos->fd);
>> +                     free(pos);
>> +             }
>> +     }
>> +
>> +     session->header.data_size = split->header_written;
>> +     perf_session__write_header(session, session->evlist, header_fd, true);
>> +
>> +     close(header_fd);
>> +     ret = 0;
>> +out:
>> +     free(output);
>> +     return ret;
>> +}
>> +
>> +int data_cmd_split(int argc, const char **argv, const char *prefix __maybe_unused)
>> +{
>> +     bool force = false;
>> +     struct perf_session *session;
>> +     struct perf_data_file file = {
>> +             .mode  = PERF_DATA_MODE_READ,
>> +     };
>> +     struct data_split split = {
>> +             .tool = {
>> +                     .sample         = split_sample_event,
>> +                     .fork           = split_other_events,
>> +                     .comm           = split_other_events,
>> +                     .exit           = split_other_events,
>> +                     .mmap           = split_other_events,
>> +                     .mmap2          = split_other_events,
>> +                     .lost           = split_other_events,
>> +                     .throttle       = split_other_events,
>> +                     .unthrottle     = split_other_events,
>> +             },
>> +     };
>> +     const char * const split_usage[] = {
>> +             "perf data split [<options>]",
>> +             NULL
>> +     };
>> +     const struct option split_options[] = {
>> +     OPT_STRING('i', "input", &input_name, "file", "input file name"),
>> +     OPT_STRING('o', "output", &output_name, "file", "output directory name"),
>> +     OPT_BOOLEAN('f', "force", &force, "don't complain, do it"),
>> +     OPT_INCR('v', "verbose", &verbose, "be more verbose"),
>> +     OPT_END()
>> +     };
>> +
>> +     argc = parse_options(argc, argv, split_options, split_usage, 0);
>> +     if (argc)
>> +             usage_with_options(split_usage, split_options);
>> +
>> +     file.path = input_name;
>> +     file.force = force;
>> +     session = perf_session__new(&file, false, &split.tool);
>> +     if (session == NULL)
>> +             return -1;
>> +
>> +     split.session = session;
>> +     symbol__init(&session->header.env);
>> +
>> +     __data_cmd_split(&split);
>> +
>> +     perf_session__delete(session);
>> +     return 0;
>> +}
>> +
>>  int cmd_data(int argc, const char **argv, const char *prefix)
>>  {
>>       struct data_cmd *cmd;
>> --
>> 2.1.3



-- 
Thanks,
Namhyung
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ