lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <84ed17c4-c31e-ba94-0f5b-30bff707c7cf@linux.intel.com>
Date:   Wed, 31 Oct 2018 12:19:10 -0400
From:   "Liang, Kan" <kan.liang@...ux.intel.com>
To:     Arnaldo Carvalho de Melo <acme@...nel.org>,
        David Miller <davem@...emloft.net>
Cc:     linux-kernel@...r.kernel.org, Kan Liang <kan.liang@...el.com>,
        David Ahern <dsahern@...il.com>,
        Ying Huang <ying.huang@...el.com>
Subject: Re: [PATCH RFC] perf: Go back to by-hand proc mmap parsing and kill
 timeout



On 10/31/2018 9:21 AM, Arnaldo Carvalho de Melo wrote:
> Em Tue, Oct 30, 2018 at 10:54:16PM -0700, David Miller escreveu:
>>
>> This goes back to by-hand parsing of the proc mmap file, and removes
>> the timeout.
>>
>> In my measurements this makes the parsing about twice as fast.
>>
>> Profiling thread synthesizing shows that most of the time is spent in
>> the sscanf() call.
>>

I did some performance test to evaluate the time cost on proc mmap 
parsing with different options/patches.

The benchmark I used is case-small-allocs from vm-scalability.

I did the test on two machines.
M1: Skylake server (224 CPUs)
M2: Knights Landing (272 CPUs)

With three different options for perf top.
P1: baseline test. No David's patch. "perf top"
P2: No David's patch. Set a big timeout. The proc map processing will 
not be truncated. "perf top --proc-map-timeout 10000"
P3: With David's patch. "perf top"

The elapsed time is the time cost for synthesize_threads.

On M1
	Elapsed time (Seconds)
P1	~6
P2	8-24*
P3	6-11*
*The time cost is not stable. I did several tests. Here lists a range of 
the results.

On M2
	Elapsed time (Seconds)
P1	41
P2	282
P3	176
I only did test one time.

The conclusion,
- David's patch significantly improve the time cost of proc mmap parsing.
- On some platform, especially Knights Landing, the time cost is still 
huge even with David's patch.
- I think timeout is still needed. We have to give user warning/hints if 
the processing time is too long. For example, with P3 on M2, I can only 
see a black screen during the ~3 mins processing time.

Thanks,
Kan

>> Processing samples is critical for perf top and perf record -a startup
>> performance, and perf's ability to keep up with high sample rates.
> 
> I'm going to split this patch into two, one that makes it go back to
> by-hand proc mmap parsing, and the other killing the mmap-timeout.
> 
> Thanks,
> 
> - Arnaldo
>   
>> Signed-off-by: David S. Miller <davem@...emloft.net>
>>
>> diff --git a/tools/perf/builtin-kvm.c b/tools/perf/builtin-kvm.c
>> index 2b1ef704169f..d4b5177c3c22 100644
>> --- a/tools/perf/builtin-kvm.c
>> +++ b/tools/perf/builtin-kvm.c
>> @@ -1364,8 +1364,6 @@ static int kvm_events_live(struct perf_kvm_stat *kvm,
>>   			"show events other than"
>>   			" HLT (x86 only) or Wait state (s390 only)"
>>   			" that take longer than duration usecs"),
>> -		OPT_UINTEGER(0, "proc-map-timeout", &kvm->opts.proc_map_timeout,
>> -				"per thread proc mmap processing timeout in ms"),
>>   		OPT_END()
>>   	};
>>   	const char * const live_usage[] = {
>> @@ -1394,7 +1392,6 @@ static int kvm_events_live(struct perf_kvm_stat *kvm,
>>   	kvm->opts.target.uses_mmap = false;
>>   	kvm->opts.target.uid_str = NULL;
>>   	kvm->opts.target.uid = UINT_MAX;
>> -	kvm->opts.proc_map_timeout = 500;
>>   
>>   	symbol__init(NULL);
>>   	disable_buildid_cache();
>> @@ -1453,8 +1450,7 @@ static int kvm_events_live(struct perf_kvm_stat *kvm,
>>   	perf_session__set_id_hdr_size(kvm->session);
>>   	ordered_events__set_copy_on_queue(&kvm->session->ordered_events, true);
>>   	machine__synthesize_threads(&kvm->session->machines.host, &kvm->opts.target,
>> -				    kvm->evlist->threads, false,
>> -				    kvm->opts.proc_map_timeout, 1);
>> +				    kvm->evlist->threads, false, 1);
>>   	err = kvm_live_open_events(kvm);
>>   	if (err)
>>   		goto out;
>> diff --git a/tools/perf/builtin-record.c b/tools/perf/builtin-record.c
>> index 0980dfe3396b..4259ea42a8ef 100644
>> --- a/tools/perf/builtin-record.c
>> +++ b/tools/perf/builtin-record.c
>> @@ -633,8 +633,7 @@ static int record__synthesize_workload(struct record *rec, bool tail)
>>   	err = perf_event__synthesize_thread_map(&rec->tool, thread_map,
>>   						 process_synthesized_event,
>>   						 &rec->session->machines.host,
>> -						 rec->opts.sample_address,
>> -						 rec->opts.proc_map_timeout);
>> +						 rec->opts.sample_address);
>>   	thread_map__put(thread_map);
>>   	return err;
>>   }
>> @@ -848,8 +847,7 @@ static int record__synthesize(struct record *rec, bool tail)
>>   	}
>>   
>>   	err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
>> -					    process_synthesized_event, opts->sample_address,
>> -					    opts->proc_map_timeout, 1);
>> +					    process_synthesized_event, opts->sample_address, 1);
>>   out:
>>   	return err;
>>   }
>> @@ -1521,7 +1519,6 @@ static struct record record = {
>>   			.uses_mmap   = true,
>>   			.default_per_cpu = true,
>>   		},
>> -		.proc_map_timeout     = 500,
>>   	},
>>   	.tool = {
>>   		.sample		= process_sample_event,
>> @@ -1651,8 +1648,6 @@ static struct option __record_options[] = {
>>   	parse_clockid),
>>   	OPT_STRING_OPTARG('S', "snapshot", &record.opts.auxtrace_snapshot_opts,
>>   			  "opts", "AUX area tracing Snapshot Mode", ""),
>> -	OPT_UINTEGER(0, "proc-map-timeout", &record.opts.proc_map_timeout,
>> -			"per thread proc mmap processing timeout in ms"),
>>   	OPT_BOOLEAN(0, "namespaces", &record.opts.record_namespaces,
>>   		    "Record namespaces events"),
>>   	OPT_BOOLEAN(0, "switch-events", &record.opts.record_switch_events,
>> diff --git a/tools/perf/builtin-top.c b/tools/perf/builtin-top.c
>> index d21d8751e749..4eec44d815bd 100644
>> --- a/tools/perf/builtin-top.c
>> +++ b/tools/perf/builtin-top.c
>> @@ -1096,7 +1096,6 @@ static int __cmd_top(struct perf_top *top)
>>   
>>   	machine__synthesize_threads(&top->session->machines.host, &opts->target,
>>   				    top->evlist->threads, false,
>> -				    opts->proc_map_timeout,
>>   				    top->nr_threads_synthesize);
>>   
>>   	if (top->nr_threads_synthesize > 1)
>> @@ -1256,7 +1255,6 @@ int cmd_top(int argc, const char **argv)
>>   			.target		= {
>>   				.uses_mmap   = true,
>>   			},
>> -			.proc_map_timeout    = 500,
>>   			.overwrite	= 1,
>>   		},
>>   		.max_stack	     = sysctl__max_stack(),
>> @@ -1360,8 +1358,6 @@ int cmd_top(int argc, const char **argv)
>>   	OPT_STRING('w', "column-widths", &symbol_conf.col_width_list_str,
>>   		   "width[,width...]",
>>   		   "don't try to adjust column width, use these fixed values"),
>> -	OPT_UINTEGER(0, "proc-map-timeout", &opts->proc_map_timeout,
>> -			"per thread proc mmap processing timeout in ms"),
>>   	OPT_CALLBACK_NOOPT('b', "branch-any", &opts->branch_stack,
>>   		     "branch any", "sample any taken branches",
>>   		     parse_branch_stack),
>> diff --git a/tools/perf/builtin-trace.c b/tools/perf/builtin-trace.c
>> index 90289f31dd87..0b68e0538107 100644
>> --- a/tools/perf/builtin-trace.c
>> +++ b/tools/perf/builtin-trace.c
>> @@ -1240,8 +1240,7 @@ static int trace__symbols_init(struct trace *trace, struct perf_evlist *evlist)
>>   		goto out;
>>   
>>   	err = __machine__synthesize_threads(trace->host, &trace->tool, &trace->opts.target,
>> -					    evlist->threads, trace__tool_process, false,
>> -					    trace->opts.proc_map_timeout, 1);
>> +					    evlist->threads, trace__tool_process, false, 1);
>>   out:
>>   	if (err)
>>   		symbol__exit();
>> @@ -3242,7 +3241,6 @@ int cmd_trace(int argc, const char **argv)
>>   			.user_interval = ULLONG_MAX,
>>   			.no_buffering  = true,
>>   			.mmap_pages    = UINT_MAX,
>> -			.proc_map_timeout  = 500,
>>   		},
>>   		.output = stderr,
>>   		.show_comm = true,
>> @@ -3310,8 +3308,6 @@ int cmd_trace(int argc, const char **argv)
>>   		     "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)),
>>   	OPT_BOOLEAN(0, "print-sample", &trace.print_sample,
>>   			"print the PERF_RECORD_SAMPLE PERF_SAMPLE_ info, for debugging"),
>> -	OPT_UINTEGER(0, "proc-map-timeout", &trace.opts.proc_map_timeout,
>> -			"per thread proc mmap processing timeout in ms"),
>>   	OPT_CALLBACK('G', "cgroup", &trace, "name", "monitor event in cgroup name only",
>>   		     trace__parse_cgroups),
>>   	OPT_UINTEGER('D', "delay", &trace.opts.initial_delay,
>> diff --git a/tools/perf/perf.h b/tools/perf/perf.h
>> index 21bf7f5a3cf5..0e2aafd8bb6e 100644
>> --- a/tools/perf/perf.h
>> +++ b/tools/perf/perf.h
>> @@ -81,7 +81,6 @@ struct record_opts {
>>   	unsigned     initial_delay;
>>   	bool         use_clockid;
>>   	clockid_t    clockid;
>> -	unsigned int proc_map_timeout;
>>   };
>>   
>>   struct option;
>> diff --git a/tools/perf/tests/code-reading.c b/tools/perf/tests/code-reading.c
>> index 6b049f3f5cf4..dbf2c69944d2 100644
>> --- a/tools/perf/tests/code-reading.c
>> +++ b/tools/perf/tests/code-reading.c
>> @@ -599,7 +599,7 @@ static int do_test_code_reading(bool try_kcore)
>>   	}
>>   
>>   	ret = perf_event__synthesize_thread_map(NULL, threads,
>> -						perf_event__process, machine, false, 500);
>> +						perf_event__process, machine, false);
>>   	if (ret < 0) {
>>   		pr_debug("perf_event__synthesize_thread_map failed\n");
>>   		goto out_err;
>> diff --git a/tools/perf/tests/dwarf-unwind.c b/tools/perf/tests/dwarf-unwind.c
>> index 2f008067d989..7c8d2e422401 100644
>> --- a/tools/perf/tests/dwarf-unwind.c
>> +++ b/tools/perf/tests/dwarf-unwind.c
>> @@ -34,7 +34,7 @@ static int init_live_machine(struct machine *machine)
>>   	pid_t pid = getpid();
>>   
>>   	return perf_event__synthesize_mmap_events(NULL, &event, pid, pid,
>> -						  mmap_handler, machine, true, 500);
>> +						  mmap_handler, machine, true);
>>   }
>>   
>>   /*
>> diff --git a/tools/perf/tests/mmap-thread-lookup.c b/tools/perf/tests/mmap-thread-lookup.c
>> index b1af2499a3c9..5ede9b561d32 100644
>> --- a/tools/perf/tests/mmap-thread-lookup.c
>> +++ b/tools/perf/tests/mmap-thread-lookup.c
>> @@ -132,7 +132,7 @@ static int synth_all(struct machine *machine)
>>   {
>>   	return perf_event__synthesize_threads(NULL,
>>   					      perf_event__process,
>> -					      machine, 0, 500, 1);
>> +					      machine, 0, 1);
>>   }
>>   
>>   static int synth_process(struct machine *machine)
>> @@ -144,7 +144,7 @@ static int synth_process(struct machine *machine)
>>   
>>   	err = perf_event__synthesize_thread_map(NULL, map,
>>   						perf_event__process,
>> -						machine, 0, 500);
>> +						machine, 0);
>>   
>>   	thread_map__put(map);
>>   	return err;
>> diff --git a/tools/perf/util/event.c b/tools/perf/util/event.c
>> index bc646185f8d9..92b09b2ba6bf 100644
>> --- a/tools/perf/util/event.c
>> +++ b/tools/perf/util/event.c
>> @@ -317,19 +317,39 @@ static int perf_event__synthesize_fork(struct perf_tool *tool,
>>   	return 0;
>>   }
>>   
>> +static int dec(char ch)
>> +{
>> +	if ((ch >= '0') && (ch <= '9'))
>> +		return ch - '0';
>> +	return -1;
>> +}
>> +
>> +static int dec2u64(const char *ptr, u64 *long_val)
>> +{
>> +	const char *p = ptr;
>> +
>> +	*long_val = 0;
>> +	while (*p) {
>> +		const int dec_val = dec(*p);
>> +
>> +		if (dec_val < 0)
>> +			break;
>> +
>> +		*long_val = (*long_val * 10) + dec_val;
>> +		p++;
>> +	}
>> +	return p - ptr;
>> +}
>> +
>>   int perf_event__synthesize_mmap_events(struct perf_tool *tool,
>>   				       union perf_event *event,
>>   				       pid_t pid, pid_t tgid,
>>   				       perf_event__handler_t process,
>>   				       struct machine *machine,
>> -				       bool mmap_data,
>> -				       unsigned int proc_map_timeout)
>> +				       bool mmap_data)
>>   {
>>   	char filename[PATH_MAX];
>>   	FILE *fp;
>> -	unsigned long long t;
>> -	bool truncation = false;
>> -	unsigned long long timeout = proc_map_timeout * 1000000ULL;
>>   	int rc = 0;
>>   	const char *hugetlbfs_mnt = hugetlbfs__mountpoint();
>>   	int hugetlbfs_mnt_len = hugetlbfs_mnt ? strlen(hugetlbfs_mnt) : 0;
>> @@ -350,87 +370,99 @@ int perf_event__synthesize_mmap_events(struct perf_tool *tool,
>>   	}
>>   
>>   	event->header.type = PERF_RECORD_MMAP2;
>> -	t = rdclock();
>>   
>>   	while (1) {
>> -		char bf[BUFSIZ];
>> -		char prot[5];
>> -		char execname[PATH_MAX];
>> +		char bf[BUFSIZ], *pbf = bf;
>>   		char anonstr[] = "//anon";
>> -		unsigned int ino;
>> +		char *execname;
>>   		size_t size;
>>   		ssize_t n;
>> +		u64 tmp;
>>   
>>   		if (fgets(bf, sizeof(bf), fp) == NULL)
>>   			break;
>>   
>> -		if ((rdclock() - t) > timeout) {
>> -			pr_warning("Reading %s time out. "
>> -				   "You may want to increase "
>> -				   "the time limit by --proc-map-timeout\n",
>> -				   filename);
>> -			truncation = true;
>> -			goto out;
>> -		}
>> -
>> -		/* ensure null termination since stack will be reused. */
>> -		strcpy(execname, "");
>> -
>>   		/* 00400000-0040c000 r-xp 00000000 fd:01 41038  /bin/cat */
>> -		n = sscanf(bf, "%"PRIx64"-%"PRIx64" %s %"PRIx64" %x:%x %u %[^\n]\n",
>> -		       &event->mmap2.start, &event->mmap2.len, prot,
>> -		       &event->mmap2.pgoff, &event->mmap2.maj,
>> -		       &event->mmap2.min,
>> -		       &ino, execname);
>> -
>> -		/*
>> - 		 * Anon maps don't have the execname.
>> - 		 */
>> -		if (n < 7)
>> +		n = hex2u64(pbf, &event->mmap2.start);
>> +		if (n < 0)
>>   			continue;
>> -
>> -		event->mmap2.ino = (u64)ino;
>> -
>> -		/*
>> -		 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
>> -		 */
>> -		if (machine__is_host(machine))
>> -			event->header.misc = PERF_RECORD_MISC_USER;
>> -		else
>> -			event->header.misc = PERF_RECORD_MISC_GUEST_USER;
>> +		pbf += n + 1;
>> +		n = hex2u64(pbf, &event->mmap2.len);
>> +		if (n < 0)
>> +			continue;
>> +		pbf += n + 1;
>>   
>>   		/* map protection and flags bits */
>>   		event->mmap2.prot = 0;
>>   		event->mmap2.flags = 0;
>> -		if (prot[0] == 'r')
>> +		if (pbf[0] == 'r')
>>   			event->mmap2.prot |= PROT_READ;
>> -		if (prot[1] == 'w')
>> +		if (pbf[1] == 'w')
>>   			event->mmap2.prot |= PROT_WRITE;
>> -		if (prot[2] == 'x')
>> +		if (pbf[2] == 'x')
>>   			event->mmap2.prot |= PROT_EXEC;
>>   
>> -		if (prot[3] == 's')
>> +		if (pbf[3] == 's')
>>   			event->mmap2.flags |= MAP_SHARED;
>>   		else
>>   			event->mmap2.flags |= MAP_PRIVATE;
>>   
>> -		if (prot[2] != 'x') {
>> -			if (!mmap_data || prot[0] != 'r')
>> +		if (pbf[2] != 'x') {
>> +			if (!mmap_data || pbf[0] != 'r')
>>   				continue;
>>   
>>   			event->header.misc |= PERF_RECORD_MISC_MMAP_DATA;
>>   		}
>>   
>> -out:
>> -		if (truncation)
>> -			event->header.misc |= PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT;
>> +		pbf += 5;
>> +		n = hex2u64(pbf, &event->mmap2.pgoff);
>> +		if (n < 0)
>> +			continue;
>> +		pbf += n + 1;
>> +
>> +		n = hex2u64(pbf, &tmp);
>> +		if (n < 0)
>> +			continue;
>> +		event->mmap2.maj = tmp;
>> +		pbf += n + 1;
>> +
>> +		n = hex2u64(pbf, &tmp);
>> +		if (n < 0)
>> +			continue;
>> +		event->mmap2.min = tmp;
>> +		pbf += n + 1;
>>   
>> -		if (!strcmp(execname, ""))
>> -			strcpy(execname, anonstr);
>> +		n = dec2u64(pbf, &event->mmap2.ino);
>> +		if (n < 0)
>> +			continue;
>> +		pbf += n;
>> +
>> +		execname = strchr(pbf, '/');
>> +		if (!execname)
>> +			execname = strchr(pbf, '[');
>> +
>> +		/*
>> +		 * Anon map, skip.
>> +		 */
>> +		if (!execname)
>> +			continue;
>> +
>> +		pbf = strchr(execname, '\n');
>> +		if (!pbf)
>> +			continue;
>> +		*pbf = '\0';
>> +
>> +		/*
>> +		 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
>> +		 */
>> +		if (machine__is_host(machine))
>> +			event->header.misc = PERF_RECORD_MISC_USER;
>> +		else
>> +			event->header.misc = PERF_RECORD_MISC_GUEST_USER;
>>   
>>   		if (hugetlbfs_mnt_len &&
>>   		    !strncmp(execname, hugetlbfs_mnt, hugetlbfs_mnt_len)) {
>> -			strcpy(execname, anonstr);
>> +			execname = anonstr;
>>   			event->mmap2.flags |= MAP_HUGETLB;
>>   		}
>>   
>> @@ -449,9 +481,6 @@ int perf_event__synthesize_mmap_events(struct perf_tool *tool,
>>   			rc = -1;
>>   			break;
>>   		}
>> -
>> -		if (truncation)
>> -			break;
>>   	}
>>   
>>   	fclose(fp);
>> @@ -520,8 +549,7 @@ static int __event__synthesize_thread(union perf_event *comm_event,
>>   				      perf_event__handler_t process,
>>   				      struct perf_tool *tool,
>>   				      struct machine *machine,
>> -				      bool mmap_data,
>> -				      unsigned int proc_map_timeout)
>> +				      bool mmap_data)
>>   {
>>   	char filename[PATH_MAX];
>>   	DIR *tasks;
>> @@ -547,8 +575,7 @@ static int __event__synthesize_thread(union perf_event *comm_event,
>>   		 */
>>   		if (pid == tgid &&
>>   		    perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
>> -						       process, machine, mmap_data,
>> -						       proc_map_timeout))
>> +						       process, machine, mmap_data))
>>   			return -1;
>>   
>>   		return 0;
>> @@ -597,7 +624,7 @@ static int __event__synthesize_thread(union perf_event *comm_event,
>>   		if (_pid == pid) {
>>   			/* process the parent's maps too */
>>   			rc = perf_event__synthesize_mmap_events(tool, mmap_event, pid, tgid,
>> -						process, machine, mmap_data, proc_map_timeout);
>> +						process, machine, mmap_data);
>>   			if (rc)
>>   				break;
>>   		}
>> @@ -611,8 +638,7 @@ int perf_event__synthesize_thread_map(struct perf_tool *tool,
>>   				      struct thread_map *threads,
>>   				      perf_event__handler_t process,
>>   				      struct machine *machine,
>> -				      bool mmap_data,
>> -				      unsigned int proc_map_timeout)
>> +				      bool mmap_data)
>>   {
>>   	union perf_event *comm_event, *mmap_event, *fork_event;
>>   	union perf_event *namespaces_event;
>> @@ -642,7 +668,7 @@ int perf_event__synthesize_thread_map(struct perf_tool *tool,
>>   					       fork_event, namespaces_event,
>>   					       thread_map__pid(threads, thread), 0,
>>   					       process, tool, machine,
>> -					       mmap_data, proc_map_timeout)) {
>> +					       mmap_data)) {
>>   			err = -1;
>>   			break;
>>   		}
>> @@ -668,7 +694,7 @@ int perf_event__synthesize_thread_map(struct perf_tool *tool,
>>   						       fork_event, namespaces_event,
>>   						       comm_event->comm.pid, 0,
>>   						       process, tool, machine,
>> -						       mmap_data, proc_map_timeout)) {
>> +						       mmap_data)) {
>>   				err = -1;
>>   				break;
>>   			}
>> @@ -689,7 +715,6 @@ static int __perf_event__synthesize_threads(struct perf_tool *tool,
>>   					    perf_event__handler_t process,
>>   					    struct machine *machine,
>>   					    bool mmap_data,
>> -					    unsigned int proc_map_timeout,
>>   					    struct dirent **dirent,
>>   					    int start,
>>   					    int num)
>> @@ -733,8 +758,7 @@ static int __perf_event__synthesize_threads(struct perf_tool *tool,
>>   		 */
>>   		__event__synthesize_thread(comm_event, mmap_event, fork_event,
>>   					   namespaces_event, pid, 1, process,
>> -					   tool, machine, mmap_data,
>> -					   proc_map_timeout);
>> +					   tool, machine, mmap_data);
>>   	}
>>   	err = 0;
>>   
>> @@ -754,7 +778,6 @@ struct synthesize_threads_arg {
>>   	perf_event__handler_t process;
>>   	struct machine *machine;
>>   	bool mmap_data;
>> -	unsigned int proc_map_timeout;
>>   	struct dirent **dirent;
>>   	int num;
>>   	int start;
>> @@ -766,7 +789,7 @@ static void *synthesize_threads_worker(void *arg)
>>   
>>   	__perf_event__synthesize_threads(args->tool, args->process,
>>   					 args->machine, args->mmap_data,
>> -					 args->proc_map_timeout, args->dirent,
>> +					 args->dirent,
>>   					 args->start, args->num);
>>   	return NULL;
>>   }
>> @@ -775,7 +798,6 @@ int perf_event__synthesize_threads(struct perf_tool *tool,
>>   				   perf_event__handler_t process,
>>   				   struct machine *machine,
>>   				   bool mmap_data,
>> -				   unsigned int proc_map_timeout,
>>   				   unsigned int nr_threads_synthesize)
>>   {
>>   	struct synthesize_threads_arg *args = NULL;
>> @@ -805,7 +827,6 @@ int perf_event__synthesize_threads(struct perf_tool *tool,
>>   	if (thread_nr <= 1) {
>>   		err = __perf_event__synthesize_threads(tool, process,
>>   						       machine, mmap_data,
>> -						       proc_map_timeout,
>>   						       dirent, base, n);
>>   		goto free_dirent;
>>   	}
>> @@ -827,7 +848,6 @@ int perf_event__synthesize_threads(struct perf_tool *tool,
>>   		args[i].process = process;
>>   		args[i].machine = machine;
>>   		args[i].mmap_data = mmap_data;
>> -		args[i].proc_map_timeout = proc_map_timeout;
>>   		args[i].dirent = dirent;
>>   	}
>>   	for (i = 0; i < m; i++) {
>> diff --git a/tools/perf/util/event.h b/tools/perf/util/event.h
>> index bfa60bcafbde..45a3b47fc82c 100644
>> --- a/tools/perf/util/event.h
>> +++ b/tools/perf/util/event.h
>> @@ -412,7 +412,6 @@ struct events_stats {
>>   	u32 nr_unknown_id;
>>   	u32 nr_unprocessable_samples;
>>   	u32 nr_auxtrace_errors[PERF_AUXTRACE_ERROR_MAX];
>> -	u32 nr_proc_map_timeout;
>>   };
>>   
>>   enum {
>> @@ -669,8 +668,7 @@ typedef int (*perf_event__handler_t)(struct perf_tool *tool,
>>   int perf_event__synthesize_thread_map(struct perf_tool *tool,
>>   				      struct thread_map *threads,
>>   				      perf_event__handler_t process,
>> -				      struct machine *machine, bool mmap_data,
>> -				      unsigned int proc_map_timeout);
>> +				      struct machine *machine, bool mmap_data);
>>   int perf_event__synthesize_thread_map2(struct perf_tool *tool,
>>   				      struct thread_map *threads,
>>   				      perf_event__handler_t process,
>> @@ -682,7 +680,6 @@ int perf_event__synthesize_cpu_map(struct perf_tool *tool,
>>   int perf_event__synthesize_threads(struct perf_tool *tool,
>>   				   perf_event__handler_t process,
>>   				   struct machine *machine, bool mmap_data,
>> -				   unsigned int proc_map_timeout,
>>   				   unsigned int nr_threads_synthesize);
>>   int perf_event__synthesize_kernel_mmap(struct perf_tool *tool,
>>   				       perf_event__handler_t process,
>> @@ -797,8 +794,7 @@ int perf_event__synthesize_mmap_events(struct perf_tool *tool,
>>   				       pid_t pid, pid_t tgid,
>>   				       perf_event__handler_t process,
>>   				       struct machine *machine,
>> -				       bool mmap_data,
>> -				       unsigned int proc_map_timeout);
>> +				       bool mmap_data);
>>   
>>   int perf_event__synthesize_extra_kmaps(struct perf_tool *tool,
>>   				       perf_event__handler_t process,
>> diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
>> index 111ae858cbcb..6427a74ebc4a 100644
>> --- a/tools/perf/util/machine.c
>> +++ b/tools/perf/util/machine.c
>> @@ -2443,15 +2443,13 @@ int machines__for_each_thread(struct machines *machines,
>>   int __machine__synthesize_threads(struct machine *machine, struct perf_tool *tool,
>>   				  struct target *target, struct thread_map *threads,
>>   				  perf_event__handler_t process, bool data_mmap,
>> -				  unsigned int proc_map_timeout,
>>   				  unsigned int nr_threads_synthesize)
>>   {
>>   	if (target__has_task(target))
>> -		return perf_event__synthesize_thread_map(tool, threads, process, machine, data_mmap, proc_map_timeout);
>> +		return perf_event__synthesize_thread_map(tool, threads, process, machine, data_mmap);
>>   	else if (target__has_cpu(target))
>>   		return perf_event__synthesize_threads(tool, process,
>>   						      machine, data_mmap,
>> -						      proc_map_timeout,
>>   						      nr_threads_synthesize);
>>   	/* command specified */
>>   	return 0;
>> diff --git a/tools/perf/util/machine.h b/tools/perf/util/machine.h
>> index d856b85862e2..688d84ba823a 100644
>> --- a/tools/perf/util/machine.h
>> +++ b/tools/perf/util/machine.h
>> @@ -247,17 +247,14 @@ int machines__for_each_thread(struct machines *machines,
>>   int __machine__synthesize_threads(struct machine *machine, struct perf_tool *tool,
>>   				  struct target *target, struct thread_map *threads,
>>   				  perf_event__handler_t process, bool data_mmap,
>> -				  unsigned int proc_map_timeout,
>>   				  unsigned int nr_threads_synthesize);
>>   static inline
>>   int machine__synthesize_threads(struct machine *machine, struct target *target,
>>   				struct thread_map *threads, bool data_mmap,
>> -				unsigned int proc_map_timeout,
>>   				unsigned int nr_threads_synthesize)
>>   {
>>   	return __machine__synthesize_threads(machine, NULL, target, threads,
>>   					     perf_event__process, data_mmap,
>> -					     proc_map_timeout,
>>   					     nr_threads_synthesize);
>>   }
>>   
>> diff --git a/tools/perf/util/session.c b/tools/perf/util/session.c
>> index 7d2c8ce6cfad..33d2c50cdc18 100644
>> --- a/tools/perf/util/session.c
>> +++ b/tools/perf/util/session.c
>> @@ -1261,8 +1261,6 @@ static int machines__deliver_event(struct machines *machines,
>>   	case PERF_RECORD_MMAP:
>>   		return tool->mmap(tool, event, sample, machine);
>>   	case PERF_RECORD_MMAP2:
>> -		if (event->header.misc & PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT)
>> -			++evlist->stats.nr_proc_map_timeout;
>>   		return tool->mmap2(tool, event, sample, machine);
>>   	case PERF_RECORD_COMM:
>>   		return tool->comm(tool, event, sample, machine);
>> @@ -1640,17 +1638,6 @@ static void perf_session__warn_about_errors(const struct perf_session *session)
>>   	perf_session__warn_order(session);
>>   
>>   	events_stats__auxtrace_error_warn(stats);
>> -
>> -	if (stats->nr_proc_map_timeout != 0) {
>> -		ui__warning("%d map information files for pre-existing threads were\n"
>> -			    "not processed, if there are samples for addresses they\n"
>> -			    "will not be resolved, you may find out which are these\n"
>> -			    "threads by running with -v and redirecting the output\n"
>> -			    "to a file.\n"
>> -			    "The time limit to process proc map is too short?\n"
>> -			    "Increase it by --proc-map-timeout\n",
>> -			    stats->nr_proc_map_timeout);
>> -	}
>>   }
>>   
>>   static int perf_session__flush_thread_stack(struct thread *thread,

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ