lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Wed, 24 May 2017 15:21:27 +0900 From: Namhyung Kim <namhyung@...nel.org> To: Ingo Molnar <mingo@...nel.org> Cc: LKML <linux-kernel@...r.kernel.org>, kernel-team@....com, Arnaldo Carvalho de Melo <acme@...nel.org>, Jiri Olsa <jolsa@...nel.org>, Milian Wolff <milian.wolff@...b.com>, Yao Jin <yao.jin@...ux.intel.com>, Arnaldo Carvalho de Melo <acme@...hat.com>, David Ahern <dsahern@...il.com>, Peter Zijlstra <a.p.zijlstra@...llo.nl> Subject: [PATCH 5/7] perf report: always honor callchain order for inlined nodes From: Milian Wolff <milian.wolff@...b.com> So far, the inlined nodes where only reversed when we built perf against libbfd. If that was not available, the addr2line fallback code path was missing the inline_list__reverse call. Now we always add the nodes in the correct order within inline_list__append. This removes the need to reverse the list and also ensures that all callers construct the list in the right order. Cc: Arnaldo Carvalho de Melo <acme@...hat.com> Cc: David Ahern <dsahern@...il.com> Cc: Peter Zijlstra <a.p.zijlstra@...llo.nl> Cc: Yao Jin <yao.jin@...ux.intel.com> Signed-off-by: Milian Wolff <milian.wolff@...b.com> Signed-off-by: Namhyung Kim <namhyung@...nel.org> --- tools/perf/util/srcline.c | 18 ++++-------------- 1 file changed, 4 insertions(+), 14 deletions(-) diff --git a/tools/perf/util/srcline.c b/tools/perf/util/srcline.c index 5e376d64d59e..6af0364cad06 100644 --- a/tools/perf/util/srcline.c +++ b/tools/perf/util/srcline.c @@ -56,7 +56,10 @@ static int inline_list__append(char *filename, char *funcname, int line_nr, } } - list_add_tail(&ilist->list, &node->val); + if (callchain_param.order == ORDER_CALLEE) + list_add_tail(&ilist->list, &node->val); + else + list_add(&ilist->list, &node->val); return 0; } @@ -200,14 +203,6 @@ static void addr2line_cleanup(struct a2l_data *a2l) #define MAX_INLINE_NEST 1024 -static void inline_list__reverse(struct inline_node *node) -{ - struct inline_list *ilist, *n; - - list_for_each_entry_safe_reverse(ilist, n, &node->val, list) - list_move_tail(&ilist->list, &node->val); -} - static int addr2line(const char *dso_name, u64 addr, char **file, unsigned int *line, struct dso *dso, bool unwind_inlines, struct inline_node *node) @@ -250,11 +245,6 @@ static int addr2line(const char *dso_name, u64 addr, ret = 1; } } - - if ((node != NULL) && - (callchain_param.order != ORDER_CALLEE)) { - inline_list__reverse(node); - } } if (file) { -- 2.13.0
Powered by blists - more mailing lists