lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:   Thu,  5 Dec 2019 16:12:26 -0800
From:   Daniel Xu <dxu@...uu.xyz>
To:     ast@...nel.org, daniel@...earbox.net, yhs@...com, kafai@...com,
        songliubraving@...com, andriin@...com, netdev@...r.kernel.org,
        bpf@...r.kernel.org
Cc:     Daniel Xu <dxu@...uu.xyz>, peterz@...radead.org, mingo@...hat.com,
        acme@...nel.org, linux-kernel@...r.kernel.org, kernel-team@...com
Subject: [PATCH bpf] bpf: Add LBR data to BPF_PROG_TYPE_PERF_EVENT prog context

Last-branch-record is an intel CPU feature that can be configured to
record certain branches that are taken during code execution. This data
is particularly interesting for profile guided optimizations. perf has
had LBR support for a while but the data collection can be a bit coarse
grained.

We (Facebook) have recently run a lot of experiments with feeding
filtered LBR data to various PGO pipelines. We've seen really good
results (+2.5% throughput with lower cpu util and lower latency) by
feeding high request latency LBR branches to the compiler on a
request-oriented service. We used bpf to read a special request context
ID (which is how we associate branches with latency) from a fixed
userspace address. Reading from the fixed address is why bpf support is
useful.

Aside from this particular use case, having LBR data available to bpf
progs can be useful to get stack traces out of userspace applications
that omit frame pointers.

This patch adds support for LBR data to bpf perf progs.

Some notes:
* We use `__u64 entries[BPF_MAX_LBR_ENTRIES * 3]` instead of
  `struct perf_branch_entry[BPF_MAX_LBR_ENTRIES]` because checkpatch.pl
  warns about including a uapi header from another uapi header

* We define BPF_MAX_LBR_ENTRIES as 32 (instead of using the value from
  arch/x86/events/perf_events.h) because including arch specific headers
  seems wrong and could introduce circular header includes.

Signed-off-by: Daniel Xu <dxu@...uu.xyz>
---
 include/uapi/linux/bpf_perf_event.h |  5 ++++
 kernel/trace/bpf_trace.c            | 39 +++++++++++++++++++++++++++++
 2 files changed, 44 insertions(+)

diff --git a/include/uapi/linux/bpf_perf_event.h b/include/uapi/linux/bpf_perf_event.h
index eb1b9d21250c..dc87e3d50390 100644
--- a/include/uapi/linux/bpf_perf_event.h
+++ b/include/uapi/linux/bpf_perf_event.h
@@ -10,10 +10,15 @@
 
 #include <asm/bpf_perf_event.h>
 
+#define BPF_MAX_LBR_ENTRIES 32
+
 struct bpf_perf_event_data {
 	bpf_user_pt_regs_t regs;
 	__u64 sample_period;
 	__u64 addr;
+	__u64 nr_lbr;
+	/* Cast to struct perf_branch_entry* before using */
+	__u64 entries[BPF_MAX_LBR_ENTRIES * 3];
 };
 
 #endif /* _UAPI__LINUX_BPF_PERF_EVENT_H__ */
diff --git a/kernel/trace/bpf_trace.c b/kernel/trace/bpf_trace.c
index ffc91d4935ac..96ba7995b3d7 100644
--- a/kernel/trace/bpf_trace.c
+++ b/kernel/trace/bpf_trace.c
@@ -1259,6 +1259,14 @@ static bool pe_prog_is_valid_access(int off, int size, enum bpf_access_type type
 		if (!bpf_ctx_narrow_access_ok(off, size, size_u64))
 			return false;
 		break;
+	case bpf_ctx_range(struct bpf_perf_event_data, nr_lbr):
+		bpf_ctx_record_field_size(info, size_u64);
+		if (!bpf_ctx_narrow_access_ok(off, size, size_u64))
+			return false;
+		break;
+	case bpf_ctx_range(struct bpf_perf_event_data, entries):
+		/* No narrow loads */
+		break;
 	default:
 		if (size != sizeof(long))
 			return false;
@@ -1273,6 +1281,7 @@ static u32 pe_prog_convert_ctx_access(enum bpf_access_type type,
 				      struct bpf_prog *prog, u32 *target_size)
 {
 	struct bpf_insn *insn = insn_buf;
+	int off;
 
 	switch (si->off) {
 	case offsetof(struct bpf_perf_event_data, sample_period):
@@ -1291,6 +1300,36 @@ static u32 pe_prog_convert_ctx_access(enum bpf_access_type type,
 				      bpf_target_off(struct perf_sample_data, addr, 8,
 						     target_size));
 		break;
+	case offsetof(struct bpf_perf_event_data, nr_lbr):
+		/* Load struct perf_sample_data* */
+		*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct bpf_perf_event_data_kern,
+						       data), si->dst_reg, si->src_reg,
+				      offsetof(struct bpf_perf_event_data_kern, data));
+		/* Load struct perf_branch_stack* */
+		*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct perf_sample_data, br_stack),
+				      si->dst_reg, si->dst_reg,
+				      offsetof(struct perf_sample_data, br_stack));
+		/* Load nr */
+		*insn++ = BPF_LDX_MEM(BPF_DW, si->dst_reg, si->dst_reg,
+				      bpf_target_off(struct perf_branch_stack, nr, 8,
+						     target_size));
+		break;
+	case bpf_ctx_range(struct bpf_perf_event_data, entries):
+		off = si->off;
+		off -= offsetof(struct bpf_perf_event_data, entries);
+
+		/* Load struct perf_sample_data* */
+		*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct bpf_perf_event_data_kern,
+						       data), si->dst_reg, si->src_reg,
+				      offsetof(struct bpf_perf_event_data_kern, data));
+		/* Load struct perf_branch_stack* */
+		*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct perf_sample_data, br_stack),
+				      si->dst_reg, si->dst_reg,
+				      offsetof(struct perf_sample_data, br_stack));
+		/* Load requested memory */
+		*insn++ = BPF_LDX_MEM(BPF_SIZE(si->code), si->dst_reg, si->dst_reg,
+				      offsetof(struct perf_branch_stack, entries) + off);
+		break;
 	default:
 		*insn++ = BPF_LDX_MEM(BPF_FIELD_SIZEOF(struct bpf_perf_event_data_kern,
 						       regs), si->dst_reg, si->src_reg,
-- 
2.24.0

Powered by blists - more mailing lists