lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Thu, 20 Feb 2020 21:45:20 +0100 From: Thomas Gleixner <tglx@...utronix.de> To: LKML <linux-kernel@...r.kernel.org> Cc: David Miller <davem@...emloft.net>, bpf@...r.kernel.org, netdev@...r.kernel.org, Alexei Starovoitov <ast@...nel.org>, Daniel Borkmann <daniel@...earbox.net>, Sebastian Sewior <bigeasy@...utronix.de>, Peter Zijlstra <peterz@...radead.org>, Clark Williams <williams@...hat.com>, Steven Rostedt <rostedt@...dmis.org>, Juri Lelli <juri.lelli@...hat.com>, Ingo Molnar <mingo@...nel.org>, Mathieu Desnoyers <mathieu.desnoyers@...icios.com>, Vinicius Costa Gomes <vinicius.gomes@...el.com>, Jakub Kicinski <kuba@...nel.org> Subject: [patch V2 03/20] bpf/tracing: Remove redundant preempt_disable() in __bpf_trace_run() __bpf_trace_run() disables preemption around the BPF_PROG_RUN() invocation. This is redundant because __bpf_trace_run() is invoked from a trace point via __DO_TRACE() which already disables preemption _before_ invoking any of the functions which are attached to a trace point. Remove it. Signed-off-by: Thomas Gleixner <tglx@...utronix.de> --- kernel/trace/bpf_trace.c | 2 -- 1 file changed, 2 deletions(-) --- a/kernel/trace/bpf_trace.c +++ b/kernel/trace/bpf_trace.c @@ -1476,9 +1476,7 @@ static __always_inline void __bpf_trace_run(struct bpf_prog *prog, u64 *args) { rcu_read_lock(); - preempt_disable(); (void) BPF_PROG_RUN(prog, args); - preempt_enable(); rcu_read_unlock(); }
Powered by blists - more mailing lists