lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Thu, 10 Aug 2017 05:03:19 -0700 From: "tip-bot for leilei.lin" <tipbot@...or.com> To: linux-tip-commits@...r.kernel.org Cc: hpa@...or.com, mingo@...nel.org, leilei.lin@...baba-inc.com, tglx@...utronix.de, linux-kernel@...r.kernel.org, torvalds@...ux-foundation.org, peterz@...radead.org Subject: [tip:perf/core] perf/core: Reduce context switch overhead Commit-ID: fdccc3fb7a42ea4e4cd77d2fb8fa3a45c66ec0bf Gitweb: http://git.kernel.org/tip/fdccc3fb7a42ea4e4cd77d2fb8fa3a45c66ec0bf Author: leilei.lin <leilei.lin@...baba-inc.com> AuthorDate: Wed, 9 Aug 2017 08:29:21 +0800 Committer: Ingo Molnar <mingo@...nel.org> CommitDate: Thu, 10 Aug 2017 12:08:40 +0200 perf/core: Reduce context switch overhead Skip most of the PMU context switching overhead when ctx->nr_events is 0. 50% performance overhead was observed under an extreme testcase. Signed-off-by: leilei.lin <leilei.lin@...baba-inc.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org> Cc: Linus Torvalds <torvalds@...ux-foundation.org> Cc: Peter Zijlstra <peterz@...radead.org> Cc: Thomas Gleixner <tglx@...utronix.de> Cc: acme@...nel.org Cc: alexander.shishkin@...ux.intel.com Cc: eranian@...il.com Cc: jolsa@...hat.com Cc: linxiulei@...il.com Cc: yang_oliver@...mail.com Link: http://lkml.kernel.org/r/20170809002921.69813-1-leilei.lin@alibaba-inc.com [ Rewrote the changelog. ] Signed-off-by: Ingo Molnar <mingo@...nel.org> --- kernel/events/core.c | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/kernel/events/core.c b/kernel/events/core.c index ee20d4c..d704e23 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -3211,6 +3211,13 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx, return; perf_ctx_lock(cpuctx, ctx); + /* + * We must check ctx->nr_events while holding ctx->lock, such + * that we serialize against perf_install_in_context(). + */ + if (!ctx->nr_events) + goto unlock; + perf_pmu_disable(ctx->pmu); /* * We want to keep the following priority order: @@ -3224,6 +3231,8 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx, cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE); perf_event_sched_in(cpuctx, ctx, task); perf_pmu_enable(ctx->pmu); + +unlock: perf_ctx_unlock(cpuctx, ctx); }
Powered by blists - more mailing lists