[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <tip-fdccc3fb7a42ea4e4cd77d2fb8fa3a45c66ec0bf@git.kernel.org>
Date:   Thu, 10 Aug 2017 05:03:19 -0700
From:   "tip-bot for leilei.lin" <tipbot@...or.com>
To:     linux-tip-commits@...r.kernel.org
Cc:     hpa@...or.com, mingo@...nel.org, leilei.lin@...baba-inc.com,
        tglx@...utronix.de, linux-kernel@...r.kernel.org,
        torvalds@...ux-foundation.org, peterz@...radead.org
Subject: [tip:perf/core] perf/core: Reduce context switch overhead
Commit-ID:  fdccc3fb7a42ea4e4cd77d2fb8fa3a45c66ec0bf
Gitweb:     http://git.kernel.org/tip/fdccc3fb7a42ea4e4cd77d2fb8fa3a45c66ec0bf
Author:     leilei.lin <leilei.lin@...baba-inc.com>
AuthorDate: Wed, 9 Aug 2017 08:29:21 +0800
Committer:  Ingo Molnar <mingo@...nel.org>
CommitDate: Thu, 10 Aug 2017 12:08:40 +0200
perf/core: Reduce context switch overhead
Skip most of the PMU context switching overhead when ctx->nr_events is 0.
50% performance overhead was observed under an extreme testcase.
Signed-off-by: leilei.lin <leilei.lin@...baba-inc.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Peter Zijlstra <peterz@...radead.org>
Cc: Thomas Gleixner <tglx@...utronix.de>
Cc: acme@...nel.org
Cc: alexander.shishkin@...ux.intel.com
Cc: eranian@...il.com
Cc: jolsa@...hat.com
Cc: linxiulei@...il.com
Cc: yang_oliver@...mail.com
Link: http://lkml.kernel.org/r/20170809002921.69813-1-leilei.lin@alibaba-inc.com
[ Rewrote the changelog. ]
Signed-off-by: Ingo Molnar <mingo@...nel.org>
---
 kernel/events/core.c | 9 +++++++++
 1 file changed, 9 insertions(+)
diff --git a/kernel/events/core.c b/kernel/events/core.c
index ee20d4c..d704e23 100644
--- a/kernel/events/core.c
+++ b/kernel/events/core.c
@@ -3211,6 +3211,13 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx,
 		return;
 
 	perf_ctx_lock(cpuctx, ctx);
+	/*
+	 * We must check ctx->nr_events while holding ctx->lock, such
+	 * that we serialize against perf_install_in_context().
+	 */
+	if (!ctx->nr_events)
+		goto unlock;
+
 	perf_pmu_disable(ctx->pmu);
 	/*
 	 * We want to keep the following priority order:
@@ -3224,6 +3231,8 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx,
 		cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE);
 	perf_event_sched_in(cpuctx, ctx, task);
 	perf_pmu_enable(ctx->pmu);
+
+unlock:
 	perf_ctx_unlock(cpuctx, ctx);
 }
 
Powered by blists - more mailing lists
 
