[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAJZ5v0iovirufvvDQVGbpgdfHftULGRNksvBCsn1kd8+9ntpZw@mail.gmail.com>
Date: Tue, 19 Apr 2022 18:27:39 +0200
From: "Rafael J. Wysocki" <rafael@...nel.org>
To: Thomas Gleixner <tglx@...utronix.de>
Cc: LKML <linux-kernel@...r.kernel.org>,
"the arch/x86 maintainers" <x86@...nel.org>,
"Rafael J. Wysocki" <rafael@...nel.org>,
Linux PM <linux-pm@...r.kernel.org>,
Eric Dumazet <edumazet@...gle.com>,
"Paul E. McKenney" <paulmck@...nel.org>
Subject: Re: [patch 07/10] x86/aperfmperf: Make parts of the frequency
invariance code unconditional
On Fri, Apr 15, 2022 at 9:20 PM Thomas Gleixner <tglx@...utronix.de> wrote:
>
> The frequency invariance support is currently limited to x86/64 and SMP,
> which is the vast majority of machines.
>
> arch_scale_freq_tick() is called every tick on all CPUs and reads the APERF
> and MPERF MSRs. The CPU frequency getters function do the same via dedicated
> IPIs.
>
> While it could be argued that on systems where frequency invariance support
> is disabled (32bit, !SMP) the per tick read of the APERF and MPERF MSRs can
> be avoided, it does not make sense to keep the extra code and the resulting
> runtime issues of mass IPIs around.
>
> As a first step split out the non frequency invariance specific
> initialization code and the read MSR portion of arch_scale_freq_tick(). The
> rest of the code is still conditional and guarded with a static key.
>
> Signed-off-by: Thomas Gleixner <tglx@...utronix.de>
All good AFAICS:
Reviewed-by: Rafael J. Wysocki <rafael.j.wysocki@...el.com>
> ---
> arch/x86/include/asm/cpu.h | 2 +
> arch/x86/include/asm/topology.h | 4 --
> arch/x86/kernel/cpu/aperfmperf.c | 63 +++++++++++++++++++++++----------------
> arch/x86/kernel/smpboot.c | 3 -
> 4 files changed, 41 insertions(+), 31 deletions(-)
>
> --- a/arch/x86/include/asm/cpu.h
> +++ b/arch/x86/include/asm/cpu.h
> @@ -36,6 +36,8 @@ extern int _debug_hotplug_cpu(int cpu, i
> #endif
> #endif
>
> +extern void ap_init_aperfmperf(void);
> +
> int mwait_usable(const struct cpuinfo_x86 *);
>
> unsigned int x86_family(unsigned int sig);
> --- a/arch/x86/include/asm/topology.h
> +++ b/arch/x86/include/asm/topology.h
> @@ -217,13 +217,9 @@ extern void arch_scale_freq_tick(void);
>
> extern void arch_set_max_freq_ratio(bool turbo_disabled);
> extern void freq_invariance_set_perf_ratio(u64 ratio, bool turbo_disabled);
> -extern void bp_init_freq_invariance(void);
> -extern void ap_init_freq_invariance(void);
> #else
> static inline void arch_set_max_freq_ratio(bool turbo_disabled) { }
> static inline void freq_invariance_set_perf_ratio(u64 ratio, bool turbo_disabled) { }
> -static inline void bp_init_freq_invariance(void) { }
> -static inline void ap_init_freq_invariance(void) { }
> #endif
>
> #ifdef CONFIG_ACPI_CPPC_LIB
> --- a/arch/x86/kernel/cpu/aperfmperf.c
> +++ b/arch/x86/kernel/cpu/aperfmperf.c
> @@ -17,6 +17,7 @@
> #include <linux/smp.h>
> #include <linux/syscore_ops.h>
>
> +#include <asm/cpu.h>
> #include <asm/cpu_device_id.h>
> #include <asm/intel-family.h>
>
> @@ -164,6 +165,17 @@ unsigned int arch_freq_get_on_cpu(int cp
> return per_cpu(samples.khz, cpu);
> }
>
> +static void init_counter_refs(void)
> +{
> + u64 aperf, mperf;
> +
> + rdmsrl(MSR_IA32_APERF, aperf);
> + rdmsrl(MSR_IA32_MPERF, mperf);
> +
> + this_cpu_write(cpu_samples.aperf, aperf);
> + this_cpu_write(cpu_samples.mperf, mperf);
> +}
> +
> #if defined(CONFIG_X86_64) && defined(CONFIG_SMP)
> /*
> * APERF/MPERF frequency ratio computation.
> @@ -405,17 +417,6 @@ static bool __init intel_set_max_freq_ra
> return true;
> }
>
> -static void init_counter_refs(void)
> -{
> - u64 aperf, mperf;
> -
> - rdmsrl(MSR_IA32_APERF, aperf);
> - rdmsrl(MSR_IA32_MPERF, mperf);
> -
> - this_cpu_write(cpu_samples.aperf, aperf);
> - this_cpu_write(cpu_samples.mperf, mperf);
> -}
> -
> #ifdef CONFIG_PM_SLEEP
> static struct syscore_ops freq_invariance_syscore_ops = {
> .resume = init_counter_refs,
> @@ -447,13 +448,8 @@ void freq_invariance_set_perf_ratio(u64
> freq_invariance_enable();
> }
>
> -void __init bp_init_freq_invariance(void)
> +static void __init bp_init_freq_invariance(void)
> {
> - if (!cpu_feature_enabled(X86_FEATURE_APERFMPERF))
> - return;
> -
> - init_counter_refs();
> -
> if (boot_cpu_data.x86_vendor != X86_VENDOR_INTEL)
> return;
>
> @@ -461,12 +457,6 @@ void __init bp_init_freq_invariance(void
> freq_invariance_enable();
> }
>
> -void ap_init_freq_invariance(void)
> -{
> - if (cpu_feature_enabled(X86_FEATURE_APERFMPERF))
> - init_counter_refs();
> -}
> -
> static void disable_freq_invariance_workfn(struct work_struct *work)
> {
> static_branch_disable(&arch_scale_freq_key);
> @@ -481,6 +471,9 @@ static void scale_freq_tick(u64 acnt, u6
> {
> u64 freq_scale;
>
> + if (!arch_scale_freq_invariant())
> + return;
> +
> if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt))
> goto error;
>
> @@ -501,13 +494,17 @@ static void scale_freq_tick(u64 acnt, u6
> pr_warn("Scheduler frequency invariance went wobbly, disabling!\n");
> schedule_work(&disable_freq_invariance_work);
> }
> +#else
> +static inline void bp_init_freq_invariance(void) { }
> +static inline void scale_freq_tick(u64 acnt, u64 mcnt) { }
> +#endif /* CONFIG_X86_64 && CONFIG_SMP */
>
> void arch_scale_freq_tick(void)
> {
> struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
> u64 acnt, mcnt, aperf, mperf;
>
> - if (!arch_scale_freq_invariant())
> + if (!cpu_feature_enabled(X86_FEATURE_APERFMPERF))
> return;
>
> rdmsrl(MSR_IA32_APERF, aperf);
> @@ -520,4 +517,20 @@ void arch_scale_freq_tick(void)
>
> scale_freq_tick(acnt, mcnt);
> }
> -#endif /* CONFIG_X86_64 && CONFIG_SMP */
> +
> +static int __init bp_init_aperfmperf(void)
> +{
> + if (!cpu_feature_enabled(X86_FEATURE_APERFMPERF))
> + return 0;
> +
> + init_counter_refs();
> + bp_init_freq_invariance();
> + return 0;
> +}
> +early_initcall(bp_init_aperfmperf);
> +
> +void ap_init_aperfmperf(void)
> +{
> + if (cpu_feature_enabled(X86_FEATURE_APERFMPERF))
> + init_counter_refs();
> +}
> --- a/arch/x86/kernel/smpboot.c
> +++ b/arch/x86/kernel/smpboot.c
> @@ -186,7 +186,7 @@ static void smp_callin(void)
> */
> set_cpu_sibling_map(raw_smp_processor_id());
>
> - ap_init_freq_invariance();
> + ap_init_aperfmperf();
>
> /*
> * Get our bogomips.
> @@ -1396,7 +1396,6 @@ void __init native_smp_prepare_cpus(unsi
> {
> smp_prepare_cpus_common();
>
> - bp_init_freq_invariance();
> smp_sanity_check();
>
> switch (apic_intr_mode) {
>
Powered by blists - more mailing lists