[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87frtizigf.fsf@BLR-5CG11610CF.amd.com>
Date: Wed, 12 Jun 2024 13:44:08 +0530
From: Gautham R.Shenoy <gautham.shenoy@....com>
To: Perry Yuan <perry.yuan@....com>, <Mario.Limonciello@....com>
CC: <rafael.j.wysocki@...el.com>, <viresh.kumar@...aro.org>,
<Ray.Huang@....com>, <Borislav.Petkov@....com>, <Alexander.Deucher@....com>,
<Xinmei.Huang@....com>, <Xiaojian.Du@....com>, <Li.Meng@....com>,
<linux-pm@...r.kernel.org>, <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v3 07/10] cpufreq: amd-pstate: switch boot_cpu_has() to
cpu_feature_enabled()
Perry Yuan <perry.yuan@....com> writes:
> replace the usage of the deprecated boot_cpu_has() function with
> the modern cpu_feature_enabled() function. The switch to cpu_feature_enabled()
> ensures compatibility with the latest CPU feature detection mechanisms and
> improves code maintainability.
>
> Acked-by: Mario Limonciello <mario.limonciello@....com>
> Suggested-by: Borislav Petkov (AMD) <bp@...en8.de>
> Signed-off-by: Perry Yuan <perry.yuan@....com>
Reviewed-by: Gautham R. Shenoy <gautham.shenoy@....com>
> ---
> drivers/cpufreq/amd-pstate.c | 24 ++++++++++++------------
> 1 file changed, 12 insertions(+), 12 deletions(-)
>
> diff --git a/drivers/cpufreq/amd-pstate.c b/drivers/cpufreq/amd-pstate.c
> index 6b9fc24001f2..cb59de71b6ee 100644
> --- a/drivers/cpufreq/amd-pstate.c
> +++ b/drivers/cpufreq/amd-pstate.c
> @@ -157,7 +157,7 @@ static int __init dmi_matched_7k62_bios_bug(const struct dmi_system_id *dmi)
> * broken BIOS lack of nominal_freq and lowest_freq capabilities
> * definition in ACPI tables
> */
> - if (boot_cpu_has(X86_FEATURE_ZEN2)) {
> + if (cpu_feature_enabled(X86_FEATURE_ZEN2)) {
> quirks = dmi->driver_data;
> pr_info("Overriding nominal and lowest frequencies for %s\n", dmi->ident);
> return 1;
> @@ -199,7 +199,7 @@ static s16 amd_pstate_get_epp(struct amd_cpudata *cpudata, u64 cppc_req_cached)
> u64 epp;
> int ret;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> if (!cppc_req_cached) {
> epp = rdmsrl_on_cpu(cpudata->cpu, MSR_AMD_CPPC_REQ,
> &cppc_req_cached);
> @@ -252,7 +252,7 @@ static int amd_pstate_set_epp(struct amd_cpudata *cpudata, u32 epp)
> int ret;
> struct cppc_perf_ctrls perf_ctrls;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> u64 value = READ_ONCE(cpudata->cppc_req_cached);
>
> value &= ~GENMASK_ULL(31, 24);
> @@ -753,7 +753,7 @@ static int amd_pstate_get_highest_perf(int cpu, u32 *highest_perf)
> {
> int ret;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> u64 cap1;
>
> ret = rdmsrl_safe_on_cpu(cpu, MSR_AMD_CPPC_CAP1, &cap1);
> @@ -988,7 +988,7 @@ static int amd_pstate_cpu_init(struct cpufreq_policy *policy)
> /* It will be updated by governor */
> policy->cur = policy->cpuinfo.min_freq;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC))
> + if (cpu_feature_enabled(X86_FEATURE_CPPC))
> policy->fast_switch_possible = true;
>
> ret = freq_qos_add_request(&policy->constraints, &cpudata->req[0],
> @@ -1221,7 +1221,7 @@ static int amd_pstate_change_mode_without_dvr_change(int mode)
>
> cppc_state = mode;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC) || cppc_state == AMD_PSTATE_ACTIVE)
> + if (cpu_feature_enabled(X86_FEATURE_CPPC) || cppc_state == AMD_PSTATE_ACTIVE)
> return 0;
>
> for_each_present_cpu(cpu) {
> @@ -1450,7 +1450,7 @@ static int amd_pstate_epp_cpu_init(struct cpufreq_policy *policy)
> else
> policy->policy = CPUFREQ_POLICY_POWERSAVE;
>
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> ret = rdmsrl_on_cpu(cpudata->cpu, MSR_AMD_CPPC_REQ, &value);
> if (ret)
> return ret;
> @@ -1540,7 +1540,7 @@ static void amd_pstate_epp_update_limit(struct cpufreq_policy *policy)
> epp = 0;
>
> /* Set initial EPP value */
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> value &= ~GENMASK_ULL(31, 24);
> value |= (u64)epp << 24;
> }
> @@ -1579,7 +1579,7 @@ static void amd_pstate_epp_reenable(struct amd_cpudata *cpudata)
> value = READ_ONCE(cpudata->cppc_req_cached);
> max_perf = READ_ONCE(cpudata->highest_perf);
>
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> wrmsrl_on_cpu(cpudata->cpu, MSR_AMD_CPPC_REQ, value);
> } else {
> perf_ctrls.max_perf = max_perf;
> @@ -1613,7 +1613,7 @@ static void amd_pstate_epp_offline(struct cpufreq_policy *policy)
> value = READ_ONCE(cpudata->cppc_req_cached);
>
> mutex_lock(&amd_pstate_limits_lock);
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> cpudata->epp_policy = CPUFREQ_POLICY_UNKNOWN;
>
> /* Set max perf same as min perf */
> @@ -1815,7 +1815,7 @@ static int __init amd_pstate_init(void)
> */
> if (amd_pstate_acpi_pm_profile_undefined() ||
> amd_pstate_acpi_pm_profile_server() ||
> - !boot_cpu_has(X86_FEATURE_CPPC)) {
> + !cpu_feature_enabled(X86_FEATURE_CPPC)) {
> pr_info("driver load is disabled, boot with specific mode to enable this\n");
> return -ENODEV;
> }
> @@ -1834,7 +1834,7 @@ static int __init amd_pstate_init(void)
> }
>
> /* capability check */
> - if (boot_cpu_has(X86_FEATURE_CPPC)) {
> + if (cpu_feature_enabled(X86_FEATURE_CPPC)) {
> pr_debug("AMD CPPC MSR based functionality is supported\n");
> if (cppc_state != AMD_PSTATE_ACTIVE)
> current_pstate_driver->adjust_perf = amd_pstate_adjust_perf;
> --
> 2.34.1
Powered by blists - more mailing lists