lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <a4cff15e-2da9-4d00-85e8-d18e7e0f67fb@amd.com>
Date: Wed, 5 Feb 2025 11:46:17 -0600
From: Mario Limonciello <mario.limonciello@....com>
To: Dhananjay Ugwekar <dhananjay.ugwekar@....com>, gautham.shenoy@....com,
 rafael@...nel.org, viresh.kumar@...aro.org
Cc: linux-kernel@...r.kernel.org, linux-pm@...r.kernel.org
Subject: Re: [PATCH 06/12] cpufreq/amd-pstate: Convert all perf values to u8

On 2/5/2025 05:25, Dhananjay Ugwekar wrote:
> All perf values are always within 0-255 range, hence convert their
> datatype to u8 everywhere.
> 
> Signed-off-by: Dhananjay Ugwekar <dhananjay.ugwekar@....com>

Reviewed-by: Mario Limonciello <mario.limonciello@....com>

I'll queue this for 6.15.

> ---
>   drivers/cpufreq/amd-pstate-trace.h | 46 +++++++++++------------
>   drivers/cpufreq/amd-pstate.c       | 60 +++++++++++++++---------------
>   drivers/cpufreq/amd-pstate.h       | 18 ++++-----
>   3 files changed, 62 insertions(+), 62 deletions(-)
> 
> diff --git a/drivers/cpufreq/amd-pstate-trace.h b/drivers/cpufreq/amd-pstate-trace.h
> index 8d692415d905..f457d4af2c62 100644
> --- a/drivers/cpufreq/amd-pstate-trace.h
> +++ b/drivers/cpufreq/amd-pstate-trace.h
> @@ -24,9 +24,9 @@
>   
>   TRACE_EVENT(amd_pstate_perf,
>   
> -	TP_PROTO(unsigned long min_perf,
> -		 unsigned long target_perf,
> -		 unsigned long capacity,
> +	TP_PROTO(u8 min_perf,
> +		 u8 target_perf,
> +		 u8 capacity,
>   		 u64 freq,
>   		 u64 mperf,
>   		 u64 aperf,
> @@ -47,9 +47,9 @@ TRACE_EVENT(amd_pstate_perf,
>   		),
>   
>   	TP_STRUCT__entry(
> -		__field(unsigned long, min_perf)
> -		__field(unsigned long, target_perf)
> -		__field(unsigned long, capacity)
> +		__field(u8, min_perf)
> +		__field(u8, target_perf)
> +		__field(u8, capacity)
>   		__field(unsigned long long, freq)
>   		__field(unsigned long long, mperf)
>   		__field(unsigned long long, aperf)
> @@ -70,10 +70,10 @@ TRACE_EVENT(amd_pstate_perf,
>   		__entry->fast_switch = fast_switch;
>   		),
>   
> -	TP_printk("amd_min_perf=%lu amd_des_perf=%lu amd_max_perf=%lu freq=%llu mperf=%llu aperf=%llu tsc=%llu cpu_id=%u fast_switch=%s",
> -		  (unsigned long)__entry->min_perf,
> -		  (unsigned long)__entry->target_perf,
> -		  (unsigned long)__entry->capacity,
> +	TP_printk("amd_min_perf=%hhu amd_des_perf=%hhu amd_max_perf=%hhu freq=%llu mperf=%llu aperf=%llu tsc=%llu cpu_id=%u fast_switch=%s",
> +		  (u8)__entry->min_perf,
> +		  (u8)__entry->target_perf,
> +		  (u8)__entry->capacity,
>   		  (unsigned long long)__entry->freq,
>   		  (unsigned long long)__entry->mperf,
>   		  (unsigned long long)__entry->aperf,
> @@ -86,10 +86,10 @@ TRACE_EVENT(amd_pstate_perf,
>   TRACE_EVENT(amd_pstate_epp_perf,
>   
>   	TP_PROTO(unsigned int cpu_id,
> -		 unsigned int highest_perf,
> -		 unsigned int epp,
> -		 unsigned int min_perf,
> -		 unsigned int max_perf,
> +		 u8 highest_perf,
> +		 u8 epp,
> +		 u8 min_perf,
> +		 u8 max_perf,
>   		 bool boost
>   		 ),
>   
> @@ -102,10 +102,10 @@ TRACE_EVENT(amd_pstate_epp_perf,
>   
>   	TP_STRUCT__entry(
>   		__field(unsigned int, cpu_id)
> -		__field(unsigned int, highest_perf)
> -		__field(unsigned int, epp)
> -		__field(unsigned int, min_perf)
> -		__field(unsigned int, max_perf)
> +		__field(u8, highest_perf)
> +		__field(u8, epp)
> +		__field(u8, min_perf)
> +		__field(u8, max_perf)
>   		__field(bool, boost)
>   		),
>   
> @@ -118,12 +118,12 @@ TRACE_EVENT(amd_pstate_epp_perf,
>   		__entry->boost = boost;
>   		),
>   
> -	TP_printk("cpu%u: [%u<->%u]/%u, epp=%u, boost=%u",
> +	TP_printk("cpu%u: [%hhu<->%hhu]/%hhu, epp=%hhu, boost=%u",
>   		  (unsigned int)__entry->cpu_id,
> -		  (unsigned int)__entry->min_perf,
> -		  (unsigned int)__entry->max_perf,
> -		  (unsigned int)__entry->highest_perf,
> -		  (unsigned int)__entry->epp,
> +		  (u8)__entry->min_perf,
> +		  (u8)__entry->max_perf,
> +		  (u8)__entry->highest_perf,
> +		  (u8)__entry->epp,
>   		  (bool)__entry->boost
>   		 )
>   );
> diff --git a/drivers/cpufreq/amd-pstate.c b/drivers/cpufreq/amd-pstate.c
> index e179e929b941..dd4f23fa2587 100644
> --- a/drivers/cpufreq/amd-pstate.c
> +++ b/drivers/cpufreq/amd-pstate.c
> @@ -186,7 +186,7 @@ static inline int get_mode_idx_from_str(const char *str, size_t size)
>   static DEFINE_MUTEX(amd_pstate_limits_lock);
>   static DEFINE_MUTEX(amd_pstate_driver_lock);
>   
> -static s16 msr_get_epp(struct amd_cpudata *cpudata)
> +static u8 msr_get_epp(struct amd_cpudata *cpudata)
>   {
>   	u64 value;
>   	int ret;
> @@ -207,7 +207,7 @@ static inline s16 amd_pstate_get_epp(struct amd_cpudata *cpudata)
>   	return static_call(amd_pstate_get_epp)(cpudata);
>   }
>   
> -static s16 shmem_get_epp(struct amd_cpudata *cpudata)
> +static u8 shmem_get_epp(struct amd_cpudata *cpudata)
>   {
>   	u64 epp;
>   	int ret;
> @@ -218,11 +218,11 @@ static s16 shmem_get_epp(struct amd_cpudata *cpudata)
>   		return ret;
>   	}
>   
> -	return (s16)(epp & 0xff);
> +	return FIELD_GET(AMD_CPPC_EPP_PERF_MASK, epp);
>   }
>   
> -static int msr_update_perf(struct amd_cpudata *cpudata, u32 min_perf,
> -			   u32 des_perf, u32 max_perf, u32 epp, bool fast_switch)
> +static int msr_update_perf(struct amd_cpudata *cpudata, u8 min_perf,
> +			   u8 des_perf, u8 max_perf, u8 epp, bool fast_switch)
>   {
>   	u64 value, prev;
>   
> @@ -257,15 +257,15 @@ static int msr_update_perf(struct amd_cpudata *cpudata, u32 min_perf,
>   DEFINE_STATIC_CALL(amd_pstate_update_perf, msr_update_perf);
>   
>   static inline int amd_pstate_update_perf(struct amd_cpudata *cpudata,
> -					  u32 min_perf, u32 des_perf,
> -					  u32 max_perf, u32 epp,
> +					  u8 min_perf, u8 des_perf,
> +					  u8 max_perf, u8 epp,
>   					  bool fast_switch)
>   {
>   	return static_call(amd_pstate_update_perf)(cpudata, min_perf, des_perf,
>   						   max_perf, epp, fast_switch);
>   }
>   
> -static int msr_set_epp(struct amd_cpudata *cpudata, u32 epp)
> +static int msr_set_epp(struct amd_cpudata *cpudata, u8 epp)
>   {
>   	u64 value, prev;
>   	int ret;
> @@ -292,12 +292,12 @@ static int msr_set_epp(struct amd_cpudata *cpudata, u32 epp)
>   
>   DEFINE_STATIC_CALL(amd_pstate_set_epp, msr_set_epp);
>   
> -static inline int amd_pstate_set_epp(struct amd_cpudata *cpudata, u32 epp)
> +static inline int amd_pstate_set_epp(struct amd_cpudata *cpudata, u8 epp)
>   {
>   	return static_call(amd_pstate_set_epp)(cpudata, epp);
>   }
>   
> -static int shmem_set_epp(struct amd_cpudata *cpudata, u32 epp)
> +static int shmem_set_epp(struct amd_cpudata *cpudata, u8 epp)
>   {
>   	int ret;
>   	struct cppc_perf_ctrls perf_ctrls;
> @@ -320,7 +320,7 @@ static int amd_pstate_set_energy_pref_index(struct cpufreq_policy *policy,
>   					    int pref_index)
>   {
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	int epp;
> +	u8 epp;
>   
>   	if (!pref_index)
>   		epp = cpudata->epp_default;
> @@ -479,8 +479,8 @@ static inline int amd_pstate_init_perf(struct amd_cpudata *cpudata)
>   	return static_call(amd_pstate_init_perf)(cpudata);
>   }
>   
> -static int shmem_update_perf(struct amd_cpudata *cpudata, u32 min_perf,
> -			     u32 des_perf, u32 max_perf, u32 epp, bool fast_switch)
> +static int shmem_update_perf(struct amd_cpudata *cpudata, u8 min_perf,
> +			     u8 des_perf, u8 max_perf, u8 epp, bool fast_switch)
>   {
>   	struct cppc_perf_ctrls perf_ctrls;
>   
> @@ -531,14 +531,14 @@ static inline bool amd_pstate_sample(struct amd_cpudata *cpudata)
>   	return true;
>   }
>   
> -static void amd_pstate_update(struct amd_cpudata *cpudata, u32 min_perf,
> -			      u32 des_perf, u32 max_perf, bool fast_switch, int gov_flags)
> +static void amd_pstate_update(struct amd_cpudata *cpudata, u8 min_perf,
> +			      u8 des_perf, u8 max_perf, bool fast_switch, int gov_flags)
>   {
>   	unsigned long max_freq;
>   	struct cpufreq_policy *policy = cpufreq_cpu_get(cpudata->cpu);
> -	u32 nominal_perf = READ_ONCE(cpudata->nominal_perf);
> +	u8 nominal_perf = READ_ONCE(cpudata->nominal_perf);
>   
> -	des_perf = clamp_t(unsigned long, des_perf, min_perf, max_perf);
> +	des_perf = clamp_t(u8, des_perf, min_perf, max_perf);
>   
>   	max_freq = READ_ONCE(cpudata->max_limit_freq);
>   	policy->cur = div_u64(des_perf * max_freq, max_perf);
> @@ -550,7 +550,7 @@ static void amd_pstate_update(struct amd_cpudata *cpudata, u32 min_perf,
>   
>   	/* limit the max perf when core performance boost feature is disabled */
>   	if (!cpudata->boost_supported)
> -		max_perf = min_t(unsigned long, nominal_perf, max_perf);
> +		max_perf = min_t(u8, nominal_perf, max_perf);
>   
>   	if (trace_amd_pstate_perf_enabled() && amd_pstate_sample(cpudata)) {
>   		trace_amd_pstate_perf(min_perf, des_perf, max_perf, cpudata->freq,
> @@ -591,7 +591,8 @@ static int amd_pstate_verify(struct cpufreq_policy_data *policy_data)
>   
>   static int amd_pstate_update_min_max_limit(struct cpufreq_policy *policy)
>   {
> -	u32 max_limit_perf, min_limit_perf, max_perf, max_freq;
> +	u8 max_limit_perf, min_limit_perf, max_perf;
> +	u32 max_freq;
>   	struct amd_cpudata *cpudata = policy->driver_data;
>   
>   	max_perf = READ_ONCE(cpudata->highest_perf);
> @@ -615,7 +616,7 @@ static int amd_pstate_update_freq(struct cpufreq_policy *policy,
>   {
>   	struct cpufreq_freqs freqs;
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	unsigned long des_perf, cap_perf;
> +	u8 des_perf, cap_perf;
>   
>   	if (!cpudata->max_freq)
>   		return -ENODEV;
> @@ -670,8 +671,7 @@ static void amd_pstate_adjust_perf(unsigned int cpu,
>   				   unsigned long target_perf,
>   				   unsigned long capacity)
>   {
> -	unsigned long max_perf, min_perf, des_perf,
> -		      cap_perf, min_limit_perf;
> +	u8 max_perf, min_perf, des_perf, cap_perf, min_limit_perf;
>   	struct cpufreq_policy *policy = cpufreq_cpu_get(cpu);
>   	struct amd_cpudata *cpudata;
>   
> @@ -904,8 +904,8 @@ static int amd_pstate_init_freq(struct amd_cpudata *cpudata)
>   {
>   	int ret;
>   	u32 min_freq, max_freq;
> -	u32 highest_perf, nominal_perf, nominal_freq;
> -	u32 lowest_nonlinear_perf, lowest_nonlinear_freq;
> +	u8 highest_perf, nominal_perf, lowest_nonlinear_perf;
> +	u32 nominal_freq, lowest_nonlinear_freq;
>   	struct cppc_perf_caps cppc_perf;
>   
>   	ret = cppc_get_perf_caps(cpudata->cpu, &cppc_perf);
> @@ -1112,7 +1112,7 @@ static ssize_t show_amd_pstate_lowest_nonlinear_freq(struct cpufreq_policy *poli
>   static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy,
>   					    char *buf)
>   {
> -	u32 perf;
> +	u8 perf;
>   	struct amd_cpudata *cpudata = policy->driver_data;
>   
>   	perf = READ_ONCE(cpudata->highest_perf);
> @@ -1123,7 +1123,7 @@ static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy,
>   static ssize_t show_amd_pstate_prefcore_ranking(struct cpufreq_policy *policy,
>   						char *buf)
>   {
> -	u32 perf;
> +	u8 perf;
>   	struct amd_cpudata *cpudata = policy->driver_data;
>   
>   	perf = READ_ONCE(cpudata->prefcore_ranking);
> @@ -1186,7 +1186,7 @@ static ssize_t show_energy_performance_preference(
>   				struct cpufreq_policy *policy, char *buf)
>   {
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	int preference;
> +	u8 preference;
>   
>   	switch (cpudata->epp_cached) {
>   	case AMD_CPPC_EPP_PERFORMANCE:
> @@ -1548,7 +1548,7 @@ static void amd_pstate_epp_cpu_exit(struct cpufreq_policy *policy)
>   static int amd_pstate_epp_update_limit(struct cpufreq_policy *policy)
>   {
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	u32 epp;
> +	u8 epp;
>   
>   	amd_pstate_update_min_max_limit(policy);
>   
> @@ -1597,7 +1597,7 @@ static int amd_pstate_epp_set_policy(struct cpufreq_policy *policy)
>   static int amd_pstate_epp_reenable(struct cpufreq_policy *policy)
>   {
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	u64 max_perf;
> +	u8 max_perf;
>   	int ret;
>   
>   	ret = amd_pstate_cppc_enable(true);
> @@ -1634,7 +1634,7 @@ static int amd_pstate_epp_cpu_online(struct cpufreq_policy *policy)
>   static int amd_pstate_epp_cpu_offline(struct cpufreq_policy *policy)
>   {
>   	struct amd_cpudata *cpudata = policy->driver_data;
> -	int min_perf;
> +	u8 min_perf;
>   
>   	if (cpudata->suspended)
>   		return 0;
> diff --git a/drivers/cpufreq/amd-pstate.h b/drivers/cpufreq/amd-pstate.h
> index 9747e3be6cee..19d405c6d805 100644
> --- a/drivers/cpufreq/amd-pstate.h
> +++ b/drivers/cpufreq/amd-pstate.h
> @@ -70,13 +70,13 @@ struct amd_cpudata {
>   	struct	freq_qos_request req[2];
>   	u64	cppc_req_cached;
>   
> -	u32	highest_perf;
> -	u32	nominal_perf;
> -	u32	lowest_nonlinear_perf;
> -	u32	lowest_perf;
> -	u32     prefcore_ranking;
> -	u32     min_limit_perf;
> -	u32     max_limit_perf;
> +	u8	highest_perf;
> +	u8	nominal_perf;
> +	u8	lowest_nonlinear_perf;
> +	u8	lowest_perf;
> +	u8	prefcore_ranking;
> +	u8	min_limit_perf;
> +	u8	max_limit_perf;
>   	u32     min_limit_freq;
>   	u32     max_limit_freq;
>   
> @@ -93,11 +93,11 @@ struct amd_cpudata {
>   	bool	hw_prefcore;
>   
>   	/* EPP feature related attributes*/
> -	s16	epp_cached;
> +	u8	epp_cached;
>   	u32	policy;
>   	u64	cppc_cap1_cached;
>   	bool	suspended;
> -	s16	epp_default;
> +	u8	epp_default;
>   };
>   
>   /*


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ