lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 02 Mar 2016 15:59:17 -0800
From:	"Christopher Hall" <christopher.s.hall@...el.com>
To:	tglx@...utronix.de, luto@...capital.net, john.stultz@...aro.org,
	"Christopher S. Hall" <christopher.s.hall@...el.com>
Cc:	hpa@...or.com, x86@...nel.org, linux-kernel@...r.kernel.org,
	peterz@...radead.org
Subject: Re: [PATCH v7 6/8] x86: tsc: Always Running Timer (ART) correlated
 clocksource

Hi John,

There was no response from Andy on this. Should I assume that this will  
have to wait until 4.7? Is the plan to hold off on this, the ptp and  
e1000e patches? Thanks.

Chris

On Tue, 01 Mar 2016 17:11:13 -0800, Christopher Hall  
<christopher.s.hall@...el.com> wrote:
> Andy,
>
> On Mon, 29 Feb 2016 06:33:47 -0800, Christopher S. Hall  
> <christopher.s.hall@...el.com> wrote:
>
> Do you have any comment on this? John needs your ACK. Thanks.
>
> Chris
>
>> *** Commit message below:
>>
>> On modern Intel systems TSC is derived from the new Always Running Timer
>> (ART). ART can be captured simultaneous to the capture of
>> audio and network device clocks, allowing a correlation between  
>> timebases
>> to be constructed. Upon capture, the driver converts the captured ART
>> value to the appropriate system clock using the correlated clocksource
>> mechanism.
>>
>> On systems that support ART a new CPUID leaf (0x15) returns parameters
>> “m” and “n” such that:
>>
>> TSC_value = (ART_value * m) / n + k [n >= 1]
>>
>> [k is an offset that can adjusted by a privileged agent. The
>> IA32_TSC_ADJUST MSR is an example of an interface to adjust k.
>> See 17.14.4 of the Intel SDM for more details]
>>
>> Reviewed-by: Thomas Gleixner <tglx@...utronix.de>
>> Signed-off-by: Christopher S. Hall <christopher.s.hall@...el.com>
>> [jstultz: Tweaked to fix build issue, also reworked math for
>> 64bit division on 32bit systems]
>> Signed-off-by: John Stultz <john.stultz@...aro.org>
>> ---
>>  arch/x86/include/asm/cpufeature.h |  2 +-
>>  arch/x86/include/asm/tsc.h        |  2 ++
>>  arch/x86/kernel/tsc.c             | 58  
>> +++++++++++++++++++++++++++++++++++++++
>>  3 files changed, 61 insertions(+), 1 deletion(-)
>>
>> diff --git a/arch/x86/include/asm/cpufeature.h  
>> b/arch/x86/include/asm/cpufeature.h
>> index 7ad8c94..ff557b4 100644
>> --- a/arch/x86/include/asm/cpufeature.h
>> +++ b/arch/x86/include/asm/cpufeature.h
>> @@ -85,7 +85,7 @@
>>  #define X86_FEATURE_P4		( 3*32+ 7) /* "" P4 */
>>  #define X86_FEATURE_CONSTANT_TSC ( 3*32+ 8) /* TSC ticks at a constant  
>> rate */
>>  #define X86_FEATURE_UP		( 3*32+ 9) /* smp kernel running on up */
>> -/* free, was #define X86_FEATURE_FXSAVE_LEAK ( 3*32+10) * "" FXSAVE  
>> leaks FOP/FIP/FOP */
>> +#define X86_FEATURE_ART		(3*32+10) /* Platform has always running  
>> timer (ART) */
>>  #define X86_FEATURE_ARCH_PERFMON ( 3*32+11) /* Intel Architectural  
>> PerfMon */
>>  #define X86_FEATURE_PEBS	( 3*32+12) /* Precise-Event Based Sampling */
>>  #define X86_FEATURE_BTS		( 3*32+13) /* Branch Trace Store */
>> diff --git a/arch/x86/include/asm/tsc.h b/arch/x86/include/asm/tsc.h
>> index 6d7c547..174c421 100644
>> --- a/arch/x86/include/asm/tsc.h
>> +++ b/arch/x86/include/asm/tsc.h
>> @@ -29,6 +29,8 @@ static inline cycles_t get_cycles(void)
>>  	return rdtsc();
>>  }
>> +extern struct system_counterval_t convert_art_to_tsc(cycle_t art);
>> +
>>  extern void tsc_init(void);
>>  extern void mark_tsc_unstable(char *reason);
>>  extern int unsynchronized_tsc(void);
>> diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
>> index 3d743da..a10cff1 100644
>> --- a/arch/x86/kernel/tsc.c
>> +++ b/arch/x86/kernel/tsc.c
>> @@ -43,6 +43,11 @@ static DEFINE_STATIC_KEY_FALSE(__use_tsc);
>> int tsc_clocksource_reliable;
>> +static u32 art_to_tsc_numerator;
>> +static u32 art_to_tsc_denominator;
>> +static u64 art_to_tsc_offset;
>> +struct clocksource *art_related_clocksource;
>> +
>>  /*
>>   * Use a ring-buffer like data structure, where a writer advances the  
>> head by
>>   * writing a new data entry and a reader advances the tail when it  
>> observes a
>> @@ -949,6 +954,36 @@ static struct notifier_block  
>> time_cpufreq_notifier_block = {
>>  	.notifier_call  = time_cpufreq_notifier
>>  };
>> +#define ART_CPUID_LEAF (0x15)
>> +#define ART_MIN_DENOMINATOR (1)
>> +
>> +
>> +/*
>> + * If ART is present detect the numerator:denominator to convert to TSC
>> + */
>> +static void detect_art(void)
>> +{
>> +	unsigned int unused[2];
>> +
>> +	if (boot_cpu_data.cpuid_level < ART_CPUID_LEAF)
>> +		return;
>> +
>> +	cpuid(ART_CPUID_LEAF, &art_to_tsc_denominator,
>> +	      &art_to_tsc_numerator, unused, unused+1);
>> +
>> +	/* Don't enable ART in a VM, non-stop TSC required */
>> +	if (boot_cpu_has(X86_FEATURE_HYPERVISOR) ||
>> +	    !boot_cpu_has(X86_FEATURE_NONSTOP_TSC) ||
>> +	    art_to_tsc_denominator < ART_MIN_DENOMINATOR)
>> +		return;
>> +
>> +	if (rdmsrl_safe(MSR_IA32_TSC_ADJUST, &art_to_tsc_offset))
>> +		return;
>> +
>> +	/* Make this sticky over multiple CPU init calls */
>> +	setup_force_cpu_cap(X86_FEATURE_ART);
>> +}
>> +
>>  static int __init cpufreq_tsc(void)
>>  {
>>  	if (!cpu_has_tsc)
>> @@ -1071,6 +1106,25 @@ int unsynchronized_tsc(void)
>>  	return 0;
>>  }
>> +/*
>> + * Convert ART to TSC given numerator/denominator found in detect_art()
>> + */
>> +struct system_counterval_t convert_art_to_tsc(cycle_t art)
>> +{
>> +	u64 tmp, res, rem;
>> +
>> +	rem = do_div(art, art_to_tsc_denominator);
>> +
>> +	res = art * art_to_tsc_numerator;
>> +	tmp = rem * art_to_tsc_numerator;
>> +
>> +	do_div(tmp, art_to_tsc_denominator);
>> +	res += tmp + art_to_tsc_offset;
>> +
>> +	return (struct system_counterval_t) {.cs = art_related_clocksource,
>> +			.cycles = res};
>> +}
>> +EXPORT_SYMBOL(convert_art_to_tsc);
>> static void tsc_refine_calibration_work(struct work_struct *work);
>>  static DECLARE_DELAYED_WORK(tsc_irqwork, tsc_refine_calibration_work);
>> @@ -1142,6 +1196,8 @@ static void tsc_refine_calibration_work(struct  
>> work_struct *work)
>>  		(unsigned long)tsc_khz % 1000);
>> out:
>> +	if (boot_cpu_has(X86_FEATURE_ART))
>> +		art_related_clocksource = &clocksource_tsc;
>>  	clocksource_register_khz(&clocksource_tsc, tsc_khz);
>>  }
>> @@ -1235,6 +1291,8 @@ void __init tsc_init(void)
>>  		mark_tsc_unstable("TSCs unsynchronized");
>> 	check_system_tsc_reliable();
>> +
>> +	detect_art();
>>  }
>> #ifdef CONFIG_SMP

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ