lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <aacf7d03-faa7-c17d-b763-c80b293106ea@arm.com>
Date:   Wed, 24 Jan 2018 10:58:22 +0000
From:   Marc Zyngier <marc.zyngier@....com>
To:     Khuong Dinh <kdinh@....com>, linux-arm-kernel@...ts.infradead.org,
        will.deacon@....com, catalin.marinas@....com
Cc:     msalter@...hat.com, jcm@...hat.com, lorenzo.pieralisi@....com,
        ard.biesheuvel@...aro.org, linux-kernel@...r.kernel.org,
        christoffer.dall@...aro.org, patches@....com
Subject: Re: [PATCH] arm64: turn off xgene branch prediction while in kernel
 space

Khuong,

On 24/01/18 02:13, Khuong Dinh wrote:
> Aliasing attacks against CPU branch predictors can allow an attacker to
> redirect speculative control flow on some CPUs and potentially divulge
> information from one context to another.
> 
> This patch only supports for XGene processors.
> 
> Signed-off-by: Mark Salter <msalter@...hat.com>
> Signed-off-by: Khuong Dinh <kdinh@....com>
> ---
>  arch/arm64/include/asm/cpucaps.h |    3 ++-
>  arch/arm64/include/asm/fixmap.h  |    4 ++++
>  arch/arm64/kernel/cpu_errata.c   |   18 ++++++++++++++++++
>  arch/arm64/kernel/entry.S        |   28 ++++++++++++++++++++++++++++
>  arch/arm64/kernel/smp.c          |   34 ++++++++++++++++++++++++++++++++++
>  5 files changed, 86 insertions(+), 1 deletions(-)
> 
> diff --git a/arch/arm64/include/asm/cpucaps.h b/arch/arm64/include/asm/cpucaps.h
> index bb26382..dc9ada1 100644
> --- a/arch/arm64/include/asm/cpucaps.h
> +++ b/arch/arm64/include/asm/cpucaps.h
> @@ -45,7 +45,8 @@
>  #define ARM64_HARDEN_BRANCH_PREDICTOR		24
>  #define ARM64_HARDEN_BP_POST_GUEST_EXIT		25
>  #define ARM64_HAS_RAS_EXTN			26
> +#define ARM64_XGENE_HARDEN_BRANCH_PREDICTOR	27
>  

Why isn't this using the infrastructure that is already in place?

> -#define ARM64_NCAPS				27
> +#define ARM64_NCAPS				28
>  
>  #endif /* __ASM_CPUCAPS_H */
> diff --git a/arch/arm64/include/asm/fixmap.h b/arch/arm64/include/asm/fixmap.h
> index ec1e6d6..d5400ca 100644
> --- a/arch/arm64/include/asm/fixmap.h
> +++ b/arch/arm64/include/asm/fixmap.h
> @@ -63,6 +63,10 @@ enum fixed_addresses {
>  	FIX_ENTRY_TRAMP_TEXT,
>  #define TRAMP_VALIAS		(__fix_to_virt(FIX_ENTRY_TRAMP_TEXT))
>  #endif /* CONFIG_UNMAP_KERNEL_AT_EL0 */
> +
> +#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
> +	FIX_BOOT_CPU_BP_CTLREG,
> +#endif /* CONFIG_HARDEN_BRANCH_PREDICTOR */
>  	__end_of_permanent_fixed_addresses,
>  
>  	/*
> diff --git a/arch/arm64/kernel/cpu_errata.c b/arch/arm64/kernel/cpu_errata.c
> index ed68818..1554014 100644
> --- a/arch/arm64/kernel/cpu_errata.c
> +++ b/arch/arm64/kernel/cpu_errata.c
> @@ -53,6 +53,18 @@
>  		(arm64_ftr_reg_ctrel0.sys_val & arm64_ftr_reg_ctrel0.strict_mask);
>  }
>  
> +#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
> +static bool is_xgene_cpu(const struct arm64_cpu_capabilities *entry, int scope)
> +{
> +	unsigned int midr = read_cpuid_id();
> +	unsigned int variant = MIDR_VARIANT(midr);
> +
> +	WARN_ON(scope != SCOPE_LOCAL_CPU);
> +	return MIDR_IMPLEMENTOR(midr) == ARM_CPU_IMP_APM && (variant <= 3) &&
> +					is_hyp_mode_available();
> +}

So what happens in a guest? No BP invalidation whatsoever? I don't think
that's acceptable.

> +#endif
> +
>  static int cpu_enable_trap_ctr_access(void *__unused)
>  {
>  	/* Clear SCTLR_EL1.UCT */
> @@ -369,6 +381,12 @@ static int qcom_enable_link_stack_sanitization(void *data)
>  		MIDR_ALL_VERSIONS(MIDR_CAVIUM_THUNDERX2),
>  		.enable = enable_psci_bp_hardening,
>  	},
> +	{
> +		.desc = "ARM64 XGENE branch predictors control",
> +		.capability = ARM64_XGENE_HARDEN_BRANCH_PREDICTOR,
> +		.def_scope = SCOPE_LOCAL_CPU,
> +		.matches = is_xgene_cpu,
> +	},
>  #endif
>  	{
>  	}
> diff --git a/arch/arm64/kernel/entry.S b/arch/arm64/kernel/entry.S
> index b34e717..8c7d98e 100644
> --- a/arch/arm64/kernel/entry.S
> +++ b/arch/arm64/kernel/entry.S
> @@ -62,6 +62,32 @@
>  #endif
>  	.endm
>  
> +	.macro bp_disable, tmp1, tmp2
> +#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
> +	alternative_if ARM64_XGENE_HARDEN_BRANCH_PREDICTOR
> +	adr_l	x\tmp1, bp_ctlreg
> +	mrs	x\tmp2, tpidr_el1
> +	ldr	x\tmp1, [x\tmp1, x\tmp2]
> +	ldr	w\tmp2, [x\tmp1]
> +	orr	w\tmp2, w\tmp2, #(1 << 25)
> +	str	w\tmp2, [x\tmp1]
> +	alternative_else_nop_endif
> +#endif
> +	.endm
> +
> +	.macro bp_enable, tmp1, tmp2
> +#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
> +	alternative_if ARM64_XGENE_HARDEN_BRANCH_PREDICTOR
> +	adr_l	x\tmp1, bp_ctlreg
> +	mrs	x\tmp2, tpidr_el1
> +	ldr	x\tmp1, [x\tmp1, x\tmp2]
> +	ldr	w\tmp2, [x\tmp1]
> +	and	w\tmp2, w\tmp2, #~(1 << 25)
> +	str	w\tmp2, [x\tmp1]
> +	alternative_else_nop_endif
> +#endif
> +	.endm
> +
>  /*
>   * Bad Abort numbers
>   *-----------------
> @@ -158,6 +184,7 @@ alternative_else_nop_endif
>  	stp	x28, x29, [sp, #16 * 14]
>  
>  	.if	\el == 0
> +	bp_disable 20, 21
>  	mrs	x21, sp_el0
>  	ldr_this_cpu	tsk, __entry_task, x20	// Ensure MDSCR_EL1.SS is clear,
>  	ldr	x19, [tsk, #TSK_TI_FLAGS]	// since we can unmask debug
> @@ -307,6 +334,7 @@ alternative_else_nop_endif
>  
>  	msr	elr_el1, x21			// set up the return data
>  	msr	spsr_el1, x22
> +	bp_enable 21, 22
>  	ldp	x0, x1, [sp, #16 * 0]
>  	ldp	x2, x3, [sp, #16 * 1]
>  	ldp	x4, x5, [sp, #16 * 2]

This is not what we do on other cores. Why is XGene any different?

> diff --git a/arch/arm64/kernel/smp.c b/arch/arm64/kernel/smp.c
> index 3b8ad7b..69646be 100644
> --- a/arch/arm64/kernel/smp.c
> +++ b/arch/arm64/kernel/smp.c
> @@ -85,6 +85,38 @@ enum ipi_msg_type {
>  	IPI_WAKEUP
>  };
>  
> +#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
> +DEFINE_PER_CPU_READ_MOSTLY(void __iomem *, bp_ctlreg);
> +
> +static void map_bp_ctlreg(void)
> +{
> +	if (cpus_have_const_cap(ARM64_XGENE_HARDEN_BRANCH_PREDICTOR)) {
> +		u64 mpidr = read_cpuid_mpidr();
> +		unsigned int idx;
> +		void __iomem *p;
> +		phys_addr_t pa;
> +
> +		idx =  (MPIDR_AFFINITY_LEVEL(mpidr, 1) << 1) +
> +			MPIDR_AFFINITY_LEVEL(mpidr, 0);
> +		pa = 0x7c0c0000ULL | (0x100000ULL * idx);
> +		if (smp_processor_id())
> +			p = ioremap(pa, PAGE_SIZE);
> +		else {
> +			/* boot processor uses fixmap */
> +			set_fixmap_io(FIX_BOOT_CPU_BP_CTLREG, pa);
> +			p = (void __iomem *)__fix_to_virt(
> +						FIX_BOOT_CPU_BP_CTLREG);
> +		}
> +		__this_cpu_write(bp_ctlreg, p);
> +
> +		pr_debug("%s: cpu%d idx=%d pa=0x%llx %p", __func__,
> +			 smp_processor_id(), idx, pa, p);
> +	}
> +}
> +#else
> +static inline void map_bp_ctlreg(void) {}
> +#endif
> +
>  #ifdef CONFIG_ARM64_VHE
>  
>  /* Whether the boot CPU is running in HYP mode or not*/
> @@ -224,6 +256,7 @@ asmlinkage void secondary_start_kernel(void)
>  
>  	cpu = task_cpu(current);
>  	set_my_cpu_offset(per_cpu_offset(cpu));
> +	map_bp_ctlreg();
>  
>  	/*
>  	 * All kernel threads share the same mm context; grab a
> @@ -454,6 +487,7 @@ void __init smp_prepare_boot_cpu(void)
>  	 * cpuinfo_store_boot_cpu() above.
>  	 */
>  	update_cpu_errata_workarounds();
> +	map_bp_ctlreg();
>  }
>  
>  static u64 __init of_get_cpu_mpidr(struct device_node *dn)
> 

Thanks,

	M.
-- 
Jazz is not dead. It just smells funny...

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ