[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20251218191332.35849-2-lukas.gerlach@cispa.de>
Date: Thu, 18 Dec 2025 20:13:31 +0100
From: Lukas Gerlach <lukas.gerlach@...pa.de>
To: <linux-riscv@...ts.infradead.org>
CC: <palmer@...belt.com>, <pjw@...nel.org>, <aou@...s.berkeley.edu>,
<alex@...ti.fr>, <linux-kernel@...r.kernel.org>, <daniel.weber@...pa.de>,
<michael.schwarz@...pa.de>, <marton.bognar@...euven.be>,
<jo.vanbulck@...euven.be>, Lukas Gerlach <lukas.gerlach@...pa.de>
Subject: [PATCH 1/2] riscv: Use pointer masking to limit uaccess speculation
Similarly to x86 and arm64, mitigate speculation past an access_ok()
check by masking the pointer before use.
On RISC-V, user addresses have the MSB clear while kernel addresses
have the MSB set. The uaccess_mask_ptr() function clears the MSB,
ensuring any kernel pointer becomes invalid and will fault, while
valid user pointers remain unchanged. This prevents speculative
access to kernel memory via user copy functions.
The masking is applied to __get_user, __put_user, raw_copy_from_user,
raw_copy_to_user, clear_user, and the unsafe_* variants.
Signed-off-by: Lukas Gerlach <lukas.gerlach@...pa.de>
---
arch/riscv/include/asm/uaccess.h | 41 +++++++++++++++++++++++++-------
1 file changed, 32 insertions(+), 9 deletions(-)
diff --git a/arch/riscv/include/asm/uaccess.h b/arch/riscv/include/asm/uaccess.h
index 36bba6720c26..ceee1d62ff9b 100644
--- a/arch/riscv/include/asm/uaccess.h
+++ b/arch/riscv/include/asm/uaccess.h
@@ -74,6 +74,23 @@ static inline unsigned long __untagged_addr_remote(struct mm_struct *mm, unsigne
#define __typefits(x, type, not) \
__builtin_choose_expr(sizeof(x) <= sizeof(type), (unsigned type)0, not)
+/*
+ * Sanitize a uaccess pointer such that it cannot reach any kernel address.
+ *
+ * On RISC-V, virtual addresses are sign-extended from the top implemented bit.
+ * User addresses have the MSB clear; kernel addresses have the MSB set.
+ * Clearing the MSB ensures any kernel pointer becomes non-canonical and will
+ * fault, while valid user pointers remain unchanged.
+ */
+#define uaccess_mask_ptr(ptr) ((__typeof__(ptr))__uaccess_mask_ptr(ptr))
+static inline void __user *__uaccess_mask_ptr(const void __user *ptr)
+{
+ unsigned long val = (unsigned long)ptr;
+
+ val = (val << 1) >> 1;
+ return (void __user *)val;
+}
+
/*
* The exception table consists of pairs of addresses: the first is the
* address of an instruction that is allowed to fault, and the second is
@@ -235,7 +252,8 @@ __gu_failed: \
*/
#define __get_user(x, ptr) \
({ \
- const __typeof__(*(ptr)) __user *__gu_ptr = untagged_addr(ptr); \
+ const __typeof__(*(ptr)) __user *__gu_ptr = \
+ uaccess_mask_ptr(untagged_addr(ptr)); \
long __gu_err = 0; \
__typeof__(x) __gu_val; \
\
@@ -366,7 +384,8 @@ err_label: \
*/
#define __put_user(x, ptr) \
({ \
- __typeof__(*(ptr)) __user *__gu_ptr = untagged_addr(ptr); \
+ __typeof__(*(ptr)) __user *__gu_ptr = \
+ uaccess_mask_ptr(untagged_addr(ptr)); \
__typeof__(*__gu_ptr) __val = (x); \
long __pu_err = 0; \
\
@@ -413,13 +432,15 @@ unsigned long __must_check __asm_copy_from_user(void *to,
static inline unsigned long
raw_copy_from_user(void *to, const void __user *from, unsigned long n)
{
- return __asm_copy_from_user(to, untagged_addr(from), n);
+ return __asm_copy_from_user(to,
+ uaccess_mask_ptr(untagged_addr(from)), n);
}
static inline unsigned long
raw_copy_to_user(void __user *to, const void *from, unsigned long n)
{
- return __asm_copy_to_user(untagged_addr(to), from, n);
+ return __asm_copy_to_user(
+ uaccess_mask_ptr(untagged_addr(to)), from, n);
}
extern long strncpy_from_user(char *dest, const char __user *src, long count);
@@ -434,7 +455,7 @@ unsigned long __must_check clear_user(void __user *to, unsigned long n)
{
might_fault();
return access_ok(to, n) ?
- __clear_user(untagged_addr(to), n) : n;
+ __clear_user(uaccess_mask_ptr(untagged_addr(to)), n) : n;
}
#define arch_get_kernel_nofault(dst, src, type, err_label) \
@@ -461,20 +482,22 @@ static inline void user_access_restore(unsigned long enabled) { }
* the error labels - thus the macro games.
*/
#define arch_unsafe_put_user(x, ptr, label) \
- __put_user_nocheck(x, (ptr), label)
+ __put_user_nocheck(x, uaccess_mask_ptr(ptr), label)
#define arch_unsafe_get_user(x, ptr, label) do { \
__inttype(*(ptr)) __gu_val; \
- __get_user_nocheck(__gu_val, (ptr), label); \
+ __get_user_nocheck(__gu_val, uaccess_mask_ptr(ptr), label); \
(x) = (__force __typeof__(*(ptr)))__gu_val; \
} while (0)
#define unsafe_copy_to_user(_dst, _src, _len, label) \
- if (__asm_copy_to_user_sum_enabled(_dst, _src, _len)) \
+ if (__asm_copy_to_user_sum_enabled( \
+ uaccess_mask_ptr(_dst), _src, _len)) \
goto label;
#define unsafe_copy_from_user(_dst, _src, _len, label) \
- if (__asm_copy_from_user_sum_enabled(_dst, _src, _len)) \
+ if (__asm_copy_from_user_sum_enabled( \
+ _dst, uaccess_mask_ptr(_src), _len)) \
goto label;
#else /* CONFIG_MMU */
--
2.51.0
Powered by blists - more mailing lists