[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <202109270924.D0BC211B4@keescook>
Date: Mon, 27 Sep 2021 09:27:42 -0700
From: Kees Cook <keescook@...omium.org>
To: Xiongwei Song <sxwjean@...il.com>
Cc: Xiongwei Song <sxwjean@...com>, x86@...nel.org, linux-mm@...ck.org,
tglx@...utronix.de, mingo@...hat.com, bp@...en8.de, hpa@...or.com,
luto@...nel.org, krisman@...labora.com, chang.seok.bae@...el.com,
viro@...iv.linux.org.uk, nivedita@...m.mit.edu,
adobriyan@...il.com, oleg@...hat.com, sblbir@...zon.com,
axboe@...nel.dk, laijs@...ux.alibaba.com,
dave.hansen@...ux.intel.com, peterz@...radead.org,
akpm@...ux-foundation.org, arnd@...db.de, davem@...emloft.net,
kim.phillips@....com, yazen.ghannam@....com, dave@...olabs.net,
metze@...ba.org, elver@...gle.com, ebiederm@...ssion.com,
Christophe Leroy <christophe.leroy@...roup.eu>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH RESEND 2/2] x86/mm: Randomize va with generic
arch_pick_mmap_layout()
On Mon, Sep 27, 2021 at 03:37:28PM +0800, Xiongwei Song wrote:
> Hi Experts,
>
> Any other objections on this patch?
>
> Is it worth doing so? If yes, I will update the patch based on Peter's comments.
>
> Regards,
> Xiongwei
>
> On Tue, Sep 21, 2021 at 10:35 PM <sxwjean@...com> wrote:
> >
> > From: Xiongwei Song <sxwjean@...il.com>
> >
> > The code logic of arch_pick_mmap_layout() of x86 is basiclly same as
> > arch_pick_mmap_layout() in mm/util.c. Let's delete the function and
> > the related assistant functions in x86.
> >
> > There are some differences between x86 and geneirc code:
> > - mmap_is_legacy(), there is no check for stack limit of 32 bits process
> > in x86, while generic code does, which is suitable for x86 too.
> > - arch_randomize_brk(), it only randomized the brk with SZ_32M range
> > for 32 bits and 64 bits processes, while the generic code randomizes brk
> > with SZ_1G range for 64 bits process, which is suitable for x86 too.
> > - Implement is_compat_task(), which means 32 bits process in 64 bits
> > kernel, to adapt generic implementation.
> > - The special implementation of x86 is the code with
> > CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES enabled. I assume other archs will
> > enable CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES in the future, so move the
> > implementation to the generic part.
Would it be possible to break this patch up into the separate pieces you
outline above? This code is pretty twisty, and keeping isolated changes
in separate patches can help with review.
Also, can you verify that the before/after entropy remains the same?
(There are a bunch of tools that measure various aspects of the layout
randomization, etc.)
Other than that, I'm all for consolidating this code. I've taking a few
passes at it in the past, so I welcome improvements here. There's a lot
of arch-specific stuff that doesn't need to be arch-specific. :)
Thanks for digging into this!
-Kees
> >
> > Hence select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT in x86 to use
> > generic arch_pick_mmap_layout(). Meanwhile, the ARCH_HAS_ELF_RANDOMIZE
> > can be enabled automatically, remove select for it.
> >
> > Signed-off-by: Xiongwei Song <sxwjean@...il.com>
> > ---
> > arch/x86/Kconfig | 2 +-
> > arch/x86/include/asm/compat.h | 7 +-
> > arch/x86/include/asm/processor.h | 4 +-
> > arch/x86/kernel/process.c | 5 --
> > arch/x86/mm/mmap.c | 112 -------------------------------
> > include/linux/compat.h | 4 ++
> > mm/util.c | 18 ++++-
> > 7 files changed, 29 insertions(+), 123 deletions(-)
> >
> > diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig
> > index dad7f85dcdea..c081e6ff7f11 100644
> > --- a/arch/x86/Kconfig
> > +++ b/arch/x86/Kconfig
> > @@ -72,7 +72,6 @@ config X86
> > select ARCH_HAS_DEBUG_VM_PGTABLE if !X86_PAE
> > select ARCH_HAS_DEVMEM_IS_ALLOWED
> > select ARCH_HAS_EARLY_DEBUG if KGDB
> > - select ARCH_HAS_ELF_RANDOMIZE
> > select ARCH_HAS_FAST_MULTIPLIER
> > select ARCH_HAS_FILTER_PGPROT
> > select ARCH_HAS_FORTIFY_SOURCE
> > @@ -114,6 +113,7 @@ config X86
> > select ARCH_USE_SYM_ANNOTATIONS
> > select ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
> > select ARCH_WANT_DEFAULT_BPF_JIT if X86_64
> > + select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT
> > select ARCH_WANTS_DYNAMIC_TASK_STRUCT
> > select ARCH_WANTS_NO_INSTR
> > select ARCH_WANT_HUGE_PMD_SHARE
> > diff --git a/arch/x86/include/asm/compat.h b/arch/x86/include/asm/compat.h
> > index 7516e4199b3c..c697e377644d 100644
> > --- a/arch/x86/include/asm/compat.h
> > +++ b/arch/x86/include/asm/compat.h
> > @@ -151,6 +151,11 @@ struct compat_shmid64_ds {
> > compat_ulong_t __unused5;
> > };
> >
> > +static inline int is_compat_task(void)
> > +{
> > + return IS_ENABLED(CONFIG_COMPAT) && test_thread_flag(TIF_32BIT);
> > +}
> > +
> > #ifdef CONFIG_X86_X32_ABI
> > #define COMPAT_USE_64BIT_TIME \
> > (!!(task_pt_regs(current)->orig_ax & __X32_SYSCALL_BIT))
> > @@ -165,12 +170,12 @@ static inline bool in_x32_syscall(void)
> > return false;
> > }
> >
> > +#ifdef CONFIG_COMPAT
> > static inline bool in_32bit_syscall(void)
> > {
> > return in_ia32_syscall() || in_x32_syscall();
> > }
> >
> > -#ifdef CONFIG_COMPAT
> > static inline bool in_compat_syscall(void)
> > {
> > return in_32bit_syscall();
> > diff --git a/arch/x86/include/asm/processor.h b/arch/x86/include/asm/processor.h
> > index 9ad2acaaae9b..c28a36ee6eb0 100644
> > --- a/arch/x86/include/asm/processor.h
> > +++ b/arch/x86/include/asm/processor.h
> > @@ -708,7 +708,6 @@ extern int bootloader_version;
> >
> > extern char ignore_fpu_irq;
> >
> > -#define HAVE_ARCH_PICK_MMAP_LAYOUT 1
> > #define ARCH_HAS_PREFETCHW
> > #define ARCH_HAS_SPINLOCK_PREFETCH
> >
> > @@ -785,6 +784,9 @@ extern void start_thread(struct pt_regs *regs, unsigned long new_ip,
> > */
> > #define __TASK_UNMAPPED_BASE(task_size) (PAGE_ALIGN(task_size / 3))
> > #define TASK_UNMAPPED_BASE __TASK_UNMAPPED_BASE(TASK_SIZE_LOW)
> > +#ifdef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES
> > +#define TASK_UNMAPPED_COMPAT_BASE __TASK_UNMAPPED_BASE(task_size_32bit())
> > +#endif
> >
> > #define KSTK_EIP(task) (task_pt_regs(task)->ip)
> >
> > diff --git a/arch/x86/kernel/process.c b/arch/x86/kernel/process.c
> > index 1d9463e3096b..1e747d34c18d 100644
> > --- a/arch/x86/kernel/process.c
> > +++ b/arch/x86/kernel/process.c
> > @@ -931,11 +931,6 @@ unsigned long arch_align_stack(unsigned long sp)
> > return sp & ~0xf;
> > }
> >
> > -unsigned long arch_randomize_brk(struct mm_struct *mm)
> > -{
> > - return randomize_page(mm->brk, 0x02000000);
> > -}
> > -
> > /*
> > * Called from fs/proc with a reference on @p to find the function
> > * which called into schedule(). This needs to be done carefully
> > diff --git a/arch/x86/mm/mmap.c b/arch/x86/mm/mmap.c
> > index c90c20904a60..daf65cc5e5b1 100644
> > --- a/arch/x86/mm/mmap.c
> > +++ b/arch/x86/mm/mmap.c
> > @@ -38,118 +38,6 @@ unsigned long task_size_64bit(int full_addr_space)
> > return full_addr_space ? TASK_SIZE_MAX : DEFAULT_MAP_WINDOW;
> > }
> >
> > -static unsigned long stack_maxrandom_size(unsigned long task_size)
> > -{
> > - unsigned long max = 0;
> > - if (current->flags & PF_RANDOMIZE) {
> > - max = (-1UL) & __STACK_RND_MASK(task_size == task_size_32bit());
> > - max <<= PAGE_SHIFT;
> > - }
> > -
> > - return max;
> > -}
> > -
> > -#ifdef CONFIG_COMPAT
> > -# define mmap32_rnd_bits mmap_rnd_compat_bits
> > -# define mmap64_rnd_bits mmap_rnd_bits
> > -#else
> > -# define mmap32_rnd_bits mmap_rnd_bits
> > -# define mmap64_rnd_bits mmap_rnd_bits
> > -#endif
> > -
> > -#define SIZE_128M (128 * 1024 * 1024UL)
> > -
> > -static int mmap_is_legacy(void)
> > -{
> > - if (current->personality & ADDR_COMPAT_LAYOUT)
> > - return 1;
> > -
> > - return sysctl_legacy_va_layout;
> > -}
> > -
> > -static unsigned long arch_rnd(unsigned int rndbits)
> > -{
> > - if (!(current->flags & PF_RANDOMIZE))
> > - return 0;
> > - return (get_random_long() & ((1UL << rndbits) - 1)) << PAGE_SHIFT;
> > -}
> > -
> > -unsigned long arch_mmap_rnd(void)
> > -{
> > - return arch_rnd(mmap_is_ia32() ? mmap32_rnd_bits : mmap64_rnd_bits);
> > -}
> > -
> > -static unsigned long mmap_base(unsigned long rnd, unsigned long task_size,
> > - struct rlimit *rlim_stack)
> > -{
> > - unsigned long gap = rlim_stack->rlim_cur;
> > - unsigned long pad = stack_maxrandom_size(task_size) + stack_guard_gap;
> > - unsigned long gap_min, gap_max;
> > -
> > - /* Values close to RLIM_INFINITY can overflow. */
> > - if (gap + pad > gap)
> > - gap += pad;
> > -
> > - /*
> > - * Top of mmap area (just below the process stack).
> > - * Leave an at least ~128 MB hole with possible stack randomization.
> > - */
> > - gap_min = SIZE_128M;
> > - gap_max = (task_size / 6) * 5;
> > -
> > - if (gap < gap_min)
> > - gap = gap_min;
> > - else if (gap > gap_max)
> > - gap = gap_max;
> > -
> > - return PAGE_ALIGN(task_size - gap - rnd);
> > -}
> > -
> > -static unsigned long mmap_legacy_base(unsigned long rnd,
> > - unsigned long task_size)
> > -{
> > - return __TASK_UNMAPPED_BASE(task_size) + rnd;
> > -}
> > -
> > -/*
> > - * This function, called very early during the creation of a new
> > - * process VM image, sets up which VM layout function to use:
> > - */
> > -static void arch_pick_mmap_base(unsigned long *base, unsigned long *legacy_base,
> > - unsigned long random_factor, unsigned long task_size,
> > - struct rlimit *rlim_stack)
> > -{
> > - *legacy_base = mmap_legacy_base(random_factor, task_size);
> > - if (mmap_is_legacy())
> > - *base = *legacy_base;
> > - else
> > - *base = mmap_base(random_factor, task_size, rlim_stack);
> > -}
> > -
> > -void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack)
> > -{
> > - if (mmap_is_legacy())
> > - mm->get_unmapped_area = arch_get_unmapped_area;
> > - else
> > - mm->get_unmapped_area = arch_get_unmapped_area_topdown;
> > -
> > - arch_pick_mmap_base(&mm->mmap_base, &mm->mmap_legacy_base,
> > - arch_rnd(mmap64_rnd_bits), task_size_64bit(0),
> > - rlim_stack);
> > -
> > -#ifdef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES
> > - /*
> > - * The mmap syscall mapping base decision depends solely on the
> > - * syscall type (64-bit or compat). This applies for 64bit
> > - * applications and 32bit applications. The 64bit syscall uses
> > - * mmap_base, the compat syscall uses mmap_compat_base.
> > - */
> > - arch_pick_mmap_base(&mm->mmap_compat_base, &mm->mmap_compat_legacy_base,
> > - arch_rnd(mmap32_rnd_bits), task_size_32bit(),
> > - rlim_stack);
> > -#endif
> > -}
> > -
> > unsigned long get_mmap_base(int is_legacy)
> > {
> > struct mm_struct *mm = current->mm;
> > diff --git a/include/linux/compat.h b/include/linux/compat.h
> > index 1c758b0e0359..0f7cc94f9b3f 100644
> > --- a/include/linux/compat.h
> > +++ b/include/linux/compat.h
> > @@ -946,6 +946,10 @@ static inline bool in_compat_syscall(void) { return false; }
> >
> > #endif /* CONFIG_COMPAT */
> >
> > +#ifndef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES
> > +static inline bool in_32bit_syscall(void) { return false; }
> > +#endif
> > +
> > #define BITS_PER_COMPAT_LONG (8*sizeof(compat_long_t))
> >
> > #define BITS_TO_COMPAT_LONGS(bits) DIV_ROUND_UP(bits, BITS_PER_COMPAT_LONG)
> > diff --git a/mm/util.c b/mm/util.c
> > index 4ac87f1b30f1..8932388c96a3 100644
> > --- a/mm/util.c
> > +++ b/mm/util.c
> > @@ -357,8 +357,9 @@ unsigned long arch_mmap_rnd(void)
> > {
> > unsigned long rnd;
> >
> > -#ifdef CONFIG_HAVE_ARCH_MMAP_RND_COMPAT_BITS
> > - if (is_compat_task())
> > +#if defined(CONFIG_HAVE_ARCH_MMAP_RND_COMPAT_BITS) \
> > + || defined(CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES)
> > + if (is_compat_task() || in_32bit_syscall())
> > rnd = get_random_long() & ((1UL << mmap_rnd_compat_bits) - 1);
> > else
> > #endif /* CONFIG_HAVE_ARCH_MMAP_RND_COMPAT_BITS */
> > @@ -413,13 +414,24 @@ void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack)
> > if (current->flags & PF_RANDOMIZE)
> > random_factor = arch_mmap_rnd();
> >
> > + mm->mmap_legacy_base = TASK_UNMAPPED_BASE + random_factor;
> > if (mmap_is_legacy(rlim_stack)) {
> > - mm->mmap_base = TASK_UNMAPPED_BASE + random_factor;
> > + mm->mmap_base = mm->mmap_legacy_base;
> > mm->get_unmapped_area = arch_get_unmapped_area;
> > } else {
> > mm->mmap_base = mmap_base(random_factor, rlim_stack);
> > mm->get_unmapped_area = arch_get_unmapped_area_topdown;
> > }
> > +
> > +#ifdef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES
> > + if (mmap_is_legacy(rlim_stack)) {
> > + mm->mmap_compat_legacy_base =
> > + TASK_UNMAPPED_COMPAT_BASE + random_factor;
> > + mm->mmap_compat_base = mm->mmap_compat_legacy_base;
> > + } else {
> > + mm->mmap_compat_base = mmap_base(random_factor, rlim_stack);
> > + }
> > +#endif
> > }
> > #elif defined(CONFIG_MMU) && !defined(HAVE_ARCH_PICK_MMAP_LAYOUT)
> > void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack)
> > --
> > 2.30.2
> >
--
Kees Cook
Powered by blists - more mailing lists