lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20250823161326.635281786@linutronix.de>
Date: Sat, 23 Aug 2025 18:39:04 +0200 (CEST)
From: Thomas Gleixner <tglx@...utronix.de>
To: LKML <linux-kernel@...r.kernel.org>
Cc: Jens Axboe <axboe@...nel.dk>,
 Mathieu Desnoyers <mathieu.desnoyers@...icios.com>,
 Peter Zijlstra <peterz@...radead.org>,
 "Paul E. McKenney" <paulmck@...nel.org>,
 Boqun Feng <boqun.feng@...il.com>,
 Paolo Bonzini <pbonzini@...hat.com>,
 Sean Christopherson <seanjc@...gle.com>,
 Wei Liu <wei.liu@...nel.org>,
 Dexuan Cui <decui@...rosoft.com>,
 x86@...nel.org,
 Arnd Bergmann <arnd@...db.de>,
 Heiko Carstens <hca@...ux.ibm.com>,
 Christian Borntraeger <borntraeger@...ux.ibm.com>,
 Sven Schnelle <svens@...ux.ibm.com>,
 Huacai Chen <chenhuacai@...nel.org>,
 Paul Walmsley <paul.walmsley@...ive.com>,
 Palmer Dabbelt <palmer@...belt.com>,
 linux-arch@...r.kernel.org,
 Thomas Bogendoerfer <tsbogend@...ha.franken.de>,
 Michael Ellerman <mpe@...erman.id.au>,
 Jonas Bonn <jonas@...thpole.se>
Subject: [patch V2 00/37] rseq: Optimize exit to user space

OA
This is a follow up on the initial series, which did a very basic attempt
to sanitize the RSEQ handling in the kernel:

   https://lore.kernel.org/all/20250813155941.014821755@linutronix.de

Further analysis turned up more than these initial problems:

  1) task::rseq_event_mask is a pointless bit-field despite the fact that
     the ABI flags it was meant to support have been deprecated and
     functionally disabled three years ago.

  2) task::rseq_event_mask is accumulating bits unless there is a critical
     section discovered in the user space rseq memory. This results in
     pointless invocations of the rseq user space exit handler even if
     there had nothing changed. As a matter of correctness these bits have
     to be clear when exiting to user space and therefore pristine when
     coming back into the kernel. Aside of correctness, this also avoids
     pointless evaluation of the user space memory, which is a performance
     benefit.

  3) The evaluation of critical sections does not differentiate between
     syscall and interrupt/exception exits. The current implementation
     silently tolerates and fixes up critical sections which invoked a
     syscall unless CONFIG_DEBUG_RSEQ is enabled.

     That's just wrong. If user space does that on a production kernel it
     can keep the pieces. The kernel is not there to proliferate mindless
     user space programming and letting everyone pay the performance
     penalty.

Additional findings:

  4) The decision to raise the work for exit is more than suboptimal.
     Basically every context switch does so if the task has rseq, which is
     nowadays likely as glibc makes use of it if available.

     The consequence is that a lot of exits have to process RSEQ just for
     nothing. The only reasons to do so are:

       the task was interrupted in user space and schedules

     or

       the CPU or MM CID changes in schedule() independent of the entry
       mode

     That reduces the invocation space obviously significantly.

  5) Signal handling does the RSEQ update unconditionally.

     That's wrong as the only reason to do so is when the task was
     interrupted in user space independent of a schedule event.

     The only important task in that case is to handle the critical section
     because after switching to the signal frame the original return IP is
     not longer available.

     The CPU/MM CID values do not need to be updated at that point as they
     can change again before the signal delivery goes out to user space.

     Again, if the task was in a critical section and issued a syscall then
     it can keep the pieces as that's a violation of the ABI contract.

  6) CPU and MM CID are updated unconditionally

     That's again a pointless exercise when they didn't change. Then the
     only action required is to check the critical section if and only if
     the entry came via an interrupt.

     That can obviously be avoided by caching the values written to user
     space and avoiding that path if they haven't changed

  7) The TIF_NOTIFY_RESUME mechanism is a horrorshow

     TIF_NOTIFY_RESUME is a multiplexing TIF bit and needs to invoke the
     world and some more. Depending on workloads this can be set by
     task_work, security, block and memory management. All unrelated to
     RSEQ and quite some of them are likely to cause a reschedule.
     But most of them are low frequency.

     So doing this work in the loop unconditionally is just waste. The
     correct point to do this is at the end of that loop once all other bits
     have been processed, because that's the point where the task is
     actually going out to user space.

  8) #7 caused another subtle work for nothing issue

     IO/URING and hypervisors invoke resume_user_mode_work() with a NULL
     pointer for pt_regs, which causes the RSEQ code to ignore the critical
     section check, but updating the CPU ID/ MM CID values unconditionally.

     For IO/URING this invocation is irrelevant because the IO workers can
     never go out to user space and therefore do not have RSEQ memory in
     the first place. So it's a non problem in the existing code as
     task::rseq is NULL in that case.

     Hypervisors are a different story. They need to drain task_work and
     other pending items, which are multiplexed by TIF_NOTIFY_RESUME,
     before entering guest mode.

     The invocation of resume_user_mode_work() clears TIF_NOTIFY_RESUME,
     which means if rseq would ignore that case then it could miss a CPU
     or MM CID update on the way back to user space.

     The handling of that is just a horrible and mindless hack as the event
     might be re-raised between the time the ioctl() enters guest mode and
     the actual exit to user space.

     So the obvious thing is to ignore the regs=NULL call and let the
     offending hypervisor calls check when returning from the ioctl()
     whether the event bit is set and re-raise the notification again.

  9) Code efficiency

     RSEQ aims to improve performance for user space, but it completely
     ignores the fact, that this needs to be implemented in a way which
     does not impact the performance of the kernel significantly.

     So far this did not pop up as just a few people used it, but that has
     changed because glibc started to use it widely.

     It's not so efficiOS^H^Hent as advertised:

     For a full kernel rebuild:

     	 exit to user:                  50106428
	 signal checks:                    24703
	 slowpath runs:                   943785 1.88%
	 id updates:                      968488 1.93%
	 cs checks:                       888768 1.77%
	   cs cleared:                      888768 100.00%
	   cs fixup:                             0 0.00%

     The cs_cleared/fixup numbers are relative to cs_checks, which means
     this does pointless clears even if it was clear already because that's
     what is the case with glibc. glibc is only interested in the CPU/MM
     CID values so far. And no, it's not only a store, it's the whole dance
     of spectre-v1 mitigation plus user access enable/disable.

     I really have to ask why all those people who are deeply caring about
     performance and have utilized this stuff for quite some time have not
     noticed what a performance sh*tshow this inflicts on the kernel.

     Seriously?

     Aside of that did anyone ever look at the resulting assembly code?

     While the user space counterpart is carefully crafted in hand written
     assembly for efficiency the kernel implementation is a complete and
     utter nightmare.

     C allows to implement very efficient code, but the fact that the
     compiler requires help from the programmer to turn it into actual
     efficient assembly is not a totally new finding.

     The absence of properly designed data structures, which allow to build
     efficient and comprehensible code, is amazing.

     Slapping random storage members into struct task_struct, implementing
     a conditional maze around them and then hoping that the compiler will
     create sensible code out of it, is wishful thinking at best.

     I completely understand that all the sanity checks and loops and hoops
     were required to get this actually working. But I'm seriously grumpy
     about the fact, that once the PoC code was shoved into the kernel, the
     'performance work' was completed and just the next fancy 'performance'
     features were piled on top of it.

     Once this is addressed the above picture changes completely:

          exit to user:                  50071399
	  signal checks:                      182
	  slowpath runs:                     8149 0.02%
	  fastpath runs:                   763366 1.52%
	  id updates:                       78419 0.16%
	  cs checks:                            0 0.00%
	  cs cleared:                           0
	  cs fixup:                             0

     And according to perf that very same kernel build consistently gets
     faster from:

           92.500894648 seconds time elapsed (upstream)

     to

           91.693778124 seconds time elapsed

     Not a lot but not in the noise either, i.e. >= 1%

     For a 1e9 gettid() loop this results in going from:

           49.782703878 seconds time elapsed (upstream)

     to

      	   49.327360851 seconds time elapsed

     Not a lot either, but consistently >=1% as the above.

     For actual rseq critical section usage this makes even more of a
     difference. Aside of avoiding pointless work in the kernel this does
     also not abort critical sections when not necessary, which improves
     user space performance too. The kernel selftests magically complete
     six seconds faster, which is again not a lot compared to a total run
     time of ~540s. But those tests are not really reflecting real work
     loads, they emphasize on stress testing the kernel implementation,
     which is perfectly fine. But still, the reduced amount of pointless
     kernel work is what makes the difference:

     Upstream:

	exit to user:                 736568321
	signal checks:                 62002182
	slowpath runs:                358362121 57.07%
	fastpath runs:                        0 0.00%
	id updates:                   358362129 57.07%
	cs checks:                    358362110 57.07%
	  cs cleared:                   347611246 97.01%
	  cs fixup:                      10750863  2.99%

     Upstream + simple obvious fixes:

	exit to user:                 736462043
	signal checks:                 61902154
	slowpath runs:                223394362 30.33%
	fastpath runs:                        0 0.00%
	id updates:                   285296516 38.74%
	cs checks:                    284430371 38.62%
	  cs cleared:                   277110571 97.43%
	  cs fixup:                       7319800  2.57%

     Fully patched:

        exit to user:                 736477630
	signal checks:                        0
	slowpath runs:                     1367 0.00%
	fastpath runs:                114230240 15.51%
	id updates:                   106498492 14.46%
	cs checks:                      7665511 1.04%
	  cs cleared:                     2092386 27.30%
	  cs fixup:                       5573118 72.70%

     Perf confirms:

       46,242,994,988,432      cycles
        8,390,996,405,969      instructions

           1949.923762000 seconds user
          16111.947776000 seconds sys

       versus:

       44,910,362,743,328      cycles		-3 %
        8,305,138,767,914      instructions     -1 %

           1985.342453000 seconds user		+2 %
          15653.639218000 seconds sys		-3 %

     Running the same with only looking at the kernel counts:

      1) Upstream:

         39,530,448,181,974      cycles:khH
          2,872,985,410,904      instructions:khH
            640,968,094,639      branches:khH

      2) Upstream + simple obvious fixes:
                                                            -> #1
         39,005,607,758,104      cycles:khH                 -1.5 %
          2,853,676,629,923      instructions:khH           -0.8 %
            635,538,256,875      branches:khH               -1.0 %

      3) Fully patched:
						   -> #2    -> #1
         38,786,867,052,870      cycles:khH        -0.6 %   -2.3 %
          2,784,773,121,491      instructions:khH  -2.5 %   -3.1 %
            618,677,047,160      branches:khH      -2.7 %   -3.6 %

     Looking at the kernel build that way:

      1) Upstream:

        474,191,665,063      cycles:khH
        164,176,072,767      instructions:khH
         33,037,122,843      branches:khH

      2) Upstream + simple obvious fixes:
                                                            -> #1
         473,423,488,450      cycles:khH 		    ~0.0 %          
         162,749,873,914      instructions:khH              -1.0 %
          32,264,397,605      branches:khH                  -2.5 %

      3) Fully patched:
                                                   -> #2    -> #1
         468,658,348,098      cycles:khH           -1.1 %   -1.2 %
         160,941,028,283      instructions:khH     -1.2 %   -2.0 %
          31,893,296,267      branches:khH         -1.5 %   -3.7 %

     That's pretty much in line with the 2-3% regressions observed by Jens.

     TBH, I absolutely do not care about the way how performance is
     evaluated in the entities who pushed for this, but I very much care
     about the general impact of this.

     It seems to me the performance evaluation by the $corp power users of
     RSEQ stopped at the point where a significant improvement in their
     user space was observed. It's interesting, that the resulting overhead
     in the kernel seems not that relevant for the overall performance
     evaluation. That's left to others to mop up.

     The tragedy of the commons in action.

TBH, what surprised me most when looking into this in detail, was the large
amount of low hanging fruit, which was sitting there in plain sight.

That said, this series addresses the overall problems by:

  1) Limiting the RSEQ work to the actual conditions where it is
     required. The full benefit is only available for architectures using
     the generic entry infrastructure. All others get at least the basic
     improvements.

  2) Re-implementing the whole user space handling based on proper data
     structures and by actually looking at the impact it creates in the
     fast path.

  3) Moving the actual handling of RSEQ out to the latest point in the exit
     path, where possible. This is fully inlined into the fast path to keep
     the impact confined.

     The initial attempt to make it completely independent of TIF bits and
     just handle it with a quick check unconditionally on exit to user
     space turned out to be not feasible. On workloads which are doing a
     lot of quick syscalls the extra four instructions add up
     significantly.

     So instead I ended up doing it at the end of the exit to user TIF
     work loop once when all other TIF bits have been processed. At this
     point interrupts are disabled and there is no way that the state
     can change before the task goes out to user space for real.

  Versus the limitations of #1 and #3:

   I wasted several days of my so copious time to figure out how to not
   break all the architectures, which still insist on benefiting from core
   code improvements by pulling everything and the world into their
   architecture specific hackery.

   It's more than five years now that the generic entry code infrastructure
   has been introduced for the very reason to lower the burden for core
   code developers and maintainers and to share the common functionality
   across the architecture zoo.

   Aside of the initial x86 move, which started this effort, there are only
   three architectures who actually made the effort to utilize this. Two of
   them were new ones, which were asked to use it right away.

   The only existing one, which converted over since then is S390 and I'm
   truly grateful that they improved the generic infrastructure in that
   process significantly.

   On ARM[64] there are at least serious efforts underway to move their
   code over.

   Does everybody else think that core code improvements come for free and
   the architecture specific hackery does not put any burden on others?

   Here is the hall of fame as far as RSEQ goes:

   	arch/mips/Kconfig:      select HAVE_RSEQ
	arch/openrisc/Kconfig:  select HAVE_RSEQ
	arch/powerpc/Kconfig:   select HAVE_RSEQ

   Two of them are barely maintained and shouldn't have RSEQ in the first
   place....

   While I was very forthcoming in the past to accomodate for that and went
   out of my way to enable stuff for everyone, but I'm drawing a line now.

   All extra improvements which are enabled by #1/#3 depend hard on the
   generic infrastructure.

   I know that it's quite some effort to move an architecture over, but
   it's a one time effort and investment into the future. This 'my
   architecture is special for no reason' mindset is not sustainable and
   just pushes the burden on others. There is zero justification for this.

   Not converging on common infrastructure is not only a burden for the
   core people, it's also a missed opportunity for the architectures to
   lower their own burden of chasing core improvements and implementing
   them each with a different set of bugs.

   This is not the first time this happens. There are enough other examples
   where it took ages to consolidate on common code. This just accumulates
   technical debt and needless complexity, which everyone suffers from.

   I have happily converted the four architectures, which use the generic
   entry code over, to utilize a shared generic TIF bit header so that
   adding the TIF_RSEQ bit becomes a two line change and all four get the
   benefit immediately. That was more consequent than just adding the bits
   for each of them and it makes further maintainence of core
   infrastructure simpler for all sides. See?


That said, as for the first version these patches have a pile of dependencies:

The series depends on the separately posted rseq bugfix:

   https://lore.kernel.org/lkml/87o6sj6z95.ffs@tglx/

and the uaccess generic helper series:

   https://lore.kernel.org/lkml/20250813150610.521355442@linutronix.de/

and a related futex fix in

   git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git locking/urgent

The combination of all of them and some other related fixes (rseq
selftests) are available here:

    git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/base

For your convenience all of it is also available as a conglomerate from
git:

    git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/perf

The diffstat looks large, but a lot of that is due to extensive comments
and the extra hackery to accommodate for random architecture code.

I did not yet come around to test this on anything else than x86. Help with
that would be truly appreciated.

Thanks,

	tglx

  "Additional problems are the offspring of poor solutions." - Mark Twain
	
---
 Documentation/admin-guide/kernel-parameters.txt |    4 
 arch/Kconfig                                    |    4 
 arch/loongarch/Kconfig                          |    1 
 arch/loongarch/include/asm/thread_info.h        |   76 +-
 arch/riscv/Kconfig                              |    1 
 arch/riscv/include/asm/thread_info.h            |   29 -
 arch/s390/Kconfig                               |    1 
 arch/s390/include/asm/thread_info.h             |   44 -
 arch/x86/Kconfig                                |    1 
 arch/x86/entry/syscall_32.c                     |    3 
 arch/x86/include/asm/thread_info.h              |   74 +-
 b/include/asm-generic/thread_info_tif.h         |   51 ++
 b/include/linux/rseq_entry.h                    |  601 +++++++++++++++++++++++
 b/include/linux/rseq_types.h                    |   72 ++
 drivers/hv/mshv_root_main.c                     |    2 
 fs/binfmt_elf.c                                 |    2 
 fs/exec.c                                       |    2 
 include/linux/entry-common.h                    |   38 -
 include/linux/irq-entry-common.h                |   68 ++
 include/linux/mm.h                              |   25 
 include/linux/resume_user_mode.h                |    2 
 include/linux/rseq.h                            |  216 +++++---
 include/linux/sched.h                           |   50 +
 include/linux/thread_info.h                     |    5 
 include/trace/events/rseq.h                     |    4 
 include/uapi/linux/rseq.h                       |   21 
 init/Kconfig                                    |   28 +
 kernel/entry/common.c                           |   37 -
 kernel/entry/syscall-common.c                   |    8 
 kernel/rseq.c                                   |  610 ++++++++++--------------
 kernel/sched/core.c                             |   10 
 kernel/sched/membarrier.c                       |    8 
 kernel/sched/sched.h                            |    5 
 virt/kvm/kvm_main.c                             |    3 
 34 files changed, 1406 insertions(+), 700 deletions(-)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ