Documentation/admin-guide/kernel-parameters.txt | 4 arch/Kconfig | 4 arch/loongarch/Kconfig | 1 arch/loongarch/include/asm/thread_info.h | 76 +- arch/riscv/Kconfig | 1 arch/riscv/include/asm/thread_info.h | 29 - arch/s390/Kconfig | 1 arch/s390/include/asm/thread_info.h | 44 - arch/x86/Kconfig | 1 arch/x86/entry/syscall_32.c | 3 arch/x86/include/asm/thread_info.h | 74 +- b/include/asm-generic/thread_info_tif.h | 51 ++ b/include/linux/rseq_entry.h | 601 +++++++++++++++++++++++ b/include/linux/rseq_types.h | 72 ++ drivers/hv/mshv_root_main.c | 2 fs/binfmt_elf.c | 2 fs/exec.c | 2 include/linux/entry-common.h | 38 - include/linux/irq-entry-common.h | 68 ++ include/linux/mm.h | 25 include/linux/resume_user_mode.h | 2 include/linux/rseq.h | 216 +++++--- include/linux/sched.h | 50 + include/linux/thread_info.h | 5 include/trace/events/rseq.h | 4 include/uapi/linux/rseq.h | 21 init/Kconfig | 28 + kernel/entry/common.c | 37 - kernel/entry/syscall-common.c | 8 kernel/rseq.c | 610 ++++++++++-------------- kernel/sched/core.c | 10 kernel/sched/membarrier.c | 8 kernel/sched/sched.h | 5 virt/kvm/kvm_main.c | 3 34 files changed, 1406 insertions(+), 700 deletions(-)
OA This is a follow up on the initial series, which did a very basic attempt to sanitize the RSEQ handling in the kernel: https://lore.kernel.org/all/20250813155941.014821755@linutronix.de Further analysis turned up more than these initial problems: 1) task::rseq_event_mask is a pointless bit-field despite the fact that the ABI flags it was meant to support have been deprecated and functionally disabled three years ago. 2) task::rseq_event_mask is accumulating bits unless there is a critical section discovered in the user space rseq memory. This results in pointless invocations of the rseq user space exit handler even if there had nothing changed. As a matter of correctness these bits have to be clear when exiting to user space and therefore pristine when coming back into the kernel. Aside of correctness, this also avoids pointless evaluation of the user space memory, which is a performance benefit. 3) The evaluation of critical sections does not differentiate between syscall and interrupt/exception exits. The current implementation silently tolerates and fixes up critical sections which invoked a syscall unless CONFIG_DEBUG_RSEQ is enabled. That's just wrong. If user space does that on a production kernel it can keep the pieces. The kernel is not there to proliferate mindless user space programming and letting everyone pay the performance penalty. Additional findings: 4) The decision to raise the work for exit is more than suboptimal. Basically every context switch does so if the task has rseq, which is nowadays likely as glibc makes use of it if available. The consequence is that a lot of exits have to process RSEQ just for nothing. The only reasons to do so are: the task was interrupted in user space and schedules or the CPU or MM CID changes in schedule() independent of the entry mode That reduces the invocation space obviously significantly. 5) Signal handling does the RSEQ update unconditionally. That's wrong as the only reason to do so is when the task was interrupted in user space independent of a schedule event. The only important task in that case is to handle the critical section because after switching to the signal frame the original return IP is not longer available. The CPU/MM CID values do not need to be updated at that point as they can change again before the signal delivery goes out to user space. Again, if the task was in a critical section and issued a syscall then it can keep the pieces as that's a violation of the ABI contract. 6) CPU and MM CID are updated unconditionally That's again a pointless exercise when they didn't change. Then the only action required is to check the critical section if and only if the entry came via an interrupt. That can obviously be avoided by caching the values written to user space and avoiding that path if they haven't changed 7) The TIF_NOTIFY_RESUME mechanism is a horrorshow TIF_NOTIFY_RESUME is a multiplexing TIF bit and needs to invoke the world and some more. Depending on workloads this can be set by task_work, security, block and memory management. All unrelated to RSEQ and quite some of them are likely to cause a reschedule. But most of them are low frequency. So doing this work in the loop unconditionally is just waste. The correct point to do this is at the end of that loop once all other bits have been processed, because that's the point where the task is actually going out to user space. 8) #7 caused another subtle work for nothing issue IO/URING and hypervisors invoke resume_user_mode_work() with a NULL pointer for pt_regs, which causes the RSEQ code to ignore the critical section check, but updating the CPU ID/ MM CID values unconditionally. For IO/URING this invocation is irrelevant because the IO workers can never go out to user space and therefore do not have RSEQ memory in the first place. So it's a non problem in the existing code as task::rseq is NULL in that case. Hypervisors are a different story. They need to drain task_work and other pending items, which are multiplexed by TIF_NOTIFY_RESUME, before entering guest mode. The invocation of resume_user_mode_work() clears TIF_NOTIFY_RESUME, which means if rseq would ignore that case then it could miss a CPU or MM CID update on the way back to user space. The handling of that is just a horrible and mindless hack as the event might be re-raised between the time the ioctl() enters guest mode and the actual exit to user space. So the obvious thing is to ignore the regs=NULL call and let the offending hypervisor calls check when returning from the ioctl() whether the event bit is set and re-raise the notification again. 9) Code efficiency RSEQ aims to improve performance for user space, but it completely ignores the fact, that this needs to be implemented in a way which does not impact the performance of the kernel significantly. So far this did not pop up as just a few people used it, but that has changed because glibc started to use it widely. It's not so efficiOS^H^Hent as advertised: For a full kernel rebuild: exit to user: 50106428 signal checks: 24703 slowpath runs: 943785 1.88% id updates: 968488 1.93% cs checks: 888768 1.77% cs cleared: 888768 100.00% cs fixup: 0 0.00% The cs_cleared/fixup numbers are relative to cs_checks, which means this does pointless clears even if it was clear already because that's what is the case with glibc. glibc is only interested in the CPU/MM CID values so far. And no, it's not only a store, it's the whole dance of spectre-v1 mitigation plus user access enable/disable. I really have to ask why all those people who are deeply caring about performance and have utilized this stuff for quite some time have not noticed what a performance sh*tshow this inflicts on the kernel. Seriously? Aside of that did anyone ever look at the resulting assembly code? While the user space counterpart is carefully crafted in hand written assembly for efficiency the kernel implementation is a complete and utter nightmare. C allows to implement very efficient code, but the fact that the compiler requires help from the programmer to turn it into actual efficient assembly is not a totally new finding. The absence of properly designed data structures, which allow to build efficient and comprehensible code, is amazing. Slapping random storage members into struct task_struct, implementing a conditional maze around them and then hoping that the compiler will create sensible code out of it, is wishful thinking at best. I completely understand that all the sanity checks and loops and hoops were required to get this actually working. But I'm seriously grumpy about the fact, that once the PoC code was shoved into the kernel, the 'performance work' was completed and just the next fancy 'performance' features were piled on top of it. Once this is addressed the above picture changes completely: exit to user: 50071399 signal checks: 182 slowpath runs: 8149 0.02% fastpath runs: 763366 1.52% id updates: 78419 0.16% cs checks: 0 0.00% cs cleared: 0 cs fixup: 0 And according to perf that very same kernel build consistently gets faster from: 92.500894648 seconds time elapsed (upstream) to 91.693778124 seconds time elapsed Not a lot but not in the noise either, i.e. >= 1% For a 1e9 gettid() loop this results in going from: 49.782703878 seconds time elapsed (upstream) to 49.327360851 seconds time elapsed Not a lot either, but consistently >=1% as the above. For actual rseq critical section usage this makes even more of a difference. Aside of avoiding pointless work in the kernel this does also not abort critical sections when not necessary, which improves user space performance too. The kernel selftests magically complete six seconds faster, which is again not a lot compared to a total run time of ~540s. But those tests are not really reflecting real work loads, they emphasize on stress testing the kernel implementation, which is perfectly fine. But still, the reduced amount of pointless kernel work is what makes the difference: Upstream: exit to user: 736568321 signal checks: 62002182 slowpath runs: 358362121 57.07% fastpath runs: 0 0.00% id updates: 358362129 57.07% cs checks: 358362110 57.07% cs cleared: 347611246 97.01% cs fixup: 10750863 2.99% Upstream + simple obvious fixes: exit to user: 736462043 signal checks: 61902154 slowpath runs: 223394362 30.33% fastpath runs: 0 0.00% id updates: 285296516 38.74% cs checks: 284430371 38.62% cs cleared: 277110571 97.43% cs fixup: 7319800 2.57% Fully patched: exit to user: 736477630 signal checks: 0 slowpath runs: 1367 0.00% fastpath runs: 114230240 15.51% id updates: 106498492 14.46% cs checks: 7665511 1.04% cs cleared: 2092386 27.30% cs fixup: 5573118 72.70% Perf confirms: 46,242,994,988,432 cycles 8,390,996,405,969 instructions 1949.923762000 seconds user 16111.947776000 seconds sys versus: 44,910,362,743,328 cycles -3 % 8,305,138,767,914 instructions -1 % 1985.342453000 seconds user +2 % 15653.639218000 seconds sys -3 % Running the same with only looking at the kernel counts: 1) Upstream: 39,530,448,181,974 cycles:khH 2,872,985,410,904 instructions:khH 640,968,094,639 branches:khH 2) Upstream + simple obvious fixes: -> #1 39,005,607,758,104 cycles:khH -1.5 % 2,853,676,629,923 instructions:khH -0.8 % 635,538,256,875 branches:khH -1.0 % 3) Fully patched: -> #2 -> #1 38,786,867,052,870 cycles:khH -0.6 % -2.3 % 2,784,773,121,491 instructions:khH -2.5 % -3.1 % 618,677,047,160 branches:khH -2.7 % -3.6 % Looking at the kernel build that way: 1) Upstream: 474,191,665,063 cycles:khH 164,176,072,767 instructions:khH 33,037,122,843 branches:khH 2) Upstream + simple obvious fixes: -> #1 473,423,488,450 cycles:khH ~0.0 % 162,749,873,914 instructions:khH -1.0 % 32,264,397,605 branches:khH -2.5 % 3) Fully patched: -> #2 -> #1 468,658,348,098 cycles:khH -1.1 % -1.2 % 160,941,028,283 instructions:khH -1.2 % -2.0 % 31,893,296,267 branches:khH -1.5 % -3.7 % That's pretty much in line with the 2-3% regressions observed by Jens. TBH, I absolutely do not care about the way how performance is evaluated in the entities who pushed for this, but I very much care about the general impact of this. It seems to me the performance evaluation by the $corp power users of RSEQ stopped at the point where a significant improvement in their user space was observed. It's interesting, that the resulting overhead in the kernel seems not that relevant for the overall performance evaluation. That's left to others to mop up. The tragedy of the commons in action. TBH, what surprised me most when looking into this in detail, was the large amount of low hanging fruit, which was sitting there in plain sight. That said, this series addresses the overall problems by: 1) Limiting the RSEQ work to the actual conditions where it is required. The full benefit is only available for architectures using the generic entry infrastructure. All others get at least the basic improvements. 2) Re-implementing the whole user space handling based on proper data structures and by actually looking at the impact it creates in the fast path. 3) Moving the actual handling of RSEQ out to the latest point in the exit path, where possible. This is fully inlined into the fast path to keep the impact confined. The initial attempt to make it completely independent of TIF bits and just handle it with a quick check unconditionally on exit to user space turned out to be not feasible. On workloads which are doing a lot of quick syscalls the extra four instructions add up significantly. So instead I ended up doing it at the end of the exit to user TIF work loop once when all other TIF bits have been processed. At this point interrupts are disabled and there is no way that the state can change before the task goes out to user space for real. Versus the limitations of #1 and #3: I wasted several days of my so copious time to figure out how to not break all the architectures, which still insist on benefiting from core code improvements by pulling everything and the world into their architecture specific hackery. It's more than five years now that the generic entry code infrastructure has been introduced for the very reason to lower the burden for core code developers and maintainers and to share the common functionality across the architecture zoo. Aside of the initial x86 move, which started this effort, there are only three architectures who actually made the effort to utilize this. Two of them were new ones, which were asked to use it right away. The only existing one, which converted over since then is S390 and I'm truly grateful that they improved the generic infrastructure in that process significantly. On ARM[64] there are at least serious efforts underway to move their code over. Does everybody else think that core code improvements come for free and the architecture specific hackery does not put any burden on others? Here is the hall of fame as far as RSEQ goes: arch/mips/Kconfig: select HAVE_RSEQ arch/openrisc/Kconfig: select HAVE_RSEQ arch/powerpc/Kconfig: select HAVE_RSEQ Two of them are barely maintained and shouldn't have RSEQ in the first place.... While I was very forthcoming in the past to accomodate for that and went out of my way to enable stuff for everyone, but I'm drawing a line now. All extra improvements which are enabled by #1/#3 depend hard on the generic infrastructure. I know that it's quite some effort to move an architecture over, but it's a one time effort and investment into the future. This 'my architecture is special for no reason' mindset is not sustainable and just pushes the burden on others. There is zero justification for this. Not converging on common infrastructure is not only a burden for the core people, it's also a missed opportunity for the architectures to lower their own burden of chasing core improvements and implementing them each with a different set of bugs. This is not the first time this happens. There are enough other examples where it took ages to consolidate on common code. This just accumulates technical debt and needless complexity, which everyone suffers from. I have happily converted the four architectures, which use the generic entry code over, to utilize a shared generic TIF bit header so that adding the TIF_RSEQ bit becomes a two line change and all four get the benefit immediately. That was more consequent than just adding the bits for each of them and it makes further maintainence of core infrastructure simpler for all sides. See? That said, as for the first version these patches have a pile of dependencies: The series depends on the separately posted rseq bugfix: https://lore.kernel.org/lkml/87o6sj6z95.ffs@tglx/ and the uaccess generic helper series: https://lore.kernel.org/lkml/20250813150610.521355442@linutronix.de/ and a related futex fix in git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git locking/urgent The combination of all of them and some other related fixes (rseq selftests) are available here: git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/base For your convenience all of it is also available as a conglomerate from git: git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/perf The diffstat looks large, but a lot of that is due to extensive comments and the extra hackery to accommodate for random architecture code. I did not yet come around to test this on anything else than x86. Help with that would be truly appreciated. Thanks, tglx "Additional problems are the offspring of poor solutions." - Mark Twain --- Documentation/admin-guide/kernel-parameters.txt | 4 arch/Kconfig | 4 arch/loongarch/Kconfig | 1 arch/loongarch/include/asm/thread_info.h | 76 +- arch/riscv/Kconfig | 1 arch/riscv/include/asm/thread_info.h | 29 - arch/s390/Kconfig | 1 arch/s390/include/asm/thread_info.h | 44 - arch/x86/Kconfig | 1 arch/x86/entry/syscall_32.c | 3 arch/x86/include/asm/thread_info.h | 74 +- b/include/asm-generic/thread_info_tif.h | 51 ++ b/include/linux/rseq_entry.h | 601 +++++++++++++++++++++++ b/include/linux/rseq_types.h | 72 ++ drivers/hv/mshv_root_main.c | 2 fs/binfmt_elf.c | 2 fs/exec.c | 2 include/linux/entry-common.h | 38 - include/linux/irq-entry-common.h | 68 ++ include/linux/mm.h | 25 include/linux/resume_user_mode.h | 2 include/linux/rseq.h | 216 +++++--- include/linux/sched.h | 50 + include/linux/thread_info.h | 5 include/trace/events/rseq.h | 4 include/uapi/linux/rseq.h | 21 init/Kconfig | 28 + kernel/entry/common.c | 37 - kernel/entry/syscall-common.c | 8 kernel/rseq.c | 610 ++++++++++-------------- kernel/sched/core.c | 10 kernel/sched/membarrier.c | 8 kernel/sched/sched.h | 5 virt/kvm/kvm_main.c | 3 34 files changed, 1406 insertions(+), 700 deletions(-)
On 2025-08-23 12:39, Thomas Gleixner wrote: > OA > This is a follow up on the initial series, which did a very basic attempt > to sanitize the RSEQ handling in the kernel: > > https://lore.kernel.org/all/20250813155941.014821755@linutronix.de > > Further analysis turned up more than these initial problems: Thanks Thomas for looking into this. Sorry for the delayed reply, I was on vacation. > > 1) task::rseq_event_mask is a pointless bit-field despite the fact that > the ABI flags it was meant to support have been deprecated and > functionally disabled three years ago. Yes, this should be converted to a simple boolean now. > > 2) task::rseq_event_mask is accumulating bits unless there is a critical > section discovered in the user space rseq memory. This results in > pointless invocations of the rseq user space exit handler even if > there had nothing changed. As a matter of correctness these bits have > to be clear when exiting to user space and therefore pristine when > coming back into the kernel. Aside of correctness, this also avoids > pointless evaluation of the user space memory, which is a performance > benefit. Thanks for catching this, that's indeed not the intended behavior. > > 3) The evaluation of critical sections does not differentiate between > syscall and interrupt/exception exits. The current implementation > silently tolerates and fixes up critical sections which invoked a > syscall unless CONFIG_DEBUG_RSEQ is enabled. > > That's just wrong. If user space does that on a production kernel it > can keep the pieces. The kernel is not there to proliferate mindless > user space programming and letting everyone pay the performance > penalty. Agreed. There is no point in supporting a userspace behavior on production kernels that is prevented on debug kernels, especially if this adds overhead to production kernels. > > Additional findings: > > 4) The decision to raise the work for exit is more than suboptimal. Terminology-wise, just making sure we are on the same page: here you are talking about "exit to usermode", and *not* the exit(2) system call. I'm clarifying because I know mm people care about fork/clone/exit syscall overhead as well. > Basically every context switch does so if the task has rseq, which is > nowadays likely as glibc makes use of it if available. Correct. > > The consequence is that a lot of exits have to process RSEQ just for > nothing. The only reasons to do so are: > > the task was interrupted in user space and schedules interrupted or takes a trap/exception (I will assume you consider traps and exceptions as an interrupt classes within this discussion). > > or > > the CPU or MM CID changes in schedule() independent of the entry > mode or the numa node id, which is typically tied to the CPU number, except on powerpc AFAIR where numa node id to cpu mapping can be reconfigured dynamically. > > That reduces the invocation space obviously significantly. Yes. > > 5) Signal handling does the RSEQ update unconditionally. > > That's wrong as the only reason to do so is when the task was > interrupted in user space independent of a schedule event. > > The only important task in that case is to handle the critical section > because after switching to the signal frame the original return IP is > not longer available. > > The CPU/MM CID values do not need to be updated at that point as they > can change again before the signal delivery goes out to user space. > > Again, if the task was in a critical section and issued a syscall then > it can keep the pieces as that's a violation of the ABI contract. The key thing here is that the state of the cpu/mm cid/numa node id fields are updated before returning to the userspace signal handler, and that a rseq critical section within a signal handler works. From your description here we can indeed do less work on signal delivery and still meet those requirements. That's good. > > 6) CPU and MM CID are updated unconditionally > > That's again a pointless exercise when they didn't change. Then the > only action required is to check the critical section if and only if > the entry came via an interrupt. > > That can obviously be avoided by caching the values written to user > space and avoiding that path if they haven't changed This is a good performance improvement, I agree. Note that this will likely break Google's tcmalloc hack of re-using the cpu_id_start field as a way to get notified about preemption. But they were warned not to do that, and it breaks the documented userspace ABI contract, so they will need to adapt. This situation is documented here: commit 7d5265ffcd8b rseq: Validate read-only fields under DEBUG_RSEQ config > > 7) The TIF_NOTIFY_RESUME mechanism is a horrorshow > > TIF_NOTIFY_RESUME is a multiplexing TIF bit and needs to invoke the > world and some more. Depending on workloads this can be set by > task_work, security, block and memory management. All unrelated to > RSEQ and quite some of them are likely to cause a reschedule. > But most of them are low frequency. > > So doing this work in the loop unconditionally is just waste. The > correct point to do this is at the end of that loop once all other bits > have been processed, because that's the point where the task is > actually going out to user space. Note that the rseq work can trigger a page fault and a reschedule, which makes me whether moving this work out of the work loop can be OK ? I'll need to have a closer look at the relevant patches to understand this better. I suspect that the underlying problem here is not that the notify resume adds too much overhead, but rather that we're setting the TIF_NOTIFY_RESUME bit way more often than is good for us. Initially I focused on minimizing the scheduler overhead, at the expense of doing more work than strictly needed on exit to usermode. But of course reality is not as simple, and we need to fine the right balance. If your goal is to have fewer calls to the resume notifier triggered by rseq, we could perhaps change the way rseq_set_notify_resume() is implemented to make it conditional on either: - migration, - preemption AND current->rseq->rseq_cs userspace value is set OR would require a page fault to be read. - preemption AND mm_cid changes. I'm not sure what to do about the powerpc numa node id reconfiguration fringe use-case though. > > 8) #7 caused another subtle work for nothing issue > > IO/URING and hypervisors invoke resume_user_mode_work() with a NULL > pointer for pt_regs, which causes the RSEQ code to ignore the critical > section check, but updating the CPU ID/ MM CID values unconditionally. > > For IO/URING this invocation is irrelevant because the IO workers can > never go out to user space and therefore do not have RSEQ memory in > the first place. So it's a non problem in the existing code as > task::rseq is NULL in that case. OK > > Hypervisors are a different story. They need to drain task_work and > other pending items, which are multiplexed by TIF_NOTIFY_RESUME, > before entering guest mode. > > The invocation of resume_user_mode_work() clears TIF_NOTIFY_RESUME, > which means if rseq would ignore that case then it could miss a CPU > or MM CID update on the way back to user space. > > The handling of that is just a horrible and mindless hack as the event > might be re-raised between the time the ioctl() enters guest mode and > the actual exit to user space. > > So the obvious thing is to ignore the regs=NULL call and let the > offending hypervisor calls check when returning from the ioctl() > whether the event bit is set and re-raise the notification again. Is this a useless work issue or a correctness issue ? Also, AFAIU, your proposed approach will ensure that the rseq fields are up to date when returning to userspace from the ioctl in the host process, but there are no guarantees about having up to date fields when running the guest VM. I'm fine with this, but I think it needs to be clearly spelled out. > > 9) Code efficiency > > RSEQ aims to improve performance for user space, but it completely > ignores the fact, that this needs to be implemented in a way which > does not impact the performance of the kernel significantly. > > So far this did not pop up as just a few people used it, but that has > changed because glibc started to use it widely. The history of incremental rseq upstreaming into the kernel and then glibc can help understand how this situation came to be: Linux commit d7822b1e24f2 ("rseq: Introduce restartable sequences system call") (2018) This includes microbenchmarks of specific use-cases, which clearly do not stress neither the scheduler nor exit to usermode. It includes jemalloc memory allocator (Facebook) latency benchmarks, which depend heavily on the userspace fast-paths. It includes hackbench results, cover the scheduling and return to usermode overhead. Given that there was no libc integration back then, the hackbench results do not include any rseq-triggered resume notifier work. Based on the rseq use at that point in time, the overhead of rseq was acceptable for it to be merged. glibc commit 95e114a0919d ("nptl: Add rseq registration") (2021) Florian Weimer added this feature to glibc. At this point the overhead you observe now should have become visible. This integration and performance testing was done by the glibc maintainers and distribution vendors who packaged this libc. [...] > That said, this series addresses the overall problems by: > > 1) Limiting the RSEQ work to the actual conditions where it is > required. The full benefit is only available for architectures using > the generic entry infrastructure. All others get at least the basic > improvements. > > 2) Re-implementing the whole user space handling based on proper data > structures and by actually looking at the impact it creates in the > fast path. > > 3) Moving the actual handling of RSEQ out to the latest point in the exit > path, where possible. This is fully inlined into the fast path to keep > the impact confined. > > The initial attempt to make it completely independent of TIF bits and > just handle it with a quick check unconditionally on exit to user > space turned out to be not feasible. On workloads which are doing a > lot of quick syscalls the extra four instructions add up > significantly. > > So instead I ended up doing it at the end of the exit to user TIF > work loop once when all other TIF bits have been processed. At this > point interrupts are disabled and there is no way that the state > can change before the task goes out to user space for real. I'll need to review what happens in case rseq needs to take a page fault after your proposed changes. More discussion to come around the specific patch in the series. > > Versus the limitations of #1 and #3: > > I wasted several days of my so copious time to figure out how to not > break all the architectures, which still insist on benefiting from core > code improvements by pulling everything and the world into their > architecture specific hackery. > > It's more than five years now that the generic entry code infrastructure > has been introduced for the very reason to lower the burden for core > code developers and maintainers and to share the common functionality > across the architecture zoo. > > Aside of the initial x86 move, which started this effort, there are only > three architectures who actually made the effort to utilize this. Two of > them were new ones, which were asked to use it right away. > > The only existing one, which converted over since then is S390 and I'm > truly grateful that they improved the generic infrastructure in that > process significantly. > > On ARM[64] there are at least serious efforts underway to move their > code over. > > Does everybody else think that core code improvements come for free and > the architecture specific hackery does not put any burden on others? > > Here is the hall of fame as far as RSEQ goes: > > arch/mips/Kconfig: select HAVE_RSEQ > arch/openrisc/Kconfig: select HAVE_RSEQ > arch/powerpc/Kconfig: select HAVE_RSEQ > > Two of them are barely maintained and shouldn't have RSEQ in the first > place.... > > While I was very forthcoming in the past to accomodate for that and went > out of my way to enable stuff for everyone, but I'm drawing a line now. > > All extra improvements which are enabled by #1/#3 depend hard on the > generic infrastructure. > > I know that it's quite some effort to move an architecture over, but > it's a one time effort and investment into the future. This 'my > architecture is special for no reason' mindset is not sustainable and > just pushes the burden on others. There is zero justification for this. > > Not converging on common infrastructure is not only a burden for the > core people, it's also a missed opportunity for the architectures to > lower their own burden of chasing core improvements and implementing > them each with a different set of bugs. > > This is not the first time this happens. There are enough other examples > where it took ages to consolidate on common code. This just accumulates > technical debt and needless complexity, which everyone suffers from. > > I have happily converted the four architectures, which use the generic > entry code over, to utilize a shared generic TIF bit header so that > adding the TIF_RSEQ bit becomes a two line change and all four get the > benefit immediately. That was more consequent than just adding the bits > for each of them and it makes further maintainence of core > infrastructure simpler for all sides. See? Can we make RSEQ depend on CONFIG_GENERIC_ENTRY (if that is the correct option) ? This would force additional architectures to move to the generic entry infrastructure if they want to benefit from rseq. I'll start looking into the series now. Thanks, Mathieu > > > That said, as for the first version these patches have a pile of dependencies: > > The series depends on the separately posted rseq bugfix: > > https://lore.kernel.org/lkml/87o6sj6z95.ffs@tglx/ > > and the uaccess generic helper series: > > https://lore.kernel.org/lkml/20250813150610.521355442@linutronix.de/ > > and a related futex fix in > > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git locking/urgent > > The combination of all of them and some other related fixes (rseq > selftests) are available here: > > git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/base > > For your convenience all of it is also available as a conglomerate from > git: > > git://git.kernel.org/pub/scm/linux/kernel/git/tglx/devel.git rseq/perf > > The diffstat looks large, but a lot of that is due to extensive comments > and the extra hackery to accommodate for random architecture code. > > I did not yet come around to test this on anything else than x86. Help with > that would be truly appreciated. > > Thanks, > > tglx > > "Additional problems are the offspring of poor solutions." - Mark Twain > > --- > Documentation/admin-guide/kernel-parameters.txt | 4 > arch/Kconfig | 4 > arch/loongarch/Kconfig | 1 > arch/loongarch/include/asm/thread_info.h | 76 +- > arch/riscv/Kconfig | 1 > arch/riscv/include/asm/thread_info.h | 29 - > arch/s390/Kconfig | 1 > arch/s390/include/asm/thread_info.h | 44 - > arch/x86/Kconfig | 1 > arch/x86/entry/syscall_32.c | 3 > arch/x86/include/asm/thread_info.h | 74 +- > b/include/asm-generic/thread_info_tif.h | 51 ++ > b/include/linux/rseq_entry.h | 601 +++++++++++++++++++++++ > b/include/linux/rseq_types.h | 72 ++ > drivers/hv/mshv_root_main.c | 2 > fs/binfmt_elf.c | 2 > fs/exec.c | 2 > include/linux/entry-common.h | 38 - > include/linux/irq-entry-common.h | 68 ++ > include/linux/mm.h | 25 > include/linux/resume_user_mode.h | 2 > include/linux/rseq.h | 216 +++++--- > include/linux/sched.h | 50 + > include/linux/thread_info.h | 5 > include/trace/events/rseq.h | 4 > include/uapi/linux/rseq.h | 21 > init/Kconfig | 28 + > kernel/entry/common.c | 37 - > kernel/entry/syscall-common.c | 8 > kernel/rseq.c | 610 ++++++++++-------------- > kernel/sched/core.c | 10 > kernel/sched/membarrier.c | 8 > kernel/sched/sched.h | 5 > virt/kvm/kvm_main.c | 3 > 34 files changed, 1406 insertions(+), 700 deletions(-) -- Mathieu Desnoyers EfficiOS Inc. https://www.efficios.com
© 2016 - 2025 Red Hat, Inc.