Currently on systems where the timer interrupt (or any other
fast-at-human-scale periodic interrupt) is active then it is impossible
to step any code with interrupts unlocked because we will always end up
stepping into the timer interrupt instead of stepping the user code.
The common user's goal while single stepping is that when they step then
the system will stop at PC+4 or PC+I for a branch that gets taken
relative to the instruction they are stepping. So, fix broken single step
implementation via skipping single stepping into interrupt handlers.
The methodology is when we receive an interrupt from EL1, check if we
are single stepping (pstate.SS). If yes then we save MDSCR_EL1.SS and
clear the register bit if it was set. Then unmask only D and leave I set.
On return from the interrupt, set D and restore MDSCR_EL1.SS. Along with
this skip reschedule if we were stepping.
Suggested-by: Will Deacon <will@kernel.org>
Signed-off-by: Sumit Garg <sumit.garg@linaro.org>
Tested-by: Douglas Anderson <dianders@chromium.org>
Acked-by: Daniel Thompson <daniel.thompson@linaro.org>
Tested-by: Daniel Thompson <daniel.thompson@linaro.org>
---
arch/arm64/kernel/entry-common.c | 22 ++++++++++++++++++++--
1 file changed, 20 insertions(+), 2 deletions(-)
diff --git a/arch/arm64/kernel/entry-common.c b/arch/arm64/kernel/entry-common.c
index cce1167199e3..568481f66977 100644
--- a/arch/arm64/kernel/entry-common.c
+++ b/arch/arm64/kernel/entry-common.c
@@ -231,11 +231,15 @@ DEFINE_STATIC_KEY_TRUE(sk_dynamic_irqentry_exit_cond_resched);
#define need_irq_preemption() (IS_ENABLED(CONFIG_PREEMPTION))
#endif
-static void __sched arm64_preempt_schedule_irq(void)
+static void __sched arm64_preempt_schedule_irq(struct pt_regs *regs)
{
if (!need_irq_preemption())
return;
+ /* Don't reschedule in case we are single stepping */
+ if (regs->pstate & DBG_SPSR_SS)
+ return;
+
/*
* Note: thread_info::preempt_count includes both thread_info::count
* and thread_info::need_resched, and is not equivalent to
@@ -471,19 +475,33 @@ static __always_inline void __el1_irq(struct pt_regs *regs,
do_interrupt_handler(regs, handler);
irq_exit_rcu();
- arm64_preempt_schedule_irq();
+ arm64_preempt_schedule_irq(regs);
exit_to_kernel_mode(regs);
}
+
static void noinstr el1_interrupt(struct pt_regs *regs,
void (*handler)(struct pt_regs *))
{
+ unsigned long mdscr;
+
+ /* Disable single stepping within interrupt handler */
+ if (regs->pstate & DBG_SPSR_SS) {
+ mdscr = read_sysreg(mdscr_el1);
+ write_sysreg(mdscr & ~DBG_MDSCR_SS, mdscr_el1);
+ }
+
write_sysreg(DAIF_PROCCTX_NOIRQ, daif);
if (IS_ENABLED(CONFIG_ARM64_PSEUDO_NMI) && !interrupts_enabled(regs))
__el1_pnmi(regs, handler);
else
__el1_irq(regs, handler);
+
+ if (regs->pstate & DBG_SPSR_SS) {
+ write_sysreg(DAIF_PROCCTX_NOIRQ | PSR_D_BIT, daif);
+ write_sysreg(mdscr, mdscr_el1);
+ }
}
asmlinkage void noinstr el1h_64_irq_handler(struct pt_regs *regs)
--
2.34.1
On Thu, 2023-02-02 at 13:01 +0530, Sumit Garg wrote:
> Currently on systems where the timer interrupt (or any other
> fast-at-human-scale periodic interrupt) is active then it is impossible
> to step any code with interrupts unlocked because we will always end up
> stepping into the timer interrupt instead of stepping the user code.
>
> The common user's goal while single stepping is that when they step then
> the system will stop at PC+4 or PC+I for a branch that gets taken
> relative to the instruction they are stepping. So, fix broken single step
> implementation via skipping single stepping into interrupt handlers.
>
> The methodology is when we receive an interrupt from EL1, check if we
> are single stepping (pstate.SS). If yes then we save MDSCR_EL1.SS and
> clear the register bit if it was set. Then unmask only D and leave I set.
> On return from the interrupt, set D and restore MDSCR_EL1.SS. Along with
> this skip reschedule if we were stepping.
>
> Suggested-by: Will Deacon <will@kernel.org>
> Signed-off-by: Sumit Garg <sumit.garg@linaro.org>
> Tested-by: Douglas Anderson <dianders@chromium.org>
> Acked-by: Daniel Thompson <daniel.thompson@linaro.org>
> Tested-by: Daniel Thompson <daniel.thompson@linaro.org>
Hi Sumit,
I was wondering what the status of this patch is. [1] sounds like there were
remaining concerns, but I'm not sure about the details.
The current state is very unsatisfactory, as I need to apply this patch every
time I want to use KGDB on arm64. It is not only required for single-stepping,
but it also fixes 'continue' endlessly hitting the same breakpoint (until the
breakpoint is deleted or disabled).
Best,
Matthias
[1] https://lkml.org/lkml/2023/2/22/49
> ---
> arch/arm64/kernel/entry-common.c | 22 ++++++++++++++++++++--
> 1 file changed, 20 insertions(+), 2 deletions(-)
>
> diff --git a/arch/arm64/kernel/entry-common.c b/arch/arm64/kernel/entry-common.c
> index cce1167199e3..568481f66977 100644
> --- a/arch/arm64/kernel/entry-common.c
> +++ b/arch/arm64/kernel/entry-common.c
> @@ -231,11 +231,15 @@ DEFINE_STATIC_KEY_TRUE(sk_dynamic_irqentry_exit_cond_resched);
> #define need_irq_preemption() (IS_ENABLED(CONFIG_PREEMPTION))
> #endif
>
> -static void __sched arm64_preempt_schedule_irq(void)
> +static void __sched arm64_preempt_schedule_irq(struct pt_regs *regs)
> {
> if (!need_irq_preemption())
> return;
>
> + /* Don't reschedule in case we are single stepping */
> + if (regs->pstate & DBG_SPSR_SS)
> + return;
> +
> /*
> * Note: thread_info::preempt_count includes both thread_info::count
> * and thread_info::need_resched, and is not equivalent to
> @@ -471,19 +475,33 @@ static __always_inline void __el1_irq(struct pt_regs *regs,
> do_interrupt_handler(regs, handler);
> irq_exit_rcu();
>
> - arm64_preempt_schedule_irq();
> + arm64_preempt_schedule_irq(regs);
>
> exit_to_kernel_mode(regs);
> }
> +
> static void noinstr el1_interrupt(struct pt_regs *regs,
> void (*handler)(struct pt_regs *))
> {
> + unsigned long mdscr;
> +
> + /* Disable single stepping within interrupt handler */
> + if (regs->pstate & DBG_SPSR_SS) {
> + mdscr = read_sysreg(mdscr_el1);
> + write_sysreg(mdscr & ~DBG_MDSCR_SS, mdscr_el1);
> + }
> +
> write_sysreg(DAIF_PROCCTX_NOIRQ, daif);
>
> if (IS_ENABLED(CONFIG_ARM64_PSEUDO_NMI) && !interrupts_enabled(regs))
> __el1_pnmi(regs, handler);
> else
> __el1_irq(regs, handler);
> +
> + if (regs->pstate & DBG_SPSR_SS) {
> + write_sysreg(DAIF_PROCCTX_NOIRQ | PSR_D_BIT, daif);
> + write_sysreg(mdscr, mdscr_el1);
> + }
> }
>
> asmlinkage void noinstr el1h_64_irq_handler(struct pt_regs *regs)
--
TQ-Systems GmbH | Mühlstraße 2, Gut Delling | 82229 Seefeld, Germany
Amtsgericht München, HRB 105018
Geschäftsführer: Detlef Schneider, Rüdiger Stahl, Stefan Schneider
https://www.tq-group.com/
© 2016 - 2026 Red Hat, Inc.