From: Andy Chiu <andy.chiu@sifive.com>
Runtime code patching must be done at a naturally aligned address, or we
may execute on a partial instruction.
We have encountered problems traced back to static jump functions during
the test. We switched the tracer randomly for every 1~5 seconds on a
dual-core QEMU setup and found the kernel sucking at a static branch
where it jumps to itself.
The reason is that the static branch was 2-byte but not 4-byte aligned.
Then, the kernel would patch the instruction, either J or NOP, with two
half-word stores if the machine does not have efficient unaligned
accesses. Thus, moments exist where half of the NOP mixes with the other
half of the J when transitioning the branch. In our particular case, on
a little-endian machine, the upper half of the NOP was mixed with the
lower part of the J when enabling the branch, resulting in a jump that
jumped to itself. Conversely, it would result in a HINT instruction when
disabling the branch, but it might not be observable.
ARM64 does not have this problem since all instructions must be 4-byte
aligned.
Fixes: ebc00dde8a97 ("riscv: Add jump-label implementation")
Link: https://lore.kernel.org/linux-riscv/20220913094252.3555240-6-andy.chiu@sifive.com/
Signed-off-by: Andy Chiu <andy.chiu@sifive.com>
Reviewed-by: Greentime Hu <greentime.hu@sifive.com>
Signed-off-by: Guo Ren <guoren@kernel.org>
---
arch/riscv/include/asm/jump_label.h | 2 ++
1 file changed, 2 insertions(+)
diff --git a/arch/riscv/include/asm/jump_label.h b/arch/riscv/include/asm/jump_label.h
index 6d58bbb5da46..14a5ea8d8ef0 100644
--- a/arch/riscv/include/asm/jump_label.h
+++ b/arch/riscv/include/asm/jump_label.h
@@ -18,6 +18,7 @@ static __always_inline bool arch_static_branch(struct static_key * const key,
const bool branch)
{
asm_volatile_goto(
+ " .align 2 \n\t"
" .option push \n\t"
" .option norelax \n\t"
" .option norvc \n\t"
@@ -39,6 +40,7 @@ static __always_inline bool arch_static_branch_jump(struct static_key * const ke
const bool branch)
{
asm_volatile_goto(
+ " .align 2 \n\t"
" .option push \n\t"
" .option norelax \n\t"
" .option norvc \n\t"
--
2.36.1
guoren@kernel.org writes: > From: Andy Chiu <andy.chiu@sifive.com> > > Runtime code patching must be done at a naturally aligned address, or we > may execute on a partial instruction. > > We have encountered problems traced back to static jump functions during > the test. We switched the tracer randomly for every 1~5 seconds on a > dual-core QEMU setup and found the kernel sucking at a static branch > where it jumps to itself. > > The reason is that the static branch was 2-byte but not 4-byte aligned. > Then, the kernel would patch the instruction, either J or NOP, with two > half-word stores if the machine does not have efficient unaligned > accesses. Thus, moments exist where half of the NOP mixes with the other > half of the J when transitioning the branch. In our particular case, on > a little-endian machine, the upper half of the NOP was mixed with the > lower part of the J when enabling the branch, resulting in a jump that > jumped to itself. Conversely, it would result in a HINT instruction when > disabling the branch, but it might not be observable. > > ARM64 does not have this problem since all instructions must be 4-byte > aligned. Reviewed-by: Björn Töpel <bjorn@kernel.org> Nice catch! And I guess this is an issue for kprobes as well, no? I.e. in general replacing 32b insns with an ebreak. This is only valid for natural aligned 32b insns? @Guo I don't see the point of doing a series for this, and asking the maintainers to "pick this patch to stable, and the other for next". Isn't that just more work for the maintainers/reviewers? Björn
On Mon, Jan 30, 2023 at 7:57 PM Björn Töpel <bjorn@kernel.org> wrote: > > guoren@kernel.org writes: > > > From: Andy Chiu <andy.chiu@sifive.com> > > > > Runtime code patching must be done at a naturally aligned address, or we > > may execute on a partial instruction. > > > > We have encountered problems traced back to static jump functions during > > the test. We switched the tracer randomly for every 1~5 seconds on a > > dual-core QEMU setup and found the kernel sucking at a static branch > > where it jumps to itself. > > > > The reason is that the static branch was 2-byte but not 4-byte aligned. > > Then, the kernel would patch the instruction, either J or NOP, with two > > half-word stores if the machine does not have efficient unaligned > > accesses. Thus, moments exist where half of the NOP mixes with the other > > half of the J when transitioning the branch. In our particular case, on > > a little-endian machine, the upper half of the NOP was mixed with the > > lower part of the J when enabling the branch, resulting in a jump that > > jumped to itself. Conversely, it would result in a HINT instruction when > > disabling the branch, but it might not be observable. > > > > ARM64 does not have this problem since all instructions must be 4-byte > > aligned. > > Reviewed-by: Björn Töpel <bjorn@kernel.org> > > Nice catch! And I guess this is an issue for kprobes as well, no? > I.e. in general replacing 32b insns with an ebreak. This is only valid > for natural aligned 32b insns? > > @Guo I don't see the point of doing a series for this, and asking the > maintainers to "pick this patch to stable, and the other for > next". Isn't that just more work for the maintainers/reviewers? If these two patches are separated, they are all fixup that issue and competed with each other. Making my patch an optimization patch must depend on it. That's why I put them in one series. > > > Björn -- Best Regards Guo Ren
© 2016 - 2025 Red Hat, Inc.