arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- 1 file changed, 26 insertions(+), 22 deletions(-)
Floating point instructions in userspace can crash some arm kernels
built with clang/LLD 17.0.6:
BUG: unsupported FP instruction in kernel mode
FPEXC == 0xc0000780
Internal error: Oops - undefined instruction: 0 [#1] ARM
CPU: 0 PID: 196 Comm: vfp-reproducer Not tainted 6.10.0 #1
Hardware name: BCM2835
PC is at vfp_support_entry+0xc8/0x2cc
LR is at do_undefinstr+0xa8/0x250
pc : [<c0101d50>] lr : [<c010a80c>] psr: a0000013
sp : dc8d1f68 ip : 60000013 fp : bedea19c
r10: ec532b17 r9 : 00000010 r8 : 0044766c
r7 : c0000780 r6 : ec532b17 r5 : c1c13800 r4 : dc8d1fb0
r3 : c10072c4 r2 : c0101c88 r1 : ec532b17 r0 : 0044766c
Flags: NzCv IRQs on FIQs on Mode SVC_32 ISA ARM Segment none
Control: 00c5387d Table: 0251c008 DAC: 00000051
Register r0 information: non-paged memory
Register r1 information: vmalloc memory
Register r2 information: non-slab/vmalloc memory
Register r3 information: non-slab/vmalloc memory
Register r4 information: 2-page vmalloc region
Register r5 information: slab kmalloc-cg-2k
Register r6 information: vmalloc memory
Register r7 information: non-slab/vmalloc memory
Register r8 information: non-paged memory
Register r9 information: zero-size pointer
Register r10 information: vmalloc memory
Register r11 information: non-paged memory
Register r12 information: non-paged memory
Process vfp-reproducer (pid: 196, stack limit = 0x61aaaf8b)
Stack: (0xdc8d1f68 to 0xdc8d2000)
1f60: 0000081f b6f69300 0000000f c10073f4 c10072c4 dc8d1fb0
1f80: ec532b17 0c532b17 0044766c b6f9ccd8 00000000 c010a80c 00447670 60000010
1fa0: ffffffff c1c13800 00c5387d c0100f10 b6f68af8 00448fc0 00000000 bedea188
1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c
1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff 00000000 00000000
Call trace:
[<c0101d50>] (vfp_support_entry) from [<c010a80c>] (do_undefinstr+0xa8/0x250)
[<c010a80c>] (do_undefinstr) from [<c0100f10>] (__und_usr+0x70/0x80)
Exception stack(0xdc8d1fb0 to 0xdc8d1ff8)
1fa0: b6f68af8 00448fc0 00000000 bedea188
1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c
1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff
Code: 0a000061 e3877202 e594003c e3a09010 (eef16a10)
---[ end trace 0000000000000000 ]---
Kernel panic - not syncing: Fatal exception in interrupt
---[ end Kernel panic - not syncing: Fatal exception in interrupt ]---
This is a minimal userspace reproducer on a Raspberry Pi Zero W:
#include <stdio.h>
#include <math.h>
int main(void)
{
double v = 1.0;
printf("%f\n", NAN + *(volatile double *)&v);
return 0;
}
Another way to consistently trigger the oops is:
calvin@raspberry-pi-zero-w ~$ python -c "import json"
The bug reproduces only when the kernel is built with DYNAMIC_DEBUG=n,
because the pr_debug() calls act as barriers even when not activated.
This is the output from the same kernel source built with the same
compiler and DYNAMIC_DEBUG=y, where the userspace reproducer works as
expected:
VFP: bounce: trigger ec532b17 fpexc c0000780
VFP: emulate: INST=0xee377b06 SCR=0x00000000
VFP: bounce: trigger eef1fa10 fpexc c0000780
VFP: emulate: INST=0xeeb40b40 SCR=0x00000000
VFP: raising exceptions 30000000
calvin@raspberry-pi-zero-w ~$ ./vfp-reproducer
nan
Crudely grepping for vmsr/vmrs instructions in the otherwise nearly
idential text for vfp_support_entry() makes the problem obvious:
vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc
vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0
vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7
vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc
vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr
vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0
vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid
vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr
vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst
vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2
vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc
vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0
vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7
vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr
vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM!
vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0
vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid
vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst
vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc
vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2
I think LLVM's reordering is valid as the code is currently written: the
compiler doesn't know the instructions have side effects in hardware.
Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be
reordered with respect to each other. The original compiler now produces
working kernels on my hardware with DYNAMIC_DEBUG=n.
This is the relevant piece of the diff of the vfp_support_entry() text,
from the original oopsing kernel to a working kernel with this patch:
vmrs r0, fpscr
tst r0, #4096
bne 0xc0101d48
tst r0, #458752
beq 0xc0101ecc
orr r7, r7, #536870912
ldr r0, [r4, #0x3c]
mov r9, #16
-vmrs r6, fpscr
orr r9, r9, #251658240
add r0, r0, #4
str r0, [r4, #0x3c]
mvn r0, #159
sub r0, r0, #-1207959552
and r0, r7, r0
vmsr fpexc, r0
vmrs r0, fpsid
+vmrs r6, fpscr
and r0, r0, #983040
cmp r0, #65536
bne 0xc0101d88
Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code")
Signed-off-by: Calvin Owens <calvin@wbinvd.org>
---
arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++-------------------
1 file changed, 26 insertions(+), 22 deletions(-)
diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h
index 3c7938fd40aa..32090b0fb250 100644
--- a/arch/arm/vfp/vfpinstr.h
+++ b/arch/arm/vfp/vfpinstr.h
@@ -64,33 +64,37 @@
#ifdef CONFIG_AS_VFP_VMRS_FPINST
-#define fmrx(_vfp_) ({ \
- u32 __v; \
- asm(".fpu vfpv2\n" \
- "vmrs %0, " #_vfp_ \
- : "=r" (__v) : : "cc"); \
- __v; \
- })
-
-#define fmxr(_vfp_,_var_) \
- asm(".fpu vfpv2\n" \
- "vmsr " #_vfp_ ", %0" \
- : : "r" (_var_) : "cc")
+#define fmrx(_vfp_) ({ \
+ u32 __v; \
+ asm volatile (".fpu vfpv2\n" \
+ "vmrs %0, " #_vfp_ \
+ : "=r" (__v) : : "cc"); \
+ __v; \
+})
+
+#define fmxr(_vfp_, _var_) ({ \
+ asm volatile (".fpu vfpv2\n" \
+ "vmsr " #_vfp_ ", %0" \
+ : : "r" (_var_) : "cc"); \
+})
#else
#define vfpreg(_vfp_) #_vfp_
-#define fmrx(_vfp_) ({ \
- u32 __v; \
- asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \
- : "=r" (__v) : : "cc"); \
- __v; \
- })
-
-#define fmxr(_vfp_,_var_) \
- asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \
- : : "r" (_var_) : "cc")
+#define fmrx(_vfp_) ({ \
+ u32 __v; \
+ asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \
+ "cr0, 0 @ fmrx %0, " #_vfp_ \
+ : "=r" (__v) : : "cc"); \
+ __v; \
+})
+
+#define fmxr(_vfp_, _var_) ({ \
+ asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \
+ "cr0, 0 @ fmxr " #_vfp_ ", %0" \
+ : : "r" (_var_) : "cc"); \
+})
#endif
--
2.39.2
(cc Arnd, Nathan, zhuqiuer)
On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote:
>
> Floating point instructions in userspace can crash some arm kernels
> built with clang/LLD 17.0.6:
>
> BUG: unsupported FP instruction in kernel mode
> FPEXC == 0xc0000780
> Internal error: Oops - undefined instruction: 0 [#1] ARM
> CPU: 0 PID: 196 Comm: vfp-reproducer Not tainted 6.10.0 #1
> Hardware name: BCM2835
> PC is at vfp_support_entry+0xc8/0x2cc
> LR is at do_undefinstr+0xa8/0x250
> pc : [<c0101d50>] lr : [<c010a80c>] psr: a0000013
> sp : dc8d1f68 ip : 60000013 fp : bedea19c
> r10: ec532b17 r9 : 00000010 r8 : 0044766c
> r7 : c0000780 r6 : ec532b17 r5 : c1c13800 r4 : dc8d1fb0
> r3 : c10072c4 r2 : c0101c88 r1 : ec532b17 r0 : 0044766c
> Flags: NzCv IRQs on FIQs on Mode SVC_32 ISA ARM Segment none
> Control: 00c5387d Table: 0251c008 DAC: 00000051
> Register r0 information: non-paged memory
> Register r1 information: vmalloc memory
> Register r2 information: non-slab/vmalloc memory
> Register r3 information: non-slab/vmalloc memory
> Register r4 information: 2-page vmalloc region
> Register r5 information: slab kmalloc-cg-2k
> Register r6 information: vmalloc memory
> Register r7 information: non-slab/vmalloc memory
> Register r8 information: non-paged memory
> Register r9 information: zero-size pointer
> Register r10 information: vmalloc memory
> Register r11 information: non-paged memory
> Register r12 information: non-paged memory
> Process vfp-reproducer (pid: 196, stack limit = 0x61aaaf8b)
> Stack: (0xdc8d1f68 to 0xdc8d2000)
> 1f60: 0000081f b6f69300 0000000f c10073f4 c10072c4 dc8d1fb0
> 1f80: ec532b17 0c532b17 0044766c b6f9ccd8 00000000 c010a80c 00447670 60000010
> 1fa0: ffffffff c1c13800 00c5387d c0100f10 b6f68af8 00448fc0 00000000 bedea188
> 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c
> 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff 00000000 00000000
> Call trace:
> [<c0101d50>] (vfp_support_entry) from [<c010a80c>] (do_undefinstr+0xa8/0x250)
> [<c010a80c>] (do_undefinstr) from [<c0100f10>] (__und_usr+0x70/0x80)
> Exception stack(0xdc8d1fb0 to 0xdc8d1ff8)
> 1fa0: b6f68af8 00448fc0 00000000 bedea188
> 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c
> 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff
> Code: 0a000061 e3877202 e594003c e3a09010 (eef16a10)
> ---[ end trace 0000000000000000 ]---
> Kernel panic - not syncing: Fatal exception in interrupt
> ---[ end Kernel panic - not syncing: Fatal exception in interrupt ]---
>
> This is a minimal userspace reproducer on a Raspberry Pi Zero W:
>
> #include <stdio.h>
> #include <math.h>
>
> int main(void)
> {
> double v = 1.0;
> printf("%f\n", NAN + *(volatile double *)&v);
> return 0;
> }
>
> Another way to consistently trigger the oops is:
>
> calvin@raspberry-pi-zero-w ~$ python -c "import json"
>
> The bug reproduces only when the kernel is built with DYNAMIC_DEBUG=n,
> because the pr_debug() calls act as barriers even when not activated.
>
> This is the output from the same kernel source built with the same
> compiler and DYNAMIC_DEBUG=y, where the userspace reproducer works as
> expected:
>
> VFP: bounce: trigger ec532b17 fpexc c0000780
> VFP: emulate: INST=0xee377b06 SCR=0x00000000
> VFP: bounce: trigger eef1fa10 fpexc c0000780
> VFP: emulate: INST=0xeeb40b40 SCR=0x00000000
> VFP: raising exceptions 30000000
>
> calvin@raspberry-pi-zero-w ~$ ./vfp-reproducer
> nan
>
> Crudely grepping for vmsr/vmrs instructions in the otherwise nearly
> idential text for vfp_support_entry() makes the problem obvious:
>
> vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc
> vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0
> vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7
> vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc
> vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr
> vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0
> vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid
> vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr
> vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst
> vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2
>
> vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc
> vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0
> vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7
> vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr
> vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM!
> vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0
> vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid
> vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst
> vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc
> vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2
>
> I think LLVM's reordering is valid as the code is currently written: the
> compiler doesn't know the instructions have side effects in hardware.
>
> Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be
> reordered with respect to each other. The original compiler now produces
> working kernels on my hardware with DYNAMIC_DEBUG=n.
>
> This is the relevant piece of the diff of the vfp_support_entry() text,
> from the original oopsing kernel to a working kernel with this patch:
>
> vmrs r0, fpscr
> tst r0, #4096
> bne 0xc0101d48
> tst r0, #458752
> beq 0xc0101ecc
> orr r7, r7, #536870912
> ldr r0, [r4, #0x3c]
> mov r9, #16
> -vmrs r6, fpscr
> orr r9, r9, #251658240
> add r0, r0, #4
> str r0, [r4, #0x3c]
> mvn r0, #159
> sub r0, r0, #-1207959552
> and r0, r7, r0
> vmsr fpexc, r0
> vmrs r0, fpsid
> +vmrs r6, fpscr
> and r0, r0, #983040
> cmp r0, #65536
> bne 0xc0101d88
>
> Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code")
> Signed-off-by: Calvin Owens <calvin@wbinvd.org>
> ---
> arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++-------------------
> 1 file changed, 26 insertions(+), 22 deletions(-)
>
Thanks for the patch, and for the excellent analysis.
Note that this fix has been proposed in the past, as well as another
one addressing the same issue, but we've been incredibly sloppy
getting it merged.
https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/
https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/
What both of us appear to have missed is that there are two versions
of these routines, which should either be dropped (as they are
obsolete now that the minimum binutils version is 2.25) or fixed up as
well, as you do below.
Anyone have any thoughts on using a memory clobber as opposed to
volatile? I think volatile means that the access cannot be elided, but
it is unclear to me whether that implies any ordering. A 'memory'
clobber implies that globally visible state is updated, which seems
like a stronger guarantee to me.
In any case, let's work together to get /some/ version of this fix merged asap.
> diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h
> index 3c7938fd40aa..32090b0fb250 100644
> --- a/arch/arm/vfp/vfpinstr.h
> +++ b/arch/arm/vfp/vfpinstr.h
> @@ -64,33 +64,37 @@
>
> #ifdef CONFIG_AS_VFP_VMRS_FPINST
>
> -#define fmrx(_vfp_) ({ \
> - u32 __v; \
> - asm(".fpu vfpv2\n" \
> - "vmrs %0, " #_vfp_ \
> - : "=r" (__v) : : "cc"); \
> - __v; \
> - })
> -
> -#define fmxr(_vfp_,_var_) \
> - asm(".fpu vfpv2\n" \
> - "vmsr " #_vfp_ ", %0" \
> - : : "r" (_var_) : "cc")
> +#define fmrx(_vfp_) ({ \
> + u32 __v; \
> + asm volatile (".fpu vfpv2\n" \
> + "vmrs %0, " #_vfp_ \
> + : "=r" (__v) : : "cc"); \
> + __v; \
> +})
> +
> +#define fmxr(_vfp_, _var_) ({ \
> + asm volatile (".fpu vfpv2\n" \
> + "vmsr " #_vfp_ ", %0" \
> + : : "r" (_var_) : "cc"); \
> +})
>
> #else
>
> #define vfpreg(_vfp_) #_vfp_
>
> -#define fmrx(_vfp_) ({ \
> - u32 __v; \
> - asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \
> - : "=r" (__v) : : "cc"); \
> - __v; \
> - })
> -
> -#define fmxr(_vfp_,_var_) \
> - asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \
> - : : "r" (_var_) : "cc")
> +#define fmrx(_vfp_) ({ \
> + u32 __v; \
> + asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \
> + "cr0, 0 @ fmrx %0, " #_vfp_ \
> + : "=r" (__v) : : "cc"); \
> + __v; \
> +})
> +
> +#define fmxr(_vfp_, _var_) ({ \
> + asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \
> + "cr0, 0 @ fmxr " #_vfp_ ", %0" \
> + : : "r" (_var_) : "cc"); \
> +})
>
> #endif
>
> --
> 2.39.2
>
On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote:
> (cc Arnd, Nathan, zhuqiuer)
>
> On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote:
> > <snip>
> >
> > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly
> > idential text for vfp_support_entry() makes the problem obvious:
> >
> > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc
> > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0
> > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7
> > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc
> > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr
> > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0
> > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid
> > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr
> > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst
> > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2
> >
> > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc
> > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0
> > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7
> > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr
> > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM!
> > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0
> > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid
> > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst
> > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc
> > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2
> >
> > I think LLVM's reordering is valid as the code is currently written: the
> > compiler doesn't know the instructions have side effects in hardware.
> >
> > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be
> > reordered with respect to each other. The original compiler now produces
> > working kernels on my hardware with DYNAMIC_DEBUG=n.
> >
> > This is the relevant piece of the diff of the vfp_support_entry() text,
> > from the original oopsing kernel to a working kernel with this patch:
> >
> > vmrs r0, fpscr
> > tst r0, #4096
> > bne 0xc0101d48
> > tst r0, #458752
> > beq 0xc0101ecc
> > orr r7, r7, #536870912
> > ldr r0, [r4, #0x3c]
> > mov r9, #16
> > -vmrs r6, fpscr
> > orr r9, r9, #251658240
> > add r0, r0, #4
> > str r0, [r4, #0x3c]
> > mvn r0, #159
> > sub r0, r0, #-1207959552
> > and r0, r7, r0
> > vmsr fpexc, r0
> > vmrs r0, fpsid
> > +vmrs r6, fpscr
> > and r0, r0, #983040
> > cmp r0, #65536
> > bne 0xc0101d88
> >
> > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code")
> > Signed-off-by: Calvin Owens <calvin@wbinvd.org>
> > ---
> > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++-------------------
> > 1 file changed, 26 insertions(+), 22 deletions(-)
> >
>
> Thanks for the patch, and for the excellent analysis.
>
> Note that this fix has been proposed in the past, as well as another
> one addressing the same issue, but we've been incredibly sloppy
> getting it merged.
>
> https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/
> https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/
Ah sorry for missing that, I searched for the symptom not the cure.
> What both of us appear to have missed is that there are two versions
> of these routines, which should either be dropped (as they are
> obsolete now that the minimum binutils version is 2.25) or fixed up as
> well, as you do below.
>
> Anyone have any thoughts on using a memory clobber as opposed to
> volatile? I think volatile means that the access cannot be elided, but
> it is unclear to me whether that implies any ordering. A 'memory'
> clobber implies that globally visible state is updated, which seems
> like a stronger guarantee to me.
My thinking was that if 'asm volatile' is sufficient, it will suppress
less optimization than the clobber, since the clobber might force the
compiler to assume unrelated memory could have been modified when it
really never is. But I'm not sure about that.
Out of curiousity, I tried it both ways with the same compiler just now,
the only tiny difference in the emitted vfp_support_entry() is here:
--- /volatile 2024-07-28 13:28:59.890505404 -0700
+++ /memclobber 2024-07-28 13:28:59.890505404 -0700
str r0, [r5, #0x4]
vmrs r7, fpexc
tst r7, #1073741824
bne 0xc0101d28
orr r7, r7, #1073741824
bic r0, r7, #-2147483648
vmsr fpexc, r0
+ldr r8, [pc, #0x26c]
ldr r0, [r5, #0x8]
-ldr r8, [pc, #0x268]
add r6, r5, #224
ldr r0, [r8, r0, lsl #2]
cmp r0, r6
beq 0xc0101d18
Thanks,
Calvin
> In any case, let's work together to get /some/ version of this fix merged asap.
>
> > diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h
> > index 3c7938fd40aa..32090b0fb250 100644
> > --- a/arch/arm/vfp/vfpinstr.h
> > +++ b/arch/arm/vfp/vfpinstr.h
> > @@ -64,33 +64,37 @@
> >
> > #ifdef CONFIG_AS_VFP_VMRS_FPINST
> >
> > -#define fmrx(_vfp_) ({ \
> > - u32 __v; \
> > - asm(".fpu vfpv2\n" \
> > - "vmrs %0, " #_vfp_ \
> > - : "=r" (__v) : : "cc"); \
> > - __v; \
> > - })
> > -
> > -#define fmxr(_vfp_,_var_) \
> > - asm(".fpu vfpv2\n" \
> > - "vmsr " #_vfp_ ", %0" \
> > - : : "r" (_var_) : "cc")
> > +#define fmrx(_vfp_) ({ \
> > + u32 __v; \
> > + asm volatile (".fpu vfpv2\n" \
> > + "vmrs %0, " #_vfp_ \
> > + : "=r" (__v) : : "cc"); \
> > + __v; \
> > +})
> > +
> > +#define fmxr(_vfp_, _var_) ({ \
> > + asm volatile (".fpu vfpv2\n" \
> > + "vmsr " #_vfp_ ", %0" \
> > + : : "r" (_var_) : "cc"); \
> > +})
> >
> > #else
> >
> > #define vfpreg(_vfp_) #_vfp_
> >
> > -#define fmrx(_vfp_) ({ \
> > - u32 __v; \
> > - asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \
> > - : "=r" (__v) : : "cc"); \
> > - __v; \
> > - })
> > -
> > -#define fmxr(_vfp_,_var_) \
> > - asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \
> > - : : "r" (_var_) : "cc")
> > +#define fmrx(_vfp_) ({ \
> > + u32 __v; \
> > + asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \
> > + "cr0, 0 @ fmrx %0, " #_vfp_ \
> > + : "=r" (__v) : : "cc"); \
> > + __v; \
> > +})
> > +
> > +#define fmxr(_vfp_, _var_) ({ \
> > + asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \
> > + "cr0, 0 @ fmxr " #_vfp_ ", %0" \
> > + : : "r" (_var_) : "cc"); \
> > +})
> >
> > #endif
> >
> > --
> > 2.39.2
> >
On Sun, 28 Jul 2024 at 23:29, Calvin Owens <calvin@wbinvd.org> wrote:
>
> On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote:
> > (cc Arnd, Nathan, zhuqiuer)
> >
> > On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote:
> > > <snip>
> > >
> > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly
> > > idential text for vfp_support_entry() makes the problem obvious:
> > >
> > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc
> > > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0
> > > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7
> > > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc
> > > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr
> > > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0
> > > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid
> > > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr
> > > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst
> > > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2
> > >
> > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc
> > > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0
> > > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7
> > > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr
> > > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM!
> > > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0
> > > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid
> > > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst
> > > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc
> > > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2
> > >
> > > I think LLVM's reordering is valid as the code is currently written: the
> > > compiler doesn't know the instructions have side effects in hardware.
> > >
> > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be
> > > reordered with respect to each other. The original compiler now produces
> > > working kernels on my hardware with DYNAMIC_DEBUG=n.
> > >
> > > This is the relevant piece of the diff of the vfp_support_entry() text,
> > > from the original oopsing kernel to a working kernel with this patch:
> > >
> > > vmrs r0, fpscr
> > > tst r0, #4096
> > > bne 0xc0101d48
> > > tst r0, #458752
> > > beq 0xc0101ecc
> > > orr r7, r7, #536870912
> > > ldr r0, [r4, #0x3c]
> > > mov r9, #16
> > > -vmrs r6, fpscr
> > > orr r9, r9, #251658240
> > > add r0, r0, #4
> > > str r0, [r4, #0x3c]
> > > mvn r0, #159
> > > sub r0, r0, #-1207959552
> > > and r0, r7, r0
> > > vmsr fpexc, r0
> > > vmrs r0, fpsid
> > > +vmrs r6, fpscr
> > > and r0, r0, #983040
> > > cmp r0, #65536
> > > bne 0xc0101d88
> > >
> > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code")
> > > Signed-off-by: Calvin Owens <calvin@wbinvd.org>
> > > ---
> > > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++-------------------
> > > 1 file changed, 26 insertions(+), 22 deletions(-)
> > >
> >
> > Thanks for the patch, and for the excellent analysis.
> >
> > Note that this fix has been proposed in the past, as well as another
> > one addressing the same issue, but we've been incredibly sloppy
> > getting it merged.
> >
> > https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/
> > https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/
>
> Ah sorry for missing that, I searched for the symptom not the cure.
>
> > What both of us appear to have missed is that there are two versions
> > of these routines, which should either be dropped (as they are
> > obsolete now that the minimum binutils version is 2.25) or fixed up as
> > well, as you do below.
> >
> > Anyone have any thoughts on using a memory clobber as opposed to
> > volatile? I think volatile means that the access cannot be elided, but
> > it is unclear to me whether that implies any ordering. A 'memory'
> > clobber implies that globally visible state is updated, which seems
> > like a stronger guarantee to me.
>
> My thinking was that if 'asm volatile' is sufficient, it will suppress
> less optimization than the clobber, since the clobber might force the
> compiler to assume unrelated memory could have been modified when it
> really never is. But I'm not sure about that.
>
> Out of curiousity, I tried it both ways with the same compiler just now,
> the only tiny difference in the emitted vfp_support_entry() is here:
>
> --- /volatile 2024-07-28 13:28:59.890505404 -0700
> +++ /memclobber 2024-07-28 13:28:59.890505404 -0700
> str r0, [r5, #0x4]
> vmrs r7, fpexc
> tst r7, #1073741824
> bne 0xc0101d28
> orr r7, r7, #1073741824
> bic r0, r7, #-2147483648
> vmsr fpexc, r0
> +ldr r8, [pc, #0x26c]
> ldr r0, [r5, #0x8]
> -ldr r8, [pc, #0x268]
> add r6, r5, #224
> ldr r0, [r8, r0, lsl #2]
> cmp r0, r6
> beq 0xc0101d18
>
Right. So it doesn't matter in practice - good to know.
So in the 'volatile' case, I guess we are relying on the compiler not
reordering those with respect to each other, which should be
sufficient to ensure that we do not access VFP status register before
enabling the VFP unit via the control register, as all are accessed
using inline asm.
In any case, this should go into the patch system.
https://www.armlinux.org.uk/developer/patches/section.php?section=0
Note to self and other: follow-up with a patch that removes
CONFIG_AS_VFP_VMRS_FPINST, which is no longer needed.
On Monday 07/29 at 09:12 +0200, Ard Biesheuvel wrote:
> On Sun, 28 Jul 2024 at 23:29, Calvin Owens <calvin@wbinvd.org> wrote:
> >
> > On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote:
> > > (cc Arnd, Nathan, zhuqiuer)
> > >
> > > On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote:
> > > > <snip>
> > > >
> > > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly
> > > > idential text for vfp_support_entry() makes the problem obvious:
> > > >
> > > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc
> > > > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0
> > > > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7
> > > > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc
> > > > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr
> > > > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0
> > > > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid
> > > > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr
> > > > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst
> > > > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2
> > > >
> > > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc
> > > > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0
> > > > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7
> > > > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr
> > > > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM!
> > > > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0
> > > > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid
> > > > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst
> > > > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc
> > > > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2
> > > >
> > > > I think LLVM's reordering is valid as the code is currently written: the
> > > > compiler doesn't know the instructions have side effects in hardware.
> > > >
> > > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be
> > > > reordered with respect to each other. The original compiler now produces
> > > > working kernels on my hardware with DYNAMIC_DEBUG=n.
> > > >
> > > > This is the relevant piece of the diff of the vfp_support_entry() text,
> > > > from the original oopsing kernel to a working kernel with this patch:
> > > >
> > > > vmrs r0, fpscr
> > > > tst r0, #4096
> > > > bne 0xc0101d48
> > > > tst r0, #458752
> > > > beq 0xc0101ecc
> > > > orr r7, r7, #536870912
> > > > ldr r0, [r4, #0x3c]
> > > > mov r9, #16
> > > > -vmrs r6, fpscr
> > > > orr r9, r9, #251658240
> > > > add r0, r0, #4
> > > > str r0, [r4, #0x3c]
> > > > mvn r0, #159
> > > > sub r0, r0, #-1207959552
> > > > and r0, r7, r0
> > > > vmsr fpexc, r0
> > > > vmrs r0, fpsid
> > > > +vmrs r6, fpscr
> > > > and r0, r0, #983040
> > > > cmp r0, #65536
> > > > bne 0xc0101d88
> > > >
> > > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code")
> > > > Signed-off-by: Calvin Owens <calvin@wbinvd.org>
> > > > ---
> > > > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++-------------------
> > > > 1 file changed, 26 insertions(+), 22 deletions(-)
> > > >
> > >
> > > Thanks for the patch, and for the excellent analysis.
> > >
> > > Note that this fix has been proposed in the past, as well as another
> > > one addressing the same issue, but we've been incredibly sloppy
> > > getting it merged.
> > >
> > > https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/
> > > https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/
> >
> > Ah sorry for missing that, I searched for the symptom not the cure.
> >
> > > What both of us appear to have missed is that there are two versions
> > > of these routines, which should either be dropped (as they are
> > > obsolete now that the minimum binutils version is 2.25) or fixed up as
> > > well, as you do below.
> > >
> > > Anyone have any thoughts on using a memory clobber as opposed to
> > > volatile? I think volatile means that the access cannot be elided, but
> > > it is unclear to me whether that implies any ordering. A 'memory'
> > > clobber implies that globally visible state is updated, which seems
> > > like a stronger guarantee to me.
> >
> > My thinking was that if 'asm volatile' is sufficient, it will suppress
> > less optimization than the clobber, since the clobber might force the
> > compiler to assume unrelated memory could have been modified when it
> > really never is. But I'm not sure about that.
> >
> > Out of curiousity, I tried it both ways with the same compiler just now,
> > the only tiny difference in the emitted vfp_support_entry() is here:
> >
> > --- /volatile 2024-07-28 13:28:59.890505404 -0700
> > +++ /memclobber 2024-07-28 13:28:59.890505404 -0700
> > str r0, [r5, #0x4]
> > vmrs r7, fpexc
> > tst r7, #1073741824
> > bne 0xc0101d28
> > orr r7, r7, #1073741824
> > bic r0, r7, #-2147483648
> > vmsr fpexc, r0
> > +ldr r8, [pc, #0x26c]
> > ldr r0, [r5, #0x8]
> > -ldr r8, [pc, #0x268]
> > add r6, r5, #224
> > ldr r0, [r8, r0, lsl #2]
> > cmp r0, r6
> > beq 0xc0101d18
> >
>
> Right. So it doesn't matter in practice - good to know.
>
> So in the 'volatile' case, I guess we are relying on the compiler not
> reordering those with respect to each other, which should be
> sufficient to ensure that we do not access VFP status register before
> enabling the VFP unit via the control register, as all are accessed
> using inline asm.
>
> In any case, this should go into the patch system.
> https://www.armlinux.org.uk/developer/patches/section.php?section=0
Done:
https://www.armlinux.org.uk/developer/patches/viewpatch.php?id=9410/1
Thanks,
Calvin
> Note to self and other: follow-up with a patch that removes
> CONFIG_AS_VFP_VMRS_FPINST, which is no longer needed.
© 2016 - 2026 Red Hat, Inc.