1 | The following changes since commit 67e41fe0cfb62e6cdfa659f0155417d17e5274ea: | 1 | The following changes since commit c586691e676214eb7edf6a468e84e7ce3b314d43: |
---|---|---|---|
2 | 2 | ||
3 | Merge tag 'pull-ppc-20220104' of https://github.com/legoater/qemu into staging (2022-01-04 07:23:27 -0800) | 3 | Merge tag 'pull-target-arm-20230502-2' of https://git.linaro.org/people/pmaydell/qemu-arm into staging (2023-05-02 16:38:29 +0100) |
4 | 4 | ||
5 | are available in the Git repository at: | 5 | are available in the Git repository at: |
6 | 6 | ||
7 | https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220104 | 7 | https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20230502-2 |
8 | 8 | ||
9 | for you to fetch changes up to d7478d4229f0a2b2817a55487e6b17081099fae4: | 9 | for you to fetch changes up to 129f1f9ee7df77d367d961b3c25353612d33cd83: |
10 | 10 | ||
11 | common-user: Fix tail calls to safe_syscall_set_errno_tail (2022-01-04 15:41:03 -0800) | 11 | tcg: Introduce tcg_out_movext2 (2023-05-02 13:05:45 -0700) |
12 | 12 | ||
13 | ---------------------------------------------------------------- | 13 | ---------------------------------------------------------------- |
14 | Fix for safe_syscall_base. | 14 | Misc tcg-related patch queue. |
15 | Fix for folding of vector add/sub. | 15 | |
16 | Fix build on loongarch64 with gcc 8. | 16 | v2: Update bitops.h rotate patch. |
17 | Remove decl for qemu_run_machine_init_done_notifiers. | ||
18 | 17 | ||
19 | ---------------------------------------------------------------- | 18 | ---------------------------------------------------------------- |
20 | Philippe Mathieu-Daudé (1): | 19 | Dickon Hood (1): |
21 | linux-user: Fix trivial build error on loongarch64 hosts | 20 | qemu/bitops.h: Limit rotate amounts |
22 | 21 | ||
23 | Richard Henderson (2): | 22 | Kiran Ostrolenk (1): |
24 | tcg/optimize: Fix folding of vector ops | 23 | qemu/host-utils.h: Add clz and ctz functions for lower-bit integers |
25 | common-user: Fix tail calls to safe_syscall_set_errno_tail | ||
26 | 24 | ||
27 | Xiaoyao Li (1): | 25 | Nazar Kazakov (2): |
28 | sysemu: Cleanup qemu_run_machine_init_done_notifiers() | 26 | tcg: Add tcg_gen_gvec_andcs |
27 | tcg: Add tcg_gen_gvec_rotrs | ||
29 | 28 | ||
30 | include/sysemu/sysemu.h | 1 - | 29 | Richard Henderson (7): |
31 | linux-user/host/loongarch64/host-signal.h | 4 +-- | 30 | softmmu: Tidy dirtylimit_dirty_ring_full_time |
32 | tcg/optimize.c | 49 +++++++++++++++++++++++------- | 31 | qemu/int128: Re-shuffle Int128Alias members |
33 | common-user/host/i386/safe-syscall.inc.S | 1 + | 32 | migration/xbzrle: Use __attribute__((target)) for avx512 |
34 | common-user/host/mips/safe-syscall.inc.S | 1 + | 33 | accel/tcg: Add cpu_ld*_code_mmu |
35 | common-user/host/x86_64/safe-syscall.inc.S | 1 + | 34 | tcg/loongarch64: Conditionalize tcg_out_exts_i32_i64 |
36 | 6 files changed, 42 insertions(+), 15 deletions(-) | 35 | tcg/mips: Conditionalize tcg_out_exts_i32_i64 |
36 | tcg: Introduce tcg_out_movext2 | ||
37 | 37 | ||
38 | Weiwei Li (1): | ||
39 | accel/tcg: Uncache the host address for instruction fetch when tlb size < 1 | ||
40 | |||
41 | meson.build | 5 +-- | ||
42 | accel/tcg/tcg-runtime.h | 1 + | ||
43 | include/exec/cpu_ldst.h | 9 ++++++ | ||
44 | include/qemu/bitops.h | 16 +++++----- | ||
45 | include/qemu/host-utils.h | 54 +++++++++++++++++++++++++++++++ | ||
46 | include/qemu/int128.h | 4 +-- | ||
47 | include/tcg/tcg-op-gvec.h | 4 +++ | ||
48 | accel/tcg/cputlb.c | 53 ++++++++++++++++++++++++++++++ | ||
49 | accel/tcg/tcg-runtime-gvec.c | 11 +++++++ | ||
50 | accel/tcg/user-exec.c | 58 +++++++++++++++++++++++++++++++++ | ||
51 | migration/xbzrle.c | 9 +++--- | ||
52 | softmmu/dirtylimit.c | 15 ++++++--- | ||
53 | tcg/tcg-op-gvec.c | 28 ++++++++++++++++ | ||
54 | tcg/tcg.c | 69 +++++++++++++++++++++++++++++++++++++--- | ||
55 | tcg/arm/tcg-target.c.inc | 44 +++++++++++-------------- | ||
56 | tcg/i386/tcg-target.c.inc | 19 +++++------ | ||
57 | tcg/loongarch64/tcg-target.c.inc | 4 ++- | ||
58 | tcg/mips/tcg-target.c.inc | 4 ++- | ||
59 | 18 files changed, 339 insertions(+), 68 deletions(-) | diff view generated by jsdifflib |
1 | Bitwise operations are easy to fold, because the operation is | 1 | From: Dickon Hood <dickon.hood@codethink.co.uk> |
---|---|---|---|
2 | identical regardless of element size. But add and sub need | ||
3 | extra element size info that is not currently propagated. | ||
4 | 2 | ||
5 | Fixes: 2f9f08ba43d | 3 | Rotates have been fixed up to only allow for reasonable rotate amounts |
6 | Cc: qemu-stable@nongnu.org | 4 | (ie, no rotates >7 on an 8b value etc.) This fixes a problem with riscv |
7 | Resolves: https://gitlab.com/qemu-project/qemu/-/issues/799 | 5 | vector rotate instructions. |
8 | Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | 6 | |
7 | Signed-off-by: Dickon Hood <dickon.hood@codethink.co.uk> | ||
8 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
9 | Message-Id: <20230428144757.57530-9-lawrence.hunter@codethink.co.uk> | ||
10 | [rth: Mask shifts in both directions.] | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 11 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
10 | --- | 12 | --- |
11 | tcg/optimize.c | 49 ++++++++++++++++++++++++++++++++++++++----------- | 13 | include/qemu/bitops.h | 16 ++++++++-------- |
12 | 1 file changed, 38 insertions(+), 11 deletions(-) | 14 | 1 file changed, 8 insertions(+), 8 deletions(-) |
13 | 15 | ||
14 | diff --git a/tcg/optimize.c b/tcg/optimize.c | 16 | diff --git a/include/qemu/bitops.h b/include/qemu/bitops.h |
15 | index XXXXXXX..XXXXXXX 100644 | 17 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/tcg/optimize.c | 18 | --- a/include/qemu/bitops.h |
17 | +++ b/tcg/optimize.c | 19 | +++ b/include/qemu/bitops.h |
18 | @@ -XXX,XX +XXX,XX @@ static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y) | 20 | @@ -XXX,XX +XXX,XX @@ static inline unsigned long find_first_zero_bit(const unsigned long *addr, |
19 | CASE_OP_32_64(mul): | 21 | */ |
20 | return x * y; | 22 | static inline uint8_t rol8(uint8_t word, unsigned int shift) |
21 | 23 | { | |
22 | - CASE_OP_32_64(and): | 24 | - return (word << shift) | (word >> ((8 - shift) & 7)); |
23 | + CASE_OP_32_64_VEC(and): | 25 | + return (word << (shift & 7)) | (word >> (-shift & 7)); |
24 | return x & y; | ||
25 | |||
26 | - CASE_OP_32_64(or): | ||
27 | + CASE_OP_32_64_VEC(or): | ||
28 | return x | y; | ||
29 | |||
30 | - CASE_OP_32_64(xor): | ||
31 | + CASE_OP_32_64_VEC(xor): | ||
32 | return x ^ y; | ||
33 | |||
34 | case INDEX_op_shl_i32: | ||
35 | @@ -XXX,XX +XXX,XX @@ static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y) | ||
36 | case INDEX_op_rotl_i64: | ||
37 | return rol64(x, y & 63); | ||
38 | |||
39 | - CASE_OP_32_64(not): | ||
40 | + CASE_OP_32_64_VEC(not): | ||
41 | return ~x; | ||
42 | |||
43 | CASE_OP_32_64(neg): | ||
44 | return -x; | ||
45 | |||
46 | - CASE_OP_32_64(andc): | ||
47 | + CASE_OP_32_64_VEC(andc): | ||
48 | return x & ~y; | ||
49 | |||
50 | - CASE_OP_32_64(orc): | ||
51 | + CASE_OP_32_64_VEC(orc): | ||
52 | return x | ~y; | ||
53 | |||
54 | CASE_OP_32_64(eqv): | ||
55 | @@ -XXX,XX +XXX,XX @@ static bool fold_const2(OptContext *ctx, TCGOp *op) | ||
56 | return false; | ||
57 | } | 26 | } |
58 | 27 | ||
59 | +static bool fold_commutative(OptContext *ctx, TCGOp *op) | 28 | /** |
60 | +{ | 29 | @@ -XXX,XX +XXX,XX @@ static inline uint8_t rol8(uint8_t word, unsigned int shift) |
61 | + swap_commutative(op->args[0], &op->args[1], &op->args[2]); | 30 | */ |
62 | + return false; | 31 | static inline uint8_t ror8(uint8_t word, unsigned int shift) |
63 | +} | ||
64 | + | ||
65 | static bool fold_const2_commutative(OptContext *ctx, TCGOp *op) | ||
66 | { | 32 | { |
67 | swap_commutative(op->args[0], &op->args[1], &op->args[2]); | 33 | - return (word >> shift) | (word << ((8 - shift) & 7)); |
68 | @@ -XXX,XX +XXX,XX @@ static bool fold_add(OptContext *ctx, TCGOp *op) | 34 | + return (word >> (shift & 7)) | (word << (-shift & 7)); |
69 | return false; | ||
70 | } | 35 | } |
71 | 36 | ||
72 | +/* We cannot as yet do_constant_folding with vectors. */ | 37 | /** |
73 | +static bool fold_add_vec(OptContext *ctx, TCGOp *op) | 38 | @@ -XXX,XX +XXX,XX @@ static inline uint8_t ror8(uint8_t word, unsigned int shift) |
74 | +{ | 39 | */ |
75 | + if (fold_commutative(ctx, op) || | 40 | static inline uint16_t rol16(uint16_t word, unsigned int shift) |
76 | + fold_xi_to_x(ctx, op, 0)) { | ||
77 | + return true; | ||
78 | + } | ||
79 | + return false; | ||
80 | +} | ||
81 | + | ||
82 | static bool fold_addsub2(OptContext *ctx, TCGOp *op, bool add) | ||
83 | { | 41 | { |
84 | if (arg_is_const(op->args[2]) && arg_is_const(op->args[3]) && | 42 | - return (word << shift) | (word >> ((16 - shift) & 15)); |
85 | @@ -XXX,XX +XXX,XX @@ static bool fold_sub_to_neg(OptContext *ctx, TCGOp *op) | 43 | + return (word << (shift & 15)) | (word >> (-shift & 15)); |
86 | return false; | ||
87 | } | 44 | } |
88 | 45 | ||
89 | -static bool fold_sub(OptContext *ctx, TCGOp *op) | 46 | /** |
90 | +/* We cannot as yet do_constant_folding with vectors. */ | 47 | @@ -XXX,XX +XXX,XX @@ static inline uint16_t rol16(uint16_t word, unsigned int shift) |
91 | +static bool fold_sub_vec(OptContext *ctx, TCGOp *op) | 48 | */ |
49 | static inline uint16_t ror16(uint16_t word, unsigned int shift) | ||
92 | { | 50 | { |
93 | - if (fold_const2(ctx, op) || | 51 | - return (word >> shift) | (word << ((16 - shift) & 15)); |
94 | - fold_xx_to_i(ctx, op, 0) || | 52 | + return (word >> (shift & 15)) | (word << (-shift & 15)); |
95 | + if (fold_xx_to_i(ctx, op, 0) || | ||
96 | fold_xi_to_x(ctx, op, 0) || | ||
97 | fold_sub_to_neg(ctx, op)) { | ||
98 | return true; | ||
99 | @@ -XXX,XX +XXX,XX @@ static bool fold_sub(OptContext *ctx, TCGOp *op) | ||
100 | return false; | ||
101 | } | 53 | } |
102 | 54 | ||
103 | +static bool fold_sub(OptContext *ctx, TCGOp *op) | 55 | /** |
104 | +{ | 56 | @@ -XXX,XX +XXX,XX @@ static inline uint16_t ror16(uint16_t word, unsigned int shift) |
105 | + return fold_const2(ctx, op) || fold_sub_vec(ctx, op); | 57 | */ |
106 | +} | 58 | static inline uint32_t rol32(uint32_t word, unsigned int shift) |
107 | + | ||
108 | static bool fold_sub2(OptContext *ctx, TCGOp *op) | ||
109 | { | 59 | { |
110 | return fold_addsub2(ctx, op, false); | 60 | - return (word << shift) | (word >> ((32 - shift) & 31)); |
111 | @@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s) | 61 | + return (word << (shift & 31)) | (word >> (-shift & 31)); |
112 | * Sorted alphabetically by opcode as much as possible. | 62 | } |
113 | */ | 63 | |
114 | switch (opc) { | 64 | /** |
115 | - CASE_OP_32_64_VEC(add): | 65 | @@ -XXX,XX +XXX,XX @@ static inline uint32_t rol32(uint32_t word, unsigned int shift) |
116 | + CASE_OP_32_64(add): | 66 | */ |
117 | done = fold_add(&ctx, op); | 67 | static inline uint32_t ror32(uint32_t word, unsigned int shift) |
118 | break; | 68 | { |
119 | + case INDEX_op_add_vec: | 69 | - return (word >> shift) | (word << ((32 - shift) & 31)); |
120 | + done = fold_add_vec(&ctx, op); | 70 | + return (word >> (shift & 31)) | (word << (-shift & 31)); |
121 | + break; | 71 | } |
122 | CASE_OP_32_64(add2): | 72 | |
123 | done = fold_add2(&ctx, op); | 73 | /** |
124 | break; | 74 | @@ -XXX,XX +XXX,XX @@ static inline uint32_t ror32(uint32_t word, unsigned int shift) |
125 | @@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s) | 75 | */ |
126 | CASE_OP_32_64(sextract): | 76 | static inline uint64_t rol64(uint64_t word, unsigned int shift) |
127 | done = fold_sextract(&ctx, op); | 77 | { |
128 | break; | 78 | - return (word << shift) | (word >> ((64 - shift) & 63)); |
129 | - CASE_OP_32_64_VEC(sub): | 79 | + return (word << (shift & 63)) | (word >> (-shift & 63)); |
130 | + CASE_OP_32_64(sub): | 80 | } |
131 | done = fold_sub(&ctx, op); | 81 | |
132 | break; | 82 | /** |
133 | + case INDEX_op_sub_vec: | 83 | @@ -XXX,XX +XXX,XX @@ static inline uint64_t rol64(uint64_t word, unsigned int shift) |
134 | + done = fold_sub_vec(&ctx, op); | 84 | */ |
135 | + break; | 85 | static inline uint64_t ror64(uint64_t word, unsigned int shift) |
136 | CASE_OP_32_64(sub2): | 86 | { |
137 | done = fold_sub2(&ctx, op); | 87 | - return (word >> shift) | (word << ((64 - shift) & 63)); |
138 | break; | 88 | + return (word >> (shift & 63)) | (word << (-shift & 63)); |
89 | } | ||
90 | |||
91 | /** | ||
139 | -- | 92 | -- |
140 | 2.25.1 | 93 | 2.34.1 |
141 | |||
142 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
2 | 1 | ||
3 | When building using GCC 8.3.0 on loongarch64 (Loongnix) we get: | ||
4 | |||
5 | In file included from ../linux-user/signal.c:33: | ||
6 | ../linux-user/host/loongarch64/host-signal.h: In function ‘host_signal_write’: | ||
7 | ../linux-user/host/loongarch64/host-signal.h:57:9: error: a label can only be part of a statement and a declaration is not a statement | ||
8 | uint32_t sel = (insn >> 15) & 0b11111111111; | ||
9 | ^~~~~~~~ | ||
10 | |||
11 | We don't use the 'sel' variable more than once, so drop it. | ||
12 | |||
13 | Meson output for the record: | ||
14 | |||
15 | Host machine cpu family: loongarch64 | ||
16 | Host machine cpu: loongarch64 | ||
17 | C compiler for the host machine: cc (gcc 8.3.0 "cc (Loongnix 8.3.0-6.lnd.vec.27) 8.3.0") | ||
18 | C linker for the host machine: cc ld.bfd 2.31.1-system | ||
19 | |||
20 | Fixes: ad812c3bd65 ("linux-user: Implement CPU-specific signal handler for loongarch64 hosts") | ||
21 | Reported-by: Song Gao <gaosong@loongson.cn> | ||
22 | Suggested-by: Song Gao <gaosong@loongson.cn> | ||
23 | Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
24 | Reviewed-by: WANG Xuerui <git@xen0n.name> | ||
25 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
26 | Message-Id: <20220104215027.2180972-1-f4bug@amsat.org> | ||
27 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
28 | --- | ||
29 | linux-user/host/loongarch64/host-signal.h | 4 +--- | ||
30 | 1 file changed, 1 insertion(+), 3 deletions(-) | ||
31 | |||
32 | diff --git a/linux-user/host/loongarch64/host-signal.h b/linux-user/host/loongarch64/host-signal.h | ||
33 | index XXXXXXX..XXXXXXX 100644 | ||
34 | --- a/linux-user/host/loongarch64/host-signal.h | ||
35 | +++ b/linux-user/host/loongarch64/host-signal.h | ||
36 | @@ -XXX,XX +XXX,XX @@ static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc) | ||
37 | } | ||
38 | break; | ||
39 | case 0b001110: /* indexed, atomic, bounds-checking memory operations */ | ||
40 | - uint32_t sel = (insn >> 15) & 0b11111111111; | ||
41 | - | ||
42 | - switch (sel) { | ||
43 | + switch ((insn >> 15) & 0b11111111111) { | ||
44 | case 0b00000100000: /* stx.b */ | ||
45 | case 0b00000101000: /* stx.h */ | ||
46 | case 0b00000110000: /* stx.w */ | ||
47 | -- | ||
48 | 2.25.1 | ||
49 | |||
50 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | From: Xiaoyao Li <xiaoyao.li@intel.com> | ||
2 | 1 | ||
3 | Remove qemu_run_machine_init_done_notifiers() since no implementation | ||
4 | and user. | ||
5 | |||
6 | Fixes: f66dc8737c9 ("vl: move all generic initialization out of vl.c") | ||
7 | Signed-off-by: Xiaoyao Li <xiaoyao.li@intel.com> | ||
8 | Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
9 | Message-Id: <20220104024136.1433545-1-xiaoyao.li@intel.com> | ||
10 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
11 | --- | ||
12 | include/sysemu/sysemu.h | 1 - | ||
13 | 1 file changed, 1 deletion(-) | ||
14 | |||
15 | diff --git a/include/sysemu/sysemu.h b/include/sysemu/sysemu.h | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/include/sysemu/sysemu.h | ||
18 | +++ b/include/sysemu/sysemu.h | ||
19 | @@ -XXX,XX +XXX,XX @@ extern bool qemu_uuid_set; | ||
20 | void qemu_add_exit_notifier(Notifier *notify); | ||
21 | void qemu_remove_exit_notifier(Notifier *notify); | ||
22 | |||
23 | -void qemu_run_machine_init_done_notifiers(void); | ||
24 | void qemu_add_machine_init_done_notifier(Notifier *notify); | ||
25 | void qemu_remove_machine_init_done_notifier(Notifier *notify); | ||
26 | |||
27 | -- | ||
28 | 2.25.1 | ||
29 | |||
30 | diff view generated by jsdifflib |
Deleted patch | |||
---|---|---|---|
1 | For the ABIs in which the syscall return register is not | ||
2 | also the first function argument register, move the errno | ||
3 | value into the correct place. | ||
4 | 1 | ||
5 | Fixes: a3310c0397e2 ("linux-user: Move syscall error detection into safe_syscall_base") | ||
6 | Reported-by: Laurent Vivier <laurent@vivier.eu> | ||
7 | Tested-by: Laurent Vivier <laurent@vivier.eu> | ||
8 | Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Message-Id: <20220104190454.542225-1-richard.henderson@linaro.org> | ||
11 | --- | ||
12 | common-user/host/i386/safe-syscall.inc.S | 1 + | ||
13 | common-user/host/mips/safe-syscall.inc.S | 1 + | ||
14 | common-user/host/x86_64/safe-syscall.inc.S | 1 + | ||
15 | 3 files changed, 3 insertions(+) | ||
16 | |||
17 | diff --git a/common-user/host/i386/safe-syscall.inc.S b/common-user/host/i386/safe-syscall.inc.S | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/common-user/host/i386/safe-syscall.inc.S | ||
20 | +++ b/common-user/host/i386/safe-syscall.inc.S | ||
21 | @@ -XXX,XX +XXX,XX @@ safe_syscall_end: | ||
22 | pop %ebp | ||
23 | .cfi_adjust_cfa_offset -4 | ||
24 | .cfi_restore ebp | ||
25 | + mov %eax, (%esp) | ||
26 | jmp safe_syscall_set_errno_tail | ||
27 | |||
28 | .cfi_endproc | ||
29 | diff --git a/common-user/host/mips/safe-syscall.inc.S b/common-user/host/mips/safe-syscall.inc.S | ||
30 | index XXXXXXX..XXXXXXX 100644 | ||
31 | --- a/common-user/host/mips/safe-syscall.inc.S | ||
32 | +++ b/common-user/host/mips/safe-syscall.inc.S | ||
33 | @@ -XXX,XX +XXX,XX @@ safe_syscall_end: | ||
34 | 1: USE_ALT_CP(t0) | ||
35 | SETUP_GPX(t1) | ||
36 | SETUP_GPX64(t0, t1) | ||
37 | + move a0, v0 | ||
38 | PTR_LA t9, safe_syscall_set_errno_tail | ||
39 | jr t9 | ||
40 | |||
41 | diff --git a/common-user/host/x86_64/safe-syscall.inc.S b/common-user/host/x86_64/safe-syscall.inc.S | ||
42 | index XXXXXXX..XXXXXXX 100644 | ||
43 | --- a/common-user/host/x86_64/safe-syscall.inc.S | ||
44 | +++ b/common-user/host/x86_64/safe-syscall.inc.S | ||
45 | @@ -XXX,XX +XXX,XX @@ safe_syscall_end: | ||
46 | 1: pop %rbp | ||
47 | .cfi_def_cfa_offset 8 | ||
48 | .cfi_restore rbp | ||
49 | + mov %eax, %edi | ||
50 | jmp safe_syscall_set_errno_tail | ||
51 | .cfi_endproc | ||
52 | |||
53 | -- | ||
54 | 2.25.1 | ||
55 | |||
56 | diff view generated by jsdifflib |