1 | The following changes since commit ff56877e911782dedc9a424233fd3f62369c258c: | 1 | With a couple of linux-user and target/sparc patches thrown in for good measure. |
---|---|---|---|
2 | 2 | ||
3 | Merge remote-tracking branch 'remotes/kraxel/tags/vga-20181015-pull-request' into staging (2018-10-15 15:03:45 +0100) | 3 | r~ |
4 | |||
5 | |||
6 | The following changes since commit 495de0fd82d8bb2d7035f82d9869cfeb48de2f9e: | ||
7 | |||
8 | Merge tag 'pull-trivial-patches' of https://gitlab.com/mjt0k/qemu into staging (2025-02-14 08:19:05 -0500) | ||
4 | 9 | ||
5 | are available in the Git repository at: | 10 | are available in the Git repository at: |
6 | 11 | ||
7 | https://github.com/rth7680/qemu.git tags/pull-tcg-20181016 | 12 | https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20250215 |
8 | 13 | ||
9 | for you to fetch changes up to e3e9d1ea20c75718ce7c528c588a0a497f12f750: | 14 | for you to fetch changes up to 2132751069134114814c7e1609e9cf644f077aad: |
10 | 15 | ||
11 | cputlb: read CPUTLBEntry.addr_write atomically (2018-10-16 10:04:27 -0700) | 16 | target/sparc: fake UltraSPARC T1 PCR and PIC registers (2025-02-15 12:04:13 -0800) |
12 | 17 | ||
13 | ---------------------------------------------------------------- | 18 | ---------------------------------------------------------------- |
14 | Queued tcg patches | 19 | tcg: Remove last traces of TCG_TARGET_NEED_POOL_LABELS |
20 | tcg: Cleanups after disallowing 64-on-32 | ||
21 | tcg: Introduce constraint for zero register | ||
22 | linux-user: Move TARGET_SA_RESTORER out of generic/signal.h | ||
23 | linux-user: Fix alignment when unmapping excess reservation | ||
24 | target/sparc: Fix register selection for all F*TOx and FxTO* instructions | ||
25 | target/sparc: Fix gdbstub incorrectly handling registers f32-f62 | ||
26 | target/sparc: fake UltraSPARC T1 PCR and PIC registers | ||
15 | 27 | ||
16 | ---------------------------------------------------------------- | 28 | ---------------------------------------------------------------- |
17 | Emilio G. Cota (10): | 29 | Andreas Schwab (1): |
18 | tcg: access cpu->icount_decr.u16.high with atomics | 30 | linux-user: Move TARGET_SA_RESTORER out of generic/signal.h |
19 | tcg: fix use of uninitialized variable under CONFIG_PROFILER | ||
20 | tcg: plug holes in struct TCGProfile | ||
21 | tcg: distribute tcg_time into TCG contexts | ||
22 | target/alpha: remove tlb_flush from alpha_cpu_initfn | ||
23 | target/unicore32: remove tlb_flush from uc32_init_fn | ||
24 | exec: introduce tlb_init | ||
25 | cputlb: fix assert_cpu_is_self macro | ||
26 | cputlb: serialize tlb updates with env->tlb_lock | ||
27 | cputlb: read CPUTLBEntry.addr_write atomically | ||
28 | 31 | ||
29 | Richard Henderson (11): | 32 | Artyom Tarasenko (1): |
30 | tcg: Implement CPU_LOG_TB_NOCHAIN during expansion | 33 | target/sparc: fake UltraSPARC T1 PCR and PIC registers |
31 | tcg: Add tlb_index and tlb_entry helpers | ||
32 | tcg: Split CONFIG_ATOMIC128 | ||
33 | target/i386: Convert to HAVE_CMPXCHG128 | ||
34 | target/arm: Convert to HAVE_CMPXCHG128 | ||
35 | target/arm: Check HAVE_CMPXCHG128 at translate time | ||
36 | target/ppc: Convert to HAVE_CMPXCHG128 and HAVE_ATOMIC128 | ||
37 | target/s390x: Convert to HAVE_CMPXCHG128 and HAVE_ATOMIC128 | ||
38 | target/s390x: Split do_cdsg, do_lpq, do_stpq | ||
39 | target/s390x: Skip wout, cout helpers if op helper does not return | ||
40 | target/s390x: Check HAVE_ATOMIC128 and HAVE_CMPXCHG128 at translate | ||
41 | 34 | ||
42 | accel/tcg/atomic_template.h | 20 +++- | 35 | Fabiano Rosas (1): |
43 | accel/tcg/softmmu_template.h | 64 +++++----- | 36 | elfload: Fix alignment when unmapping excess reservation |
44 | include/exec/cpu-defs.h | 3 + | ||
45 | include/exec/cpu_ldst.h | 30 ++++- | ||
46 | include/exec/cpu_ldst_template.h | 25 ++-- | ||
47 | include/exec/exec-all.h | 8 ++ | ||
48 | include/qemu/atomic128.h | 155 ++++++++++++++++++++++++ | ||
49 | include/qemu/timer.h | 1 - | ||
50 | target/ppc/helper.h | 2 +- | ||
51 | tcg/tcg.h | 20 ++-- | ||
52 | accel/tcg/cpu-exec.c | 2 +- | ||
53 | accel/tcg/cputlb.c | 235 +++++++++++++++++++----------------- | ||
54 | accel/tcg/tcg-all.c | 2 +- | ||
55 | accel/tcg/translate-all.c | 2 +- | ||
56 | accel/tcg/user-exec.c | 5 +- | ||
57 | cpus.c | 3 +- | ||
58 | exec.c | 1 + | ||
59 | monitor.c | 13 +- | ||
60 | qom/cpu.c | 2 +- | ||
61 | target/alpha/cpu.c | 1 - | ||
62 | target/arm/helper-a64.c | 251 +++++++++++++++++++-------------------- | ||
63 | target/arm/translate-a64.c | 38 +++--- | ||
64 | target/i386/mem_helper.c | 9 +- | ||
65 | target/ppc/mem_helper.c | 33 ++++- | ||
66 | target/ppc/translate.c | 115 +++++++++--------- | ||
67 | target/s390x/mem_helper.c | 202 +++++++++++++++---------------- | ||
68 | target/s390x/translate.c | 45 +++++-- | ||
69 | target/unicore32/cpu.c | 2 - | ||
70 | tcg/tcg-op.c | 9 +- | ||
71 | tcg/tcg.c | 25 +++- | ||
72 | configure | 19 +++ | ||
73 | 31 files changed, 830 insertions(+), 512 deletions(-) | ||
74 | create mode 100644 include/qemu/atomic128.h | ||
75 | 37 | ||
38 | Mikael Szreder (2): | ||
39 | target/sparc: Fix register selection for all F*TOx and FxTO* instructions | ||
40 | target/sparc: Fix gdbstub incorrectly handling registers f32-f62 | ||
41 | |||
42 | Richard Henderson (19): | ||
43 | tcg: Remove last traces of TCG_TARGET_NEED_POOL_LABELS | ||
44 | tcg: Remove TCG_OVERSIZED_GUEST | ||
45 | tcg: Drop support for two address registers in gen_ldst | ||
46 | tcg: Merge INDEX_op_qemu_*_{a32,a64}_* | ||
47 | tcg/arm: Drop addrhi from prepare_host_addr | ||
48 | tcg/i386: Drop addrhi from prepare_host_addr | ||
49 | tcg/mips: Drop addrhi from prepare_host_addr | ||
50 | tcg/ppc: Drop addrhi from prepare_host_addr | ||
51 | tcg: Replace addr{lo,hi}_reg with addr_reg in TCGLabelQemuLdst | ||
52 | plugins: Fix qemu_plugin_read_memory_vaddr parameters | ||
53 | accel/tcg: Fix tlb_set_page_with_attrs, tlb_set_page | ||
54 | include/exec: Change vaddr to uintptr_t | ||
55 | include/exec: Use uintptr_t in CPUTLBEntry | ||
56 | tcg: Introduce the 'z' constraint for a hardware zero register | ||
57 | tcg/aarch64: Use 'z' constraint | ||
58 | tcg/loongarch64: Use 'z' constraint | ||
59 | tcg/mips: Use 'z' constraint | ||
60 | tcg/riscv: Use 'z' constraint | ||
61 | tcg/sparc64: Use 'z' constraint | ||
62 | |||
63 | include/exec/tlb-common.h | 10 +- | ||
64 | include/exec/vaddr.h | 16 ++-- | ||
65 | include/qemu/atomic.h | 18 +--- | ||
66 | include/tcg/oversized-guest.h | 23 ----- | ||
67 | include/tcg/tcg-opc.h | 28 ++---- | ||
68 | include/tcg/tcg.h | 3 +- | ||
69 | linux-user/aarch64/target_signal.h | 2 + | ||
70 | linux-user/arm/target_signal.h | 2 + | ||
71 | linux-user/generic/signal.h | 1 - | ||
72 | linux-user/i386/target_signal.h | 2 + | ||
73 | linux-user/m68k/target_signal.h | 1 + | ||
74 | linux-user/microblaze/target_signal.h | 2 + | ||
75 | linux-user/ppc/target_signal.h | 2 + | ||
76 | linux-user/s390x/target_signal.h | 2 + | ||
77 | linux-user/sh4/target_signal.h | 2 + | ||
78 | linux-user/x86_64/target_signal.h | 2 + | ||
79 | linux-user/xtensa/target_signal.h | 2 + | ||
80 | tcg/aarch64/tcg-target-con-set.h | 12 +-- | ||
81 | tcg/aarch64/tcg-target.h | 2 + | ||
82 | tcg/loongarch64/tcg-target-con-set.h | 15 ++- | ||
83 | tcg/loongarch64/tcg-target-con-str.h | 1 - | ||
84 | tcg/loongarch64/tcg-target.h | 2 + | ||
85 | tcg/mips/tcg-target-con-set.h | 26 +++--- | ||
86 | tcg/mips/tcg-target-con-str.h | 1 - | ||
87 | tcg/mips/tcg-target.h | 2 + | ||
88 | tcg/riscv/tcg-target-con-set.h | 10 +- | ||
89 | tcg/riscv/tcg-target-con-str.h | 1 - | ||
90 | tcg/riscv/tcg-target.h | 2 + | ||
91 | tcg/sparc64/tcg-target-con-set.h | 12 +-- | ||
92 | tcg/sparc64/tcg-target-con-str.h | 1 - | ||
93 | tcg/sparc64/tcg-target.h | 3 +- | ||
94 | tcg/tci/tcg-target.h | 1 - | ||
95 | accel/tcg/cputlb.c | 32 ++----- | ||
96 | accel/tcg/tcg-all.c | 9 +- | ||
97 | linux-user/elfload.c | 4 +- | ||
98 | plugins/api.c | 2 +- | ||
99 | target/arm/ptw.c | 34 ------- | ||
100 | target/riscv/cpu_helper.c | 13 +-- | ||
101 | target/sparc/gdbstub.c | 18 +++- | ||
102 | target/sparc/translate.c | 19 ++++ | ||
103 | tcg/optimize.c | 21 ++--- | ||
104 | tcg/tcg-op-ldst.c | 103 +++++---------------- | ||
105 | tcg/tcg.c | 97 +++++++++---------- | ||
106 | tcg/tci.c | 119 +++++------------------- | ||
107 | docs/devel/multi-thread-tcg.rst | 1 - | ||
108 | docs/devel/tcg-ops.rst | 4 +- | ||
109 | target/sparc/insns.decode | 19 ++-- | ||
110 | tcg/aarch64/tcg-target.c.inc | 86 +++++++---------- | ||
111 | tcg/arm/tcg-target.c.inc | 104 ++++++--------------- | ||
112 | tcg/i386/tcg-target.c.inc | 125 +++++++------------------ | ||
113 | tcg/loongarch64/tcg-target.c.inc | 72 ++++++--------- | ||
114 | tcg/mips/tcg-target.c.inc | 169 +++++++++++----------------------- | ||
115 | tcg/ppc/tcg-target.c.inc | 164 ++++++++------------------------- | ||
116 | tcg/riscv/tcg-target.c.inc | 56 +++++------ | ||
117 | tcg/s390x/tcg-target.c.inc | 40 +++----- | ||
118 | tcg/sparc64/tcg-target.c.inc | 45 ++++----- | ||
119 | tcg/tci/tcg-target.c.inc | 60 +++--------- | ||
120 | 57 files changed, 536 insertions(+), 1089 deletions(-) | ||
121 | delete mode 100644 include/tcg/oversized-guest.h | diff view generated by jsdifflib |
1 | When op raises an exception, it may not have initialized the output | 1 | These should have been removed with the rest. There are |
---|---|---|---|
2 | temps that would be written back by wout or cout. | 2 | a couple of hosts which can emit guest_base into the |
3 | constant pool: aarch64, mips64, ppc64, riscv64. | ||
3 | 4 | ||
4 | Reviewed-by: David Hildenbrand <david@redhat.com> | 5 | Fixes: a417ef835058 ("tcg: Remove TCG_TARGET_NEED_LDST_LABELS and TCG_TARGET_NEED_POOL_LABELS") |
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
6 | --- | 7 | --- |
7 | target/s390x/translate.c | 20 +++++++++++++++----- | 8 | tcg/tci/tcg-target.h | 1 - |
8 | 1 file changed, 15 insertions(+), 5 deletions(-) | 9 | tcg/tcg.c | 4 ---- |
10 | 2 files changed, 5 deletions(-) | ||
9 | 11 | ||
10 | diff --git a/target/s390x/translate.c b/target/s390x/translate.c | 12 | diff --git a/tcg/tci/tcg-target.h b/tcg/tci/tcg-target.h |
11 | index XXXXXXX..XXXXXXX 100644 | 13 | index XXXXXXX..XXXXXXX 100644 |
12 | --- a/target/s390x/translate.c | 14 | --- a/tcg/tci/tcg-target.h |
13 | +++ b/target/s390x/translate.c | 15 | +++ b/tcg/tci/tcg-target.h |
14 | @@ -XXX,XX +XXX,XX @@ struct DisasInsn { | 16 | @@ -XXX,XX +XXX,XX @@ typedef enum { |
15 | 17 | } TCGReg; | |
16 | const char *name; | 18 | |
17 | 19 | #define HAVE_TCG_QEMU_TB_EXEC | |
18 | + /* Pre-process arguments before HELP_OP. */ | 20 | -#define TCG_TARGET_NEED_POOL_LABELS |
19 | void (*help_in1)(DisasContext *, DisasFields *, DisasOps *); | 21 | |
20 | void (*help_in2)(DisasContext *, DisasFields *, DisasOps *); | 22 | #endif /* TCG_TARGET_H */ |
21 | void (*help_prep)(DisasContext *, DisasFields *, DisasOps *); | 23 | diff --git a/tcg/tcg.c b/tcg/tcg.c |
22 | + | 24 | index XXXXXXX..XXXXXXX 100644 |
23 | + /* | 25 | --- a/tcg/tcg.c |
24 | + * Post-process output after HELP_OP. | 26 | +++ b/tcg/tcg.c |
25 | + * Note that these are not called if HELP_OP returns DISAS_NORETURN. | 27 | @@ -XXX,XX +XXX,XX @@ void tcg_prologue_init(void) |
26 | + */ | 28 | tcg_qemu_tb_exec = (tcg_prologue_fn *)tcg_splitwx_to_rx(s->code_ptr); |
27 | void (*help_wout)(DisasContext *, DisasFields *, DisasOps *); | 29 | #endif |
28 | void (*help_cout)(DisasContext *, DisasOps *); | 30 | |
29 | + | 31 | -#ifdef TCG_TARGET_NEED_POOL_LABELS |
30 | + /* Implement the operation itself. */ | 32 | s->pool_labels = NULL; |
31 | DisasJumpType (*help_op)(DisasContext *, DisasOps *); | 33 | -#endif |
32 | 34 | ||
33 | uint64_t data; | 35 | qemu_thread_jit_write(); |
34 | @@ -XXX,XX +XXX,XX @@ static DisasJumpType translate_one(CPUS390XState *env, DisasContext *s) | 36 | /* Generate the prologue. */ |
35 | if (insn->help_op) { | 37 | tcg_target_qemu_prologue(s); |
36 | ret = insn->help_op(s, &o); | 38 | |
39 | -#ifdef TCG_TARGET_NEED_POOL_LABELS | ||
40 | /* Allow the prologue to put e.g. guest_base into a pool entry. */ | ||
41 | { | ||
42 | int result = tcg_out_pool_finalize(s); | ||
43 | tcg_debug_assert(result == 0); | ||
37 | } | 44 | } |
38 | - if (insn->help_wout) { | 45 | -#endif |
39 | - insn->help_wout(s, &f, &o); | 46 | |
40 | - } | 47 | prologue_size = tcg_current_code_size(s); |
41 | - if (insn->help_cout) { | 48 | perf_report_prologue(s->code_gen_ptr, prologue_size); |
42 | - insn->help_cout(s, &o); | ||
43 | + if (ret != DISAS_NORETURN) { | ||
44 | + if (insn->help_wout) { | ||
45 | + insn->help_wout(s, &f, &o); | ||
46 | + } | ||
47 | + if (insn->help_cout) { | ||
48 | + insn->help_cout(s, &o); | ||
49 | + } | ||
50 | } | ||
51 | |||
52 | /* Free any temporaries created by the helpers. */ | ||
53 | -- | 49 | -- |
54 | 2.17.2 | 50 | 2.43.0 |
55 | |||
56 | diff view generated by jsdifflib |
1 | GCC7+ will no longer advertise support for 16-byte __atomic operations | 1 | This is now prohibited in configuration. |
---|---|---|---|
2 | if only cmpxchg is supported, as for x86_64. Fortunately, x86_64 still | 2 | |
3 | has support for __sync_compare_and_swap_16 and we can make use of that. | 3 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> |
4 | AArch64 does not have, nor ever has had such support, so open-code it. | ||
5 | |||
6 | Reviewed-by: Emilio G. Cota <cota@braap.org> | ||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
8 | --- | 5 | --- |
9 | accel/tcg/atomic_template.h | 20 ++++- | 6 | include/qemu/atomic.h | 18 +++-------------- |
10 | include/qemu/atomic128.h | 155 ++++++++++++++++++++++++++++++++++++ | 7 | include/tcg/oversized-guest.h | 23 ---------------------- |
11 | tcg/tcg.h | 16 ++-- | 8 | accel/tcg/cputlb.c | 7 ------- |
12 | accel/tcg/cputlb.c | 3 +- | 9 | accel/tcg/tcg-all.c | 9 ++++----- |
13 | accel/tcg/user-exec.c | 5 +- | 10 | target/arm/ptw.c | 34 --------------------------------- |
14 | configure | 19 +++++ | 11 | target/riscv/cpu_helper.c | 13 +------------ |
15 | 6 files changed, 204 insertions(+), 14 deletions(-) | 12 | docs/devel/multi-thread-tcg.rst | 1 - |
16 | create mode 100644 include/qemu/atomic128.h | 13 | 7 files changed, 8 insertions(+), 97 deletions(-) |
17 | 14 | delete mode 100644 include/tcg/oversized-guest.h | |
18 | diff --git a/accel/tcg/atomic_template.h b/accel/tcg/atomic_template.h | 15 | |
19 | index XXXXXXX..XXXXXXX 100644 | 16 | diff --git a/include/qemu/atomic.h b/include/qemu/atomic.h |
20 | --- a/accel/tcg/atomic_template.h | 17 | index XXXXXXX..XXXXXXX 100644 |
21 | +++ b/accel/tcg/atomic_template.h | 18 | --- a/include/qemu/atomic.h |
22 | @@ -XXX,XX +XXX,XX @@ ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr, | 19 | +++ b/include/qemu/atomic.h |
23 | DATA_TYPE ret; | 20 | @@ -XXX,XX +XXX,XX @@ |
24 | 21 | */ | |
25 | ATOMIC_TRACE_RMW; | 22 | #define signal_barrier() __atomic_signal_fence(__ATOMIC_SEQ_CST) |
26 | +#if DATA_SIZE == 16 | 23 | |
27 | + ret = atomic16_cmpxchg(haddr, cmpv, newv); | 24 | -/* Sanity check that the size of an atomic operation isn't "overly large". |
28 | +#else | 25 | +/* |
29 | ret = atomic_cmpxchg__nocheck(haddr, cmpv, newv); | 26 | + * Sanity check that the size of an atomic operation isn't "overly large". |
30 | +#endif | 27 | * Despite the fact that e.g. i686 has 64-bit atomic operations, we do not |
31 | ATOMIC_MMU_CLEANUP; | 28 | * want to use them because we ought not need them, and this lets us do a |
32 | return ret; | 29 | * bit of sanity checking that other 32-bit hosts might build. |
33 | } | 30 | - * |
34 | 31 | - * That said, we have a problem on 64-bit ILP32 hosts in that in order to | |
35 | #if DATA_SIZE >= 16 | 32 | - * sync with TCG_OVERSIZED_GUEST, this must match TCG_TARGET_REG_BITS. |
36 | +#if HAVE_ATOMIC128 | 33 | - * We'd prefer not want to pull in everything else TCG related, so handle |
37 | ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr EXTRA_ARGS) | 34 | - * those few cases by hand. |
38 | { | 35 | - * |
39 | ATOMIC_MMU_DECLS; | 36 | - * Note that x32 is fully detected with __x86_64__ + _ILP32, and that for |
40 | DATA_TYPE val, *haddr = ATOMIC_MMU_LOOKUP; | 37 | - * Sparc we always force the use of sparcv9 in configure. MIPS n32 (ILP32) & |
41 | 38 | - * n64 (LP64) ABIs are both detected using __mips64. | |
42 | ATOMIC_TRACE_LD; | 39 | */ |
43 | - __atomic_load(haddr, &val, __ATOMIC_RELAXED); | 40 | -#if defined(__x86_64__) || defined(__sparc__) || defined(__mips64) |
44 | + val = atomic16_read(haddr); | 41 | -# define ATOMIC_REG_SIZE 8 |
45 | ATOMIC_MMU_CLEANUP; | 42 | -#else |
46 | return val; | 43 | -# define ATOMIC_REG_SIZE sizeof(void *) |
47 | } | 44 | -#endif |
48 | @@ -XXX,XX +XXX,XX @@ void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr, | 45 | +#define ATOMIC_REG_SIZE sizeof(void *) |
49 | DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP; | 46 | |
50 | 47 | /* Weak atomic operations prevent the compiler moving other | |
51 | ATOMIC_TRACE_ST; | 48 | * loads/stores past the atomic operation load/store. However there is |
52 | - __atomic_store(haddr, &val, __ATOMIC_RELAXED); | 49 | diff --git a/include/tcg/oversized-guest.h b/include/tcg/oversized-guest.h |
53 | + atomic16_set(haddr, val); | 50 | deleted file mode 100644 |
54 | ATOMIC_MMU_CLEANUP; | ||
55 | } | ||
56 | +#endif | ||
57 | #else | ||
58 | ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr, | ||
59 | ABI_TYPE val EXTRA_ARGS) | ||
60 | @@ -XXX,XX +XXX,XX @@ ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr, | ||
61 | DATA_TYPE ret; | ||
62 | |||
63 | ATOMIC_TRACE_RMW; | ||
64 | +#if DATA_SIZE == 16 | ||
65 | + ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv)); | ||
66 | +#else | ||
67 | ret = atomic_cmpxchg__nocheck(haddr, BSWAP(cmpv), BSWAP(newv)); | ||
68 | +#endif | ||
69 | ATOMIC_MMU_CLEANUP; | ||
70 | return BSWAP(ret); | ||
71 | } | ||
72 | |||
73 | #if DATA_SIZE >= 16 | ||
74 | +#if HAVE_ATOMIC128 | ||
75 | ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr EXTRA_ARGS) | ||
76 | { | ||
77 | ATOMIC_MMU_DECLS; | ||
78 | DATA_TYPE val, *haddr = ATOMIC_MMU_LOOKUP; | ||
79 | |||
80 | ATOMIC_TRACE_LD; | ||
81 | - __atomic_load(haddr, &val, __ATOMIC_RELAXED); | ||
82 | + val = atomic16_read(haddr); | ||
83 | ATOMIC_MMU_CLEANUP; | ||
84 | return BSWAP(val); | ||
85 | } | ||
86 | @@ -XXX,XX +XXX,XX @@ void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr, | ||
87 | |||
88 | ATOMIC_TRACE_ST; | ||
89 | val = BSWAP(val); | ||
90 | - __atomic_store(haddr, &val, __ATOMIC_RELAXED); | ||
91 | + atomic16_set(haddr, val); | ||
92 | ATOMIC_MMU_CLEANUP; | ||
93 | } | ||
94 | +#endif | ||
95 | #else | ||
96 | ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr, | ||
97 | ABI_TYPE val EXTRA_ARGS) | ||
98 | diff --git a/include/qemu/atomic128.h b/include/qemu/atomic128.h | ||
99 | new file mode 100644 | ||
100 | index XXXXXXX..XXXXXXX | 51 | index XXXXXXX..XXXXXXX |
101 | --- /dev/null | 52 | --- a/include/tcg/oversized-guest.h |
102 | +++ b/include/qemu/atomic128.h | 53 | +++ /dev/null |
103 | @@ -XXX,XX +XXX,XX @@ | 54 | @@ -XXX,XX +XXX,XX @@ |
104 | +/* | 55 | -/* SPDX-License-Identifier: MIT */ |
105 | + * Simple interface for 128-bit atomic operations. | 56 | -/* |
106 | + * | 57 | - * Define TCG_OVERSIZED_GUEST |
107 | + * Copyright (C) 2018 Linaro, Ltd. | 58 | - * Copyright (c) 2008 Fabrice Bellard |
108 | + * | 59 | - */ |
109 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. | 60 | - |
110 | + * See the COPYING file in the top-level directory. | 61 | -#ifndef EXEC_TCG_OVERSIZED_GUEST_H |
111 | + * | 62 | -#define EXEC_TCG_OVERSIZED_GUEST_H |
112 | + * See docs/devel/atomics.txt for discussion about the guarantees each | 63 | - |
113 | + * atomic primitive is meant to provide. | 64 | -#include "tcg-target-reg-bits.h" |
114 | + */ | 65 | -#include "cpu-param.h" |
115 | + | 66 | - |
116 | +#ifndef QEMU_ATOMIC128_H | 67 | -/* |
117 | +#define QEMU_ATOMIC128_H | 68 | - * Oversized TCG guests make things like MTTCG hard |
118 | + | 69 | - * as we can't use atomics for cputlb updates. |
119 | +/* | 70 | - */ |
120 | + * GCC is a house divided about supporting large atomic operations. | 71 | -#if TARGET_LONG_BITS > TCG_TARGET_REG_BITS |
121 | + * | 72 | -#define TCG_OVERSIZED_GUEST 1 |
122 | + * For hosts that only have large compare-and-swap, a legalistic reading | 73 | -#else |
123 | + * of the C++ standard means that one cannot implement __atomic_read on | 74 | -#define TCG_OVERSIZED_GUEST 0 |
124 | + * read-only memory, and thus all atomic operations must synchronize | 75 | -#endif |
125 | + * through libatomic. | 76 | - |
126 | + * | 77 | -#endif |
127 | + * See https://gcc.gnu.org/bugzilla/show_bug.cgi?id=80878 | ||
128 | + * | ||
129 | + * This interpretation is not especially helpful for QEMU. | ||
130 | + * For softmmu, all RAM is always read/write from the hypervisor. | ||
131 | + * For user-only, if the guest doesn't implement such an __atomic_read | ||
132 | + * then the host need not worry about it either. | ||
133 | + * | ||
134 | + * Moreover, using libatomic is not an option, because its interface is | ||
135 | + * built for std::atomic<T>, and requires that *all* accesses to such an | ||
136 | + * object go through the library. In our case we do not have an object | ||
137 | + * in the C/C++ sense, but a view of memory as seen by the guest. | ||
138 | + * The guest may issue a large atomic operation and then access those | ||
139 | + * pieces using word-sized accesses. From the hypervisor, we have no | ||
140 | + * way to connect those two actions. | ||
141 | + * | ||
142 | + * Therefore, special case each platform. | ||
143 | + */ | ||
144 | + | ||
145 | +#if defined(CONFIG_ATOMIC128) | ||
146 | +static inline Int128 atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new) | ||
147 | +{ | ||
148 | + return atomic_cmpxchg__nocheck(ptr, cmp, new); | ||
149 | +} | ||
150 | +# define HAVE_CMPXCHG128 1 | ||
151 | +#elif defined(CONFIG_CMPXCHG128) | ||
152 | +static inline Int128 atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new) | ||
153 | +{ | ||
154 | + return __sync_val_compare_and_swap_16(ptr, cmp, new); | ||
155 | +} | ||
156 | +# define HAVE_CMPXCHG128 1 | ||
157 | +#elif defined(__aarch64__) | ||
158 | +/* Through gcc 8, aarch64 has no support for 128-bit at all. */ | ||
159 | +static inline Int128 atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new) | ||
160 | +{ | ||
161 | + uint64_t cmpl = int128_getlo(cmp), cmph = int128_gethi(cmp); | ||
162 | + uint64_t newl = int128_getlo(new), newh = int128_gethi(new); | ||
163 | + uint64_t oldl, oldh; | ||
164 | + uint32_t tmp; | ||
165 | + | ||
166 | + asm("0: ldaxp %[oldl], %[oldh], %[mem]\n\t" | ||
167 | + "cmp %[oldl], %[cmpl]\n\t" | ||
168 | + "ccmp %[oldh], %[cmph], #0, eq\n\t" | ||
169 | + "b.ne 1f\n\t" | ||
170 | + "stlxp %w[tmp], %[newl], %[newh], %[mem]\n\t" | ||
171 | + "cbnz %w[tmp], 0b\n" | ||
172 | + "1:" | ||
173 | + : [mem] "+m"(*ptr), [tmp] "=&r"(tmp), | ||
174 | + [oldl] "=&r"(oldl), [oldh] "=r"(oldh) | ||
175 | + : [cmpl] "r"(cmpl), [cmph] "r"(cmph), | ||
176 | + [newl] "r"(newl), [newh] "r"(newh) | ||
177 | + : "memory", "cc"); | ||
178 | + | ||
179 | + return int128_make128(oldl, oldh); | ||
180 | +} | ||
181 | +# define HAVE_CMPXCHG128 1 | ||
182 | +#else | ||
183 | +/* Fallback definition that must be optimized away, or error. */ | ||
184 | +Int128 __attribute__((error("unsupported atomic"))) | ||
185 | + atomic16_cmpxchg(Int128 *ptr, Int128 cmp, Int128 new); | ||
186 | +# define HAVE_CMPXCHG128 0 | ||
187 | +#endif /* Some definition for HAVE_CMPXCHG128 */ | ||
188 | + | ||
189 | + | ||
190 | +#if defined(CONFIG_ATOMIC128) | ||
191 | +static inline Int128 atomic16_read(Int128 *ptr) | ||
192 | +{ | ||
193 | + return atomic_read__nocheck(ptr); | ||
194 | +} | ||
195 | + | ||
196 | +static inline void atomic16_set(Int128 *ptr, Int128 val) | ||
197 | +{ | ||
198 | + atomic_set__nocheck(ptr, val); | ||
199 | +} | ||
200 | + | ||
201 | +# define HAVE_ATOMIC128 1 | ||
202 | +#elif !defined(CONFIG_USER_ONLY) && defined(__aarch64__) | ||
203 | +/* We can do better than cmpxchg for AArch64. */ | ||
204 | +static inline Int128 atomic16_read(Int128 *ptr) | ||
205 | +{ | ||
206 | + uint64_t l, h; | ||
207 | + uint32_t tmp; | ||
208 | + | ||
209 | + /* The load must be paired with the store to guarantee not tearing. */ | ||
210 | + asm("0: ldxp %[l], %[h], %[mem]\n\t" | ||
211 | + "stxp %w[tmp], %[l], %[h], %[mem]\n\t" | ||
212 | + "cbnz %w[tmp], 0b" | ||
213 | + : [mem] "+m"(*ptr), [tmp] "=r"(tmp), [l] "=r"(l), [h] "=r"(h)); | ||
214 | + | ||
215 | + return int128_make128(l, h); | ||
216 | +} | ||
217 | + | ||
218 | +static inline void atomic16_set(Int128 *ptr, Int128 val) | ||
219 | +{ | ||
220 | + uint64_t l = int128_getlo(val), h = int128_gethi(val); | ||
221 | + uint64_t t1, t2; | ||
222 | + | ||
223 | + /* Load into temporaries to acquire the exclusive access lock. */ | ||
224 | + asm("0: ldxp %[t1], %[t2], %[mem]\n\t" | ||
225 | + "stxp %w[t1], %[l], %[h], %[mem]\n\t" | ||
226 | + "cbnz %w[t1], 0b" | ||
227 | + : [mem] "+m"(*ptr), [t1] "=&r"(t1), [t2] "=&r"(t2) | ||
228 | + : [l] "r"(l), [h] "r"(h)); | ||
229 | +} | ||
230 | + | ||
231 | +# define HAVE_ATOMIC128 1 | ||
232 | +#elif !defined(CONFIG_USER_ONLY) && HAVE_CMPXCHG128 | ||
233 | +static inline Int128 atomic16_read(Int128 *ptr) | ||
234 | +{ | ||
235 | + /* Maybe replace 0 with 0, returning the old value. */ | ||
236 | + return atomic16_cmpxchg(ptr, 0, 0); | ||
237 | +} | ||
238 | + | ||
239 | +static inline void atomic16_set(Int128 *ptr, Int128 val) | ||
240 | +{ | ||
241 | + Int128 old = *ptr, cmp; | ||
242 | + do { | ||
243 | + cmp = old; | ||
244 | + old = atomic16_cmpxchg(ptr, cmp, val); | ||
245 | + } while (old != cmp); | ||
246 | +} | ||
247 | + | ||
248 | +# define HAVE_ATOMIC128 1 | ||
249 | +#else | ||
250 | +/* Fallback definitions that must be optimized away, or error. */ | ||
251 | +Int128 __attribute__((error("unsupported atomic"))) | ||
252 | + atomic16_read(Int128 *ptr); | ||
253 | +void __attribute__((error("unsupported atomic"))) | ||
254 | + atomic16_set(Int128 *ptr, Int128 val); | ||
255 | +# define HAVE_ATOMIC128 0 | ||
256 | +#endif /* Some definition for HAVE_ATOMIC128 */ | ||
257 | + | ||
258 | +#endif /* QEMU_ATOMIC128_H */ | ||
259 | diff --git a/tcg/tcg.h b/tcg/tcg.h | ||
260 | index XXXXXXX..XXXXXXX 100644 | ||
261 | --- a/tcg/tcg.h | ||
262 | +++ b/tcg/tcg.h | ||
263 | @@ -XXX,XX +XXX,XX @@ | ||
264 | #include "qemu/queue.h" | ||
265 | #include "tcg-mo.h" | ||
266 | #include "tcg-target.h" | ||
267 | +#include "qemu/int128.h" | ||
268 | |||
269 | /* XXX: make safe guess about sizes */ | ||
270 | #define MAX_OP_PER_INSTR 266 | ||
271 | @@ -XXX,XX +XXX,XX @@ GEN_ATOMIC_HELPER_ALL(xchg) | ||
272 | #undef GEN_ATOMIC_HELPER | ||
273 | #endif /* CONFIG_SOFTMMU */ | ||
274 | |||
275 | -#ifdef CONFIG_ATOMIC128 | ||
276 | -#include "qemu/int128.h" | ||
277 | - | ||
278 | -/* These aren't really a "proper" helpers because TCG cannot manage Int128. | ||
279 | - However, use the same format as the others, for use by the backends. */ | ||
280 | +/* | ||
281 | + * These aren't really a "proper" helpers because TCG cannot manage Int128. | ||
282 | + * However, use the same format as the others, for use by the backends. | ||
283 | + * | ||
284 | + * The cmpxchg functions are only defined if HAVE_CMPXCHG128; | ||
285 | + * the ld/st functions are only defined if HAVE_ATOMIC128, | ||
286 | + * as defined by <qemu/atomic128.h>. | ||
287 | + */ | ||
288 | Int128 helper_atomic_cmpxchgo_le_mmu(CPUArchState *env, target_ulong addr, | ||
289 | Int128 cmpv, Int128 newv, | ||
290 | TCGMemOpIdx oi, uintptr_t retaddr); | ||
291 | @@ -XXX,XX +XXX,XX @@ void helper_atomic_sto_le_mmu(CPUArchState *env, target_ulong addr, Int128 val, | ||
292 | void helper_atomic_sto_be_mmu(CPUArchState *env, target_ulong addr, Int128 val, | ||
293 | TCGMemOpIdx oi, uintptr_t retaddr); | ||
294 | |||
295 | -#endif /* CONFIG_ATOMIC128 */ | ||
296 | - | ||
297 | #endif /* TCG_H */ | ||
298 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | 78 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c |
299 | index XXXXXXX..XXXXXXX 100644 | 79 | index XXXXXXX..XXXXXXX 100644 |
300 | --- a/accel/tcg/cputlb.c | 80 | --- a/accel/tcg/cputlb.c |
301 | +++ b/accel/tcg/cputlb.c | 81 | +++ b/accel/tcg/cputlb.c |
302 | @@ -XXX,XX +XXX,XX @@ | 82 | @@ -XXX,XX +XXX,XX @@ |
303 | #include "exec/log.h" | 83 | #include "qemu/plugin-memory.h" |
304 | #include "exec/helper-proto.h" | 84 | #endif |
305 | #include "qemu/atomic.h" | 85 | #include "tcg/tcg-ldst.h" |
306 | +#include "qemu/atomic128.h" | 86 | -#include "tcg/oversized-guest.h" |
307 | 87 | ||
308 | /* DEBUG defines, enable DEBUG_TLB_LOG to log to the CPU_LOG_MMU target */ | 88 | /* DEBUG defines, enable DEBUG_TLB_LOG to log to the CPU_LOG_MMU target */ |
309 | /* #define DEBUG_TLB */ | 89 | /* #define DEBUG_TLB */ |
310 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | 90 | @@ -XXX,XX +XXX,XX @@ static inline uint64_t tlb_read_idx(const CPUTLBEntry *entry, |
311 | #include "atomic_template.h" | 91 | return qatomic_read(ptr); |
92 | #else | ||
93 | const uint64_t *ptr = &entry->addr_idx[access_type]; | ||
94 | -# if TCG_OVERSIZED_GUEST | ||
95 | - return *ptr; | ||
96 | -# else | ||
97 | /* ofs might correspond to .addr_write, so use qatomic_read */ | ||
98 | return qatomic_read(ptr); | ||
99 | -# endif | ||
312 | #endif | 100 | #endif |
313 | 101 | } | |
314 | -#ifdef CONFIG_ATOMIC128 | 102 | |
315 | +#if HAVE_CMPXCHG128 || HAVE_ATOMIC128 | 103 | @@ -XXX,XX +XXX,XX @@ static void tlb_reset_dirty_range_locked(CPUTLBEntry *tlb_entry, |
316 | #define DATA_SIZE 16 | 104 | uint32_t *ptr_write = (uint32_t *)&tlb_entry->addr_write; |
317 | #include "atomic_template.h" | 105 | ptr_write += HOST_BIG_ENDIAN; |
106 | qatomic_set(ptr_write, *ptr_write | TLB_NOTDIRTY); | ||
107 | -#elif TCG_OVERSIZED_GUEST | ||
108 | - tlb_entry->addr_write |= TLB_NOTDIRTY; | ||
109 | #else | ||
110 | qatomic_set(&tlb_entry->addr_write, | ||
111 | tlb_entry->addr_write | TLB_NOTDIRTY); | ||
112 | diff --git a/accel/tcg/tcg-all.c b/accel/tcg/tcg-all.c | ||
113 | index XXXXXXX..XXXXXXX 100644 | ||
114 | --- a/accel/tcg/tcg-all.c | ||
115 | +++ b/accel/tcg/tcg-all.c | ||
116 | @@ -XXX,XX +XXX,XX @@ | ||
117 | #include "exec/replay-core.h" | ||
118 | #include "system/cpu-timers.h" | ||
119 | #include "tcg/startup.h" | ||
120 | -#include "tcg/oversized-guest.h" | ||
121 | #include "qapi/error.h" | ||
122 | #include "qemu/error-report.h" | ||
123 | #include "qemu/accel.h" | ||
124 | @@ -XXX,XX +XXX,XX @@ | ||
125 | #include "hw/boards.h" | ||
318 | #endif | 126 | #endif |
319 | diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c | 127 | #include "internal-common.h" |
320 | index XXXXXXX..XXXXXXX 100644 | 128 | +#include "cpu-param.h" |
321 | --- a/accel/tcg/user-exec.c | ||
322 | +++ b/accel/tcg/user-exec.c | ||
323 | @@ -XXX,XX +XXX,XX @@ | ||
324 | #include "exec/cpu_ldst.h" | ||
325 | #include "translate-all.h" | ||
326 | #include "exec/helper-proto.h" | ||
327 | +#include "qemu/atomic128.h" | ||
328 | |||
329 | #undef EAX | ||
330 | #undef ECX | ||
331 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | ||
332 | /* The following is only callable from other helpers, and matches up | ||
333 | with the softmmu version. */ | ||
334 | |||
335 | -#ifdef CONFIG_ATOMIC128 | ||
336 | +#if HAVE_ATOMIC128 || HAVE_CMPXCHG128 | ||
337 | |||
338 | #undef EXTRA_ARGS | ||
339 | #undef ATOMIC_NAME | ||
340 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | ||
341 | |||
342 | #define DATA_SIZE 16 | ||
343 | #include "atomic_template.h" | ||
344 | -#endif /* CONFIG_ATOMIC128 */ | ||
345 | +#endif | ||
346 | diff --git a/configure b/configure | ||
347 | index XXXXXXX..XXXXXXX 100755 | ||
348 | --- a/configure | ||
349 | +++ b/configure | ||
350 | @@ -XXX,XX +XXX,XX @@ EOF | ||
351 | fi | ||
352 | fi | ||
353 | |||
354 | +cmpxchg128=no | ||
355 | +if test "$int128" = yes -a "$atomic128" = no; then | ||
356 | + cat > $TMPC << EOF | ||
357 | +int main(void) | ||
358 | +{ | ||
359 | + unsigned __int128 x = 0, y = 0; | ||
360 | + __sync_val_compare_and_swap_16(&x, y, x); | ||
361 | + return 0; | ||
362 | +} | ||
363 | +EOF | ||
364 | + if compile_prog "" "" ; then | ||
365 | + cmpxchg128=yes | ||
366 | + fi | ||
367 | +fi | ||
368 | + | 129 | + |
369 | ######################################### | 130 | |
370 | # See if 64-bit atomic operations are supported. | 131 | struct TCGState { |
371 | # Note that without __atomic builtins, we can only | 132 | AccelState parent_obj; |
372 | @@ -XXX,XX +XXX,XX @@ if test "$atomic128" = "yes" ; then | 133 | @@ -XXX,XX +XXX,XX @@ DECLARE_INSTANCE_CHECKER(TCGState, TCG_STATE, |
373 | echo "CONFIG_ATOMIC128=y" >> $config_host_mak | 134 | |
374 | fi | 135 | static bool default_mttcg_enabled(void) |
375 | 136 | { | |
376 | +if test "$cmpxchg128" = "yes" ; then | 137 | - if (icount_enabled() || TCG_OVERSIZED_GUEST) { |
377 | + echo "CONFIG_CMPXCHG128=y" >> $config_host_mak | 138 | + if (icount_enabled()) { |
378 | +fi | 139 | return false; |
379 | + | 140 | } |
380 | if test "$atomic64" = "yes" ; then | 141 | #ifdef TARGET_SUPPORTS_MTTCG |
381 | echo "CONFIG_ATOMIC64=y" >> $config_host_mak | 142 | @@ -XXX,XX +XXX,XX @@ static void tcg_set_thread(Object *obj, const char *value, Error **errp) |
382 | fi | 143 | TCGState *s = TCG_STATE(obj); |
144 | |||
145 | if (strcmp(value, "multi") == 0) { | ||
146 | - if (TCG_OVERSIZED_GUEST) { | ||
147 | - error_setg(errp, "No MTTCG when guest word size > hosts"); | ||
148 | - } else if (icount_enabled()) { | ||
149 | + if (icount_enabled()) { | ||
150 | error_setg(errp, "No MTTCG when icount is enabled"); | ||
151 | } else { | ||
152 | #ifndef TARGET_SUPPORTS_MTTCG | ||
153 | diff --git a/target/arm/ptw.c b/target/arm/ptw.c | ||
154 | index XXXXXXX..XXXXXXX 100644 | ||
155 | --- a/target/arm/ptw.c | ||
156 | +++ b/target/arm/ptw.c | ||
157 | @@ -XXX,XX +XXX,XX @@ | ||
158 | #include "internals.h" | ||
159 | #include "cpu-features.h" | ||
160 | #include "idau.h" | ||
161 | -#ifdef CONFIG_TCG | ||
162 | -# include "tcg/oversized-guest.h" | ||
163 | -#endif | ||
164 | |||
165 | typedef struct S1Translate { | ||
166 | /* | ||
167 | @@ -XXX,XX +XXX,XX @@ static uint64_t arm_casq_ptw(CPUARMState *env, uint64_t old_val, | ||
168 | ptw->out_rw = true; | ||
169 | } | ||
170 | |||
171 | -#ifdef CONFIG_ATOMIC64 | ||
172 | if (ptw->out_be) { | ||
173 | old_val = cpu_to_be64(old_val); | ||
174 | new_val = cpu_to_be64(new_val); | ||
175 | @@ -XXX,XX +XXX,XX @@ static uint64_t arm_casq_ptw(CPUARMState *env, uint64_t old_val, | ||
176 | cur_val = qatomic_cmpxchg__nocheck((uint64_t *)host, old_val, new_val); | ||
177 | cur_val = le64_to_cpu(cur_val); | ||
178 | } | ||
179 | -#else | ||
180 | - /* | ||
181 | - * We can't support the full 64-bit atomic cmpxchg on the host. | ||
182 | - * Because this is only used for FEAT_HAFDBS, which is only for AA64, | ||
183 | - * we know that TCG_OVERSIZED_GUEST is set, which means that we are | ||
184 | - * running in round-robin mode and could only race with dma i/o. | ||
185 | - */ | ||
186 | -#if !TCG_OVERSIZED_GUEST | ||
187 | -# error "Unexpected configuration" | ||
188 | -#endif | ||
189 | - bool locked = bql_locked(); | ||
190 | - if (!locked) { | ||
191 | - bql_lock(); | ||
192 | - } | ||
193 | - if (ptw->out_be) { | ||
194 | - cur_val = ldq_be_p(host); | ||
195 | - if (cur_val == old_val) { | ||
196 | - stq_be_p(host, new_val); | ||
197 | - } | ||
198 | - } else { | ||
199 | - cur_val = ldq_le_p(host); | ||
200 | - if (cur_val == old_val) { | ||
201 | - stq_le_p(host, new_val); | ||
202 | - } | ||
203 | - } | ||
204 | - if (!locked) { | ||
205 | - bql_unlock(); | ||
206 | - } | ||
207 | -#endif | ||
208 | - | ||
209 | return cur_val; | ||
210 | #else | ||
211 | /* AArch32 does not have FEAT_HADFS; non-TCG guests only use debug-mode. */ | ||
212 | diff --git a/target/riscv/cpu_helper.c b/target/riscv/cpu_helper.c | ||
213 | index XXXXXXX..XXXXXXX 100644 | ||
214 | --- a/target/riscv/cpu_helper.c | ||
215 | +++ b/target/riscv/cpu_helper.c | ||
216 | @@ -XXX,XX +XXX,XX @@ | ||
217 | #include "system/cpu-timers.h" | ||
218 | #include "cpu_bits.h" | ||
219 | #include "debug.h" | ||
220 | -#include "tcg/oversized-guest.h" | ||
221 | #include "pmp.h" | ||
222 | |||
223 | int riscv_env_mmu_index(CPURISCVState *env, bool ifetch) | ||
224 | @@ -XXX,XX +XXX,XX @@ static int get_physical_address(CPURISCVState *env, hwaddr *physical, | ||
225 | hwaddr pte_addr; | ||
226 | int i; | ||
227 | |||
228 | -#if !TCG_OVERSIZED_GUEST | ||
229 | -restart: | ||
230 | -#endif | ||
231 | + restart: | ||
232 | for (i = 0; i < levels; i++, ptshift -= ptidxbits) { | ||
233 | target_ulong idx; | ||
234 | if (i == 0) { | ||
235 | @@ -XXX,XX +XXX,XX @@ restart: | ||
236 | false, MEMTXATTRS_UNSPECIFIED); | ||
237 | if (memory_region_is_ram(mr)) { | ||
238 | target_ulong *pte_pa = qemu_map_ram_ptr(mr->ram_block, addr1); | ||
239 | -#if TCG_OVERSIZED_GUEST | ||
240 | - /* | ||
241 | - * MTTCG is not enabled on oversized TCG guests so | ||
242 | - * page table updates do not need to be atomic | ||
243 | - */ | ||
244 | - *pte_pa = pte = updated_pte; | ||
245 | -#else | ||
246 | target_ulong old_pte; | ||
247 | if (riscv_cpu_sxl(env) == MXL_RV32) { | ||
248 | old_pte = qatomic_cmpxchg((uint32_t *)pte_pa, pte, updated_pte); | ||
249 | @@ -XXX,XX +XXX,XX @@ restart: | ||
250 | goto restart; | ||
251 | } | ||
252 | pte = updated_pte; | ||
253 | -#endif | ||
254 | } else { | ||
255 | /* | ||
256 | * Misconfigured PTE in ROM (AD bits are not preset) or | ||
257 | diff --git a/docs/devel/multi-thread-tcg.rst b/docs/devel/multi-thread-tcg.rst | ||
258 | index XXXXXXX..XXXXXXX 100644 | ||
259 | --- a/docs/devel/multi-thread-tcg.rst | ||
260 | +++ b/docs/devel/multi-thread-tcg.rst | ||
261 | @@ -XXX,XX +XXX,XX @@ if: | ||
262 | |||
263 | * forced by --accel tcg,thread=single | ||
264 | * enabling --icount mode | ||
265 | -* 64 bit guests on 32 bit hosts (TCG_OVERSIZED_GUEST) | ||
266 | |||
267 | In the general case of running translated code there should be no | ||
268 | inter-vCPU dependencies and all vCPUs should be able to run at full | ||
383 | -- | 269 | -- |
384 | 2.17.2 | 270 | 2.43.0 |
385 | 271 | ||
386 | 272 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
---|---|---|---|
2 | |||
3 | When we implemented per-vCPU TCG contexts, we forgot to also | ||
4 | distribute the tcg_time counter, which has remained as a global | ||
5 | accessed without any serialization, leading to potentially missed | ||
6 | counts. | ||
7 | |||
8 | Fix it by distributing the field over the TCG contexts, embedding | ||
9 | it into TCGProfile with a field called "cpu_exec_time", which is more | ||
10 | descriptive than "tcg_time". Add a function to query this value | ||
11 | directly, and for completeness, fill in the field in | ||
12 | tcg_profile_snapshot, even though its callers do not use it. | ||
13 | |||
14 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
15 | Message-Id: <20181010144853.13005-5-cota@braap.org> | ||
16 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
17 | --- | 3 | --- |
18 | include/qemu/timer.h | 1 - | 4 | tcg/tcg-op-ldst.c | 21 +++------------------ |
19 | tcg/tcg.h | 2 ++ | 5 | tcg/tcg.c | 4 +--- |
20 | cpus.c | 3 ++- | 6 | 2 files changed, 4 insertions(+), 21 deletions(-) |
21 | monitor.c | 13 ++++++++++--- | ||
22 | tcg/tcg.c | 23 +++++++++++++++++++++++ | ||
23 | 5 files changed, 37 insertions(+), 5 deletions(-) | ||
24 | 7 | ||
25 | diff --git a/include/qemu/timer.h b/include/qemu/timer.h | 8 | diff --git a/tcg/tcg-op-ldst.c b/tcg/tcg-op-ldst.c |
26 | index XXXXXXX..XXXXXXX 100644 | 9 | index XXXXXXX..XXXXXXX 100644 |
27 | --- a/include/qemu/timer.h | 10 | --- a/tcg/tcg-op-ldst.c |
28 | +++ b/include/qemu/timer.h | 11 | +++ b/tcg/tcg-op-ldst.c |
29 | @@ -XXX,XX +XXX,XX @@ static inline int64_t profile_getclock(void) | 12 | @@ -XXX,XX +XXX,XX @@ static MemOp tcg_canonicalize_memop(MemOp op, bool is64, bool st) |
30 | return get_clock(); | 13 | static void gen_ldst(TCGOpcode opc, TCGType type, TCGTemp *vl, TCGTemp *vh, |
14 | TCGTemp *addr, MemOpIdx oi) | ||
15 | { | ||
16 | - if (TCG_TARGET_REG_BITS == 64 || tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
17 | - if (vh) { | ||
18 | - tcg_gen_op4(opc, type, temp_arg(vl), temp_arg(vh), | ||
19 | - temp_arg(addr), oi); | ||
20 | - } else { | ||
21 | - tcg_gen_op3(opc, type, temp_arg(vl), temp_arg(addr), oi); | ||
22 | - } | ||
23 | + if (vh) { | ||
24 | + tcg_gen_op4(opc, type, temp_arg(vl), temp_arg(vh), temp_arg(addr), oi); | ||
25 | } else { | ||
26 | - /* See TCGV_LOW/HIGH. */ | ||
27 | - TCGTemp *al = addr + HOST_BIG_ENDIAN; | ||
28 | - TCGTemp *ah = addr + !HOST_BIG_ENDIAN; | ||
29 | - | ||
30 | - if (vh) { | ||
31 | - tcg_gen_op5(opc, type, temp_arg(vl), temp_arg(vh), | ||
32 | - temp_arg(al), temp_arg(ah), oi); | ||
33 | - } else { | ||
34 | - tcg_gen_op4(opc, type, temp_arg(vl), | ||
35 | - temp_arg(al), temp_arg(ah), oi); | ||
36 | - } | ||
37 | + tcg_gen_op3(opc, type, temp_arg(vl), temp_arg(addr), oi); | ||
38 | } | ||
31 | } | 39 | } |
32 | 40 | ||
33 | -extern int64_t tcg_time; | ||
34 | extern int64_t dev_time; | ||
35 | #endif | ||
36 | |||
37 | diff --git a/tcg/tcg.h b/tcg/tcg.h | ||
38 | index XXXXXXX..XXXXXXX 100644 | ||
39 | --- a/tcg/tcg.h | ||
40 | +++ b/tcg/tcg.h | ||
41 | @@ -XXX,XX +XXX,XX @@ typedef struct TCGOp { | ||
42 | QEMU_BUILD_BUG_ON(NB_OPS > (1 << 8)); | ||
43 | |||
44 | typedef struct TCGProfile { | ||
45 | + int64_t cpu_exec_time; | ||
46 | int64_t tb_count1; | ||
47 | int64_t tb_count; | ||
48 | int64_t op_count; /* total insn count */ | ||
49 | @@ -XXX,XX +XXX,XX @@ int tcg_check_temp_count(void); | ||
50 | #define tcg_check_temp_count() 0 | ||
51 | #endif | ||
52 | |||
53 | +int64_t tcg_cpu_exec_time(void); | ||
54 | void tcg_dump_info(FILE *f, fprintf_function cpu_fprintf); | ||
55 | void tcg_dump_op_count(FILE *f, fprintf_function cpu_fprintf); | ||
56 | |||
57 | diff --git a/cpus.c b/cpus.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | ||
59 | --- a/cpus.c | ||
60 | +++ b/cpus.c | ||
61 | @@ -XXX,XX +XXX,XX @@ static int tcg_cpu_exec(CPUState *cpu) | ||
62 | ret = cpu_exec(cpu); | ||
63 | cpu_exec_end(cpu); | ||
64 | #ifdef CONFIG_PROFILER | ||
65 | - tcg_time += profile_getclock() - ti; | ||
66 | + atomic_set(&tcg_ctx->prof.cpu_exec_time, | ||
67 | + tcg_ctx->prof.cpu_exec_time + profile_getclock() - ti); | ||
68 | #endif | ||
69 | return ret; | ||
70 | } | ||
71 | diff --git a/monitor.c b/monitor.c | ||
72 | index XXXXXXX..XXXXXXX 100644 | ||
73 | --- a/monitor.c | ||
74 | +++ b/monitor.c | ||
75 | @@ -XXX,XX +XXX,XX @@ | ||
76 | #include "sysemu/cpus.h" | ||
77 | #include "sysemu/iothread.h" | ||
78 | #include "qemu/cutils.h" | ||
79 | +#include "tcg/tcg.h" | ||
80 | |||
81 | #if defined(TARGET_S390X) | ||
82 | #include "hw/s390x/storage-keys.h" | ||
83 | @@ -XXX,XX +XXX,XX @@ static void hmp_info_numa(Monitor *mon, const QDict *qdict) | ||
84 | |||
85 | #ifdef CONFIG_PROFILER | ||
86 | |||
87 | -int64_t tcg_time; | ||
88 | int64_t dev_time; | ||
89 | |||
90 | static void hmp_info_profile(Monitor *mon, const QDict *qdict) | ||
91 | { | ||
92 | + static int64_t last_cpu_exec_time; | ||
93 | + int64_t cpu_exec_time; | ||
94 | + int64_t delta; | ||
95 | + | ||
96 | + cpu_exec_time = tcg_cpu_exec_time(); | ||
97 | + delta = cpu_exec_time - last_cpu_exec_time; | ||
98 | + | ||
99 | monitor_printf(mon, "async time %" PRId64 " (%0.3f)\n", | ||
100 | dev_time, dev_time / (double)NANOSECONDS_PER_SECOND); | ||
101 | monitor_printf(mon, "qemu time %" PRId64 " (%0.3f)\n", | ||
102 | - tcg_time, tcg_time / (double)NANOSECONDS_PER_SECOND); | ||
103 | - tcg_time = 0; | ||
104 | + delta, delta / (double)NANOSECONDS_PER_SECOND); | ||
105 | + last_cpu_exec_time = cpu_exec_time; | ||
106 | dev_time = 0; | ||
107 | } | ||
108 | #else | ||
109 | diff --git a/tcg/tcg.c b/tcg/tcg.c | 41 | diff --git a/tcg/tcg.c b/tcg/tcg.c |
110 | index XXXXXXX..XXXXXXX 100644 | 42 | index XXXXXXX..XXXXXXX 100644 |
111 | --- a/tcg/tcg.c | 43 | --- a/tcg/tcg.c |
112 | +++ b/tcg/tcg.c | 44 | +++ b/tcg/tcg.c |
113 | @@ -XXX,XX +XXX,XX @@ | 45 | @@ -XXX,XX +XXX,XX @@ void tcg_func_start(TCGContext *s) |
114 | /* Define to jump the ELF file used to communicate with GDB. */ | 46 | s->emit_before_op = NULL; |
115 | #undef DEBUG_JIT | 47 | QSIMPLEQ_INIT(&s->labels); |
116 | 48 | ||
117 | +#include "qemu/error-report.h" | 49 | - tcg_debug_assert(s->addr_type == TCG_TYPE_I32 || |
118 | #include "qemu/cutils.h" | 50 | - s->addr_type == TCG_TYPE_I64); |
119 | #include "qemu/host-utils.h" | 51 | - |
120 | #include "qemu/timer.h" | 52 | + tcg_debug_assert(s->addr_type <= TCG_TYPE_REG); |
121 | @@ -XXX,XX +XXX,XX @@ void tcg_profile_snapshot(TCGProfile *prof, bool counters, bool table) | 53 | tcg_debug_assert(s->insn_start_words > 0); |
122 | const TCGProfile *orig = &s->prof; | ||
123 | |||
124 | if (counters) { | ||
125 | + PROF_ADD(prof, orig, cpu_exec_time); | ||
126 | PROF_ADD(prof, orig, tb_count1); | ||
127 | PROF_ADD(prof, orig, tb_count); | ||
128 | PROF_ADD(prof, orig, op_count); | ||
129 | @@ -XXX,XX +XXX,XX @@ void tcg_dump_op_count(FILE *f, fprintf_function cpu_fprintf) | ||
130 | prof.table_op_count[i]); | ||
131 | } | ||
132 | } | 54 | } |
133 | + | ||
134 | +int64_t tcg_cpu_exec_time(void) | ||
135 | +{ | ||
136 | + unsigned int n_ctxs = atomic_read(&n_tcg_ctxs); | ||
137 | + unsigned int i; | ||
138 | + int64_t ret = 0; | ||
139 | + | ||
140 | + for (i = 0; i < n_ctxs; i++) { | ||
141 | + const TCGContext *s = atomic_read(&tcg_ctxs[i]); | ||
142 | + const TCGProfile *prof = &s->prof; | ||
143 | + | ||
144 | + ret += atomic_read(&prof->cpu_exec_time); | ||
145 | + } | ||
146 | + return ret; | ||
147 | +} | ||
148 | #else | ||
149 | void tcg_dump_op_count(FILE *f, fprintf_function cpu_fprintf) | ||
150 | { | ||
151 | cpu_fprintf(f, "[TCG profiler not compiled]\n"); | ||
152 | } | ||
153 | + | ||
154 | +int64_t tcg_cpu_exec_time(void) | ||
155 | +{ | ||
156 | + error_report("%s: TCG profiler not compiled", __func__); | ||
157 | + exit(EXIT_FAILURE); | ||
158 | +} | ||
159 | #endif | ||
160 | |||
161 | 55 | ||
162 | -- | 56 | -- |
163 | 2.17.2 | 57 | 2.43.0 |
164 | 58 | ||
165 | 59 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Since 64-on-32 is now unsupported, guest addresses always |
---|---|---|---|
2 | fit in one host register. Drop the replication of opcodes. | ||
2 | 3 | ||
3 | Updates can come from other threads, so readers that do not | ||
4 | take tlb_lock must use atomic_read to avoid undefined | ||
5 | behaviour (UB). | ||
6 | |||
7 | This completes the conversion to tlb_lock. This conversion results | ||
8 | on average in no performance loss, as the following experiments | ||
9 | (run on an Intel i7-6700K CPU @ 4.00GHz) show. | ||
10 | |||
11 | 1. aarch64 bootup+shutdown test: | ||
12 | |||
13 | - Before: | ||
14 | Performance counter stats for 'taskset -c 0 ../img/aarch64/die.sh' (10 runs): | ||
15 | |||
16 | 7487.087786 task-clock (msec) # 0.998 CPUs utilized ( +- 0.12% ) | ||
17 | 31,574,905,303 cycles # 4.217 GHz ( +- 0.12% ) | ||
18 | 57,097,908,812 instructions # 1.81 insns per cycle ( +- 0.08% ) | ||
19 | 10,255,415,367 branches # 1369.747 M/sec ( +- 0.08% ) | ||
20 | 173,278,962 branch-misses # 1.69% of all branches ( +- 0.18% ) | ||
21 | |||
22 | 7.504481349 seconds time elapsed ( +- 0.14% ) | ||
23 | |||
24 | - After: | ||
25 | Performance counter stats for 'taskset -c 0 ../img/aarch64/die.sh' (10 runs): | ||
26 | |||
27 | 7462.441328 task-clock (msec) # 0.998 CPUs utilized ( +- 0.07% ) | ||
28 | 31,478,476,520 cycles # 4.218 GHz ( +- 0.07% ) | ||
29 | 57,017,330,084 instructions # 1.81 insns per cycle ( +- 0.05% ) | ||
30 | 10,251,929,667 branches # 1373.804 M/sec ( +- 0.05% ) | ||
31 | 173,023,787 branch-misses # 1.69% of all branches ( +- 0.11% ) | ||
32 | |||
33 | 7.474970463 seconds time elapsed ( +- 0.07% ) | ||
34 | |||
35 | 2. SPEC06int: | ||
36 | SPEC06int (test set) | ||
37 | [Y axis: Speedup over master] | ||
38 | 1.15 +-+----+------+------+------+------+------+-------+------+------+------+------+------+------+----+-+ | ||
39 | | | | ||
40 | 1.1 +-+.................................+++.............................+ tlb-lock-v2 (m+++x) +-+ | ||
41 | | +++ | +++ tlb-lock-v3 (spinl|ck) | | ||
42 | | +++ | | +++ +++ | | | | ||
43 | 1.05 +-+....+++...........####.........|####.+++.|......|.....###....+++...........+++....###.........+-+ | ||
44 | | ### ++#| # |# |# ***### +++### +++#+# | +++ | #|# ### | | ||
45 | 1 +-+++***+#++++####+++#++#++++++++++#++#+*+*++#++++#+#+****+#++++###++++###++++###++++#+#++++#+#+++-+ | ||
46 | | *+* # #++# *** # #### *** # * *++# ****+# *| * # ****|# |# # #|# #+# # # | | ||
47 | 0.95 +-+..*.*.#....#..#.*|*..#...#..#.*|*..#.*.*..#.*|.*.#.*++*.#.*++*+#.****.#....#+#....#.#..++#.#..+-+ | ||
48 | | * * # # # *|* # # # *|* # * * # *++* # * * # * * # * |* # ++# # # # *** # | | ||
49 | | * * # ++# # *+* # # # *|* # * * # * * # * * # * * # *++* # **** # ++# # * * # | | ||
50 | 0.9 +-+..*.*.#...|#..#.*.*..#.++#..#.*|*..#.*.*..#.*..*.#.*..*.#.*..*.#.*..*.#.*.|*.#...|#.#..*.*.#..+-+ | ||
51 | | * * # *** # * * # |# # *+* # * * # * * # * * # * * # * * # *++* # |# # * * # | | ||
52 | 0.85 +-+..*.*.#..*|*..#.*.*..#.***..#.*.*..#.*.*..#.*..*.#.*..*.#.*..*.#.*..*.#.*..*.#.****.#..*.*.#..+-+ | ||
53 | | * * # *+* # * * # *|* # * * # * * # * * # * * # * * # * * # * * # * |* # * * # | | ||
54 | | * * # * * # * * # *+* # * * # * * # * * # * * # * * # * * # * * # * |* # * * # | | ||
55 | 0.8 +-+..*.*.#..*.*..#.*.*..#.*.*..#.*.*..#.*.*..#.*..*.#.*..*.#.*..*.#.*..*.#.*..*.#.*++*.#..*.*.#..+-+ | ||
56 | | * * # * * # * * # * * # * * # * * # * * # * * # * * # * * # * * # * * # * * # | | ||
57 | 0.75 +-+--***##--***###-***###-***###-***###-***###-****##-****##-****##-****##-****##-****##--***##--+-+ | ||
58 | 400.perlben401.bzip2403.gcc429.m445.gob456.hmme45462.libqua464.h26471.omnet473483.xalancbmkgeomean | ||
59 | |||
60 | png: https://imgur.com/a/BHzpPTW | ||
61 | |||
62 | Notes: | ||
63 | - tlb-lock-v2 corresponds to an implementation with a mutex. | ||
64 | - tlb-lock-v3 corresponds to the current implementation, i.e. | ||
65 | a spinlock and a single lock acquisition in tlb_set_page_with_attrs. | ||
66 | |||
67 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
68 | Message-Id: <20181016153840.25877-1-cota@braap.org> | ||
69 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
70 | --- | 5 | --- |
71 | accel/tcg/softmmu_template.h | 12 ++++++------ | 6 | include/tcg/tcg-opc.h | 28 ++------ |
72 | include/exec/cpu_ldst.h | 11 ++++++++++- | 7 | tcg/optimize.c | 21 ++---- |
73 | include/exec/cpu_ldst_template.h | 2 +- | 8 | tcg/tcg-op-ldst.c | 82 +++++---------------- |
74 | accel/tcg/cputlb.c | 19 +++++++++++++------ | 9 | tcg/tcg.c | 42 ++++------- |
75 | 4 files changed, 30 insertions(+), 14 deletions(-) | 10 | tcg/tci.c | 119 ++++++------------------------- |
11 | tcg/aarch64/tcg-target.c.inc | 36 ++++------ | ||
12 | tcg/arm/tcg-target.c.inc | 40 +++-------- | ||
13 | tcg/i386/tcg-target.c.inc | 69 ++++-------------- | ||
14 | tcg/loongarch64/tcg-target.c.inc | 36 ++++------ | ||
15 | tcg/mips/tcg-target.c.inc | 51 +++---------- | ||
16 | tcg/ppc/tcg-target.c.inc | 68 ++++-------------- | ||
17 | tcg/riscv/tcg-target.c.inc | 24 +++---- | ||
18 | tcg/s390x/tcg-target.c.inc | 36 ++++------ | ||
19 | tcg/sparc64/tcg-target.c.inc | 24 +++---- | ||
20 | tcg/tci/tcg-target.c.inc | 60 ++++------------ | ||
21 | 15 files changed, 177 insertions(+), 559 deletions(-) | ||
76 | 22 | ||
77 | diff --git a/accel/tcg/softmmu_template.h b/accel/tcg/softmmu_template.h | 23 | diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h |
78 | index XXXXXXX..XXXXXXX 100644 | 24 | index XXXXXXX..XXXXXXX 100644 |
79 | --- a/accel/tcg/softmmu_template.h | 25 | --- a/include/tcg/tcg-opc.h |
80 | +++ b/accel/tcg/softmmu_template.h | 26 | +++ b/include/tcg/tcg-opc.h |
81 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | 27 | @@ -XXX,XX +XXX,XX @@ DEF(goto_ptr, 0, 1, 0, TCG_OPF_BB_EXIT | TCG_OPF_BB_END) |
82 | uintptr_t mmu_idx = get_mmuidx(oi); | 28 | DEF(plugin_cb, 0, 0, 1, TCG_OPF_NOT_PRESENT) |
83 | uintptr_t index = tlb_index(env, mmu_idx, addr); | 29 | DEF(plugin_mem_cb, 0, 1, 1, TCG_OPF_NOT_PRESENT) |
84 | CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | 30 | |
85 | - target_ulong tlb_addr = entry->addr_write; | 31 | -/* Replicate ld/st ops for 32 and 64-bit guest addresses. */ |
86 | + target_ulong tlb_addr = tlb_addr_write(entry); | 32 | -DEF(qemu_ld_a32_i32, 1, 1, 1, |
87 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | 33 | +DEF(qemu_ld_i32, 1, 1, 1, |
88 | uintptr_t haddr; | 34 | TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) |
89 | 35 | -DEF(qemu_st_a32_i32, 0, 1 + 1, 1, | |
90 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | 36 | +DEF(qemu_st_i32, 0, 1 + 1, 1, |
91 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, MMU_DATA_STORE, | 37 | TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) |
92 | mmu_idx, retaddr); | 38 | -DEF(qemu_ld_a32_i64, DATA64_ARGS, 1, 1, |
93 | } | 39 | +DEF(qemu_ld_i64, DATA64_ARGS, 1, 1, |
94 | - tlb_addr = entry->addr_write & ~TLB_INVALID_MASK; | 40 | TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) |
95 | + tlb_addr = tlb_addr_write(entry) & ~TLB_INVALID_MASK; | 41 | -DEF(qemu_st_a32_i64, 0, DATA64_ARGS + 1, 1, |
42 | - TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
43 | - | ||
44 | -DEF(qemu_ld_a64_i32, 1, DATA64_ARGS, 1, | ||
45 | - TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
46 | -DEF(qemu_st_a64_i32, 0, 1 + DATA64_ARGS, 1, | ||
47 | - TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
48 | -DEF(qemu_ld_a64_i64, DATA64_ARGS, DATA64_ARGS, 1, | ||
49 | - TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
50 | -DEF(qemu_st_a64_i64, 0, DATA64_ARGS + DATA64_ARGS, 1, | ||
51 | +DEF(qemu_st_i64, 0, DATA64_ARGS + 1, 1, | ||
52 | TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
53 | |||
54 | /* Only used by i386 to cope with stupid register constraints. */ | ||
55 | -DEF(qemu_st8_a32_i32, 0, 1 + 1, 1, | ||
56 | - TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
57 | -DEF(qemu_st8_a64_i32, 0, 1 + DATA64_ARGS, 1, | ||
58 | +DEF(qemu_st8_i32, 0, 1 + 1, 1, | ||
59 | TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
60 | |||
61 | /* Only for 64-bit hosts at the moment. */ | ||
62 | -DEF(qemu_ld_a32_i128, 2, 1, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
63 | -DEF(qemu_ld_a64_i128, 2, 1, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
64 | -DEF(qemu_st_a32_i128, 0, 3, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
65 | -DEF(qemu_st_a64_i128, 0, 3, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
66 | +DEF(qemu_ld_i128, 2, 1, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
67 | +DEF(qemu_st_i128, 0, 3, 1, TCG_OPF_CALL_CLOBBER | TCG_OPF_SIDE_EFFECTS) | ||
68 | |||
69 | /* Host vector support. */ | ||
70 | |||
71 | diff --git a/tcg/optimize.c b/tcg/optimize.c | ||
72 | index XXXXXXX..XXXXXXX 100644 | ||
73 | --- a/tcg/optimize.c | ||
74 | +++ b/tcg/optimize.c | ||
75 | @@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s) | ||
76 | CASE_OP_32_64_VEC(orc): | ||
77 | done = fold_orc(&ctx, op); | ||
78 | break; | ||
79 | - case INDEX_op_qemu_ld_a32_i32: | ||
80 | - case INDEX_op_qemu_ld_a64_i32: | ||
81 | + case INDEX_op_qemu_ld_i32: | ||
82 | done = fold_qemu_ld_1reg(&ctx, op); | ||
83 | break; | ||
84 | - case INDEX_op_qemu_ld_a32_i64: | ||
85 | - case INDEX_op_qemu_ld_a64_i64: | ||
86 | + case INDEX_op_qemu_ld_i64: | ||
87 | if (TCG_TARGET_REG_BITS == 64) { | ||
88 | done = fold_qemu_ld_1reg(&ctx, op); | ||
89 | break; | ||
90 | } | ||
91 | QEMU_FALLTHROUGH; | ||
92 | - case INDEX_op_qemu_ld_a32_i128: | ||
93 | - case INDEX_op_qemu_ld_a64_i128: | ||
94 | + case INDEX_op_qemu_ld_i128: | ||
95 | done = fold_qemu_ld_2reg(&ctx, op); | ||
96 | break; | ||
97 | - case INDEX_op_qemu_st8_a32_i32: | ||
98 | - case INDEX_op_qemu_st8_a64_i32: | ||
99 | - case INDEX_op_qemu_st_a32_i32: | ||
100 | - case INDEX_op_qemu_st_a64_i32: | ||
101 | - case INDEX_op_qemu_st_a32_i64: | ||
102 | - case INDEX_op_qemu_st_a64_i64: | ||
103 | - case INDEX_op_qemu_st_a32_i128: | ||
104 | - case INDEX_op_qemu_st_a64_i128: | ||
105 | + case INDEX_op_qemu_st8_i32: | ||
106 | + case INDEX_op_qemu_st_i32: | ||
107 | + case INDEX_op_qemu_st_i64: | ||
108 | + case INDEX_op_qemu_st_i128: | ||
109 | done = fold_qemu_st(&ctx, op); | ||
110 | break; | ||
111 | CASE_OP_32_64(rem): | ||
112 | diff --git a/tcg/tcg-op-ldst.c b/tcg/tcg-op-ldst.c | ||
113 | index XXXXXXX..XXXXXXX 100644 | ||
114 | --- a/tcg/tcg-op-ldst.c | ||
115 | +++ b/tcg/tcg-op-ldst.c | ||
116 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i32_int(TCGv_i32 val, TCGTemp *addr, | ||
117 | MemOp orig_memop; | ||
118 | MemOpIdx orig_oi, oi; | ||
119 | TCGv_i64 copy_addr; | ||
120 | - TCGOpcode opc; | ||
121 | |||
122 | tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD); | ||
123 | orig_memop = memop = tcg_canonicalize_memop(memop, 0, 0); | ||
124 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i32_int(TCGv_i32 val, TCGTemp *addr, | ||
96 | } | 125 | } |
97 | 126 | ||
98 | /* Handle an IO access. */ | 127 | copy_addr = plugin_maybe_preserve_addr(addr); |
99 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | 128 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { |
100 | cannot evict the first. */ | 129 | - opc = INDEX_op_qemu_ld_a32_i32; |
101 | page2 = (addr + DATA_SIZE) & TARGET_PAGE_MASK; | 130 | - } else { |
102 | entry2 = tlb_entry(env, mmu_idx, page2); | 131 | - opc = INDEX_op_qemu_ld_a64_i32; |
103 | - if (!tlb_hit_page(entry2->addr_write, page2) | 132 | - } |
104 | + if (!tlb_hit_page(tlb_addr_write(entry2), page2) | 133 | - gen_ldst(opc, TCG_TYPE_I32, tcgv_i32_temp(val), NULL, addr, oi); |
105 | && !VICTIM_TLB_HIT(addr_write, page2)) { | 134 | + gen_ldst(INDEX_op_qemu_ld_i32, TCG_TYPE_I32, |
106 | tlb_fill(ENV_GET_CPU(env), page2, DATA_SIZE, MMU_DATA_STORE, | 135 | + tcgv_i32_temp(val), NULL, addr, oi); |
107 | mmu_idx, retaddr); | 136 | plugin_gen_mem_callbacks_i32(val, copy_addr, addr, orig_oi, |
108 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | 137 | QEMU_PLUGIN_MEM_R); |
109 | uintptr_t mmu_idx = get_mmuidx(oi); | 138 | |
110 | uintptr_t index = tlb_index(env, mmu_idx, addr); | 139 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i32_int(TCGv_i32 val, TCGTemp *addr, |
111 | CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
112 | - target_ulong tlb_addr = entry->addr_write; | ||
113 | + target_ulong tlb_addr = tlb_addr_write(entry); | ||
114 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | ||
115 | uintptr_t haddr; | ||
116 | |||
117 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
118 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, MMU_DATA_STORE, | ||
119 | mmu_idx, retaddr); | ||
120 | } | ||
121 | - tlb_addr = entry->addr_write & ~TLB_INVALID_MASK; | ||
122 | + tlb_addr = tlb_addr_write(entry) & ~TLB_INVALID_MASK; | ||
123 | } | 140 | } |
124 | 141 | ||
125 | /* Handle an IO access. */ | 142 | if (TCG_TARGET_HAS_qemu_st8_i32 && (memop & MO_SIZE) == MO_8) { |
126 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | 143 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { |
127 | cannot evict the first. */ | 144 | - opc = INDEX_op_qemu_st8_a32_i32; |
128 | page2 = (addr + DATA_SIZE) & TARGET_PAGE_MASK; | 145 | - } else { |
129 | entry2 = tlb_entry(env, mmu_idx, page2); | 146 | - opc = INDEX_op_qemu_st8_a64_i32; |
130 | - if (!tlb_hit_page(entry2->addr_write, page2) | 147 | - } |
131 | + if (!tlb_hit_page(tlb_addr_write(entry2), page2) | 148 | + opc = INDEX_op_qemu_st8_i32; |
132 | && !VICTIM_TLB_HIT(addr_write, page2)) { | 149 | } else { |
133 | tlb_fill(ENV_GET_CPU(env), page2, DATA_SIZE, MMU_DATA_STORE, | 150 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { |
134 | mmu_idx, retaddr); | 151 | - opc = INDEX_op_qemu_st_a32_i32; |
135 | diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h | 152 | - } else { |
136 | index XXXXXXX..XXXXXXX 100644 | 153 | - opc = INDEX_op_qemu_st_a64_i32; |
137 | --- a/include/exec/cpu_ldst.h | 154 | - } |
138 | +++ b/include/exec/cpu_ldst.h | 155 | + opc = INDEX_op_qemu_st_i32; |
139 | @@ -XXX,XX +XXX,XX @@ extern __thread uintptr_t helper_retaddr; | 156 | } |
140 | /* The memory helpers for tcg-generated code need tcg_target_long etc. */ | 157 | gen_ldst(opc, TCG_TYPE_I32, tcgv_i32_temp(val), NULL, addr, oi); |
141 | #include "tcg.h" | 158 | plugin_gen_mem_callbacks_i32(val, NULL, addr, orig_oi, QEMU_PLUGIN_MEM_W); |
142 | 159 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i64_int(TCGv_i64 val, TCGTemp *addr, | |
143 | +static inline target_ulong tlb_addr_write(const CPUTLBEntry *entry) | 160 | MemOp orig_memop; |
144 | +{ | 161 | MemOpIdx orig_oi, oi; |
145 | +#if TCG_OVERSIZED_GUEST | 162 | TCGv_i64 copy_addr; |
146 | + return entry->addr_write; | 163 | - TCGOpcode opc; |
147 | +#else | 164 | |
148 | + return atomic_read(&entry->addr_write); | 165 | if (TCG_TARGET_REG_BITS == 32 && (memop & MO_SIZE) < MO_64) { |
149 | +#endif | 166 | tcg_gen_qemu_ld_i32_int(TCGV_LOW(val), addr, idx, memop); |
150 | +} | 167 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i64_int(TCGv_i64 val, TCGTemp *addr, |
151 | + | 168 | } |
152 | /* Find the TLB index corresponding to the mmu_idx + address pair. */ | 169 | |
153 | static inline uintptr_t tlb_index(CPUArchState *env, uintptr_t mmu_idx, | 170 | copy_addr = plugin_maybe_preserve_addr(addr); |
154 | target_ulong addr) | 171 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { |
155 | @@ -XXX,XX +XXX,XX @@ static inline void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr, | 172 | - opc = INDEX_op_qemu_ld_a32_i64; |
156 | tlb_addr = tlbentry->addr_read; | 173 | - } else { |
157 | break; | 174 | - opc = INDEX_op_qemu_ld_a64_i64; |
158 | case 1: | 175 | - } |
159 | - tlb_addr = tlbentry->addr_write; | 176 | - gen_ldst_i64(opc, val, addr, oi); |
160 | + tlb_addr = tlb_addr_write(tlbentry); | 177 | + gen_ldst_i64(INDEX_op_qemu_ld_i64, val, addr, oi); |
161 | break; | 178 | plugin_gen_mem_callbacks_i64(val, copy_addr, addr, orig_oi, |
162 | case 2: | 179 | QEMU_PLUGIN_MEM_R); |
163 | tlb_addr = tlbentry->addr_code; | 180 | |
164 | diff --git a/include/exec/cpu_ldst_template.h b/include/exec/cpu_ldst_template.h | 181 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i64_int(TCGv_i64 val, TCGTemp *addr, |
165 | index XXXXXXX..XXXXXXX 100644 | ||
166 | --- a/include/exec/cpu_ldst_template.h | ||
167 | +++ b/include/exec/cpu_ldst_template.h | ||
168 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_st, SUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | ||
169 | addr = ptr; | ||
170 | mmu_idx = CPU_MMU_INDEX; | ||
171 | entry = tlb_entry(env, mmu_idx, addr); | ||
172 | - if (unlikely(entry->addr_write != | ||
173 | + if (unlikely(tlb_addr_write(entry) != | ||
174 | (addr & (TARGET_PAGE_MASK | (DATA_SIZE - 1))))) { | ||
175 | oi = make_memop_idx(SHIFT, mmu_idx); | ||
176 | glue(glue(helper_ret_st, SUFFIX), MMUSUFFIX)(env, addr, v, oi, | ||
177 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
178 | index XXXXXXX..XXXXXXX 100644 | ||
179 | --- a/accel/tcg/cputlb.c | ||
180 | +++ b/accel/tcg/cputlb.c | ||
181 | @@ -XXX,XX +XXX,XX @@ static inline bool tlb_hit_page_anyprot(CPUTLBEntry *tlb_entry, | ||
182 | target_ulong page) | ||
183 | { | 182 | { |
184 | return tlb_hit_page(tlb_entry->addr_read, page) || | 183 | TCGv_i64 swap = NULL; |
185 | - tlb_hit_page(tlb_entry->addr_write, page) || | 184 | MemOpIdx orig_oi, oi; |
186 | + tlb_hit_page(tlb_addr_write(tlb_entry), page) || | 185 | - TCGOpcode opc; |
187 | tlb_hit_page(tlb_entry->addr_code, page); | 186 | |
187 | if (TCG_TARGET_REG_BITS == 32 && (memop & MO_SIZE) < MO_64) { | ||
188 | tcg_gen_qemu_st_i32_int(TCGV_LOW(val), addr, idx, memop); | ||
189 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i64_int(TCGv_i64 val, TCGTemp *addr, | ||
190 | oi = make_memop_idx(memop, idx); | ||
191 | } | ||
192 | |||
193 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
194 | - opc = INDEX_op_qemu_st_a32_i64; | ||
195 | - } else { | ||
196 | - opc = INDEX_op_qemu_st_a64_i64; | ||
197 | - } | ||
198 | - gen_ldst_i64(opc, val, addr, oi); | ||
199 | + gen_ldst_i64(INDEX_op_qemu_st_i64, val, addr, oi); | ||
200 | plugin_gen_mem_callbacks_i64(val, NULL, addr, orig_oi, QEMU_PLUGIN_MEM_W); | ||
201 | |||
202 | if (swap) { | ||
203 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
204 | { | ||
205 | MemOpIdx orig_oi; | ||
206 | TCGv_i64 ext_addr = NULL; | ||
207 | - TCGOpcode opc; | ||
208 | |||
209 | check_max_alignment(memop_alignment_bits(memop)); | ||
210 | tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD); | ||
211 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
212 | hi = TCGV128_HIGH(val); | ||
213 | } | ||
214 | |||
215 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
216 | - opc = INDEX_op_qemu_ld_a32_i128; | ||
217 | - } else { | ||
218 | - opc = INDEX_op_qemu_ld_a64_i128; | ||
219 | - } | ||
220 | - gen_ldst(opc, TCG_TYPE_I128, tcgv_i64_temp(lo), | ||
221 | + gen_ldst(INDEX_op_qemu_ld_i128, TCG_TYPE_I128, tcgv_i64_temp(lo), | ||
222 | tcgv_i64_temp(hi), addr, oi); | ||
223 | |||
224 | if (need_bswap) { | ||
225 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
226 | canonicalize_memop_i128_as_i64(mop, memop); | ||
227 | need_bswap = (mop[0] ^ memop) & MO_BSWAP; | ||
228 | |||
229 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
230 | - opc = INDEX_op_qemu_ld_a32_i64; | ||
231 | - } else { | ||
232 | - opc = INDEX_op_qemu_ld_a64_i64; | ||
233 | - } | ||
234 | - | ||
235 | /* | ||
236 | * Since there are no global TCGv_i128, there is no visible state | ||
237 | * changed if the second load faults. Load directly into the two | ||
238 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
239 | y = TCGV128_LOW(val); | ||
240 | } | ||
241 | |||
242 | - gen_ldst_i64(opc, x, addr, make_memop_idx(mop[0], idx)); | ||
243 | + gen_ldst_i64(INDEX_op_qemu_ld_i64, x, addr, | ||
244 | + make_memop_idx(mop[0], idx)); | ||
245 | |||
246 | if (need_bswap) { | ||
247 | tcg_gen_bswap64_i64(x, x); | ||
248 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
249 | addr_p8 = tcgv_i64_temp(t); | ||
250 | } | ||
251 | |||
252 | - gen_ldst_i64(opc, y, addr_p8, make_memop_idx(mop[1], idx)); | ||
253 | + gen_ldst_i64(INDEX_op_qemu_ld_i64, y, addr_p8, | ||
254 | + make_memop_idx(mop[1], idx)); | ||
255 | tcg_temp_free_internal(addr_p8); | ||
256 | |||
257 | if (need_bswap) { | ||
258 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
259 | { | ||
260 | MemOpIdx orig_oi; | ||
261 | TCGv_i64 ext_addr = NULL; | ||
262 | - TCGOpcode opc; | ||
263 | |||
264 | check_max_alignment(memop_alignment_bits(memop)); | ||
265 | tcg_gen_req_mo(TCG_MO_ST_LD | TCG_MO_ST_ST); | ||
266 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
267 | hi = TCGV128_HIGH(val); | ||
268 | } | ||
269 | |||
270 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
271 | - opc = INDEX_op_qemu_st_a32_i128; | ||
272 | - } else { | ||
273 | - opc = INDEX_op_qemu_st_a64_i128; | ||
274 | - } | ||
275 | - gen_ldst(opc, TCG_TYPE_I128, tcgv_i64_temp(lo), | ||
276 | - tcgv_i64_temp(hi), addr, oi); | ||
277 | + gen_ldst(INDEX_op_qemu_st_i128, TCG_TYPE_I128, | ||
278 | + tcgv_i64_temp(lo), tcgv_i64_temp(hi), addr, oi); | ||
279 | |||
280 | if (need_bswap) { | ||
281 | tcg_temp_free_i64(lo); | ||
282 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
283 | |||
284 | canonicalize_memop_i128_as_i64(mop, memop); | ||
285 | |||
286 | - if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
287 | - opc = INDEX_op_qemu_st_a32_i64; | ||
288 | - } else { | ||
289 | - opc = INDEX_op_qemu_st_a64_i64; | ||
290 | - } | ||
291 | - | ||
292 | if ((memop & MO_BSWAP) == MO_LE) { | ||
293 | x = TCGV128_LOW(val); | ||
294 | y = TCGV128_HIGH(val); | ||
295 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
296 | x = b; | ||
297 | } | ||
298 | |||
299 | - gen_ldst_i64(opc, x, addr, make_memop_idx(mop[0], idx)); | ||
300 | + gen_ldst_i64(INDEX_op_qemu_st_i64, x, addr, | ||
301 | + make_memop_idx(mop[0], idx)); | ||
302 | |||
303 | if (tcg_ctx->addr_type == TCG_TYPE_I32) { | ||
304 | TCGv_i32 t = tcg_temp_ebb_new_i32(); | ||
305 | @@ -XXX,XX +XXX,XX @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr, | ||
306 | |||
307 | if (b) { | ||
308 | tcg_gen_bswap64_i64(b, y); | ||
309 | - gen_ldst_i64(opc, b, addr_p8, make_memop_idx(mop[1], idx)); | ||
310 | + gen_ldst_i64(INDEX_op_qemu_st_i64, b, addr_p8, | ||
311 | + make_memop_idx(mop[1], idx)); | ||
312 | tcg_temp_free_i64(b); | ||
313 | } else { | ||
314 | - gen_ldst_i64(opc, y, addr_p8, make_memop_idx(mop[1], idx)); | ||
315 | + gen_ldst_i64(INDEX_op_qemu_st_i64, y, addr_p8, | ||
316 | + make_memop_idx(mop[1], idx)); | ||
317 | } | ||
318 | tcg_temp_free_internal(addr_p8); | ||
319 | } else { | ||
320 | diff --git a/tcg/tcg.c b/tcg/tcg.c | ||
321 | index XXXXXXX..XXXXXXX 100644 | ||
322 | --- a/tcg/tcg.c | ||
323 | +++ b/tcg/tcg.c | ||
324 | @@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op, TCGType type, unsigned flags) | ||
325 | case INDEX_op_exit_tb: | ||
326 | case INDEX_op_goto_tb: | ||
327 | case INDEX_op_goto_ptr: | ||
328 | - case INDEX_op_qemu_ld_a32_i32: | ||
329 | - case INDEX_op_qemu_ld_a64_i32: | ||
330 | - case INDEX_op_qemu_st_a32_i32: | ||
331 | - case INDEX_op_qemu_st_a64_i32: | ||
332 | - case INDEX_op_qemu_ld_a32_i64: | ||
333 | - case INDEX_op_qemu_ld_a64_i64: | ||
334 | - case INDEX_op_qemu_st_a32_i64: | ||
335 | - case INDEX_op_qemu_st_a64_i64: | ||
336 | + case INDEX_op_qemu_ld_i32: | ||
337 | + case INDEX_op_qemu_st_i32: | ||
338 | + case INDEX_op_qemu_ld_i64: | ||
339 | + case INDEX_op_qemu_st_i64: | ||
340 | return true; | ||
341 | |||
342 | - case INDEX_op_qemu_st8_a32_i32: | ||
343 | - case INDEX_op_qemu_st8_a64_i32: | ||
344 | + case INDEX_op_qemu_st8_i32: | ||
345 | return TCG_TARGET_HAS_qemu_st8_i32; | ||
346 | |||
347 | - case INDEX_op_qemu_ld_a32_i128: | ||
348 | - case INDEX_op_qemu_ld_a64_i128: | ||
349 | - case INDEX_op_qemu_st_a32_i128: | ||
350 | - case INDEX_op_qemu_st_a64_i128: | ||
351 | + case INDEX_op_qemu_ld_i128: | ||
352 | + case INDEX_op_qemu_st_i128: | ||
353 | return TCG_TARGET_HAS_qemu_ldst_i128; | ||
354 | |||
355 | case INDEX_op_mov_i32: | ||
356 | @@ -XXX,XX +XXX,XX @@ void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs) | ||
357 | } | ||
358 | i = 1; | ||
359 | break; | ||
360 | - case INDEX_op_qemu_ld_a32_i32: | ||
361 | - case INDEX_op_qemu_ld_a64_i32: | ||
362 | - case INDEX_op_qemu_st_a32_i32: | ||
363 | - case INDEX_op_qemu_st_a64_i32: | ||
364 | - case INDEX_op_qemu_st8_a32_i32: | ||
365 | - case INDEX_op_qemu_st8_a64_i32: | ||
366 | - case INDEX_op_qemu_ld_a32_i64: | ||
367 | - case INDEX_op_qemu_ld_a64_i64: | ||
368 | - case INDEX_op_qemu_st_a32_i64: | ||
369 | - case INDEX_op_qemu_st_a64_i64: | ||
370 | - case INDEX_op_qemu_ld_a32_i128: | ||
371 | - case INDEX_op_qemu_ld_a64_i128: | ||
372 | - case INDEX_op_qemu_st_a32_i128: | ||
373 | - case INDEX_op_qemu_st_a64_i128: | ||
374 | + case INDEX_op_qemu_ld_i32: | ||
375 | + case INDEX_op_qemu_st_i32: | ||
376 | + case INDEX_op_qemu_st8_i32: | ||
377 | + case INDEX_op_qemu_ld_i64: | ||
378 | + case INDEX_op_qemu_st_i64: | ||
379 | + case INDEX_op_qemu_ld_i128: | ||
380 | + case INDEX_op_qemu_st_i128: | ||
381 | { | ||
382 | const char *s_al, *s_op, *s_at; | ||
383 | MemOpIdx oi = op->args[k++]; | ||
384 | diff --git a/tcg/tci.c b/tcg/tci.c | ||
385 | index XXXXXXX..XXXXXXX 100644 | ||
386 | --- a/tcg/tci.c | ||
387 | +++ b/tcg/tci.c | ||
388 | @@ -XXX,XX +XXX,XX @@ static void tci_args_rrrbb(uint32_t insn, TCGReg *r0, TCGReg *r1, | ||
389 | *i4 = extract32(insn, 26, 6); | ||
188 | } | 390 | } |
189 | 391 | ||
190 | @@ -XXX,XX +XXX,XX @@ static void io_writex(CPUArchState *env, CPUIOTLBEntry *iotlbentry, | 392 | -static void tci_args_rrrrr(uint32_t insn, TCGReg *r0, TCGReg *r1, |
191 | tlb_fill(cpu, addr, size, MMU_DATA_STORE, mmu_idx, retaddr); | 393 | - TCGReg *r2, TCGReg *r3, TCGReg *r4) |
192 | 394 | -{ | |
193 | entry = tlb_entry(env, mmu_idx, addr); | 395 | - *r0 = extract32(insn, 8, 4); |
194 | - tlb_addr = entry->addr_write; | 396 | - *r1 = extract32(insn, 12, 4); |
195 | + tlb_addr = tlb_addr_write(entry); | 397 | - *r2 = extract32(insn, 16, 4); |
196 | if (!(tlb_addr & ~(TARGET_PAGE_MASK | TLB_RECHECK))) { | 398 | - *r3 = extract32(insn, 20, 4); |
197 | /* RAM access */ | 399 | - *r4 = extract32(insn, 24, 4); |
198 | uintptr_t haddr = addr + entry->addend; | 400 | -} |
199 | @@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index, | 401 | - |
200 | assert_cpu_is_self(ENV_GET_CPU(env)); | 402 | static void tci_args_rrrr(uint32_t insn, |
201 | for (vidx = 0; vidx < CPU_VTLB_SIZE; ++vidx) { | 403 | TCGReg *r0, TCGReg *r1, TCGReg *r2, TCGReg *r3) |
202 | CPUTLBEntry *vtlb = &env->tlb_v_table[mmu_idx][vidx]; | 404 | { |
203 | - target_ulong cmp = *(target_ulong *)((uintptr_t)vtlb + elt_ofs); | 405 | @@ -XXX,XX +XXX,XX @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState *env, |
204 | + target_ulong cmp; | 406 | tb_ptr = ptr; |
205 | + | 407 | break; |
206 | + /* elt_ofs might correspond to .addr_write, so use atomic_read */ | 408 | |
207 | +#if TCG_OVERSIZED_GUEST | 409 | - case INDEX_op_qemu_ld_a32_i32: |
208 | + cmp = *(target_ulong *)((uintptr_t)vtlb + elt_ofs); | 410 | + case INDEX_op_qemu_ld_i32: |
209 | +#else | 411 | tci_args_rrm(insn, &r0, &r1, &oi); |
210 | + cmp = atomic_read((target_ulong *)((uintptr_t)vtlb + elt_ofs)); | 412 | - taddr = (uint32_t)regs[r1]; |
211 | +#endif | 413 | - goto do_ld_i32; |
212 | 414 | - case INDEX_op_qemu_ld_a64_i32: | |
213 | if (cmp == page) { | 415 | - if (TCG_TARGET_REG_BITS == 64) { |
214 | /* Found entry in victim tlb, swap tlb and iotlb. */ | 416 | - tci_args_rrm(insn, &r0, &r1, &oi); |
215 | @@ -XXX,XX +XXX,XX @@ void probe_write(CPUArchState *env, target_ulong addr, int size, int mmu_idx, | 417 | - taddr = regs[r1]; |
216 | uintptr_t index = tlb_index(env, mmu_idx, addr); | 418 | - } else { |
217 | CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | 419 | - tci_args_rrrr(insn, &r0, &r1, &r2, &r3); |
218 | 420 | - taddr = tci_uint64(regs[r2], regs[r1]); | |
219 | - if (!tlb_hit(entry->addr_write, addr)) { | 421 | - oi = regs[r3]; |
220 | + if (!tlb_hit(tlb_addr_write(entry), addr)) { | 422 | - } |
221 | /* TLB entry is for a different page */ | 423 | - do_ld_i32: |
222 | if (!VICTIM_TLB_HIT(addr_write, addr)) { | 424 | + taddr = regs[r1]; |
223 | tlb_fill(ENV_GET_CPU(env), addr, size, MMU_DATA_STORE, | 425 | regs[r0] = tci_qemu_ld(env, taddr, oi, tb_ptr); |
224 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | 426 | break; |
225 | size_t mmu_idx = get_mmuidx(oi); | 427 | |
226 | uintptr_t index = tlb_index(env, mmu_idx, addr); | 428 | - case INDEX_op_qemu_ld_a32_i64: |
227 | CPUTLBEntry *tlbe = tlb_entry(env, mmu_idx, addr); | 429 | - if (TCG_TARGET_REG_BITS == 64) { |
228 | - target_ulong tlb_addr = tlbe->addr_write; | 430 | - tci_args_rrm(insn, &r0, &r1, &oi); |
229 | + target_ulong tlb_addr = tlb_addr_write(tlbe); | 431 | - taddr = (uint32_t)regs[r1]; |
230 | TCGMemOp mop = get_memop(oi); | 432 | - } else { |
231 | int a_bits = get_alignment_bits(mop); | 433 | - tci_args_rrrr(insn, &r0, &r1, &r2, &r3); |
232 | int s_bits = mop & MO_SIZE; | 434 | - taddr = (uint32_t)regs[r2]; |
233 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | 435 | - oi = regs[r3]; |
234 | tlb_fill(ENV_GET_CPU(env), addr, 1 << s_bits, MMU_DATA_STORE, | 436 | - } |
235 | mmu_idx, retaddr); | 437 | - goto do_ld_i64; |
236 | } | 438 | - case INDEX_op_qemu_ld_a64_i64: |
237 | - tlb_addr = tlbe->addr_write & ~TLB_INVALID_MASK; | 439 | + case INDEX_op_qemu_ld_i64: |
238 | + tlb_addr = tlb_addr_write(tlbe) & ~TLB_INVALID_MASK; | 440 | if (TCG_TARGET_REG_BITS == 64) { |
239 | } | 441 | tci_args_rrm(insn, &r0, &r1, &oi); |
240 | 442 | taddr = regs[r1]; | |
241 | /* Notice an IO access or a needs-MMU-lookup access */ | 443 | } else { |
444 | - tci_args_rrrrr(insn, &r0, &r1, &r2, &r3, &r4); | ||
445 | - taddr = tci_uint64(regs[r3], regs[r2]); | ||
446 | - oi = regs[r4]; | ||
447 | + tci_args_rrrr(insn, &r0, &r1, &r2, &r3); | ||
448 | + taddr = regs[r2]; | ||
449 | + oi = regs[r3]; | ||
450 | } | ||
451 | - do_ld_i64: | ||
452 | tmp64 = tci_qemu_ld(env, taddr, oi, tb_ptr); | ||
453 | if (TCG_TARGET_REG_BITS == 32) { | ||
454 | tci_write_reg64(regs, r1, r0, tmp64); | ||
455 | @@ -XXX,XX +XXX,XX @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState *env, | ||
456 | } | ||
457 | break; | ||
458 | |||
459 | - case INDEX_op_qemu_st_a32_i32: | ||
460 | + case INDEX_op_qemu_st_i32: | ||
461 | tci_args_rrm(insn, &r0, &r1, &oi); | ||
462 | - taddr = (uint32_t)regs[r1]; | ||
463 | - goto do_st_i32; | ||
464 | - case INDEX_op_qemu_st_a64_i32: | ||
465 | - if (TCG_TARGET_REG_BITS == 64) { | ||
466 | - tci_args_rrm(insn, &r0, &r1, &oi); | ||
467 | - taddr = regs[r1]; | ||
468 | - } else { | ||
469 | - tci_args_rrrr(insn, &r0, &r1, &r2, &r3); | ||
470 | - taddr = tci_uint64(regs[r2], regs[r1]); | ||
471 | - oi = regs[r3]; | ||
472 | - } | ||
473 | - do_st_i32: | ||
474 | + taddr = regs[r1]; | ||
475 | tci_qemu_st(env, taddr, regs[r0], oi, tb_ptr); | ||
476 | break; | ||
477 | |||
478 | - case INDEX_op_qemu_st_a32_i64: | ||
479 | - if (TCG_TARGET_REG_BITS == 64) { | ||
480 | - tci_args_rrm(insn, &r0, &r1, &oi); | ||
481 | - tmp64 = regs[r0]; | ||
482 | - taddr = (uint32_t)regs[r1]; | ||
483 | - } else { | ||
484 | - tci_args_rrrr(insn, &r0, &r1, &r2, &r3); | ||
485 | - tmp64 = tci_uint64(regs[r1], regs[r0]); | ||
486 | - taddr = (uint32_t)regs[r2]; | ||
487 | - oi = regs[r3]; | ||
488 | - } | ||
489 | - goto do_st_i64; | ||
490 | - case INDEX_op_qemu_st_a64_i64: | ||
491 | + case INDEX_op_qemu_st_i64: | ||
492 | if (TCG_TARGET_REG_BITS == 64) { | ||
493 | tci_args_rrm(insn, &r0, &r1, &oi); | ||
494 | tmp64 = regs[r0]; | ||
495 | taddr = regs[r1]; | ||
496 | } else { | ||
497 | - tci_args_rrrrr(insn, &r0, &r1, &r2, &r3, &r4); | ||
498 | + tci_args_rrrr(insn, &r0, &r1, &r2, &r3); | ||
499 | tmp64 = tci_uint64(regs[r1], regs[r0]); | ||
500 | - taddr = tci_uint64(regs[r3], regs[r2]); | ||
501 | - oi = regs[r4]; | ||
502 | + taddr = regs[r2]; | ||
503 | + oi = regs[r3]; | ||
504 | } | ||
505 | - do_st_i64: | ||
506 | tci_qemu_st(env, taddr, tmp64, oi, tb_ptr); | ||
507 | break; | ||
508 | |||
509 | @@ -XXX,XX +XXX,XX @@ int print_insn_tci(bfd_vma addr, disassemble_info *info) | ||
510 | str_r(r3), str_r(r4), str_r(r5)); | ||
511 | break; | ||
512 | |||
513 | - case INDEX_op_qemu_ld_a32_i32: | ||
514 | - case INDEX_op_qemu_st_a32_i32: | ||
515 | - len = 1 + 1; | ||
516 | - goto do_qemu_ldst; | ||
517 | - case INDEX_op_qemu_ld_a32_i64: | ||
518 | - case INDEX_op_qemu_st_a32_i64: | ||
519 | - case INDEX_op_qemu_ld_a64_i32: | ||
520 | - case INDEX_op_qemu_st_a64_i32: | ||
521 | - len = 1 + DIV_ROUND_UP(64, TCG_TARGET_REG_BITS); | ||
522 | - goto do_qemu_ldst; | ||
523 | - case INDEX_op_qemu_ld_a64_i64: | ||
524 | - case INDEX_op_qemu_st_a64_i64: | ||
525 | - len = 2 * DIV_ROUND_UP(64, TCG_TARGET_REG_BITS); | ||
526 | - goto do_qemu_ldst; | ||
527 | - do_qemu_ldst: | ||
528 | - switch (len) { | ||
529 | - case 2: | ||
530 | - tci_args_rrm(insn, &r0, &r1, &oi); | ||
531 | - info->fprintf_func(info->stream, "%-12s %s, %s, %x", | ||
532 | - op_name, str_r(r0), str_r(r1), oi); | ||
533 | - break; | ||
534 | - case 3: | ||
535 | + case INDEX_op_qemu_ld_i64: | ||
536 | + case INDEX_op_qemu_st_i64: | ||
537 | + if (TCG_TARGET_REG_BITS == 32) { | ||
538 | tci_args_rrrr(insn, &r0, &r1, &r2, &r3); | ||
539 | info->fprintf_func(info->stream, "%-12s %s, %s, %s, %s", | ||
540 | op_name, str_r(r0), str_r(r1), | ||
541 | str_r(r2), str_r(r3)); | ||
542 | break; | ||
543 | - case 4: | ||
544 | - tci_args_rrrrr(insn, &r0, &r1, &r2, &r3, &r4); | ||
545 | - info->fprintf_func(info->stream, "%-12s %s, %s, %s, %s, %s", | ||
546 | - op_name, str_r(r0), str_r(r1), | ||
547 | - str_r(r2), str_r(r3), str_r(r4)); | ||
548 | - break; | ||
549 | - default: | ||
550 | - g_assert_not_reached(); | ||
551 | } | ||
552 | + /* fall through */ | ||
553 | + case INDEX_op_qemu_ld_i32: | ||
554 | + case INDEX_op_qemu_st_i32: | ||
555 | + tci_args_rrm(insn, &r0, &r1, &oi); | ||
556 | + info->fprintf_func(info->stream, "%-12s %s, %s, %x", | ||
557 | + op_name, str_r(r0), str_r(r1), oi); | ||
558 | break; | ||
559 | |||
560 | case 0: | ||
561 | diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc | ||
562 | index XXXXXXX..XXXXXXX 100644 | ||
563 | --- a/tcg/aarch64/tcg-target.c.inc | ||
564 | +++ b/tcg/aarch64/tcg-target.c.inc | ||
565 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
566 | tcg_out_insn(s, 3506, CSEL, ext, a0, REG0(3), REG0(4), args[5]); | ||
567 | break; | ||
568 | |||
569 | - case INDEX_op_qemu_ld_a32_i32: | ||
570 | - case INDEX_op_qemu_ld_a64_i32: | ||
571 | - case INDEX_op_qemu_ld_a32_i64: | ||
572 | - case INDEX_op_qemu_ld_a64_i64: | ||
573 | + case INDEX_op_qemu_ld_i32: | ||
574 | + case INDEX_op_qemu_ld_i64: | ||
575 | tcg_out_qemu_ld(s, a0, a1, a2, ext); | ||
576 | break; | ||
577 | - case INDEX_op_qemu_st_a32_i32: | ||
578 | - case INDEX_op_qemu_st_a64_i32: | ||
579 | - case INDEX_op_qemu_st_a32_i64: | ||
580 | - case INDEX_op_qemu_st_a64_i64: | ||
581 | + case INDEX_op_qemu_st_i32: | ||
582 | + case INDEX_op_qemu_st_i64: | ||
583 | tcg_out_qemu_st(s, REG0(0), a1, a2, ext); | ||
584 | break; | ||
585 | - case INDEX_op_qemu_ld_a32_i128: | ||
586 | - case INDEX_op_qemu_ld_a64_i128: | ||
587 | + case INDEX_op_qemu_ld_i128: | ||
588 | tcg_out_qemu_ldst_i128(s, a0, a1, a2, args[3], true); | ||
589 | break; | ||
590 | - case INDEX_op_qemu_st_a32_i128: | ||
591 | - case INDEX_op_qemu_st_a64_i128: | ||
592 | + case INDEX_op_qemu_st_i128: | ||
593 | tcg_out_qemu_ldst_i128(s, REG0(0), REG0(1), a2, args[3], false); | ||
594 | break; | ||
595 | |||
596 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
597 | case INDEX_op_movcond_i64: | ||
598 | return C_O1_I4(r, r, rC, rZ, rZ); | ||
599 | |||
600 | - case INDEX_op_qemu_ld_a32_i32: | ||
601 | - case INDEX_op_qemu_ld_a64_i32: | ||
602 | - case INDEX_op_qemu_ld_a32_i64: | ||
603 | - case INDEX_op_qemu_ld_a64_i64: | ||
604 | + case INDEX_op_qemu_ld_i32: | ||
605 | + case INDEX_op_qemu_ld_i64: | ||
606 | return C_O1_I1(r, r); | ||
607 | - case INDEX_op_qemu_ld_a32_i128: | ||
608 | - case INDEX_op_qemu_ld_a64_i128: | ||
609 | + case INDEX_op_qemu_ld_i128: | ||
610 | return C_O2_I1(r, r, r); | ||
611 | - case INDEX_op_qemu_st_a32_i32: | ||
612 | - case INDEX_op_qemu_st_a64_i32: | ||
613 | - case INDEX_op_qemu_st_a32_i64: | ||
614 | - case INDEX_op_qemu_st_a64_i64: | ||
615 | + case INDEX_op_qemu_st_i32: | ||
616 | + case INDEX_op_qemu_st_i64: | ||
617 | return C_O0_I2(rZ, r); | ||
618 | - case INDEX_op_qemu_st_a32_i128: | ||
619 | - case INDEX_op_qemu_st_a64_i128: | ||
620 | + case INDEX_op_qemu_st_i128: | ||
621 | return C_O0_I3(rZ, rZ, r); | ||
622 | |||
623 | case INDEX_op_deposit_i32: | ||
624 | diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc | ||
625 | index XXXXXXX..XXXXXXX 100644 | ||
626 | --- a/tcg/arm/tcg-target.c.inc | ||
627 | +++ b/tcg/arm/tcg-target.c.inc | ||
628 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
629 | ARITH_MOV, args[0], 0, 0); | ||
630 | break; | ||
631 | |||
632 | - case INDEX_op_qemu_ld_a32_i32: | ||
633 | + case INDEX_op_qemu_ld_i32: | ||
634 | tcg_out_qemu_ld(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
635 | break; | ||
636 | - case INDEX_op_qemu_ld_a64_i32: | ||
637 | - tcg_out_qemu_ld(s, args[0], -1, args[1], args[2], | ||
638 | - args[3], TCG_TYPE_I32); | ||
639 | - break; | ||
640 | - case INDEX_op_qemu_ld_a32_i64: | ||
641 | + case INDEX_op_qemu_ld_i64: | ||
642 | tcg_out_qemu_ld(s, args[0], args[1], args[2], -1, | ||
643 | args[3], TCG_TYPE_I64); | ||
644 | break; | ||
645 | - case INDEX_op_qemu_ld_a64_i64: | ||
646 | - tcg_out_qemu_ld(s, args[0], args[1], args[2], args[3], | ||
647 | - args[4], TCG_TYPE_I64); | ||
648 | - break; | ||
649 | |||
650 | - case INDEX_op_qemu_st_a32_i32: | ||
651 | + case INDEX_op_qemu_st_i32: | ||
652 | tcg_out_qemu_st(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
653 | break; | ||
654 | - case INDEX_op_qemu_st_a64_i32: | ||
655 | - tcg_out_qemu_st(s, args[0], -1, args[1], args[2], | ||
656 | - args[3], TCG_TYPE_I32); | ||
657 | - break; | ||
658 | - case INDEX_op_qemu_st_a32_i64: | ||
659 | + case INDEX_op_qemu_st_i64: | ||
660 | tcg_out_qemu_st(s, args[0], args[1], args[2], -1, | ||
661 | args[3], TCG_TYPE_I64); | ||
662 | break; | ||
663 | - case INDEX_op_qemu_st_a64_i64: | ||
664 | - tcg_out_qemu_st(s, args[0], args[1], args[2], args[3], | ||
665 | - args[4], TCG_TYPE_I64); | ||
666 | - break; | ||
667 | |||
668 | case INDEX_op_bswap16_i32: | ||
669 | tcg_out_bswap16(s, COND_AL, args[0], args[1], args[2]); | ||
670 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
671 | case INDEX_op_setcond2_i32: | ||
672 | return C_O1_I4(r, r, r, rI, rI); | ||
673 | |||
674 | - case INDEX_op_qemu_ld_a32_i32: | ||
675 | + case INDEX_op_qemu_ld_i32: | ||
676 | return C_O1_I1(r, q); | ||
677 | - case INDEX_op_qemu_ld_a64_i32: | ||
678 | - return C_O1_I2(r, q, q); | ||
679 | - case INDEX_op_qemu_ld_a32_i64: | ||
680 | + case INDEX_op_qemu_ld_i64: | ||
681 | return C_O2_I1(e, p, q); | ||
682 | - case INDEX_op_qemu_ld_a64_i64: | ||
683 | - return C_O2_I2(e, p, q, q); | ||
684 | - case INDEX_op_qemu_st_a32_i32: | ||
685 | + case INDEX_op_qemu_st_i32: | ||
686 | return C_O0_I2(q, q); | ||
687 | - case INDEX_op_qemu_st_a64_i32: | ||
688 | - return C_O0_I3(q, q, q); | ||
689 | - case INDEX_op_qemu_st_a32_i64: | ||
690 | + case INDEX_op_qemu_st_i64: | ||
691 | return C_O0_I3(Q, p, q); | ||
692 | - case INDEX_op_qemu_st_a64_i64: | ||
693 | - return C_O0_I4(Q, p, q, q); | ||
694 | |||
695 | case INDEX_op_st_vec: | ||
696 | return C_O0_I2(w, r); | ||
697 | diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc | ||
698 | index XXXXXXX..XXXXXXX 100644 | ||
699 | --- a/tcg/i386/tcg-target.c.inc | ||
700 | +++ b/tcg/i386/tcg-target.c.inc | ||
701 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
702 | tcg_out_modrm(s, OPC_GRP3_Ev + rexw, EXT3_NOT, a0); | ||
703 | break; | ||
704 | |||
705 | - case INDEX_op_qemu_ld_a64_i32: | ||
706 | - if (TCG_TARGET_REG_BITS == 32) { | ||
707 | - tcg_out_qemu_ld(s, a0, -1, a1, a2, args[3], TCG_TYPE_I32); | ||
708 | - break; | ||
709 | - } | ||
710 | - /* fall through */ | ||
711 | - case INDEX_op_qemu_ld_a32_i32: | ||
712 | + case INDEX_op_qemu_ld_i32: | ||
713 | tcg_out_qemu_ld(s, a0, -1, a1, -1, a2, TCG_TYPE_I32); | ||
714 | break; | ||
715 | - case INDEX_op_qemu_ld_a32_i64: | ||
716 | + case INDEX_op_qemu_ld_i64: | ||
717 | if (TCG_TARGET_REG_BITS == 64) { | ||
718 | tcg_out_qemu_ld(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); | ||
719 | } else { | ||
720 | tcg_out_qemu_ld(s, a0, a1, a2, -1, args[3], TCG_TYPE_I64); | ||
721 | } | ||
722 | break; | ||
723 | - case INDEX_op_qemu_ld_a64_i64: | ||
724 | - if (TCG_TARGET_REG_BITS == 64) { | ||
725 | - tcg_out_qemu_ld(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); | ||
726 | - } else { | ||
727 | - tcg_out_qemu_ld(s, a0, a1, a2, args[3], args[4], TCG_TYPE_I64); | ||
728 | - } | ||
729 | - break; | ||
730 | - case INDEX_op_qemu_ld_a32_i128: | ||
731 | - case INDEX_op_qemu_ld_a64_i128: | ||
732 | + case INDEX_op_qemu_ld_i128: | ||
733 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
734 | tcg_out_qemu_ld(s, a0, a1, a2, -1, args[3], TCG_TYPE_I128); | ||
735 | break; | ||
736 | |||
737 | - case INDEX_op_qemu_st_a64_i32: | ||
738 | - case INDEX_op_qemu_st8_a64_i32: | ||
739 | - if (TCG_TARGET_REG_BITS == 32) { | ||
740 | - tcg_out_qemu_st(s, a0, -1, a1, a2, args[3], TCG_TYPE_I32); | ||
741 | - break; | ||
742 | - } | ||
743 | - /* fall through */ | ||
744 | - case INDEX_op_qemu_st_a32_i32: | ||
745 | - case INDEX_op_qemu_st8_a32_i32: | ||
746 | + case INDEX_op_qemu_st_i32: | ||
747 | + case INDEX_op_qemu_st8_i32: | ||
748 | tcg_out_qemu_st(s, a0, -1, a1, -1, a2, TCG_TYPE_I32); | ||
749 | break; | ||
750 | - case INDEX_op_qemu_st_a32_i64: | ||
751 | + case INDEX_op_qemu_st_i64: | ||
752 | if (TCG_TARGET_REG_BITS == 64) { | ||
753 | tcg_out_qemu_st(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); | ||
754 | } else { | ||
755 | tcg_out_qemu_st(s, a0, a1, a2, -1, args[3], TCG_TYPE_I64); | ||
756 | } | ||
757 | break; | ||
758 | - case INDEX_op_qemu_st_a64_i64: | ||
759 | - if (TCG_TARGET_REG_BITS == 64) { | ||
760 | - tcg_out_qemu_st(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); | ||
761 | - } else { | ||
762 | - tcg_out_qemu_st(s, a0, a1, a2, args[3], args[4], TCG_TYPE_I64); | ||
763 | - } | ||
764 | - break; | ||
765 | - case INDEX_op_qemu_st_a32_i128: | ||
766 | - case INDEX_op_qemu_st_a64_i128: | ||
767 | + case INDEX_op_qemu_st_i128: | ||
768 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
769 | tcg_out_qemu_st(s, a0, a1, a2, -1, args[3], TCG_TYPE_I128); | ||
770 | break; | ||
771 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
772 | case INDEX_op_clz_i64: | ||
773 | return have_lzcnt ? C_N1_I2(r, r, rW) : C_N1_I2(r, r, r); | ||
774 | |||
775 | - case INDEX_op_qemu_ld_a32_i32: | ||
776 | + case INDEX_op_qemu_ld_i32: | ||
777 | return C_O1_I1(r, L); | ||
778 | - case INDEX_op_qemu_ld_a64_i32: | ||
779 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, L) : C_O1_I2(r, L, L); | ||
780 | |||
781 | - case INDEX_op_qemu_st_a32_i32: | ||
782 | + case INDEX_op_qemu_st_i32: | ||
783 | return C_O0_I2(L, L); | ||
784 | - case INDEX_op_qemu_st_a64_i32: | ||
785 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(L, L) : C_O0_I3(L, L, L); | ||
786 | - case INDEX_op_qemu_st8_a32_i32: | ||
787 | + case INDEX_op_qemu_st8_i32: | ||
788 | return C_O0_I2(s, L); | ||
789 | - case INDEX_op_qemu_st8_a64_i32: | ||
790 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(s, L) : C_O0_I3(s, L, L); | ||
791 | |||
792 | - case INDEX_op_qemu_ld_a32_i64: | ||
793 | + case INDEX_op_qemu_ld_i64: | ||
794 | return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, L) : C_O2_I1(r, r, L); | ||
795 | - case INDEX_op_qemu_ld_a64_i64: | ||
796 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, L) : C_O2_I2(r, r, L, L); | ||
797 | |||
798 | - case INDEX_op_qemu_st_a32_i64: | ||
799 | + case INDEX_op_qemu_st_i64: | ||
800 | return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(L, L) : C_O0_I3(L, L, L); | ||
801 | - case INDEX_op_qemu_st_a64_i64: | ||
802 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(L, L) : C_O0_I4(L, L, L, L); | ||
803 | |||
804 | - case INDEX_op_qemu_ld_a32_i128: | ||
805 | - case INDEX_op_qemu_ld_a64_i128: | ||
806 | + case INDEX_op_qemu_ld_i128: | ||
807 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
808 | return C_O2_I1(r, r, L); | ||
809 | - case INDEX_op_qemu_st_a32_i128: | ||
810 | - case INDEX_op_qemu_st_a64_i128: | ||
811 | + case INDEX_op_qemu_st_i128: | ||
812 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
813 | return C_O0_I3(L, L, L); | ||
814 | |||
815 | diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc | ||
816 | index XXXXXXX..XXXXXXX 100644 | ||
817 | --- a/tcg/loongarch64/tcg-target.c.inc | ||
818 | +++ b/tcg/loongarch64/tcg-target.c.inc | ||
819 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
820 | tcg_out_ldst(s, OPC_ST_D, a0, a1, a2); | ||
821 | break; | ||
822 | |||
823 | - case INDEX_op_qemu_ld_a32_i32: | ||
824 | - case INDEX_op_qemu_ld_a64_i32: | ||
825 | + case INDEX_op_qemu_ld_i32: | ||
826 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I32); | ||
827 | break; | ||
828 | - case INDEX_op_qemu_ld_a32_i64: | ||
829 | - case INDEX_op_qemu_ld_a64_i64: | ||
830 | + case INDEX_op_qemu_ld_i64: | ||
831 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I64); | ||
832 | break; | ||
833 | - case INDEX_op_qemu_ld_a32_i128: | ||
834 | - case INDEX_op_qemu_ld_a64_i128: | ||
835 | + case INDEX_op_qemu_ld_i128: | ||
836 | tcg_out_qemu_ldst_i128(s, a0, a1, a2, a3, true); | ||
837 | break; | ||
838 | - case INDEX_op_qemu_st_a32_i32: | ||
839 | - case INDEX_op_qemu_st_a64_i32: | ||
840 | + case INDEX_op_qemu_st_i32: | ||
841 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I32); | ||
842 | break; | ||
843 | - case INDEX_op_qemu_st_a32_i64: | ||
844 | - case INDEX_op_qemu_st_a64_i64: | ||
845 | + case INDEX_op_qemu_st_i64: | ||
846 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I64); | ||
847 | break; | ||
848 | - case INDEX_op_qemu_st_a32_i128: | ||
849 | - case INDEX_op_qemu_st_a64_i128: | ||
850 | + case INDEX_op_qemu_st_i128: | ||
851 | tcg_out_qemu_ldst_i128(s, a0, a1, a2, a3, false); | ||
852 | break; | ||
853 | |||
854 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
855 | case INDEX_op_st32_i64: | ||
856 | case INDEX_op_st_i32: | ||
857 | case INDEX_op_st_i64: | ||
858 | - case INDEX_op_qemu_st_a32_i32: | ||
859 | - case INDEX_op_qemu_st_a64_i32: | ||
860 | - case INDEX_op_qemu_st_a32_i64: | ||
861 | - case INDEX_op_qemu_st_a64_i64: | ||
862 | + case INDEX_op_qemu_st_i32: | ||
863 | + case INDEX_op_qemu_st_i64: | ||
864 | return C_O0_I2(rZ, r); | ||
865 | |||
866 | - case INDEX_op_qemu_ld_a32_i128: | ||
867 | - case INDEX_op_qemu_ld_a64_i128: | ||
868 | + case INDEX_op_qemu_ld_i128: | ||
869 | return C_N2_I1(r, r, r); | ||
870 | |||
871 | - case INDEX_op_qemu_st_a32_i128: | ||
872 | - case INDEX_op_qemu_st_a64_i128: | ||
873 | + case INDEX_op_qemu_st_i128: | ||
874 | return C_O0_I3(r, r, r); | ||
875 | |||
876 | case INDEX_op_brcond_i32: | ||
877 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
878 | case INDEX_op_ld32u_i64: | ||
879 | case INDEX_op_ld_i32: | ||
880 | case INDEX_op_ld_i64: | ||
881 | - case INDEX_op_qemu_ld_a32_i32: | ||
882 | - case INDEX_op_qemu_ld_a64_i32: | ||
883 | - case INDEX_op_qemu_ld_a32_i64: | ||
884 | - case INDEX_op_qemu_ld_a64_i64: | ||
885 | + case INDEX_op_qemu_ld_i32: | ||
886 | + case INDEX_op_qemu_ld_i64: | ||
887 | return C_O1_I1(r, r); | ||
888 | |||
889 | case INDEX_op_andc_i32: | ||
890 | diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc | ||
891 | index XXXXXXX..XXXXXXX 100644 | ||
892 | --- a/tcg/mips/tcg-target.c.inc | ||
893 | +++ b/tcg/mips/tcg-target.c.inc | ||
894 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
895 | tcg_out_setcond2(s, args[5], a0, a1, a2, args[3], args[4]); | ||
896 | break; | ||
897 | |||
898 | - case INDEX_op_qemu_ld_a64_i32: | ||
899 | - if (TCG_TARGET_REG_BITS == 32) { | ||
900 | - tcg_out_qemu_ld(s, a0, 0, a1, a2, args[3], TCG_TYPE_I32); | ||
901 | - break; | ||
902 | - } | ||
903 | - /* fall through */ | ||
904 | - case INDEX_op_qemu_ld_a32_i32: | ||
905 | + case INDEX_op_qemu_ld_i32: | ||
906 | tcg_out_qemu_ld(s, a0, 0, a1, 0, a2, TCG_TYPE_I32); | ||
907 | break; | ||
908 | - case INDEX_op_qemu_ld_a32_i64: | ||
909 | + case INDEX_op_qemu_ld_i64: | ||
910 | if (TCG_TARGET_REG_BITS == 64) { | ||
911 | tcg_out_qemu_ld(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); | ||
912 | } else { | ||
913 | tcg_out_qemu_ld(s, a0, a1, a2, 0, args[3], TCG_TYPE_I64); | ||
914 | } | ||
915 | break; | ||
916 | - case INDEX_op_qemu_ld_a64_i64: | ||
917 | - if (TCG_TARGET_REG_BITS == 64) { | ||
918 | - tcg_out_qemu_ld(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); | ||
919 | - } else { | ||
920 | - tcg_out_qemu_ld(s, a0, a1, a2, args[3], args[4], TCG_TYPE_I64); | ||
921 | - } | ||
922 | - break; | ||
923 | |||
924 | - case INDEX_op_qemu_st_a64_i32: | ||
925 | - if (TCG_TARGET_REG_BITS == 32) { | ||
926 | - tcg_out_qemu_st(s, a0, 0, a1, a2, args[3], TCG_TYPE_I32); | ||
927 | - break; | ||
928 | - } | ||
929 | - /* fall through */ | ||
930 | - case INDEX_op_qemu_st_a32_i32: | ||
931 | + case INDEX_op_qemu_st_i32: | ||
932 | tcg_out_qemu_st(s, a0, 0, a1, 0, a2, TCG_TYPE_I32); | ||
933 | break; | ||
934 | - case INDEX_op_qemu_st_a32_i64: | ||
935 | + case INDEX_op_qemu_st_i64: | ||
936 | if (TCG_TARGET_REG_BITS == 64) { | ||
937 | tcg_out_qemu_st(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); | ||
938 | } else { | ||
939 | tcg_out_qemu_st(s, a0, a1, a2, 0, args[3], TCG_TYPE_I64); | ||
940 | } | ||
941 | break; | ||
942 | - case INDEX_op_qemu_st_a64_i64: | ||
943 | - if (TCG_TARGET_REG_BITS == 64) { | ||
944 | - tcg_out_qemu_st(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); | ||
945 | - } else { | ||
946 | - tcg_out_qemu_st(s, a0, a1, a2, args[3], args[4], TCG_TYPE_I64); | ||
947 | - } | ||
948 | - break; | ||
949 | |||
950 | case INDEX_op_add2_i32: | ||
951 | tcg_out_addsub2(s, a0, a1, a2, args[3], args[4], args[5], | ||
952 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
953 | case INDEX_op_brcond2_i32: | ||
954 | return C_O0_I4(rZ, rZ, rZ, rZ); | ||
955 | |||
956 | - case INDEX_op_qemu_ld_a32_i32: | ||
957 | + case INDEX_op_qemu_ld_i32: | ||
958 | return C_O1_I1(r, r); | ||
959 | - case INDEX_op_qemu_ld_a64_i32: | ||
960 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O1_I2(r, r, r); | ||
961 | - case INDEX_op_qemu_st_a32_i32: | ||
962 | + case INDEX_op_qemu_st_i32: | ||
963 | return C_O0_I2(rZ, r); | ||
964 | - case INDEX_op_qemu_st_a64_i32: | ||
965 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(rZ, r) : C_O0_I3(rZ, r, r); | ||
966 | - case INDEX_op_qemu_ld_a32_i64: | ||
967 | + case INDEX_op_qemu_ld_i64: | ||
968 | return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I1(r, r, r); | ||
969 | - case INDEX_op_qemu_ld_a64_i64: | ||
970 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I2(r, r, r, r); | ||
971 | - case INDEX_op_qemu_st_a32_i64: | ||
972 | + case INDEX_op_qemu_st_i64: | ||
973 | return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(rZ, r) : C_O0_I3(rZ, rZ, r); | ||
974 | - case INDEX_op_qemu_st_a64_i64: | ||
975 | - return (TCG_TARGET_REG_BITS == 64 ? C_O0_I2(rZ, r) | ||
976 | - : C_O0_I4(rZ, rZ, r, r)); | ||
977 | |||
978 | default: | ||
979 | return C_NotImplemented; | ||
980 | diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc | ||
981 | index XXXXXXX..XXXXXXX 100644 | ||
982 | --- a/tcg/ppc/tcg-target.c.inc | ||
983 | +++ b/tcg/ppc/tcg-target.c.inc | ||
984 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
985 | tcg_out32(s, MODUD | TAB(args[0], args[1], args[2])); | ||
986 | break; | ||
987 | |||
988 | - case INDEX_op_qemu_ld_a64_i32: | ||
989 | - if (TCG_TARGET_REG_BITS == 32) { | ||
990 | - tcg_out_qemu_ld(s, args[0], -1, args[1], args[2], | ||
991 | - args[3], TCG_TYPE_I32); | ||
992 | - break; | ||
993 | - } | ||
994 | - /* fall through */ | ||
995 | - case INDEX_op_qemu_ld_a32_i32: | ||
996 | + case INDEX_op_qemu_ld_i32: | ||
997 | tcg_out_qemu_ld(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
998 | break; | ||
999 | - case INDEX_op_qemu_ld_a32_i64: | ||
1000 | + case INDEX_op_qemu_ld_i64: | ||
1001 | if (TCG_TARGET_REG_BITS == 64) { | ||
1002 | tcg_out_qemu_ld(s, args[0], -1, args[1], -1, | ||
1003 | args[2], TCG_TYPE_I64); | ||
1004 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1005 | args[3], TCG_TYPE_I64); | ||
1006 | } | ||
1007 | break; | ||
1008 | - case INDEX_op_qemu_ld_a64_i64: | ||
1009 | - if (TCG_TARGET_REG_BITS == 64) { | ||
1010 | - tcg_out_qemu_ld(s, args[0], -1, args[1], -1, | ||
1011 | - args[2], TCG_TYPE_I64); | ||
1012 | - } else { | ||
1013 | - tcg_out_qemu_ld(s, args[0], args[1], args[2], args[3], | ||
1014 | - args[4], TCG_TYPE_I64); | ||
1015 | - } | ||
1016 | - break; | ||
1017 | - case INDEX_op_qemu_ld_a32_i128: | ||
1018 | - case INDEX_op_qemu_ld_a64_i128: | ||
1019 | + case INDEX_op_qemu_ld_i128: | ||
1020 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
1021 | tcg_out_qemu_ldst_i128(s, args[0], args[1], args[2], args[3], true); | ||
1022 | break; | ||
1023 | |||
1024 | - case INDEX_op_qemu_st_a64_i32: | ||
1025 | - if (TCG_TARGET_REG_BITS == 32) { | ||
1026 | - tcg_out_qemu_st(s, args[0], -1, args[1], args[2], | ||
1027 | - args[3], TCG_TYPE_I32); | ||
1028 | - break; | ||
1029 | - } | ||
1030 | - /* fall through */ | ||
1031 | - case INDEX_op_qemu_st_a32_i32: | ||
1032 | + case INDEX_op_qemu_st_i32: | ||
1033 | tcg_out_qemu_st(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
1034 | break; | ||
1035 | - case INDEX_op_qemu_st_a32_i64: | ||
1036 | + case INDEX_op_qemu_st_i64: | ||
1037 | if (TCG_TARGET_REG_BITS == 64) { | ||
1038 | tcg_out_qemu_st(s, args[0], -1, args[1], -1, | ||
1039 | args[2], TCG_TYPE_I64); | ||
1040 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1041 | args[3], TCG_TYPE_I64); | ||
1042 | } | ||
1043 | break; | ||
1044 | - case INDEX_op_qemu_st_a64_i64: | ||
1045 | - if (TCG_TARGET_REG_BITS == 64) { | ||
1046 | - tcg_out_qemu_st(s, args[0], -1, args[1], -1, | ||
1047 | - args[2], TCG_TYPE_I64); | ||
1048 | - } else { | ||
1049 | - tcg_out_qemu_st(s, args[0], args[1], args[2], args[3], | ||
1050 | - args[4], TCG_TYPE_I64); | ||
1051 | - } | ||
1052 | - break; | ||
1053 | - case INDEX_op_qemu_st_a32_i128: | ||
1054 | - case INDEX_op_qemu_st_a64_i128: | ||
1055 | + case INDEX_op_qemu_st_i128: | ||
1056 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); | ||
1057 | tcg_out_qemu_ldst_i128(s, args[0], args[1], args[2], args[3], false); | ||
1058 | break; | ||
1059 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1060 | case INDEX_op_sub2_i32: | ||
1061 | return C_O2_I4(r, r, rI, rZM, r, r); | ||
1062 | |||
1063 | - case INDEX_op_qemu_ld_a32_i32: | ||
1064 | + case INDEX_op_qemu_ld_i32: | ||
1065 | return C_O1_I1(r, r); | ||
1066 | - case INDEX_op_qemu_ld_a64_i32: | ||
1067 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O1_I2(r, r, r); | ||
1068 | - case INDEX_op_qemu_ld_a32_i64: | ||
1069 | + case INDEX_op_qemu_ld_i64: | ||
1070 | return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I1(r, r, r); | ||
1071 | - case INDEX_op_qemu_ld_a64_i64: | ||
1072 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I2(r, r, r, r); | ||
1073 | |||
1074 | - case INDEX_op_qemu_st_a32_i32: | ||
1075 | + case INDEX_op_qemu_st_i32: | ||
1076 | return C_O0_I2(r, r); | ||
1077 | - case INDEX_op_qemu_st_a64_i32: | ||
1078 | + case INDEX_op_qemu_st_i64: | ||
1079 | return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I3(r, r, r); | ||
1080 | - case INDEX_op_qemu_st_a32_i64: | ||
1081 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I3(r, r, r); | ||
1082 | - case INDEX_op_qemu_st_a64_i64: | ||
1083 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I4(r, r, r, r); | ||
1084 | |||
1085 | - case INDEX_op_qemu_ld_a32_i128: | ||
1086 | - case INDEX_op_qemu_ld_a64_i128: | ||
1087 | + case INDEX_op_qemu_ld_i128: | ||
1088 | return C_N1O1_I1(o, m, r); | ||
1089 | - case INDEX_op_qemu_st_a32_i128: | ||
1090 | - case INDEX_op_qemu_st_a64_i128: | ||
1091 | + case INDEX_op_qemu_st_i128: | ||
1092 | return C_O0_I3(o, m, r); | ||
1093 | |||
1094 | case INDEX_op_add_vec: | ||
1095 | diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc | ||
1096 | index XXXXXXX..XXXXXXX 100644 | ||
1097 | --- a/tcg/riscv/tcg-target.c.inc | ||
1098 | +++ b/tcg/riscv/tcg-target.c.inc | ||
1099 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1100 | args[3], const_args[3], args[4], const_args[4]); | ||
1101 | break; | ||
1102 | |||
1103 | - case INDEX_op_qemu_ld_a32_i32: | ||
1104 | - case INDEX_op_qemu_ld_a64_i32: | ||
1105 | + case INDEX_op_qemu_ld_i32: | ||
1106 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I32); | ||
1107 | break; | ||
1108 | - case INDEX_op_qemu_ld_a32_i64: | ||
1109 | - case INDEX_op_qemu_ld_a64_i64: | ||
1110 | + case INDEX_op_qemu_ld_i64: | ||
1111 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I64); | ||
1112 | break; | ||
1113 | - case INDEX_op_qemu_st_a32_i32: | ||
1114 | - case INDEX_op_qemu_st_a64_i32: | ||
1115 | + case INDEX_op_qemu_st_i32: | ||
1116 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I32); | ||
1117 | break; | ||
1118 | - case INDEX_op_qemu_st_a32_i64: | ||
1119 | - case INDEX_op_qemu_st_a64_i64: | ||
1120 | + case INDEX_op_qemu_st_i64: | ||
1121 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I64); | ||
1122 | break; | ||
1123 | |||
1124 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1125 | case INDEX_op_sub2_i64: | ||
1126 | return C_O2_I4(r, r, rZ, rZ, rM, rM); | ||
1127 | |||
1128 | - case INDEX_op_qemu_ld_a32_i32: | ||
1129 | - case INDEX_op_qemu_ld_a64_i32: | ||
1130 | - case INDEX_op_qemu_ld_a32_i64: | ||
1131 | - case INDEX_op_qemu_ld_a64_i64: | ||
1132 | + case INDEX_op_qemu_ld_i32: | ||
1133 | + case INDEX_op_qemu_ld_i64: | ||
1134 | return C_O1_I1(r, r); | ||
1135 | - case INDEX_op_qemu_st_a32_i32: | ||
1136 | - case INDEX_op_qemu_st_a64_i32: | ||
1137 | - case INDEX_op_qemu_st_a32_i64: | ||
1138 | - case INDEX_op_qemu_st_a64_i64: | ||
1139 | + case INDEX_op_qemu_st_i32: | ||
1140 | + case INDEX_op_qemu_st_i64: | ||
1141 | return C_O0_I2(rZ, r); | ||
1142 | |||
1143 | case INDEX_op_st_vec: | ||
1144 | diff --git a/tcg/s390x/tcg-target.c.inc b/tcg/s390x/tcg-target.c.inc | ||
1145 | index XXXXXXX..XXXXXXX 100644 | ||
1146 | --- a/tcg/s390x/tcg-target.c.inc | ||
1147 | +++ b/tcg/s390x/tcg-target.c.inc | ||
1148 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1149 | args[2], const_args[2], args[3], const_args[3], args[4]); | ||
1150 | break; | ||
1151 | |||
1152 | - case INDEX_op_qemu_ld_a32_i32: | ||
1153 | - case INDEX_op_qemu_ld_a64_i32: | ||
1154 | + case INDEX_op_qemu_ld_i32: | ||
1155 | tcg_out_qemu_ld(s, args[0], args[1], args[2], TCG_TYPE_I32); | ||
1156 | break; | ||
1157 | - case INDEX_op_qemu_ld_a32_i64: | ||
1158 | - case INDEX_op_qemu_ld_a64_i64: | ||
1159 | + case INDEX_op_qemu_ld_i64: | ||
1160 | tcg_out_qemu_ld(s, args[0], args[1], args[2], TCG_TYPE_I64); | ||
1161 | break; | ||
1162 | - case INDEX_op_qemu_st_a32_i32: | ||
1163 | - case INDEX_op_qemu_st_a64_i32: | ||
1164 | + case INDEX_op_qemu_st_i32: | ||
1165 | tcg_out_qemu_st(s, args[0], args[1], args[2], TCG_TYPE_I32); | ||
1166 | break; | ||
1167 | - case INDEX_op_qemu_st_a32_i64: | ||
1168 | - case INDEX_op_qemu_st_a64_i64: | ||
1169 | + case INDEX_op_qemu_st_i64: | ||
1170 | tcg_out_qemu_st(s, args[0], args[1], args[2], TCG_TYPE_I64); | ||
1171 | break; | ||
1172 | - case INDEX_op_qemu_ld_a32_i128: | ||
1173 | - case INDEX_op_qemu_ld_a64_i128: | ||
1174 | + case INDEX_op_qemu_ld_i128: | ||
1175 | tcg_out_qemu_ldst_i128(s, args[0], args[1], args[2], args[3], true); | ||
1176 | break; | ||
1177 | - case INDEX_op_qemu_st_a32_i128: | ||
1178 | - case INDEX_op_qemu_st_a64_i128: | ||
1179 | + case INDEX_op_qemu_st_i128: | ||
1180 | tcg_out_qemu_ldst_i128(s, args[0], args[1], args[2], args[3], false); | ||
1181 | break; | ||
1182 | |||
1183 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1184 | case INDEX_op_ctpop_i64: | ||
1185 | return C_O1_I1(r, r); | ||
1186 | |||
1187 | - case INDEX_op_qemu_ld_a32_i32: | ||
1188 | - case INDEX_op_qemu_ld_a64_i32: | ||
1189 | - case INDEX_op_qemu_ld_a32_i64: | ||
1190 | - case INDEX_op_qemu_ld_a64_i64: | ||
1191 | + case INDEX_op_qemu_ld_i32: | ||
1192 | + case INDEX_op_qemu_ld_i64: | ||
1193 | return C_O1_I1(r, r); | ||
1194 | - case INDEX_op_qemu_st_a32_i64: | ||
1195 | - case INDEX_op_qemu_st_a64_i64: | ||
1196 | - case INDEX_op_qemu_st_a32_i32: | ||
1197 | - case INDEX_op_qemu_st_a64_i32: | ||
1198 | + case INDEX_op_qemu_st_i64: | ||
1199 | + case INDEX_op_qemu_st_i32: | ||
1200 | return C_O0_I2(r, r); | ||
1201 | - case INDEX_op_qemu_ld_a32_i128: | ||
1202 | - case INDEX_op_qemu_ld_a64_i128: | ||
1203 | + case INDEX_op_qemu_ld_i128: | ||
1204 | return C_O2_I1(o, m, r); | ||
1205 | - case INDEX_op_qemu_st_a32_i128: | ||
1206 | - case INDEX_op_qemu_st_a64_i128: | ||
1207 | + case INDEX_op_qemu_st_i128: | ||
1208 | return C_O0_I3(o, m, r); | ||
1209 | |||
1210 | case INDEX_op_deposit_i32: | ||
1211 | diff --git a/tcg/sparc64/tcg-target.c.inc b/tcg/sparc64/tcg-target.c.inc | ||
1212 | index XXXXXXX..XXXXXXX 100644 | ||
1213 | --- a/tcg/sparc64/tcg-target.c.inc | ||
1214 | +++ b/tcg/sparc64/tcg-target.c.inc | ||
1215 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1216 | tcg_out_arithi(s, a1, a0, 32, SHIFT_SRLX); | ||
1217 | break; | ||
1218 | |||
1219 | - case INDEX_op_qemu_ld_a32_i32: | ||
1220 | - case INDEX_op_qemu_ld_a64_i32: | ||
1221 | + case INDEX_op_qemu_ld_i32: | ||
1222 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I32); | ||
1223 | break; | ||
1224 | - case INDEX_op_qemu_ld_a32_i64: | ||
1225 | - case INDEX_op_qemu_ld_a64_i64: | ||
1226 | + case INDEX_op_qemu_ld_i64: | ||
1227 | tcg_out_qemu_ld(s, a0, a1, a2, TCG_TYPE_I64); | ||
1228 | break; | ||
1229 | - case INDEX_op_qemu_st_a32_i32: | ||
1230 | - case INDEX_op_qemu_st_a64_i32: | ||
1231 | + case INDEX_op_qemu_st_i32: | ||
1232 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I32); | ||
1233 | break; | ||
1234 | - case INDEX_op_qemu_st_a32_i64: | ||
1235 | - case INDEX_op_qemu_st_a64_i64: | ||
1236 | + case INDEX_op_qemu_st_i64: | ||
1237 | tcg_out_qemu_st(s, a0, a1, a2, TCG_TYPE_I64); | ||
1238 | break; | ||
1239 | |||
1240 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1241 | case INDEX_op_extu_i32_i64: | ||
1242 | case INDEX_op_extract_i64: | ||
1243 | case INDEX_op_sextract_i64: | ||
1244 | - case INDEX_op_qemu_ld_a32_i32: | ||
1245 | - case INDEX_op_qemu_ld_a64_i32: | ||
1246 | - case INDEX_op_qemu_ld_a32_i64: | ||
1247 | - case INDEX_op_qemu_ld_a64_i64: | ||
1248 | + case INDEX_op_qemu_ld_i32: | ||
1249 | + case INDEX_op_qemu_ld_i64: | ||
1250 | return C_O1_I1(r, r); | ||
1251 | |||
1252 | case INDEX_op_st8_i32: | ||
1253 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1254 | case INDEX_op_st_i32: | ||
1255 | case INDEX_op_st32_i64: | ||
1256 | case INDEX_op_st_i64: | ||
1257 | - case INDEX_op_qemu_st_a32_i32: | ||
1258 | - case INDEX_op_qemu_st_a64_i32: | ||
1259 | - case INDEX_op_qemu_st_a32_i64: | ||
1260 | - case INDEX_op_qemu_st_a64_i64: | ||
1261 | + case INDEX_op_qemu_st_i32: | ||
1262 | + case INDEX_op_qemu_st_i64: | ||
1263 | return C_O0_I2(rZ, r); | ||
1264 | |||
1265 | case INDEX_op_add_i32: | ||
1266 | diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc | ||
1267 | index XXXXXXX..XXXXXXX 100644 | ||
1268 | --- a/tcg/tci/tcg-target.c.inc | ||
1269 | +++ b/tcg/tci/tcg-target.c.inc | ||
1270 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
1271 | case INDEX_op_setcond2_i32: | ||
1272 | return C_O1_I4(r, r, r, r, r); | ||
1273 | |||
1274 | - case INDEX_op_qemu_ld_a32_i32: | ||
1275 | + case INDEX_op_qemu_ld_i32: | ||
1276 | return C_O1_I1(r, r); | ||
1277 | - case INDEX_op_qemu_ld_a64_i32: | ||
1278 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O1_I2(r, r, r); | ||
1279 | - case INDEX_op_qemu_ld_a32_i64: | ||
1280 | + case INDEX_op_qemu_ld_i64: | ||
1281 | return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I1(r, r, r); | ||
1282 | - case INDEX_op_qemu_ld_a64_i64: | ||
1283 | - return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I2(r, r, r, r); | ||
1284 | - case INDEX_op_qemu_st_a32_i32: | ||
1285 | + case INDEX_op_qemu_st_i32: | ||
1286 | return C_O0_I2(r, r); | ||
1287 | - case INDEX_op_qemu_st_a64_i32: | ||
1288 | + case INDEX_op_qemu_st_i64: | ||
1289 | return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I3(r, r, r); | ||
1290 | - case INDEX_op_qemu_st_a32_i64: | ||
1291 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I3(r, r, r); | ||
1292 | - case INDEX_op_qemu_st_a64_i64: | ||
1293 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(r, r) : C_O0_I4(r, r, r, r); | ||
1294 | |||
1295 | default: | ||
1296 | return C_NotImplemented; | ||
1297 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op_rrrbb(TCGContext *s, TCGOpcode op, TCGReg r0, | ||
1298 | tcg_out32(s, insn); | ||
1299 | } | ||
1300 | |||
1301 | -static void tcg_out_op_rrrrr(TCGContext *s, TCGOpcode op, TCGReg r0, | ||
1302 | - TCGReg r1, TCGReg r2, TCGReg r3, TCGReg r4) | ||
1303 | -{ | ||
1304 | - tcg_insn_unit insn = 0; | ||
1305 | - | ||
1306 | - insn = deposit32(insn, 0, 8, op); | ||
1307 | - insn = deposit32(insn, 8, 4, r0); | ||
1308 | - insn = deposit32(insn, 12, 4, r1); | ||
1309 | - insn = deposit32(insn, 16, 4, r2); | ||
1310 | - insn = deposit32(insn, 20, 4, r3); | ||
1311 | - insn = deposit32(insn, 24, 4, r4); | ||
1312 | - tcg_out32(s, insn); | ||
1313 | -} | ||
1314 | - | ||
1315 | static void tcg_out_op_rrrr(TCGContext *s, TCGOpcode op, | ||
1316 | TCGReg r0, TCGReg r1, TCGReg r2, TCGReg r3) | ||
1317 | { | ||
1318 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
1319 | tcg_out_op_rrrr(s, opc, args[0], args[1], args[2], args[3]); | ||
1320 | break; | ||
1321 | |||
1322 | - case INDEX_op_qemu_ld_a32_i32: | ||
1323 | - case INDEX_op_qemu_st_a32_i32: | ||
1324 | - tcg_out_op_rrm(s, opc, args[0], args[1], args[2]); | ||
1325 | - break; | ||
1326 | - case INDEX_op_qemu_ld_a64_i32: | ||
1327 | - case INDEX_op_qemu_st_a64_i32: | ||
1328 | - case INDEX_op_qemu_ld_a32_i64: | ||
1329 | - case INDEX_op_qemu_st_a32_i64: | ||
1330 | - if (TCG_TARGET_REG_BITS == 64) { | ||
1331 | - tcg_out_op_rrm(s, opc, args[0], args[1], args[2]); | ||
1332 | - } else { | ||
1333 | + case INDEX_op_qemu_ld_i64: | ||
1334 | + case INDEX_op_qemu_st_i64: | ||
1335 | + if (TCG_TARGET_REG_BITS == 32) { | ||
1336 | tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_TMP, args[3]); | ||
1337 | tcg_out_op_rrrr(s, opc, args[0], args[1], args[2], TCG_REG_TMP); | ||
1338 | + break; | ||
1339 | } | ||
1340 | - break; | ||
1341 | - case INDEX_op_qemu_ld_a64_i64: | ||
1342 | - case INDEX_op_qemu_st_a64_i64: | ||
1343 | - if (TCG_TARGET_REG_BITS == 64) { | ||
1344 | - tcg_out_op_rrm(s, opc, args[0], args[1], args[2]); | ||
1345 | + /* fall through */ | ||
1346 | + case INDEX_op_qemu_ld_i32: | ||
1347 | + case INDEX_op_qemu_st_i32: | ||
1348 | + if (TCG_TARGET_REG_BITS == 64 && s->addr_type == TCG_TYPE_I32) { | ||
1349 | + tcg_out_ext32u(s, TCG_REG_TMP, args[1]); | ||
1350 | + tcg_out_op_rrm(s, opc, args[0], TCG_REG_TMP, args[2]); | ||
1351 | } else { | ||
1352 | - tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_TMP, args[4]); | ||
1353 | - tcg_out_op_rrrrr(s, opc, args[0], args[1], | ||
1354 | - args[2], args[3], TCG_REG_TMP); | ||
1355 | + tcg_out_op_rrm(s, opc, args[0], args[1], args[2]); | ||
1356 | } | ||
1357 | break; | ||
1358 | |||
242 | -- | 1359 | -- |
243 | 2.17.2 | 1360 | 2.43.0 |
244 | |||
245 | diff view generated by jsdifflib |
1 | Reviewed-by: David Hildenbrand <david@redhat.com> | 1 | The guest address will now always be TCG_TYPE_I32. |
---|---|---|---|
2 | |||
2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
3 | --- | 4 | --- |
4 | target/s390x/mem_helper.c | 40 +++++++++++++++++++-------------------- | 5 | tcg/arm/tcg-target.c.inc | 63 ++++++++++++++-------------------------- |
5 | target/s390x/translate.c | 25 +++++++++++++++++------- | 6 | 1 file changed, 21 insertions(+), 42 deletions(-) |
6 | 2 files changed, 38 insertions(+), 27 deletions(-) | ||
7 | 7 | ||
8 | diff --git a/target/s390x/mem_helper.c b/target/s390x/mem_helper.c | 8 | diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc |
9 | index XXXXXXX..XXXXXXX 100644 | 9 | index XXXXXXX..XXXXXXX 100644 |
10 | --- a/target/s390x/mem_helper.c | 10 | --- a/tcg/arm/tcg-target.c.inc |
11 | +++ b/target/s390x/mem_helper.c | 11 | +++ b/tcg/arm/tcg-target.c.inc |
12 | @@ -XXX,XX +XXX,XX @@ void HELPER(cdsg_parallel)(CPUS390XState *env, uint64_t addr, | 12 | @@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb) |
13 | Int128 oldv; | 13 | #define MIN_TLB_MASK_TABLE_OFS -256 |
14 | bool fail; | 14 | |
15 | 15 | static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | |
16 | - if (!HAVE_CMPXCHG128) { | 16 | - TCGReg addrlo, TCGReg addrhi, |
17 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 17 | - MemOpIdx oi, bool is_ld) |
18 | - } | 18 | + TCGReg addr, MemOpIdx oi, bool is_ld) |
19 | + assert(HAVE_CMPXCHG128); | ||
20 | |||
21 | mem_idx = cpu_mmu_index(env, false); | ||
22 | oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
23 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(lpq_parallel)(CPUS390XState *env, uint64_t addr) | ||
24 | { | 19 | { |
25 | uintptr_t ra = GETPC(); | 20 | TCGLabelQemuLdst *ldst = NULL; |
26 | uint64_t hi, lo; | 21 | MemOp opc = get_memop(oi); |
27 | + int mem_idx; | 22 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
28 | + TCGMemOpIdx oi; | 23 | if (tcg_use_softmmu) { |
29 | + Int128 v; | 24 | *h = (HostAddress){ |
30 | 25 | .cond = COND_AL, | |
31 | - if (HAVE_ATOMIC128) { | 26 | - .base = addrlo, |
32 | - int mem_idx = cpu_mmu_index(env, false); | 27 | + .base = addr, |
33 | - TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 28 | .index = TCG_REG_R1, |
34 | - Int128 v = helper_atomic_ldo_be_mmu(env, addr, oi, ra); | 29 | .index_scratch = true, |
35 | - hi = int128_gethi(v); | 30 | }; |
36 | - lo = int128_getlo(v); | 31 | } else { |
37 | - } else { | 32 | *h = (HostAddress){ |
38 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 33 | .cond = COND_AL, |
39 | - } | 34 | - .base = addrlo, |
40 | + assert(HAVE_ATOMIC128); | 35 | + .base = addr, |
41 | + | 36 | .index = guest_base ? TCG_REG_GUEST_BASE : -1, |
42 | + mem_idx = cpu_mmu_index(env, false); | 37 | .index_scratch = false, |
43 | + oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 38 | }; |
44 | + v = helper_atomic_ldo_be_mmu(env, addr, oi, ra); | 39 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
45 | + hi = int128_gethi(v); | 40 | ldst = new_ldst_label(s); |
46 | + lo = int128_getlo(v); | 41 | ldst->is_ld = is_ld; |
47 | 42 | ldst->oi = oi; | |
48 | env->retxl = lo; | 43 | - ldst->addrlo_reg = addrlo; |
49 | return hi; | 44 | - ldst->addrhi_reg = addrhi; |
50 | @@ -XXX,XX +XXX,XX @@ void HELPER(stpq_parallel)(CPUS390XState *env, uint64_t addr, | 45 | + ldst->addrlo_reg = addr; |
51 | uint64_t low, uint64_t high) | 46 | |
47 | /* Load cpu->neg.tlb.f[mmu_idx].{mask,table} into {r0,r1}. */ | ||
48 | QEMU_BUILD_BUG_ON(offsetof(CPUTLBDescFast, mask) != 0); | ||
49 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
50 | tcg_out_ldrd_8(s, COND_AL, TCG_REG_R0, TCG_AREG0, fast_off); | ||
51 | |||
52 | /* Extract the tlb index from the address into R0. */ | ||
53 | - tcg_out_dat_reg(s, COND_AL, ARITH_AND, TCG_REG_R0, TCG_REG_R0, addrlo, | ||
54 | + tcg_out_dat_reg(s, COND_AL, ARITH_AND, TCG_REG_R0, TCG_REG_R0, addr, | ||
55 | SHIFT_IMM_LSR(s->page_bits - CPU_TLB_ENTRY_BITS)); | ||
56 | |||
57 | /* | ||
58 | * Add the tlb_table pointer, creating the CPUTLBEntry address in R1. | ||
59 | - * Load the tlb comparator into R2/R3 and the fast path addend into R1. | ||
60 | + * Load the tlb comparator into R2 and the fast path addend into R1. | ||
61 | */ | ||
62 | QEMU_BUILD_BUG_ON(HOST_BIG_ENDIAN); | ||
63 | if (cmp_off == 0) { | ||
64 | - if (s->addr_type == TCG_TYPE_I32) { | ||
65 | - tcg_out_ld32_rwb(s, COND_AL, TCG_REG_R2, | ||
66 | - TCG_REG_R1, TCG_REG_R0); | ||
67 | - } else { | ||
68 | - tcg_out_ldrd_rwb(s, COND_AL, TCG_REG_R2, | ||
69 | - TCG_REG_R1, TCG_REG_R0); | ||
70 | - } | ||
71 | + tcg_out_ld32_rwb(s, COND_AL, TCG_REG_R2, TCG_REG_R1, TCG_REG_R0); | ||
72 | } else { | ||
73 | tcg_out_dat_reg(s, COND_AL, ARITH_ADD, | ||
74 | TCG_REG_R1, TCG_REG_R1, TCG_REG_R0, 0); | ||
75 | - if (s->addr_type == TCG_TYPE_I32) { | ||
76 | - tcg_out_ld32_12(s, COND_AL, TCG_REG_R2, TCG_REG_R1, cmp_off); | ||
77 | - } else { | ||
78 | - tcg_out_ldrd_8(s, COND_AL, TCG_REG_R2, TCG_REG_R1, cmp_off); | ||
79 | - } | ||
80 | + tcg_out_ld32_12(s, COND_AL, TCG_REG_R2, TCG_REG_R1, cmp_off); | ||
81 | } | ||
82 | |||
83 | /* Load the tlb addend. */ | ||
84 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
85 | * This leaves the least significant alignment bits unchanged, and of | ||
86 | * course must be zero. | ||
87 | */ | ||
88 | - t_addr = addrlo; | ||
89 | + t_addr = addr; | ||
90 | if (a_mask < s_mask) { | ||
91 | t_addr = TCG_REG_R0; | ||
92 | tcg_out_dat_imm(s, COND_AL, ARITH_ADD, t_addr, | ||
93 | - addrlo, s_mask - a_mask); | ||
94 | + addr, s_mask - a_mask); | ||
95 | } | ||
96 | if (use_armv7_instructions && s->page_bits <= 16) { | ||
97 | tcg_out_movi32(s, COND_AL, TCG_REG_TMP, ~(s->page_mask | a_mask)); | ||
98 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
99 | } else { | ||
100 | if (a_mask) { | ||
101 | tcg_debug_assert(a_mask <= 0xff); | ||
102 | - tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addrlo, a_mask); | ||
103 | + tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addr, a_mask); | ||
104 | } | ||
105 | tcg_out_dat_reg(s, COND_AL, ARITH_MOV, TCG_REG_TMP, 0, t_addr, | ||
106 | SHIFT_IMM_LSR(s->page_bits)); | ||
107 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
108 | 0, TCG_REG_R2, TCG_REG_TMP, | ||
109 | SHIFT_IMM_LSL(s->page_bits)); | ||
110 | } | ||
111 | - | ||
112 | - if (s->addr_type != TCG_TYPE_I32) { | ||
113 | - tcg_out_dat_reg(s, COND_EQ, ARITH_CMP, 0, TCG_REG_R3, addrhi, 0); | ||
114 | - } | ||
115 | } else if (a_mask) { | ||
116 | ldst = new_ldst_label(s); | ||
117 | ldst->is_ld = is_ld; | ||
118 | ldst->oi = oi; | ||
119 | - ldst->addrlo_reg = addrlo; | ||
120 | - ldst->addrhi_reg = addrhi; | ||
121 | + ldst->addrlo_reg = addr; | ||
122 | |||
123 | /* We are expecting alignment to max out at 7 */ | ||
124 | tcg_debug_assert(a_mask <= 0xff); | ||
125 | /* tst addr, #mask */ | ||
126 | - tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addrlo, a_mask); | ||
127 | + tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addr, a_mask); | ||
128 | } | ||
129 | |||
130 | return ldst; | ||
131 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc, TCGReg datalo, | ||
132 | } | ||
133 | |||
134 | static void tcg_out_qemu_ld(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
135 | - TCGReg addrlo, TCGReg addrhi, | ||
136 | - MemOpIdx oi, TCGType data_type) | ||
137 | + TCGReg addr, MemOpIdx oi, TCGType data_type) | ||
52 | { | 138 | { |
53 | uintptr_t ra = GETPC(); | 139 | MemOp opc = get_memop(oi); |
54 | + int mem_idx; | 140 | TCGLabelQemuLdst *ldst; |
55 | + TCGMemOpIdx oi; | 141 | HostAddress h; |
56 | + Int128 v; | 142 | |
57 | 143 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, true); | |
58 | - if (HAVE_ATOMIC128) { | 144 | + ldst = prepare_host_addr(s, &h, addr, oi, true); |
59 | - int mem_idx = cpu_mmu_index(env, false); | 145 | if (ldst) { |
60 | - TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 146 | ldst->type = data_type; |
61 | - Int128 v = int128_make128(low, high); | 147 | ldst->datalo_reg = datalo; |
62 | - helper_atomic_sto_be_mmu(env, addr, v, oi, ra); | 148 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc, TCGReg datalo, |
63 | - } else { | ||
64 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
65 | - } | ||
66 | + assert(HAVE_ATOMIC128); | ||
67 | + | ||
68 | + mem_idx = cpu_mmu_index(env, false); | ||
69 | + oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
70 | + v = int128_make128(low, high); | ||
71 | + helper_atomic_sto_be_mmu(env, addr, v, oi, ra); | ||
72 | } | 149 | } |
73 | 150 | ||
74 | /* Execute instruction. This instruction executes an insn modified with | 151 | static void tcg_out_qemu_st(TCGContext *s, TCGReg datalo, TCGReg datahi, |
75 | diff --git a/target/s390x/translate.c b/target/s390x/translate.c | 152 | - TCGReg addrlo, TCGReg addrhi, |
76 | index XXXXXXX..XXXXXXX 100644 | 153 | - MemOpIdx oi, TCGType data_type) |
77 | --- a/target/s390x/translate.c | 154 | + TCGReg addr, MemOpIdx oi, TCGType data_type) |
78 | +++ b/target/s390x/translate.c | ||
79 | @@ -XXX,XX +XXX,XX @@ | ||
80 | #include "trace-tcg.h" | ||
81 | #include "exec/translator.h" | ||
82 | #include "exec/log.h" | ||
83 | +#include "qemu/atomic128.h" | ||
84 | |||
85 | |||
86 | /* Information that (most) every instruction needs to manipulate. */ | ||
87 | @@ -XXX,XX +XXX,XX @@ static DisasJumpType op_cdsg(DisasContext *s, DisasOps *o) | ||
88 | int r3 = get_field(s->fields, r3); | ||
89 | int d2 = get_field(s->fields, d2); | ||
90 | int b2 = get_field(s->fields, b2); | ||
91 | + DisasJumpType ret = DISAS_NEXT; | ||
92 | TCGv_i64 addr; | ||
93 | TCGv_i32 t_r1, t_r3; | ||
94 | |||
95 | @@ -XXX,XX +XXX,XX @@ static DisasJumpType op_cdsg(DisasContext *s, DisasOps *o) | ||
96 | addr = get_address(s, 0, b2, d2); | ||
97 | t_r1 = tcg_const_i32(r1); | ||
98 | t_r3 = tcg_const_i32(r3); | ||
99 | - if (tb_cflags(s->base.tb) & CF_PARALLEL) { | ||
100 | + if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) { | ||
101 | + gen_helper_cdsg(cpu_env, addr, t_r1, t_r3); | ||
102 | + } else if (HAVE_CMPXCHG128) { | ||
103 | gen_helper_cdsg_parallel(cpu_env, addr, t_r1, t_r3); | ||
104 | } else { | ||
105 | - gen_helper_cdsg(cpu_env, addr, t_r1, t_r3); | ||
106 | + gen_helper_exit_atomic(cpu_env); | ||
107 | + ret = DISAS_NORETURN; | ||
108 | } | ||
109 | tcg_temp_free_i64(addr); | ||
110 | tcg_temp_free_i32(t_r1); | ||
111 | tcg_temp_free_i32(t_r3); | ||
112 | |||
113 | set_cc_static(s); | ||
114 | - return DISAS_NEXT; | ||
115 | + return ret; | ||
116 | } | ||
117 | |||
118 | static DisasJumpType op_csst(DisasContext *s, DisasOps *o) | ||
119 | @@ -XXX,XX +XXX,XX @@ static DisasJumpType op_lpd(DisasContext *s, DisasOps *o) | ||
120 | |||
121 | static DisasJumpType op_lpq(DisasContext *s, DisasOps *o) | ||
122 | { | 155 | { |
123 | - if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 156 | MemOp opc = get_memop(oi); |
124 | + if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) { | 157 | TCGLabelQemuLdst *ldst; |
125 | + gen_helper_lpq(o->out, cpu_env, o->in2); | 158 | HostAddress h; |
126 | + } else if (HAVE_ATOMIC128) { | 159 | |
127 | gen_helper_lpq_parallel(o->out, cpu_env, o->in2); | 160 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, false); |
128 | } else { | 161 | + ldst = prepare_host_addr(s, &h, addr, oi, false); |
129 | - gen_helper_lpq(o->out, cpu_env, o->in2); | 162 | if (ldst) { |
130 | + gen_helper_exit_atomic(cpu_env); | 163 | ldst->type = data_type; |
131 | + return DISAS_NORETURN; | 164 | ldst->datalo_reg = datalo; |
132 | } | 165 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, |
133 | return_low128(o->out2); | 166 | break; |
134 | return DISAS_NEXT; | 167 | |
135 | @@ -XXX,XX +XXX,XX @@ static DisasJumpType op_stmh(DisasContext *s, DisasOps *o) | 168 | case INDEX_op_qemu_ld_i32: |
136 | 169 | - tcg_out_qemu_ld(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | |
137 | static DisasJumpType op_stpq(DisasContext *s, DisasOps *o) | 170 | + tcg_out_qemu_ld(s, args[0], -1, args[1], args[2], TCG_TYPE_I32); |
138 | { | 171 | break; |
139 | - if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 172 | case INDEX_op_qemu_ld_i64: |
140 | + if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) { | 173 | - tcg_out_qemu_ld(s, args[0], args[1], args[2], -1, |
141 | + gen_helper_stpq(cpu_env, o->in2, o->out2, o->out); | 174 | - args[3], TCG_TYPE_I64); |
142 | + } else if (HAVE_ATOMIC128) { | 175 | + tcg_out_qemu_ld(s, args[0], args[1], args[2], args[3], TCG_TYPE_I64); |
143 | gen_helper_stpq_parallel(cpu_env, o->in2, o->out2, o->out); | 176 | break; |
144 | } else { | 177 | |
145 | - gen_helper_stpq(cpu_env, o->in2, o->out2, o->out); | 178 | case INDEX_op_qemu_st_i32: |
146 | + gen_helper_exit_atomic(cpu_env); | 179 | - tcg_out_qemu_st(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); |
147 | + return DISAS_NORETURN; | 180 | + tcg_out_qemu_st(s, args[0], -1, args[1], args[2], TCG_TYPE_I32); |
148 | } | 181 | break; |
149 | return DISAS_NEXT; | 182 | case INDEX_op_qemu_st_i64: |
150 | } | 183 | - tcg_out_qemu_st(s, args[0], args[1], args[2], -1, |
184 | - args[3], TCG_TYPE_I64); | ||
185 | + tcg_out_qemu_st(s, args[0], args[1], args[2], args[3], TCG_TYPE_I64); | ||
186 | break; | ||
187 | |||
188 | case INDEX_op_bswap16_i32: | ||
151 | -- | 189 | -- |
152 | 2.17.2 | 190 | 2.43.0 |
153 | |||
154 | diff view generated by jsdifflib |
1 | Reviewed-by: Emilio G. Cota <cota@braap.org> | 1 | The guest address will now always fit in one register. |
---|---|---|---|
2 | |||
3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
3 | --- | 5 | --- |
4 | target/ppc/helper.h | 2 +- | 6 | tcg/i386/tcg-target.c.inc | 56 ++++++++++++++------------------------- |
5 | target/ppc/mem_helper.c | 33 ++++++++++-- | 7 | 1 file changed, 20 insertions(+), 36 deletions(-) |
6 | target/ppc/translate.c | 115 +++++++++++++++++++++------------------- | ||
7 | 3 files changed, 88 insertions(+), 62 deletions(-) | ||
8 | 8 | ||
9 | diff --git a/target/ppc/helper.h b/target/ppc/helper.h | 9 | diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc |
10 | index XXXXXXX..XXXXXXX 100644 | 10 | index XXXXXXX..XXXXXXX 100644 |
11 | --- a/target/ppc/helper.h | 11 | --- a/tcg/i386/tcg-target.c.inc |
12 | +++ b/target/ppc/helper.h | 12 | +++ b/tcg/i386/tcg-target.c.inc |
13 | @@ -XXX,XX +XXX,XX @@ DEF_HELPER_4(dscliq, void, env, fprp, fprp, i32) | 13 | @@ -XXX,XX +XXX,XX @@ static inline int setup_guest_base_seg(void) |
14 | DEF_HELPER_1(tbegin, void, env) | 14 | * is required and fill in @h with the host address for the fast path. |
15 | DEF_HELPER_FLAGS_1(fixup_thrm, TCG_CALL_NO_RWG, void, env) | 15 | */ |
16 | 16 | static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | |
17 | -#if defined(TARGET_PPC64) && defined(CONFIG_ATOMIC128) | 17 | - TCGReg addrlo, TCGReg addrhi, |
18 | +#ifdef TARGET_PPC64 | 18 | - MemOpIdx oi, bool is_ld) |
19 | DEF_HELPER_FLAGS_3(lq_le_parallel, TCG_CALL_NO_WG, i64, env, tl, i32) | 19 | + TCGReg addr, MemOpIdx oi, bool is_ld) |
20 | DEF_HELPER_FLAGS_3(lq_be_parallel, TCG_CALL_NO_WG, i64, env, tl, i32) | 20 | { |
21 | DEF_HELPER_FLAGS_5(stq_le_parallel, TCG_CALL_NO_WG, | 21 | TCGLabelQemuLdst *ldst = NULL; |
22 | diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c | 22 | MemOp opc = get_memop(oi); |
23 | index XXXXXXX..XXXXXXX 100644 | 23 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
24 | --- a/target/ppc/mem_helper.c | 24 | } else { |
25 | +++ b/target/ppc/mem_helper.c | 25 | *h = x86_guest_base; |
26 | @@ -XXX,XX +XXX,XX @@ | 26 | } |
27 | #include "exec/cpu_ldst.h" | 27 | - h->base = addrlo; |
28 | #include "tcg.h" | 28 | + h->base = addr; |
29 | #include "internal.h" | 29 | h->aa = atom_and_align_for_opc(s, opc, MO_ATOM_IFALIGN, s_bits == MO_128); |
30 | +#include "qemu/atomic128.h" | 30 | a_mask = (1 << h->aa.align) - 1; |
31 | 31 | ||
32 | //#define DEBUG_OP | 32 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
33 | 33 | ldst = new_ldst_label(s); | |
34 | @@ -XXX,XX +XXX,XX @@ target_ulong helper_lscbx(CPUPPCState *env, target_ulong addr, uint32_t reg, | 34 | ldst->is_ld = is_ld; |
35 | return i; | 35 | ldst->oi = oi; |
36 | - ldst->addrlo_reg = addrlo; | ||
37 | - ldst->addrhi_reg = addrhi; | ||
38 | + ldst->addrlo_reg = addr; | ||
39 | |||
40 | if (TCG_TARGET_REG_BITS == 64) { | ||
41 | ttype = s->addr_type; | ||
42 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
43 | } | ||
44 | } | ||
45 | |||
46 | - tcg_out_mov(s, tlbtype, TCG_REG_L0, addrlo); | ||
47 | + tcg_out_mov(s, tlbtype, TCG_REG_L0, addr); | ||
48 | tcg_out_shifti(s, SHIFT_SHR + tlbrexw, TCG_REG_L0, | ||
49 | s->page_bits - CPU_TLB_ENTRY_BITS); | ||
50 | |||
51 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
52 | * check that we don't cross pages for the complete access. | ||
53 | */ | ||
54 | if (a_mask >= s_mask) { | ||
55 | - tcg_out_mov(s, ttype, TCG_REG_L1, addrlo); | ||
56 | + tcg_out_mov(s, ttype, TCG_REG_L1, addr); | ||
57 | } else { | ||
58 | tcg_out_modrm_offset(s, OPC_LEA + trexw, TCG_REG_L1, | ||
59 | - addrlo, s_mask - a_mask); | ||
60 | + addr, s_mask - a_mask); | ||
61 | } | ||
62 | tlb_mask = s->page_mask | a_mask; | ||
63 | tgen_arithi(s, ARITH_AND + trexw, TCG_REG_L1, tlb_mask, 0); | ||
64 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
65 | ldst->label_ptr[0] = s->code_ptr; | ||
66 | s->code_ptr += 4; | ||
67 | |||
68 | - if (TCG_TARGET_REG_BITS == 32 && s->addr_type == TCG_TYPE_I64) { | ||
69 | - /* cmp 4(TCG_REG_L0), addrhi */ | ||
70 | - tcg_out_modrm_offset(s, OPC_CMP_GvEv, addrhi, | ||
71 | - TCG_REG_L0, cmp_ofs + 4); | ||
72 | - | ||
73 | - /* jne slow_path */ | ||
74 | - tcg_out_opc(s, OPC_JCC_long + JCC_JNE, 0, 0, 0); | ||
75 | - ldst->label_ptr[1] = s->code_ptr; | ||
76 | - s->code_ptr += 4; | ||
77 | - } | ||
78 | - | ||
79 | /* TLB Hit. */ | ||
80 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_L0, TCG_REG_L0, | ||
81 | offsetof(CPUTLBEntry, addend)); | ||
82 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
83 | ldst = new_ldst_label(s); | ||
84 | ldst->is_ld = is_ld; | ||
85 | ldst->oi = oi; | ||
86 | - ldst->addrlo_reg = addrlo; | ||
87 | - ldst->addrhi_reg = addrhi; | ||
88 | + ldst->addrlo_reg = addr; | ||
89 | |||
90 | /* jne slow_path */ | ||
91 | - jcc = tcg_out_cmp(s, TCG_COND_TSTNE, addrlo, a_mask, true, false); | ||
92 | + jcc = tcg_out_cmp(s, TCG_COND_TSTNE, addr, a_mask, true, false); | ||
93 | tcg_out_opc(s, OPC_JCC_long + jcc, 0, 0, 0); | ||
94 | ldst->label_ptr[0] = s->code_ptr; | ||
95 | s->code_ptr += 4; | ||
96 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
36 | } | 97 | } |
37 | 98 | ||
38 | -#if defined(TARGET_PPC64) && defined(CONFIG_ATOMIC128) | 99 | static void tcg_out_qemu_ld(TCGContext *s, TCGReg datalo, TCGReg datahi, |
39 | +#ifdef TARGET_PPC64 | 100 | - TCGReg addrlo, TCGReg addrhi, |
40 | uint64_t helper_lq_le_parallel(CPUPPCState *env, target_ulong addr, | 101 | - MemOpIdx oi, TCGType data_type) |
41 | uint32_t opidx) | 102 | + TCGReg addr, MemOpIdx oi, TCGType data_type) |
42 | { | 103 | { |
43 | - Int128 ret = helper_atomic_ldo_le_mmu(env, addr, opidx, GETPC()); | 104 | TCGLabelQemuLdst *ldst; |
44 | + Int128 ret; | 105 | HostAddress h; |
45 | + | 106 | |
46 | + /* We will have raised EXCP_ATOMIC from the translator. */ | 107 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, true); |
47 | + assert(HAVE_ATOMIC128); | 108 | + ldst = prepare_host_addr(s, &h, addr, oi, true); |
48 | + ret = helper_atomic_ldo_le_mmu(env, addr, opidx, GETPC()); | 109 | tcg_out_qemu_ld_direct(s, datalo, datahi, h, data_type, get_memop(oi)); |
49 | env->retxh = int128_gethi(ret); | 110 | |
50 | return int128_getlo(ret); | 111 | if (ldst) { |
112 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
51 | } | 113 | } |
52 | @@ -XXX,XX +XXX,XX @@ uint64_t helper_lq_le_parallel(CPUPPCState *env, target_ulong addr, | 114 | |
53 | uint64_t helper_lq_be_parallel(CPUPPCState *env, target_ulong addr, | 115 | static void tcg_out_qemu_st(TCGContext *s, TCGReg datalo, TCGReg datahi, |
54 | uint32_t opidx) | 116 | - TCGReg addrlo, TCGReg addrhi, |
117 | - MemOpIdx oi, TCGType data_type) | ||
118 | + TCGReg addr, MemOpIdx oi, TCGType data_type) | ||
55 | { | 119 | { |
56 | - Int128 ret = helper_atomic_ldo_be_mmu(env, addr, opidx, GETPC()); | 120 | TCGLabelQemuLdst *ldst; |
57 | + Int128 ret; | 121 | HostAddress h; |
58 | + | 122 | |
59 | + /* We will have raised EXCP_ATOMIC from the translator. */ | 123 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, false); |
60 | + assert(HAVE_ATOMIC128); | 124 | + ldst = prepare_host_addr(s, &h, addr, oi, false); |
61 | + ret = helper_atomic_ldo_be_mmu(env, addr, opidx, GETPC()); | 125 | tcg_out_qemu_st_direct(s, datalo, datahi, h, get_memop(oi)); |
62 | env->retxh = int128_gethi(ret); | 126 | |
63 | return int128_getlo(ret); | 127 | if (ldst) { |
64 | } | 128 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, |
65 | @@ -XXX,XX +XXX,XX @@ uint64_t helper_lq_be_parallel(CPUPPCState *env, target_ulong addr, | 129 | break; |
66 | void helper_stq_le_parallel(CPUPPCState *env, target_ulong addr, | 130 | |
67 | uint64_t lo, uint64_t hi, uint32_t opidx) | 131 | case INDEX_op_qemu_ld_i32: |
68 | { | 132 | - tcg_out_qemu_ld(s, a0, -1, a1, -1, a2, TCG_TYPE_I32); |
69 | - Int128 val = int128_make128(lo, hi); | 133 | + tcg_out_qemu_ld(s, a0, -1, a1, a2, TCG_TYPE_I32); |
70 | + Int128 val; | 134 | break; |
71 | + | 135 | case INDEX_op_qemu_ld_i64: |
72 | + /* We will have raised EXCP_ATOMIC from the translator. */ | 136 | if (TCG_TARGET_REG_BITS == 64) { |
73 | + assert(HAVE_ATOMIC128); | 137 | - tcg_out_qemu_ld(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); |
74 | + val = int128_make128(lo, hi); | 138 | + tcg_out_qemu_ld(s, a0, -1, a1, a2, TCG_TYPE_I64); |
75 | helper_atomic_sto_le_mmu(env, addr, val, opidx, GETPC()); | ||
76 | } | ||
77 | |||
78 | void helper_stq_be_parallel(CPUPPCState *env, target_ulong addr, | ||
79 | uint64_t lo, uint64_t hi, uint32_t opidx) | ||
80 | { | ||
81 | - Int128 val = int128_make128(lo, hi); | ||
82 | + Int128 val; | ||
83 | + | ||
84 | + /* We will have raised EXCP_ATOMIC from the translator. */ | ||
85 | + assert(HAVE_ATOMIC128); | ||
86 | + val = int128_make128(lo, hi); | ||
87 | helper_atomic_sto_be_mmu(env, addr, val, opidx, GETPC()); | ||
88 | } | ||
89 | |||
90 | @@ -XXX,XX +XXX,XX @@ uint32_t helper_stqcx_le_parallel(CPUPPCState *env, target_ulong addr, | ||
91 | { | ||
92 | bool success = false; | ||
93 | |||
94 | + /* We will have raised EXCP_ATOMIC from the translator. */ | ||
95 | + assert(HAVE_CMPXCHG128); | ||
96 | + | ||
97 | if (likely(addr == env->reserve_addr)) { | ||
98 | Int128 oldv, cmpv, newv; | ||
99 | |||
100 | @@ -XXX,XX +XXX,XX @@ uint32_t helper_stqcx_be_parallel(CPUPPCState *env, target_ulong addr, | ||
101 | { | ||
102 | bool success = false; | ||
103 | |||
104 | + /* We will have raised EXCP_ATOMIC from the translator. */ | ||
105 | + assert(HAVE_CMPXCHG128); | ||
106 | + | ||
107 | if (likely(addr == env->reserve_addr)) { | ||
108 | Int128 oldv, cmpv, newv; | ||
109 | |||
110 | diff --git a/target/ppc/translate.c b/target/ppc/translate.c | ||
111 | index XXXXXXX..XXXXXXX 100644 | ||
112 | --- a/target/ppc/translate.c | ||
113 | +++ b/target/ppc/translate.c | ||
114 | @@ -XXX,XX +XXX,XX @@ | ||
115 | #include "trace-tcg.h" | ||
116 | #include "exec/translator.h" | ||
117 | #include "exec/log.h" | ||
118 | +#include "qemu/atomic128.h" | ||
119 | |||
120 | |||
121 | #define CPU_SINGLE_STEP 0x1 | ||
122 | @@ -XXX,XX +XXX,XX @@ static void gen_lq(DisasContext *ctx) | ||
123 | hi = cpu_gpr[rd]; | ||
124 | |||
125 | if (tb_cflags(ctx->base.tb) & CF_PARALLEL) { | ||
126 | -#ifdef CONFIG_ATOMIC128 | ||
127 | - TCGv_i32 oi = tcg_temp_new_i32(); | ||
128 | - if (ctx->le_mode) { | ||
129 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ, ctx->mem_idx)); | ||
130 | - gen_helper_lq_le_parallel(lo, cpu_env, EA, oi); | ||
131 | + if (HAVE_ATOMIC128) { | ||
132 | + TCGv_i32 oi = tcg_temp_new_i32(); | ||
133 | + if (ctx->le_mode) { | ||
134 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ, ctx->mem_idx)); | ||
135 | + gen_helper_lq_le_parallel(lo, cpu_env, EA, oi); | ||
136 | + } else { | ||
137 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ, ctx->mem_idx)); | ||
138 | + gen_helper_lq_be_parallel(lo, cpu_env, EA, oi); | ||
139 | + } | ||
140 | + tcg_temp_free_i32(oi); | ||
141 | + tcg_gen_ld_i64(hi, cpu_env, offsetof(CPUPPCState, retxh)); | ||
142 | } else { | 139 | } else { |
143 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ, ctx->mem_idx)); | 140 | - tcg_out_qemu_ld(s, a0, a1, a2, -1, args[3], TCG_TYPE_I64); |
144 | - gen_helper_lq_be_parallel(lo, cpu_env, EA, oi); | 141 | + tcg_out_qemu_ld(s, a0, a1, a2, args[3], TCG_TYPE_I64); |
145 | + /* Restart with exclusive lock. */ | ||
146 | + gen_helper_exit_atomic(cpu_env); | ||
147 | + ctx->base.is_jmp = DISAS_NORETURN; | ||
148 | } | 142 | } |
149 | - tcg_temp_free_i32(oi); | 143 | break; |
150 | - tcg_gen_ld_i64(hi, cpu_env, offsetof(CPUPPCState, retxh)); | 144 | case INDEX_op_qemu_ld_i128: |
151 | -#else | 145 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); |
152 | - /* Restart with exclusive lock. */ | 146 | - tcg_out_qemu_ld(s, a0, a1, a2, -1, args[3], TCG_TYPE_I128); |
153 | - gen_helper_exit_atomic(cpu_env); | 147 | + tcg_out_qemu_ld(s, a0, a1, a2, args[3], TCG_TYPE_I128); |
154 | - ctx->base.is_jmp = DISAS_NORETURN; | 148 | break; |
155 | -#endif | 149 | |
156 | } else if (ctx->le_mode) { | 150 | case INDEX_op_qemu_st_i32: |
157 | tcg_gen_qemu_ld_i64(lo, EA, ctx->mem_idx, MO_LEQ); | 151 | case INDEX_op_qemu_st8_i32: |
158 | gen_addr_add(ctx, EA, EA, 8); | 152 | - tcg_out_qemu_st(s, a0, -1, a1, -1, a2, TCG_TYPE_I32); |
159 | @@ -XXX,XX +XXX,XX @@ static void gen_std(DisasContext *ctx) | 153 | + tcg_out_qemu_st(s, a0, -1, a1, a2, TCG_TYPE_I32); |
160 | hi = cpu_gpr[rs]; | 154 | break; |
161 | 155 | case INDEX_op_qemu_st_i64: | |
162 | if (tb_cflags(ctx->base.tb) & CF_PARALLEL) { | 156 | if (TCG_TARGET_REG_BITS == 64) { |
163 | -#ifdef CONFIG_ATOMIC128 | 157 | - tcg_out_qemu_st(s, a0, -1, a1, -1, a2, TCG_TYPE_I64); |
164 | - TCGv_i32 oi = tcg_temp_new_i32(); | 158 | + tcg_out_qemu_st(s, a0, -1, a1, a2, TCG_TYPE_I64); |
165 | - if (ctx->le_mode) { | ||
166 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ, ctx->mem_idx)); | ||
167 | - gen_helper_stq_le_parallel(cpu_env, EA, lo, hi, oi); | ||
168 | + if (HAVE_ATOMIC128) { | ||
169 | + TCGv_i32 oi = tcg_temp_new_i32(); | ||
170 | + if (ctx->le_mode) { | ||
171 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ, ctx->mem_idx)); | ||
172 | + gen_helper_stq_le_parallel(cpu_env, EA, lo, hi, oi); | ||
173 | + } else { | ||
174 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ, ctx->mem_idx)); | ||
175 | + gen_helper_stq_be_parallel(cpu_env, EA, lo, hi, oi); | ||
176 | + } | ||
177 | + tcg_temp_free_i32(oi); | ||
178 | } else { | ||
179 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ, ctx->mem_idx)); | ||
180 | - gen_helper_stq_be_parallel(cpu_env, EA, lo, hi, oi); | ||
181 | + /* Restart with exclusive lock. */ | ||
182 | + gen_helper_exit_atomic(cpu_env); | ||
183 | + ctx->base.is_jmp = DISAS_NORETURN; | ||
184 | } | ||
185 | - tcg_temp_free_i32(oi); | ||
186 | -#else | ||
187 | - /* Restart with exclusive lock. */ | ||
188 | - gen_helper_exit_atomic(cpu_env); | ||
189 | - ctx->base.is_jmp = DISAS_NORETURN; | ||
190 | -#endif | ||
191 | } else if (ctx->le_mode) { | ||
192 | tcg_gen_qemu_st_i64(lo, EA, ctx->mem_idx, MO_LEQ); | ||
193 | gen_addr_add(ctx, EA, EA, 8); | ||
194 | @@ -XXX,XX +XXX,XX @@ static void gen_lqarx(DisasContext *ctx) | ||
195 | hi = cpu_gpr[rd]; | ||
196 | |||
197 | if (tb_cflags(ctx->base.tb) & CF_PARALLEL) { | ||
198 | -#ifdef CONFIG_ATOMIC128 | ||
199 | - TCGv_i32 oi = tcg_temp_new_i32(); | ||
200 | - if (ctx->le_mode) { | ||
201 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ | MO_ALIGN_16, | ||
202 | - ctx->mem_idx)); | ||
203 | - gen_helper_lq_le_parallel(lo, cpu_env, EA, oi); | ||
204 | + if (HAVE_ATOMIC128) { | ||
205 | + TCGv_i32 oi = tcg_temp_new_i32(); | ||
206 | + if (ctx->le_mode) { | ||
207 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_LEQ | MO_ALIGN_16, | ||
208 | + ctx->mem_idx)); | ||
209 | + gen_helper_lq_le_parallel(lo, cpu_env, EA, oi); | ||
210 | + } else { | ||
211 | + tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ | MO_ALIGN_16, | ||
212 | + ctx->mem_idx)); | ||
213 | + gen_helper_lq_be_parallel(lo, cpu_env, EA, oi); | ||
214 | + } | ||
215 | + tcg_temp_free_i32(oi); | ||
216 | + tcg_gen_ld_i64(hi, cpu_env, offsetof(CPUPPCState, retxh)); | ||
217 | } else { | 159 | } else { |
218 | - tcg_gen_movi_i32(oi, make_memop_idx(MO_BEQ | MO_ALIGN_16, | 160 | - tcg_out_qemu_st(s, a0, a1, a2, -1, args[3], TCG_TYPE_I64); |
219 | - ctx->mem_idx)); | 161 | + tcg_out_qemu_st(s, a0, a1, a2, args[3], TCG_TYPE_I64); |
220 | - gen_helper_lq_be_parallel(lo, cpu_env, EA, oi); | ||
221 | + /* Restart with exclusive lock. */ | ||
222 | + gen_helper_exit_atomic(cpu_env); | ||
223 | + ctx->base.is_jmp = DISAS_NORETURN; | ||
224 | + tcg_temp_free(EA); | ||
225 | + return; | ||
226 | } | 162 | } |
227 | - tcg_temp_free_i32(oi); | 163 | break; |
228 | - tcg_gen_ld_i64(hi, cpu_env, offsetof(CPUPPCState, retxh)); | 164 | case INDEX_op_qemu_st_i128: |
229 | -#else | 165 | tcg_debug_assert(TCG_TARGET_REG_BITS == 64); |
230 | - /* Restart with exclusive lock. */ | 166 | - tcg_out_qemu_st(s, a0, a1, a2, -1, args[3], TCG_TYPE_I128); |
231 | - gen_helper_exit_atomic(cpu_env); | 167 | + tcg_out_qemu_st(s, a0, a1, a2, args[3], TCG_TYPE_I128); |
232 | - ctx->base.is_jmp = DISAS_NORETURN; | 168 | break; |
233 | - tcg_temp_free(EA); | 169 | |
234 | - return; | 170 | OP_32_64(mulu2): |
235 | -#endif | ||
236 | } else if (ctx->le_mode) { | ||
237 | tcg_gen_qemu_ld_i64(lo, EA, ctx->mem_idx, MO_LEQ | MO_ALIGN_16); | ||
238 | tcg_gen_mov_tl(cpu_reserve, EA); | ||
239 | @@ -XXX,XX +XXX,XX @@ static void gen_stqcx_(DisasContext *ctx) | ||
240 | hi = cpu_gpr[rs]; | ||
241 | |||
242 | if (tb_cflags(ctx->base.tb) & CF_PARALLEL) { | ||
243 | - TCGv_i32 oi = tcg_const_i32(DEF_MEMOP(MO_Q) | MO_ALIGN_16); | ||
244 | -#ifdef CONFIG_ATOMIC128 | ||
245 | - if (ctx->le_mode) { | ||
246 | - gen_helper_stqcx_le_parallel(cpu_crf[0], cpu_env, EA, lo, hi, oi); | ||
247 | + if (HAVE_CMPXCHG128) { | ||
248 | + TCGv_i32 oi = tcg_const_i32(DEF_MEMOP(MO_Q) | MO_ALIGN_16); | ||
249 | + if (ctx->le_mode) { | ||
250 | + gen_helper_stqcx_le_parallel(cpu_crf[0], cpu_env, | ||
251 | + EA, lo, hi, oi); | ||
252 | + } else { | ||
253 | + gen_helper_stqcx_be_parallel(cpu_crf[0], cpu_env, | ||
254 | + EA, lo, hi, oi); | ||
255 | + } | ||
256 | + tcg_temp_free_i32(oi); | ||
257 | } else { | ||
258 | - gen_helper_stqcx_le_parallel(cpu_crf[0], cpu_env, EA, lo, hi, oi); | ||
259 | + /* Restart with exclusive lock. */ | ||
260 | + gen_helper_exit_atomic(cpu_env); | ||
261 | + ctx->base.is_jmp = DISAS_NORETURN; | ||
262 | } | ||
263 | -#else | ||
264 | - /* Restart with exclusive lock. */ | ||
265 | - gen_helper_exit_atomic(cpu_env); | ||
266 | - ctx->base.is_jmp = DISAS_NORETURN; | ||
267 | -#endif | ||
268 | tcg_temp_free(EA); | ||
269 | - tcg_temp_free_i32(oi); | ||
270 | } else { | ||
271 | TCGLabel *lab_fail = gen_new_label(); | ||
272 | TCGLabel *lab_over = gen_new_label(); | ||
273 | -- | 171 | -- |
274 | 2.17.2 | 172 | 2.43.0 |
275 | 173 | ||
276 | 174 | diff view generated by jsdifflib |
1 | Reviewed-by: David Hildenbrand <david@redhat.com> | 1 | The guest address will now always fit in one register. |
---|---|---|---|
2 | |||
3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
3 | --- | 5 | --- |
4 | target/s390x/mem_helper.c | 128 ++++++++++++++++++-------------------- | 6 | tcg/mips/tcg-target.c.inc | 62 ++++++++++++++------------------------- |
5 | 1 file changed, 61 insertions(+), 67 deletions(-) | 7 | 1 file changed, 22 insertions(+), 40 deletions(-) |
6 | 8 | ||
7 | diff --git a/target/s390x/mem_helper.c b/target/s390x/mem_helper.c | 9 | diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc |
8 | index XXXXXXX..XXXXXXX 100644 | 10 | index XXXXXXX..XXXXXXX 100644 |
9 | --- a/target/s390x/mem_helper.c | 11 | --- a/tcg/mips/tcg-target.c.inc |
10 | +++ b/target/s390x/mem_helper.c | 12 | +++ b/tcg/mips/tcg-target.c.inc |
11 | @@ -XXX,XX +XXX,XX @@ uint32_t HELPER(trXX)(CPUS390XState *env, uint32_t r1, uint32_t r2, | 13 | @@ -XXX,XX +XXX,XX @@ bool tcg_target_has_memory_bswap(MemOp memop) |
12 | return cc; | 14 | * is required and fill in @h with the host address for the fast path. |
13 | } | 15 | */ |
14 | 16 | static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | |
15 | -static void do_cdsg(CPUS390XState *env, uint64_t addr, | 17 | - TCGReg addrlo, TCGReg addrhi, |
16 | - uint32_t r1, uint32_t r3, bool parallel) | 18 | - MemOpIdx oi, bool is_ld) |
17 | +void HELPER(cdsg)(CPUS390XState *env, uint64_t addr, | 19 | + TCGReg addr, MemOpIdx oi, bool is_ld) |
18 | + uint32_t r1, uint32_t r3) | ||
19 | { | 20 | { |
20 | uintptr_t ra = GETPC(); | 21 | TCGType addr_type = s->addr_type; |
21 | Int128 cmpv = int128_make128(env->regs[r1 + 1], env->regs[r1]); | 22 | TCGLabelQemuLdst *ldst = NULL; |
22 | Int128 newv = int128_make128(env->regs[r3 + 1], env->regs[r3]); | 23 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
23 | Int128 oldv; | 24 | ldst = new_ldst_label(s); |
24 | + uint64_t oldh, oldl; | 25 | ldst->is_ld = is_ld; |
25 | bool fail; | 26 | ldst->oi = oi; |
26 | 27 | - ldst->addrlo_reg = addrlo; | |
27 | - if (parallel) { | 28 | - ldst->addrhi_reg = addrhi; |
28 | -#if !HAVE_CMPXCHG128 | 29 | + ldst->addrlo_reg = addr; |
29 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 30 | |
30 | -#else | 31 | /* Load tlb_mask[mmu_idx] and tlb_table[mmu_idx]. */ |
31 | - int mem_idx = cpu_mmu_index(env, false); | 32 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP0, TCG_AREG0, mask_off); |
32 | - TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 33 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
33 | - oldv = helper_atomic_cmpxchgo_be_mmu(env, addr, cmpv, newv, oi, ra); | 34 | |
34 | - fail = !int128_eq(oldv, cmpv); | 35 | /* Extract the TLB index from the address into TMP3. */ |
35 | -#endif | 36 | if (TCG_TARGET_REG_BITS == 32 || addr_type == TCG_TYPE_I32) { |
36 | - } else { | 37 | - tcg_out_opc_sa(s, OPC_SRL, TCG_TMP3, addrlo, |
37 | - uint64_t oldh, oldl; | 38 | + tcg_out_opc_sa(s, OPC_SRL, TCG_TMP3, addr, |
38 | + check_alignment(env, addr, 16, ra); | 39 | s->page_bits - CPU_TLB_ENTRY_BITS); |
39 | 40 | } else { | |
40 | - check_alignment(env, addr, 16, ra); | 41 | - tcg_out_dsrl(s, TCG_TMP3, addrlo, |
41 | + oldh = cpu_ldq_data_ra(env, addr + 0, ra); | 42 | - s->page_bits - CPU_TLB_ENTRY_BITS); |
42 | + oldl = cpu_ldq_data_ra(env, addr + 8, ra); | 43 | + tcg_out_dsrl(s, TCG_TMP3, addr, s->page_bits - CPU_TLB_ENTRY_BITS); |
43 | 44 | } | |
44 | - oldh = cpu_ldq_data_ra(env, addr + 0, ra); | 45 | tcg_out_opc_reg(s, OPC_AND, TCG_TMP3, TCG_TMP3, TCG_TMP0); |
45 | - oldl = cpu_ldq_data_ra(env, addr + 8, ra); | 46 | |
47 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
48 | tcg_out_opc_imm(s, (TCG_TARGET_REG_BITS == 32 | ||
49 | || addr_type == TCG_TYPE_I32 | ||
50 | ? OPC_ADDIU : OPC_DADDIU), | ||
51 | - TCG_TMP2, addrlo, s_mask - a_mask); | ||
52 | + TCG_TMP2, addr, s_mask - a_mask); | ||
53 | tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, TCG_TMP2); | ||
54 | } else { | ||
55 | - tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, addrlo); | ||
56 | + tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, addr); | ||
57 | } | ||
58 | |||
59 | /* Zero extend a 32-bit guest address for a 64-bit host. */ | ||
60 | if (TCG_TARGET_REG_BITS == 64 && addr_type == TCG_TYPE_I32) { | ||
61 | - tcg_out_ext32u(s, TCG_TMP2, addrlo); | ||
62 | - addrlo = TCG_TMP2; | ||
63 | + tcg_out_ext32u(s, TCG_TMP2, addr); | ||
64 | + addr = TCG_TMP2; | ||
65 | } | ||
66 | |||
67 | ldst->label_ptr[0] = s->code_ptr; | ||
68 | tcg_out_opc_br(s, OPC_BNE, TCG_TMP1, TCG_TMP0); | ||
69 | |||
70 | - /* Load and test the high half tlb comparator. */ | ||
71 | - if (TCG_TARGET_REG_BITS == 32 && addr_type != TCG_TYPE_I32) { | ||
72 | - /* delay slot */ | ||
73 | - tcg_out_ldst(s, OPC_LW, TCG_TMP0, TCG_TMP3, cmp_off + HI_OFF); | ||
46 | - | 74 | - |
47 | - oldv = int128_make128(oldl, oldh); | 75 | - /* Load the tlb addend for the fast path. */ |
48 | - fail = !int128_eq(oldv, cmpv); | 76 | - tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP3, TCG_TMP3, add_off); |
49 | - if (fail) { | 77 | - |
50 | - newv = oldv; | 78 | - ldst->label_ptr[1] = s->code_ptr; |
79 | - tcg_out_opc_br(s, OPC_BNE, addrhi, TCG_TMP0); | ||
51 | - } | 80 | - } |
52 | - | 81 | - |
53 | - cpu_stq_data_ra(env, addr + 0, int128_gethi(newv), ra); | 82 | /* delay slot */ |
54 | - cpu_stq_data_ra(env, addr + 8, int128_getlo(newv), ra); | 83 | base = TCG_TMP3; |
55 | + oldv = int128_make128(oldl, oldh); | 84 | - tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_TMP3, addrlo); |
56 | + fail = !int128_eq(oldv, cmpv); | 85 | + tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_TMP3, addr); |
57 | + if (fail) { | 86 | } else { |
58 | + newv = oldv; | 87 | if (a_mask && (use_mips32r6_instructions || a_bits != s_bits)) { |
59 | } | 88 | ldst = new_ldst_label(s); |
60 | 89 | ||
61 | + cpu_stq_data_ra(env, addr + 0, int128_gethi(newv), ra); | 90 | ldst->is_ld = is_ld; |
62 | + cpu_stq_data_ra(env, addr + 8, int128_getlo(newv), ra); | 91 | ldst->oi = oi; |
63 | + | 92 | - ldst->addrlo_reg = addrlo; |
64 | env->cc_op = fail; | 93 | - ldst->addrhi_reg = addrhi; |
65 | env->regs[r1] = int128_gethi(oldv); | 94 | + ldst->addrlo_reg = addr; |
66 | env->regs[r1 + 1] = int128_getlo(oldv); | 95 | |
96 | /* We are expecting a_bits to max out at 7, much lower than ANDI. */ | ||
97 | tcg_debug_assert(a_bits < 16); | ||
98 | - tcg_out_opc_imm(s, OPC_ANDI, TCG_TMP0, addrlo, a_mask); | ||
99 | + tcg_out_opc_imm(s, OPC_ANDI, TCG_TMP0, addr, a_mask); | ||
100 | |||
101 | ldst->label_ptr[0] = s->code_ptr; | ||
102 | if (use_mips32r6_instructions) { | ||
103 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
104 | } | ||
105 | } | ||
106 | |||
107 | - base = addrlo; | ||
108 | + base = addr; | ||
109 | if (TCG_TARGET_REG_BITS == 64 && addr_type == TCG_TYPE_I32) { | ||
110 | tcg_out_ext32u(s, TCG_REG_A0, base); | ||
111 | base = TCG_REG_A0; | ||
112 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_unalign(TCGContext *s, TCGReg lo, TCGReg hi, | ||
67 | } | 113 | } |
68 | 114 | ||
69 | -void HELPER(cdsg)(CPUS390XState *env, uint64_t addr, | 115 | static void tcg_out_qemu_ld(TCGContext *s, TCGReg datalo, TCGReg datahi, |
70 | - uint32_t r1, uint32_t r3) | 116 | - TCGReg addrlo, TCGReg addrhi, |
71 | -{ | 117 | - MemOpIdx oi, TCGType data_type) |
72 | - do_cdsg(env, addr, r1, r3, false); | 118 | + TCGReg addr, MemOpIdx oi, TCGType data_type) |
73 | -} | ||
74 | - | ||
75 | void HELPER(cdsg_parallel)(CPUS390XState *env, uint64_t addr, | ||
76 | uint32_t r1, uint32_t r3) | ||
77 | { | 119 | { |
78 | - do_cdsg(env, addr, r1, r3, true); | 120 | MemOp opc = get_memop(oi); |
79 | + uintptr_t ra = GETPC(); | 121 | TCGLabelQemuLdst *ldst; |
80 | + Int128 cmpv = int128_make128(env->regs[r1 + 1], env->regs[r1]); | 122 | HostAddress h; |
81 | + Int128 newv = int128_make128(env->regs[r3 + 1], env->regs[r3]); | 123 | |
82 | + int mem_idx; | 124 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, true); |
83 | + TCGMemOpIdx oi; | 125 | + ldst = prepare_host_addr(s, &h, addr, oi, true); |
84 | + Int128 oldv; | 126 | |
85 | + bool fail; | 127 | if (use_mips32r6_instructions || h.aa.align >= (opc & MO_SIZE)) { |
86 | + | 128 | tcg_out_qemu_ld_direct(s, datalo, datahi, h.base, opc, data_type); |
87 | + if (!HAVE_CMPXCHG128) { | 129 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_unalign(TCGContext *s, TCGReg lo, TCGReg hi, |
88 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
89 | + } | ||
90 | + | ||
91 | + mem_idx = cpu_mmu_index(env, false); | ||
92 | + oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
93 | + oldv = helper_atomic_cmpxchgo_be_mmu(env, addr, cmpv, newv, oi, ra); | ||
94 | + fail = !int128_eq(oldv, cmpv); | ||
95 | + | ||
96 | + env->cc_op = fail; | ||
97 | + env->regs[r1] = int128_gethi(oldv); | ||
98 | + env->regs[r1 + 1] = int128_getlo(oldv); | ||
99 | } | 130 | } |
100 | 131 | ||
101 | static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 132 | static void tcg_out_qemu_st(TCGContext *s, TCGReg datalo, TCGReg datahi, |
102 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(lra)(CPUS390XState *env, uint64_t addr) | 133 | - TCGReg addrlo, TCGReg addrhi, |
103 | #endif | 134 | - MemOpIdx oi, TCGType data_type) |
104 | 135 | + TCGReg addr, MemOpIdx oi, TCGType data_type) | |
105 | /* load pair from quadword */ | ||
106 | -static uint64_t do_lpq(CPUS390XState *env, uint64_t addr, bool parallel) | ||
107 | +uint64_t HELPER(lpq)(CPUS390XState *env, uint64_t addr) | ||
108 | { | 136 | { |
109 | uintptr_t ra = GETPC(); | 137 | MemOp opc = get_memop(oi); |
110 | uint64_t hi, lo; | 138 | TCGLabelQemuLdst *ldst; |
111 | 139 | HostAddress h; | |
112 | - if (!parallel) { | 140 | |
113 | - check_alignment(env, addr, 16, ra); | 141 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, false); |
114 | - hi = cpu_ldq_data_ra(env, addr + 0, ra); | 142 | + ldst = prepare_host_addr(s, &h, addr, oi, false); |
115 | - lo = cpu_ldq_data_ra(env, addr + 8, ra); | 143 | |
116 | - } else if (HAVE_ATOMIC128) { | 144 | if (use_mips32r6_instructions || h.aa.align >= (opc & MO_SIZE)) { |
117 | + check_alignment(env, addr, 16, ra); | 145 | tcg_out_qemu_st_direct(s, datalo, datahi, h.base, opc); |
118 | + hi = cpu_ldq_data_ra(env, addr + 0, ra); | 146 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, |
119 | + lo = cpu_ldq_data_ra(env, addr + 8, ra); | 147 | break; |
120 | + | 148 | |
121 | + env->retxl = lo; | 149 | case INDEX_op_qemu_ld_i32: |
122 | + return hi; | 150 | - tcg_out_qemu_ld(s, a0, 0, a1, 0, a2, TCG_TYPE_I32); |
123 | +} | 151 | + tcg_out_qemu_ld(s, a0, 0, a1, a2, TCG_TYPE_I32); |
124 | + | 152 | break; |
125 | +uint64_t HELPER(lpq_parallel)(CPUS390XState *env, uint64_t addr) | 153 | case INDEX_op_qemu_ld_i64: |
126 | +{ | 154 | if (TCG_TARGET_REG_BITS == 64) { |
127 | + uintptr_t ra = GETPC(); | 155 | - tcg_out_qemu_ld(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); |
128 | + uint64_t hi, lo; | 156 | + tcg_out_qemu_ld(s, a0, 0, a1, a2, TCG_TYPE_I64); |
129 | + | 157 | } else { |
130 | + if (HAVE_ATOMIC128) { | 158 | - tcg_out_qemu_ld(s, a0, a1, a2, 0, args[3], TCG_TYPE_I64); |
131 | int mem_idx = cpu_mmu_index(env, false); | 159 | + tcg_out_qemu_ld(s, a0, a1, a2, args[3], TCG_TYPE_I64); |
132 | TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 160 | } |
133 | Int128 v = helper_atomic_ldo_be_mmu(env, addr, oi, ra); | 161 | break; |
134 | @@ -XXX,XX +XXX,XX @@ static uint64_t do_lpq(CPUS390XState *env, uint64_t addr, bool parallel) | 162 | |
135 | return hi; | 163 | case INDEX_op_qemu_st_i32: |
136 | } | 164 | - tcg_out_qemu_st(s, a0, 0, a1, 0, a2, TCG_TYPE_I32); |
137 | 165 | + tcg_out_qemu_st(s, a0, 0, a1, a2, TCG_TYPE_I32); | |
138 | -uint64_t HELPER(lpq)(CPUS390XState *env, uint64_t addr) | 166 | break; |
139 | -{ | 167 | case INDEX_op_qemu_st_i64: |
140 | - return do_lpq(env, addr, false); | 168 | if (TCG_TARGET_REG_BITS == 64) { |
141 | -} | 169 | - tcg_out_qemu_st(s, a0, 0, a1, 0, a2, TCG_TYPE_I64); |
142 | - | 170 | + tcg_out_qemu_st(s, a0, 0, a1, a2, TCG_TYPE_I64); |
143 | -uint64_t HELPER(lpq_parallel)(CPUS390XState *env, uint64_t addr) | 171 | } else { |
144 | -{ | 172 | - tcg_out_qemu_st(s, a0, a1, a2, 0, args[3], TCG_TYPE_I64); |
145 | - return do_lpq(env, addr, true); | 173 | + tcg_out_qemu_st(s, a0, a1, a2, args[3], TCG_TYPE_I64); |
146 | -} | 174 | } |
147 | - | 175 | break; |
148 | /* store pair to quadword */ | 176 | |
149 | -static void do_stpq(CPUS390XState *env, uint64_t addr, | ||
150 | - uint64_t low, uint64_t high, bool parallel) | ||
151 | +void HELPER(stpq)(CPUS390XState *env, uint64_t addr, | ||
152 | + uint64_t low, uint64_t high) | ||
153 | { | ||
154 | uintptr_t ra = GETPC(); | ||
155 | |||
156 | - if (!parallel) { | ||
157 | - check_alignment(env, addr, 16, ra); | ||
158 | - cpu_stq_data_ra(env, addr + 0, high, ra); | ||
159 | - cpu_stq_data_ra(env, addr + 8, low, ra); | ||
160 | - } else if (HAVE_ATOMIC128) { | ||
161 | + check_alignment(env, addr, 16, ra); | ||
162 | + cpu_stq_data_ra(env, addr + 0, high, ra); | ||
163 | + cpu_stq_data_ra(env, addr + 8, low, ra); | ||
164 | +} | ||
165 | + | ||
166 | +void HELPER(stpq_parallel)(CPUS390XState *env, uint64_t addr, | ||
167 | + uint64_t low, uint64_t high) | ||
168 | +{ | ||
169 | + uintptr_t ra = GETPC(); | ||
170 | + | ||
171 | + if (HAVE_ATOMIC128) { | ||
172 | int mem_idx = cpu_mmu_index(env, false); | ||
173 | TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
174 | Int128 v = int128_make128(low, high); | ||
175 | @@ -XXX,XX +XXX,XX @@ static void do_stpq(CPUS390XState *env, uint64_t addr, | ||
176 | } | ||
177 | } | ||
178 | |||
179 | -void HELPER(stpq)(CPUS390XState *env, uint64_t addr, | ||
180 | - uint64_t low, uint64_t high) | ||
181 | -{ | ||
182 | - do_stpq(env, addr, low, high, false); | ||
183 | -} | ||
184 | - | ||
185 | -void HELPER(stpq_parallel)(CPUS390XState *env, uint64_t addr, | ||
186 | - uint64_t low, uint64_t high) | ||
187 | -{ | ||
188 | - do_stpq(env, addr, low, high, true); | ||
189 | -} | ||
190 | - | ||
191 | /* Execute instruction. This instruction executes an insn modified with | ||
192 | the contents of r1. It does not change the executed instruction in memory; | ||
193 | it does not change the program counter. | ||
194 | -- | 177 | -- |
195 | 2.17.2 | 178 | 2.43.0 |
196 | 179 | ||
197 | 180 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | The guest address will now always fit in one register. |
---|---|---|---|
2 | 2 | ||
3 | As far as I can tell tlb_flush does not need to be called | 3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | this early. tlb_flush is eventually called after the CPU | ||
5 | has been realized. | ||
6 | |||
7 | This change paves the way to the introduction of tlb_init, | ||
8 | which will be called from cpu_exec_realizefn. | ||
9 | |||
10 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
11 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
12 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
13 | Message-Id: <20181009174557.16125-2-cota@braap.org> | ||
14 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
15 | --- | 5 | --- |
16 | target/alpha/cpu.c | 1 - | 6 | tcg/ppc/tcg-target.c.inc | 75 ++++++++++++---------------------------- |
17 | 1 file changed, 1 deletion(-) | 7 | 1 file changed, 23 insertions(+), 52 deletions(-) |
18 | 8 | ||
19 | diff --git a/target/alpha/cpu.c b/target/alpha/cpu.c | 9 | diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc |
20 | index XXXXXXX..XXXXXXX 100644 | 10 | index XXXXXXX..XXXXXXX 100644 |
21 | --- a/target/alpha/cpu.c | 11 | --- a/tcg/ppc/tcg-target.c.inc |
22 | +++ b/target/alpha/cpu.c | 12 | +++ b/tcg/ppc/tcg-target.c.inc |
23 | @@ -XXX,XX +XXX,XX @@ static void alpha_cpu_initfn(Object *obj) | 13 | @@ -XXX,XX +XXX,XX @@ bool tcg_target_has_memory_bswap(MemOp memop) |
24 | CPUAlphaState *env = &cpu->env; | 14 | * is required and fill in @h with the host address for the fast path. |
25 | 15 | */ | |
26 | cs->env_ptr = env; | 16 | static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
27 | - tlb_flush(cs); | 17 | - TCGReg addrlo, TCGReg addrhi, |
28 | 18 | - MemOpIdx oi, bool is_ld) | |
29 | env->lock_addr = -1; | 19 | + TCGReg addr, MemOpIdx oi, bool is_ld) |
30 | #if defined(CONFIG_USER_ONLY) | 20 | { |
21 | TCGType addr_type = s->addr_type; | ||
22 | TCGLabelQemuLdst *ldst = NULL; | ||
23 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
24 | ldst = new_ldst_label(s); | ||
25 | ldst->is_ld = is_ld; | ||
26 | ldst->oi = oi; | ||
27 | - ldst->addrlo_reg = addrlo; | ||
28 | - ldst->addrhi_reg = addrhi; | ||
29 | + ldst->addrlo_reg = addr; | ||
30 | |||
31 | /* Load tlb_mask[mmu_idx] and tlb_table[mmu_idx]. */ | ||
32 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP1, TCG_AREG0, mask_off); | ||
33 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
34 | |||
35 | /* Extract the page index, shifted into place for tlb index. */ | ||
36 | if (TCG_TARGET_REG_BITS == 32) { | ||
37 | - tcg_out_shri32(s, TCG_REG_R0, addrlo, | ||
38 | + tcg_out_shri32(s, TCG_REG_R0, addr, | ||
39 | s->page_bits - CPU_TLB_ENTRY_BITS); | ||
40 | } else { | ||
41 | - tcg_out_shri64(s, TCG_REG_R0, addrlo, | ||
42 | + tcg_out_shri64(s, TCG_REG_R0, addr, | ||
43 | s->page_bits - CPU_TLB_ENTRY_BITS); | ||
44 | } | ||
45 | tcg_out32(s, AND | SAB(TCG_REG_TMP1, TCG_REG_TMP1, TCG_REG_R0)); | ||
46 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
47 | if (a_bits < s_bits) { | ||
48 | a_bits = s_bits; | ||
49 | } | ||
50 | - tcg_out_rlw(s, RLWINM, TCG_REG_R0, addrlo, 0, | ||
51 | + tcg_out_rlw(s, RLWINM, TCG_REG_R0, addr, 0, | ||
52 | (32 - a_bits) & 31, 31 - s->page_bits); | ||
53 | } else { | ||
54 | - TCGReg t = addrlo; | ||
55 | + TCGReg t = addr; | ||
56 | |||
57 | /* | ||
58 | * If the access is unaligned, we need to make sure we fail if we | ||
59 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
60 | } | ||
61 | } | ||
62 | |||
63 | - if (TCG_TARGET_REG_BITS == 32 && addr_type != TCG_TYPE_I32) { | ||
64 | - /* Low part comparison into cr7. */ | ||
65 | - tcg_out_cmp(s, TCG_COND_EQ, TCG_REG_R0, TCG_REG_TMP2, | ||
66 | - 0, 7, TCG_TYPE_I32); | ||
67 | - | ||
68 | - /* Load the high part TLB comparator into TMP2. */ | ||
69 | - tcg_out_ld(s, TCG_TYPE_I32, TCG_REG_TMP2, TCG_REG_TMP1, | ||
70 | - cmp_off + 4 * !HOST_BIG_ENDIAN); | ||
71 | - | ||
72 | - /* Load addend, deferred for this case. */ | ||
73 | - tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP1, TCG_REG_TMP1, | ||
74 | - offsetof(CPUTLBEntry, addend)); | ||
75 | - | ||
76 | - /* High part comparison into cr6. */ | ||
77 | - tcg_out_cmp(s, TCG_COND_EQ, addrhi, TCG_REG_TMP2, | ||
78 | - 0, 6, TCG_TYPE_I32); | ||
79 | - | ||
80 | - /* Combine comparisons into cr0. */ | ||
81 | - tcg_out32(s, CRAND | BT(0, CR_EQ) | BA(6, CR_EQ) | BB(7, CR_EQ)); | ||
82 | - } else { | ||
83 | - /* Full comparison into cr0. */ | ||
84 | - tcg_out_cmp(s, TCG_COND_EQ, TCG_REG_R0, TCG_REG_TMP2, | ||
85 | - 0, 0, addr_type); | ||
86 | - } | ||
87 | + /* Full comparison into cr0. */ | ||
88 | + tcg_out_cmp(s, TCG_COND_EQ, TCG_REG_R0, TCG_REG_TMP2, 0, 0, addr_type); | ||
89 | |||
90 | /* Load a pointer into the current opcode w/conditional branch-link. */ | ||
91 | ldst->label_ptr[0] = s->code_ptr; | ||
92 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
93 | ldst = new_ldst_label(s); | ||
94 | ldst->is_ld = is_ld; | ||
95 | ldst->oi = oi; | ||
96 | - ldst->addrlo_reg = addrlo; | ||
97 | - ldst->addrhi_reg = addrhi; | ||
98 | + ldst->addrlo_reg = addr; | ||
99 | |||
100 | /* We are expecting a_bits to max out at 7, much lower than ANDI. */ | ||
101 | tcg_debug_assert(a_bits < 16); | ||
102 | - tcg_out32(s, ANDI | SAI(addrlo, TCG_REG_R0, (1 << a_bits) - 1)); | ||
103 | + tcg_out32(s, ANDI | SAI(addr, TCG_REG_R0, (1 << a_bits) - 1)); | ||
104 | |||
105 | ldst->label_ptr[0] = s->code_ptr; | ||
106 | tcg_out32(s, BC | BI(0, CR_EQ) | BO_COND_FALSE | LK); | ||
107 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
108 | |||
109 | if (TCG_TARGET_REG_BITS == 64 && addr_type == TCG_TYPE_I32) { | ||
110 | /* Zero-extend the guest address for use in the host address. */ | ||
111 | - tcg_out_ext32u(s, TCG_REG_TMP2, addrlo); | ||
112 | + tcg_out_ext32u(s, TCG_REG_TMP2, addr); | ||
113 | h->index = TCG_REG_TMP2; | ||
114 | } else { | ||
115 | - h->index = addrlo; | ||
116 | + h->index = addr; | ||
117 | } | ||
118 | |||
119 | return ldst; | ||
120 | } | ||
121 | |||
122 | static void tcg_out_qemu_ld(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
123 | - TCGReg addrlo, TCGReg addrhi, | ||
124 | - MemOpIdx oi, TCGType data_type) | ||
125 | + TCGReg addr, MemOpIdx oi, TCGType data_type) | ||
126 | { | ||
127 | MemOp opc = get_memop(oi); | ||
128 | TCGLabelQemuLdst *ldst; | ||
129 | HostAddress h; | ||
130 | |||
131 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, true); | ||
132 | + ldst = prepare_host_addr(s, &h, addr, oi, true); | ||
133 | |||
134 | if (TCG_TARGET_REG_BITS == 32 && (opc & MO_SIZE) == MO_64) { | ||
135 | if (opc & MO_BSWAP) { | ||
136 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
137 | } | ||
138 | |||
139 | static void tcg_out_qemu_st(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
140 | - TCGReg addrlo, TCGReg addrhi, | ||
141 | - MemOpIdx oi, TCGType data_type) | ||
142 | + TCGReg addr, MemOpIdx oi, TCGType data_type) | ||
143 | { | ||
144 | MemOp opc = get_memop(oi); | ||
145 | TCGLabelQemuLdst *ldst; | ||
146 | HostAddress h; | ||
147 | |||
148 | - ldst = prepare_host_addr(s, &h, addrlo, addrhi, oi, false); | ||
149 | + ldst = prepare_host_addr(s, &h, addr, oi, false); | ||
150 | |||
151 | if (TCG_TARGET_REG_BITS == 32 && (opc & MO_SIZE) == MO_64) { | ||
152 | if (opc & MO_BSWAP) { | ||
153 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ldst_i128(TCGContext *s, TCGReg datalo, TCGReg datahi, | ||
154 | uint32_t insn; | ||
155 | TCGReg index; | ||
156 | |||
157 | - ldst = prepare_host_addr(s, &h, addr_reg, -1, oi, is_ld); | ||
158 | + ldst = prepare_host_addr(s, &h, addr_reg, oi, is_ld); | ||
159 | |||
160 | /* Compose the final address, as LQ/STQ have no indexing. */ | ||
161 | index = h.index; | ||
162 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
163 | break; | ||
164 | |||
165 | case INDEX_op_qemu_ld_i32: | ||
166 | - tcg_out_qemu_ld(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
167 | + tcg_out_qemu_ld(s, args[0], -1, args[1], args[2], TCG_TYPE_I32); | ||
168 | break; | ||
169 | case INDEX_op_qemu_ld_i64: | ||
170 | if (TCG_TARGET_REG_BITS == 64) { | ||
171 | - tcg_out_qemu_ld(s, args[0], -1, args[1], -1, | ||
172 | - args[2], TCG_TYPE_I64); | ||
173 | + tcg_out_qemu_ld(s, args[0], -1, args[1], args[2], TCG_TYPE_I64); | ||
174 | } else { | ||
175 | - tcg_out_qemu_ld(s, args[0], args[1], args[2], -1, | ||
176 | + tcg_out_qemu_ld(s, args[0], args[1], args[2], | ||
177 | args[3], TCG_TYPE_I64); | ||
178 | } | ||
179 | break; | ||
180 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
181 | break; | ||
182 | |||
183 | case INDEX_op_qemu_st_i32: | ||
184 | - tcg_out_qemu_st(s, args[0], -1, args[1], -1, args[2], TCG_TYPE_I32); | ||
185 | + tcg_out_qemu_st(s, args[0], -1, args[1], args[2], TCG_TYPE_I32); | ||
186 | break; | ||
187 | case INDEX_op_qemu_st_i64: | ||
188 | if (TCG_TARGET_REG_BITS == 64) { | ||
189 | - tcg_out_qemu_st(s, args[0], -1, args[1], -1, | ||
190 | - args[2], TCG_TYPE_I64); | ||
191 | + tcg_out_qemu_st(s, args[0], -1, args[1], args[2], TCG_TYPE_I64); | ||
192 | } else { | ||
193 | - tcg_out_qemu_st(s, args[0], args[1], args[2], -1, | ||
194 | + tcg_out_qemu_st(s, args[0], args[1], args[2], | ||
195 | args[3], TCG_TYPE_I64); | ||
196 | } | ||
197 | break; | ||
31 | -- | 198 | -- |
32 | 2.17.2 | 199 | 2.43.0 |
33 | 200 | ||
34 | 201 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | There is now always only one guest address register. |
---|---|---|---|
2 | 2 | ||
3 | We forgot to initialize n in commit 15fa08f845 ("tcg: Dynamically | 3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | allocate TCGOps", 2017-12-29). | ||
5 | |||
6 | Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com> | ||
7 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
8 | Message-Id: <20181010144853.13005-3-cota@braap.org> | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
10 | --- | 5 | --- |
11 | tcg/tcg.c | 2 +- | 6 | tcg/tcg.c | 18 +++++++++--------- |
12 | 1 file changed, 1 insertion(+), 1 deletion(-) | 7 | tcg/aarch64/tcg-target.c.inc | 4 ++-- |
8 | tcg/arm/tcg-target.c.inc | 4 ++-- | ||
9 | tcg/i386/tcg-target.c.inc | 4 ++-- | ||
10 | tcg/loongarch64/tcg-target.c.inc | 4 ++-- | ||
11 | tcg/mips/tcg-target.c.inc | 4 ++-- | ||
12 | tcg/ppc/tcg-target.c.inc | 4 ++-- | ||
13 | tcg/riscv/tcg-target.c.inc | 4 ++-- | ||
14 | tcg/s390x/tcg-target.c.inc | 4 ++-- | ||
15 | tcg/sparc64/tcg-target.c.inc | 4 ++-- | ||
16 | 10 files changed, 27 insertions(+), 27 deletions(-) | ||
13 | 17 | ||
14 | diff --git a/tcg/tcg.c b/tcg/tcg.c | 18 | diff --git a/tcg/tcg.c b/tcg/tcg.c |
15 | index XXXXXXX..XXXXXXX 100644 | 19 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/tcg/tcg.c | 20 | --- a/tcg/tcg.c |
17 | +++ b/tcg/tcg.c | 21 | +++ b/tcg/tcg.c |
18 | @@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb) | 22 | @@ -XXX,XX +XXX,XX @@ struct TCGLabelQemuLdst { |
19 | 23 | bool is_ld; /* qemu_ld: true, qemu_st: false */ | |
20 | #ifdef CONFIG_PROFILER | 24 | MemOpIdx oi; |
21 | { | 25 | TCGType type; /* result type of a load */ |
22 | - int n; | 26 | - TCGReg addrlo_reg; /* reg index for low word of guest virtual addr */ |
23 | + int n = 0; | 27 | - TCGReg addrhi_reg; /* reg index for high word of guest virtual addr */ |
24 | 28 | + TCGReg addr_reg; /* reg index for guest virtual addr */ | |
25 | QTAILQ_FOREACH(op, &s->ops, link) { | 29 | TCGReg datalo_reg; /* reg index for low word to be loaded or stored */ |
26 | n++; | 30 | TCGReg datahi_reg; /* reg index for high word to be loaded or stored */ |
31 | const tcg_insn_unit *raddr; /* addr of the next IR of qemu_ld/st IR */ | ||
32 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_ld_helper_args(TCGContext *s, const TCGLabelQemuLdst *ldst, | ||
33 | */ | ||
34 | tcg_out_helper_add_mov(mov, loc + HOST_BIG_ENDIAN, | ||
35 | TCG_TYPE_I32, TCG_TYPE_I32, | ||
36 | - ldst->addrlo_reg, -1); | ||
37 | + ldst->addr_reg, -1); | ||
38 | tcg_out_helper_load_slots(s, 1, mov, parm); | ||
39 | |||
40 | tcg_out_helper_load_imm(s, loc[!HOST_BIG_ENDIAN].arg_slot, | ||
41 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_ld_helper_args(TCGContext *s, const TCGLabelQemuLdst *ldst, | ||
42 | next_arg += 2; | ||
43 | } else { | ||
44 | nmov = tcg_out_helper_add_mov(mov, loc, TCG_TYPE_I64, s->addr_type, | ||
45 | - ldst->addrlo_reg, ldst->addrhi_reg); | ||
46 | + ldst->addr_reg, -1); | ||
47 | tcg_out_helper_load_slots(s, nmov, mov, parm); | ||
48 | next_arg += nmov; | ||
49 | } | ||
50 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_st_helper_args(TCGContext *s, const TCGLabelQemuLdst *ldst, | ||
51 | |||
52 | /* Handle addr argument. */ | ||
53 | loc = &info->in[next_arg]; | ||
54 | - if (TCG_TARGET_REG_BITS == 32 && s->addr_type == TCG_TYPE_I32) { | ||
55 | + tcg_debug_assert(s->addr_type <= TCG_TYPE_REG); | ||
56 | + if (TCG_TARGET_REG_BITS == 32) { | ||
57 | /* | ||
58 | - * 32-bit host with 32-bit guest: zero-extend the guest address | ||
59 | + * 32-bit host (and thus 32-bit guest): zero-extend the guest address | ||
60 | * to 64-bits for the helper by storing the low part. Later, | ||
61 | * after we have processed the register inputs, we will load a | ||
62 | * zero for the high part. | ||
63 | */ | ||
64 | tcg_out_helper_add_mov(mov, loc + HOST_BIG_ENDIAN, | ||
65 | TCG_TYPE_I32, TCG_TYPE_I32, | ||
66 | - ldst->addrlo_reg, -1); | ||
67 | + ldst->addr_reg, -1); | ||
68 | next_arg += 2; | ||
69 | nmov += 1; | ||
70 | } else { | ||
71 | n = tcg_out_helper_add_mov(mov, loc, TCG_TYPE_I64, s->addr_type, | ||
72 | - ldst->addrlo_reg, ldst->addrhi_reg); | ||
73 | + ldst->addr_reg, -1); | ||
74 | next_arg += n; | ||
75 | nmov += n; | ||
76 | } | ||
77 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_st_helper_args(TCGContext *s, const TCGLabelQemuLdst *ldst, | ||
78 | g_assert_not_reached(); | ||
79 | } | ||
80 | |||
81 | - if (TCG_TARGET_REG_BITS == 32 && s->addr_type == TCG_TYPE_I32) { | ||
82 | + if (TCG_TARGET_REG_BITS == 32) { | ||
83 | /* Zero extend the address by loading a zero for the high part. */ | ||
84 | loc = &info->in[1 + !HOST_BIG_ENDIAN]; | ||
85 | tcg_out_helper_load_imm(s, loc->arg_slot, TCG_TYPE_I32, 0, parm); | ||
86 | diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc | ||
87 | index XXXXXXX..XXXXXXX 100644 | ||
88 | --- a/tcg/aarch64/tcg-target.c.inc | ||
89 | +++ b/tcg/aarch64/tcg-target.c.inc | ||
90 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
91 | ldst = new_ldst_label(s); | ||
92 | ldst->is_ld = is_ld; | ||
93 | ldst->oi = oi; | ||
94 | - ldst->addrlo_reg = addr_reg; | ||
95 | + ldst->addr_reg = addr_reg; | ||
96 | |||
97 | mask_type = (s->page_bits + s->tlb_dyn_max_bits > 32 | ||
98 | ? TCG_TYPE_I64 : TCG_TYPE_I32); | ||
99 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
100 | |||
101 | ldst->is_ld = is_ld; | ||
102 | ldst->oi = oi; | ||
103 | - ldst->addrlo_reg = addr_reg; | ||
104 | + ldst->addr_reg = addr_reg; | ||
105 | |||
106 | /* tst addr, #mask */ | ||
107 | tcg_out_logicali(s, I3404_ANDSI, 0, TCG_REG_XZR, addr_reg, a_mask); | ||
108 | diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc | ||
109 | index XXXXXXX..XXXXXXX 100644 | ||
110 | --- a/tcg/arm/tcg-target.c.inc | ||
111 | +++ b/tcg/arm/tcg-target.c.inc | ||
112 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
113 | ldst = new_ldst_label(s); | ||
114 | ldst->is_ld = is_ld; | ||
115 | ldst->oi = oi; | ||
116 | - ldst->addrlo_reg = addr; | ||
117 | + ldst->addr_reg = addr; | ||
118 | |||
119 | /* Load cpu->neg.tlb.f[mmu_idx].{mask,table} into {r0,r1}. */ | ||
120 | QEMU_BUILD_BUG_ON(offsetof(CPUTLBDescFast, mask) != 0); | ||
121 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
122 | ldst = new_ldst_label(s); | ||
123 | ldst->is_ld = is_ld; | ||
124 | ldst->oi = oi; | ||
125 | - ldst->addrlo_reg = addr; | ||
126 | + ldst->addr_reg = addr; | ||
127 | |||
128 | /* We are expecting alignment to max out at 7 */ | ||
129 | tcg_debug_assert(a_mask <= 0xff); | ||
130 | diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc | ||
131 | index XXXXXXX..XXXXXXX 100644 | ||
132 | --- a/tcg/i386/tcg-target.c.inc | ||
133 | +++ b/tcg/i386/tcg-target.c.inc | ||
134 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
135 | ldst = new_ldst_label(s); | ||
136 | ldst->is_ld = is_ld; | ||
137 | ldst->oi = oi; | ||
138 | - ldst->addrlo_reg = addr; | ||
139 | + ldst->addr_reg = addr; | ||
140 | |||
141 | if (TCG_TARGET_REG_BITS == 64) { | ||
142 | ttype = s->addr_type; | ||
143 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
144 | ldst = new_ldst_label(s); | ||
145 | ldst->is_ld = is_ld; | ||
146 | ldst->oi = oi; | ||
147 | - ldst->addrlo_reg = addr; | ||
148 | + ldst->addr_reg = addr; | ||
149 | |||
150 | /* jne slow_path */ | ||
151 | jcc = tcg_out_cmp(s, TCG_COND_TSTNE, addr, a_mask, true, false); | ||
152 | diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc | ||
153 | index XXXXXXX..XXXXXXX 100644 | ||
154 | --- a/tcg/loongarch64/tcg-target.c.inc | ||
155 | +++ b/tcg/loongarch64/tcg-target.c.inc | ||
156 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
157 | ldst = new_ldst_label(s); | ||
158 | ldst->is_ld = is_ld; | ||
159 | ldst->oi = oi; | ||
160 | - ldst->addrlo_reg = addr_reg; | ||
161 | + ldst->addr_reg = addr_reg; | ||
162 | |||
163 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP0, TCG_AREG0, mask_ofs); | ||
164 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP1, TCG_AREG0, table_ofs); | ||
165 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
166 | |||
167 | ldst->is_ld = is_ld; | ||
168 | ldst->oi = oi; | ||
169 | - ldst->addrlo_reg = addr_reg; | ||
170 | + ldst->addr_reg = addr_reg; | ||
171 | |||
172 | /* | ||
173 | * Without micro-architecture details, we don't know which of | ||
174 | diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc | ||
175 | index XXXXXXX..XXXXXXX 100644 | ||
176 | --- a/tcg/mips/tcg-target.c.inc | ||
177 | +++ b/tcg/mips/tcg-target.c.inc | ||
178 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
179 | ldst = new_ldst_label(s); | ||
180 | ldst->is_ld = is_ld; | ||
181 | ldst->oi = oi; | ||
182 | - ldst->addrlo_reg = addr; | ||
183 | + ldst->addr_reg = addr; | ||
184 | |||
185 | /* Load tlb_mask[mmu_idx] and tlb_table[mmu_idx]. */ | ||
186 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP0, TCG_AREG0, mask_off); | ||
187 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
188 | |||
189 | ldst->is_ld = is_ld; | ||
190 | ldst->oi = oi; | ||
191 | - ldst->addrlo_reg = addr; | ||
192 | + ldst->addr_reg = addr; | ||
193 | |||
194 | /* We are expecting a_bits to max out at 7, much lower than ANDI. */ | ||
195 | tcg_debug_assert(a_bits < 16); | ||
196 | diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc | ||
197 | index XXXXXXX..XXXXXXX 100644 | ||
198 | --- a/tcg/ppc/tcg-target.c.inc | ||
199 | +++ b/tcg/ppc/tcg-target.c.inc | ||
200 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
201 | ldst = new_ldst_label(s); | ||
202 | ldst->is_ld = is_ld; | ||
203 | ldst->oi = oi; | ||
204 | - ldst->addrlo_reg = addr; | ||
205 | + ldst->addr_reg = addr; | ||
206 | |||
207 | /* Load tlb_mask[mmu_idx] and tlb_table[mmu_idx]. */ | ||
208 | tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP1, TCG_AREG0, mask_off); | ||
209 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
210 | ldst = new_ldst_label(s); | ||
211 | ldst->is_ld = is_ld; | ||
212 | ldst->oi = oi; | ||
213 | - ldst->addrlo_reg = addr; | ||
214 | + ldst->addr_reg = addr; | ||
215 | |||
216 | /* We are expecting a_bits to max out at 7, much lower than ANDI. */ | ||
217 | tcg_debug_assert(a_bits < 16); | ||
218 | diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc | ||
219 | index XXXXXXX..XXXXXXX 100644 | ||
220 | --- a/tcg/riscv/tcg-target.c.inc | ||
221 | +++ b/tcg/riscv/tcg-target.c.inc | ||
222 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, TCGReg *pbase, | ||
223 | ldst = new_ldst_label(s); | ||
224 | ldst->is_ld = is_ld; | ||
225 | ldst->oi = oi; | ||
226 | - ldst->addrlo_reg = addr_reg; | ||
227 | + ldst->addr_reg = addr_reg; | ||
228 | |||
229 | init_setting_vtype(s); | ||
230 | |||
231 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, TCGReg *pbase, | ||
232 | ldst = new_ldst_label(s); | ||
233 | ldst->is_ld = is_ld; | ||
234 | ldst->oi = oi; | ||
235 | - ldst->addrlo_reg = addr_reg; | ||
236 | + ldst->addr_reg = addr_reg; | ||
237 | |||
238 | init_setting_vtype(s); | ||
239 | |||
240 | diff --git a/tcg/s390x/tcg-target.c.inc b/tcg/s390x/tcg-target.c.inc | ||
241 | index XXXXXXX..XXXXXXX 100644 | ||
242 | --- a/tcg/s390x/tcg-target.c.inc | ||
243 | +++ b/tcg/s390x/tcg-target.c.inc | ||
244 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
245 | ldst = new_ldst_label(s); | ||
246 | ldst->is_ld = is_ld; | ||
247 | ldst->oi = oi; | ||
248 | - ldst->addrlo_reg = addr_reg; | ||
249 | + ldst->addr_reg = addr_reg; | ||
250 | |||
251 | tcg_out_sh64(s, RSY_SRLG, TCG_TMP0, addr_reg, TCG_REG_NONE, | ||
252 | s->page_bits - CPU_TLB_ENTRY_BITS); | ||
253 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
254 | ldst = new_ldst_label(s); | ||
255 | ldst->is_ld = is_ld; | ||
256 | ldst->oi = oi; | ||
257 | - ldst->addrlo_reg = addr_reg; | ||
258 | + ldst->addr_reg = addr_reg; | ||
259 | |||
260 | tcg_debug_assert(a_mask <= 0xffff); | ||
261 | tcg_out_insn(s, RI, TMLL, addr_reg, a_mask); | ||
262 | diff --git a/tcg/sparc64/tcg-target.c.inc b/tcg/sparc64/tcg-target.c.inc | ||
263 | index XXXXXXX..XXXXXXX 100644 | ||
264 | --- a/tcg/sparc64/tcg-target.c.inc | ||
265 | +++ b/tcg/sparc64/tcg-target.c.inc | ||
266 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
267 | ldst = new_ldst_label(s); | ||
268 | ldst->is_ld = is_ld; | ||
269 | ldst->oi = oi; | ||
270 | - ldst->addrlo_reg = addr_reg; | ||
271 | + ldst->addr_reg = addr_reg; | ||
272 | ldst->label_ptr[0] = s->code_ptr; | ||
273 | |||
274 | /* bne,pn %[xi]cc, label0 */ | ||
275 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
276 | ldst = new_ldst_label(s); | ||
277 | ldst->is_ld = is_ld; | ||
278 | ldst->oi = oi; | ||
279 | - ldst->addrlo_reg = addr_reg; | ||
280 | + ldst->addr_reg = addr_reg; | ||
281 | ldst->label_ptr[0] = s->code_ptr; | ||
282 | |||
283 | /* bne,pn %icc, label0 */ | ||
27 | -- | 284 | -- |
28 | 2.17.2 | 285 | 2.43.0 |
29 | 286 | ||
30 | 287 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | The declaration uses uint64_t for addr. |
---|---|---|---|
2 | 2 | ||
3 | This plugs two 4-byte holes in 64-bit. | 3 | Fixes: 595cd9ce2ec ("plugins: add plugin API to read guest memory") |
4 | 4 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | |
5 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
6 | Message-Id: <20181010144853.13005-4-cota@braap.org> | ||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
8 | --- | 6 | --- |
9 | tcg/tcg.h | 2 +- | 7 | plugins/api.c | 2 +- |
10 | 1 file changed, 1 insertion(+), 1 deletion(-) | 8 | 1 file changed, 1 insertion(+), 1 deletion(-) |
11 | 9 | ||
12 | diff --git a/tcg/tcg.h b/tcg/tcg.h | 10 | diff --git a/plugins/api.c b/plugins/api.c |
13 | index XXXXXXX..XXXXXXX 100644 | 11 | index XXXXXXX..XXXXXXX 100644 |
14 | --- a/tcg/tcg.h | 12 | --- a/plugins/api.c |
15 | +++ b/tcg/tcg.h | 13 | +++ b/plugins/api.c |
16 | @@ -XXX,XX +XXX,XX @@ typedef struct TCGProfile { | 14 | @@ -XXX,XX +XXX,XX @@ GArray *qemu_plugin_get_registers(void) |
17 | int64_t tb_count; | 15 | return create_register_handles(regs); |
18 | int64_t op_count; /* total insn count */ | 16 | } |
19 | int op_count_max; /* max insn per TB */ | 17 | |
20 | - int64_t temp_count; | 18 | -bool qemu_plugin_read_memory_vaddr(vaddr addr, GByteArray *data, size_t len) |
21 | int temp_count_max; | 19 | +bool qemu_plugin_read_memory_vaddr(uint64_t addr, GByteArray *data, size_t len) |
22 | + int64_t temp_count; | 20 | { |
23 | int64_t del_op_count; | 21 | g_assert(current_cpu); |
24 | int64_t code_in_len; | 22 | |
25 | int64_t code_out_len; | ||
26 | -- | 23 | -- |
27 | 2.17.2 | 24 | 2.43.0 |
28 | 25 | ||
29 | 26 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | The declarations use vaddr for size. |
---|---|---|---|
2 | 2 | ||
3 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
4 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
5 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
6 | Message-Id: <20181009174557.16125-5-cota@braap.org> | ||
7 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
8 | --- | 4 | --- |
9 | accel/tcg/cputlb.c | 4 ++-- | 5 | accel/tcg/cputlb.c | 4 ++-- |
10 | 1 file changed, 2 insertions(+), 2 deletions(-) | 6 | 1 file changed, 2 insertions(+), 2 deletions(-) |
11 | 7 | ||
12 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | 8 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c |
13 | index XXXXXXX..XXXXXXX 100644 | 9 | index XXXXXXX..XXXXXXX 100644 |
14 | --- a/accel/tcg/cputlb.c | 10 | --- a/accel/tcg/cputlb.c |
15 | +++ b/accel/tcg/cputlb.c | 11 | +++ b/accel/tcg/cputlb.c |
16 | @@ -XXX,XX +XXX,XX @@ | 12 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_full(CPUState *cpu, int mmu_idx, |
17 | } \ | 13 | |
18 | } while (0) | 14 | void tlb_set_page_with_attrs(CPUState *cpu, vaddr addr, |
19 | 15 | hwaddr paddr, MemTxAttrs attrs, int prot, | |
20 | -#define assert_cpu_is_self(this_cpu) do { \ | 16 | - int mmu_idx, uint64_t size) |
21 | +#define assert_cpu_is_self(cpu) do { \ | 17 | + int mmu_idx, vaddr size) |
22 | if (DEBUG_TLB_GATE) { \ | 18 | { |
23 | - g_assert(!cpu->created || qemu_cpu_is_self(cpu)); \ | 19 | CPUTLBEntryFull full = { |
24 | + g_assert(!(cpu)->created || qemu_cpu_is_self(cpu)); \ | 20 | .phys_addr = paddr, |
25 | } \ | 21 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, vaddr addr, |
26 | } while (0) | 22 | |
27 | 23 | void tlb_set_page(CPUState *cpu, vaddr addr, | |
24 | hwaddr paddr, int prot, | ||
25 | - int mmu_idx, uint64_t size) | ||
26 | + int mmu_idx, vaddr size) | ||
27 | { | ||
28 | tlb_set_page_with_attrs(cpu, addr, paddr, MEMTXATTRS_UNSPECIFIED, | ||
29 | prot, mmu_idx, size); | ||
28 | -- | 30 | -- |
29 | 2.17.2 | 31 | 2.43.0 |
30 | |||
31 | diff view generated by jsdifflib |
1 | Reviewed-by: Emilio G. Cota <cota@braap.org> | 1 | Since we no longer support 64-bit guests on 32-bit hosts, |
---|---|---|---|
2 | Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com> | 2 | we can use a 32-bit type on a 32-bit host. |
3 | |||
4 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
4 | --- | 6 | --- |
5 | target/i386/mem_helper.c | 9 ++++----- | 7 | include/exec/vaddr.h | 16 +++++++++------- |
6 | 1 file changed, 4 insertions(+), 5 deletions(-) | 8 | 1 file changed, 9 insertions(+), 7 deletions(-) |
7 | 9 | ||
8 | diff --git a/target/i386/mem_helper.c b/target/i386/mem_helper.c | 10 | diff --git a/include/exec/vaddr.h b/include/exec/vaddr.h |
9 | index XXXXXXX..XXXXXXX 100644 | 11 | index XXXXXXX..XXXXXXX 100644 |
10 | --- a/target/i386/mem_helper.c | 12 | --- a/include/exec/vaddr.h |
11 | +++ b/target/i386/mem_helper.c | 13 | +++ b/include/exec/vaddr.h |
12 | @@ -XXX,XX +XXX,XX @@ | 14 | @@ -XXX,XX +XXX,XX @@ |
13 | #include "exec/exec-all.h" | 15 | /** |
14 | #include "exec/cpu_ldst.h" | 16 | * vaddr: |
15 | #include "qemu/int128.h" | 17 | * Type wide enough to contain any #target_ulong virtual address. |
16 | +#include "qemu/atomic128.h" | 18 | + * We do not support 64-bit guest on 32-host and detect at configure time. |
17 | #include "tcg.h" | 19 | + * Therefore, a host pointer width will always fit a guest pointer. |
18 | 20 | */ | |
19 | void helper_cmpxchg8b_unlocked(CPUX86State *env, target_ulong a0) | 21 | -typedef uint64_t vaddr; |
20 | @@ -XXX,XX +XXX,XX @@ void helper_cmpxchg16b(CPUX86State *env, target_ulong a0) | 22 | -#define VADDR_PRId PRId64 |
21 | 23 | -#define VADDR_PRIu PRIu64 | |
22 | if ((a0 & 0xf) != 0) { | 24 | -#define VADDR_PRIo PRIo64 |
23 | raise_exception_ra(env, EXCP0D_GPF, ra); | 25 | -#define VADDR_PRIx PRIx64 |
24 | - } else { | 26 | -#define VADDR_PRIX PRIX64 |
25 | -#ifndef CONFIG_ATOMIC128 | 27 | -#define VADDR_MAX UINT64_MAX |
26 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 28 | +typedef uintptr_t vaddr; |
27 | -#else | 29 | +#define VADDR_PRId PRIdPTR |
28 | + } else if (HAVE_CMPXCHG128) { | 30 | +#define VADDR_PRIu PRIuPTR |
29 | int eflags = cpu_cc_compute_all(env, CC_OP); | 31 | +#define VADDR_PRIo PRIoPTR |
30 | 32 | +#define VADDR_PRIx PRIxPTR | |
31 | Int128 cmpv = int128_make128(env->regs[R_EAX], env->regs[R_EDX]); | 33 | +#define VADDR_PRIX PRIXPTR |
32 | @@ -XXX,XX +XXX,XX @@ void helper_cmpxchg16b(CPUX86State *env, target_ulong a0) | 34 | +#define VADDR_MAX UINTPTR_MAX |
33 | eflags &= ~CC_Z; | 35 | |
34 | } | ||
35 | CC_SRC = eflags; | ||
36 | -#endif | ||
37 | + } else { | ||
38 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
39 | } | ||
40 | } | ||
41 | #endif | 36 | #endif |
42 | -- | 37 | -- |
43 | 2.17.2 | 38 | 2.43.0 |
44 | 39 | ||
45 | 40 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Since we no longer support 64-bit guests on 32-bit hosts, |
---|---|---|---|
2 | we can use a 32-bit type on a 32-bit host. This shrinks | ||
3 | the size of the structure to 16 bytes on a 32-bit host. | ||
2 | 4 | ||
3 | Currently we rely on atomic operations for cross-CPU invalidations. | 5 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | There are two cases that these atomics miss: cross-CPU invalidations | ||
5 | can race with either (1) vCPU threads flushing their TLB, which | ||
6 | happens via memset, or (2) vCPUs calling tlb_reset_dirty on their TLB, | ||
7 | which updates .addr_write with a regular store. This results in | ||
8 | undefined behaviour, since we're mixing regular and atomic ops | ||
9 | on concurrent accesses. | ||
10 | |||
11 | Fix it by using tlb_lock, a per-vCPU lock. All updaters of tlb_table | ||
12 | and the corresponding victim cache now hold the lock. | ||
13 | The readers that do not hold tlb_lock must use atomic reads when | ||
14 | reading .addr_write, since this field can be updated by other threads; | ||
15 | the conversion to atomic reads is done in the next patch. | ||
16 | |||
17 | Note that an alternative fix would be to expand the use of atomic ops. | ||
18 | However, in the case of TLB flushes this would have a huge performance | ||
19 | impact, since (1) TLB flushes can happen very frequently and (2) we | ||
20 | currently use a full memory barrier to flush each TLB entry, and a TLB | ||
21 | has many entries. Instead, acquiring the lock is barely slower than a | ||
22 | full memory barrier since it is uncontended, and with a single lock | ||
23 | acquisition we can flush the entire TLB. | ||
24 | |||
25 | Tested-by: Alex Bennée <alex.bennee@linaro.org> | ||
26 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
27 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
28 | Message-Id: <20181009174557.16125-6-cota@braap.org> | ||
29 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 6 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
30 | --- | 7 | --- |
31 | include/exec/cpu-defs.h | 3 + | 8 | include/exec/tlb-common.h | 10 +++++----- |
32 | accel/tcg/cputlb.c | 155 ++++++++++++++++++++++------------------ | 9 | accel/tcg/cputlb.c | 21 ++++----------------- |
33 | 2 files changed, 87 insertions(+), 71 deletions(-) | 10 | tcg/arm/tcg-target.c.inc | 1 - |
11 | tcg/mips/tcg-target.c.inc | 12 +++++------- | ||
12 | tcg/ppc/tcg-target.c.inc | 21 +++++---------------- | ||
13 | 5 files changed, 19 insertions(+), 46 deletions(-) | ||
34 | 14 | ||
35 | diff --git a/include/exec/cpu-defs.h b/include/exec/cpu-defs.h | 15 | diff --git a/include/exec/tlb-common.h b/include/exec/tlb-common.h |
36 | index XXXXXXX..XXXXXXX 100644 | 16 | index XXXXXXX..XXXXXXX 100644 |
37 | --- a/include/exec/cpu-defs.h | 17 | --- a/include/exec/tlb-common.h |
38 | +++ b/include/exec/cpu-defs.h | 18 | +++ b/include/exec/tlb-common.h |
39 | @@ -XXX,XX +XXX,XX @@ | 19 | @@ -XXX,XX +XXX,XX @@ |
40 | #endif | 20 | #ifndef EXEC_TLB_COMMON_H |
41 | 21 | #define EXEC_TLB_COMMON_H 1 | |
42 | #include "qemu/host-utils.h" | 22 | |
43 | +#include "qemu/thread.h" | 23 | -#define CPU_TLB_ENTRY_BITS 5 |
44 | #include "qemu/queue.h" | 24 | +#define CPU_TLB_ENTRY_BITS (HOST_LONG_BITS == 32 ? 4 : 5) |
45 | #ifdef CONFIG_TCG | 25 | |
46 | #include "tcg-target.h" | 26 | /* Minimalized TLB entry for use by TCG fast path. */ |
47 | @@ -XXX,XX +XXX,XX @@ typedef struct CPUIOTLBEntry { | 27 | typedef union CPUTLBEntry { |
48 | 28 | struct { | |
49 | #define CPU_COMMON_TLB \ | 29 | - uint64_t addr_read; |
50 | /* The meaning of the MMU modes is defined in the target code. */ \ | 30 | - uint64_t addr_write; |
51 | + /* tlb_lock serializes updates to tlb_table and tlb_v_table */ \ | 31 | - uint64_t addr_code; |
52 | + QemuSpin tlb_lock; \ | 32 | + uintptr_t addr_read; |
53 | CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE]; \ | 33 | + uintptr_t addr_write; |
54 | CPUTLBEntry tlb_v_table[NB_MMU_MODES][CPU_VTLB_SIZE]; \ | 34 | + uintptr_t addr_code; |
55 | CPUIOTLBEntry iotlb[NB_MMU_MODES][CPU_TLB_SIZE]; \ | 35 | /* |
36 | * Addend to virtual address to get host address. IO accesses | ||
37 | * use the corresponding iotlb value. | ||
38 | @@ -XXX,XX +XXX,XX @@ typedef union CPUTLBEntry { | ||
39 | * Padding to get a power of two size, as well as index | ||
40 | * access to addr_{read,write,code}. | ||
41 | */ | ||
42 | - uint64_t addr_idx[(1 << CPU_TLB_ENTRY_BITS) / sizeof(uint64_t)]; | ||
43 | + uintptr_t addr_idx[(1 << CPU_TLB_ENTRY_BITS) / sizeof(uintptr_t)]; | ||
44 | } CPUTLBEntry; | ||
45 | |||
46 | QEMU_BUILD_BUG_ON(sizeof(CPUTLBEntry) != (1 << CPU_TLB_ENTRY_BITS)); | ||
56 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | 47 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c |
57 | index XXXXXXX..XXXXXXX 100644 | 48 | index XXXXXXX..XXXXXXX 100644 |
58 | --- a/accel/tcg/cputlb.c | 49 | --- a/accel/tcg/cputlb.c |
59 | +++ b/accel/tcg/cputlb.c | 50 | +++ b/accel/tcg/cputlb.c |
60 | @@ -XXX,XX +XXX,XX @@ QEMU_BUILD_BUG_ON(NB_MMU_MODES > 16); | 51 | @@ -XXX,XX +XXX,XX @@ static inline uint64_t tlb_read_idx(const CPUTLBEntry *entry, |
61 | |||
62 | void tlb_init(CPUState *cpu) | ||
63 | { | 52 | { |
64 | + CPUArchState *env = cpu->env_ptr; | 53 | /* Do not rearrange the CPUTLBEntry structure members. */ |
65 | + | 54 | QEMU_BUILD_BUG_ON(offsetof(CPUTLBEntry, addr_read) != |
66 | + qemu_spin_init(&env->tlb_lock); | 55 | - MMU_DATA_LOAD * sizeof(uint64_t)); |
56 | + MMU_DATA_LOAD * sizeof(uintptr_t)); | ||
57 | QEMU_BUILD_BUG_ON(offsetof(CPUTLBEntry, addr_write) != | ||
58 | - MMU_DATA_STORE * sizeof(uint64_t)); | ||
59 | + MMU_DATA_STORE * sizeof(uintptr_t)); | ||
60 | QEMU_BUILD_BUG_ON(offsetof(CPUTLBEntry, addr_code) != | ||
61 | - MMU_INST_FETCH * sizeof(uint64_t)); | ||
62 | + MMU_INST_FETCH * sizeof(uintptr_t)); | ||
63 | |||
64 | -#if TARGET_LONG_BITS == 32 | ||
65 | - /* Use qatomic_read, in case of addr_write; only care about low bits. */ | ||
66 | - const uint32_t *ptr = (uint32_t *)&entry->addr_idx[access_type]; | ||
67 | - ptr += HOST_BIG_ENDIAN; | ||
68 | - return qatomic_read(ptr); | ||
69 | -#else | ||
70 | - const uint64_t *ptr = &entry->addr_idx[access_type]; | ||
71 | + const uintptr_t *ptr = &entry->addr_idx[access_type]; | ||
72 | /* ofs might correspond to .addr_write, so use qatomic_read */ | ||
73 | return qatomic_read(ptr); | ||
74 | -#endif | ||
67 | } | 75 | } |
68 | 76 | ||
69 | /* flush_all_helper: run fn across all cpus | 77 | static inline uint64_t tlb_addr_write(const CPUTLBEntry *entry) |
70 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_nocheck(CPUState *cpu) | 78 | @@ -XXX,XX +XXX,XX @@ static void tlb_reset_dirty_range_locked(CPUTLBEntry *tlb_entry, |
71 | atomic_set(&env->tlb_flush_count, env->tlb_flush_count + 1); | ||
72 | tlb_debug("(count: %zu)\n", tlb_flush_count()); | ||
73 | |||
74 | + /* | ||
75 | + * tlb_table/tlb_v_table updates from any thread must hold tlb_lock. | ||
76 | + * However, updates from the owner thread (as is the case here; see the | ||
77 | + * above assert_cpu_is_self) do not need atomic_set because all reads | ||
78 | + * that do not hold the lock are performed by the same owner thread. | ||
79 | + */ | ||
80 | + qemu_spin_lock(&env->tlb_lock); | ||
81 | memset(env->tlb_table, -1, sizeof(env->tlb_table)); | ||
82 | memset(env->tlb_v_table, -1, sizeof(env->tlb_v_table)); | ||
83 | + qemu_spin_unlock(&env->tlb_lock); | ||
84 | + | ||
85 | cpu_tb_jmp_cache_clear(cpu); | ||
86 | |||
87 | env->vtlb_index = 0; | ||
88 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_by_mmuidx_async_work(CPUState *cpu, run_on_cpu_data data) | ||
89 | |||
90 | tlb_debug("start: mmu_idx:0x%04lx\n", mmu_idx_bitmask); | ||
91 | |||
92 | + qemu_spin_lock(&env->tlb_lock); | ||
93 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
94 | |||
95 | if (test_bit(mmu_idx, &mmu_idx_bitmask)) { | ||
96 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_by_mmuidx_async_work(CPUState *cpu, run_on_cpu_data data) | ||
97 | memset(env->tlb_v_table[mmu_idx], -1, sizeof(env->tlb_v_table[0])); | ||
98 | } | ||
99 | } | ||
100 | + qemu_spin_unlock(&env->tlb_lock); | ||
101 | |||
102 | cpu_tb_jmp_cache_clear(cpu); | ||
103 | |||
104 | @@ -XXX,XX +XXX,XX @@ static inline bool tlb_hit_page_anyprot(CPUTLBEntry *tlb_entry, | ||
105 | tlb_hit_page(tlb_entry->addr_code, page); | ||
106 | } | ||
107 | |||
108 | -static inline void tlb_flush_entry(CPUTLBEntry *tlb_entry, target_ulong page) | ||
109 | +/* Called with tlb_lock held */ | ||
110 | +static inline void tlb_flush_entry_locked(CPUTLBEntry *tlb_entry, | ||
111 | + target_ulong page) | ||
112 | { | ||
113 | if (tlb_hit_page_anyprot(tlb_entry, page)) { | ||
114 | memset(tlb_entry, -1, sizeof(*tlb_entry)); | ||
115 | } | ||
116 | } | ||
117 | |||
118 | -static inline void tlb_flush_vtlb_page(CPUArchState *env, int mmu_idx, | ||
119 | - target_ulong page) | ||
120 | +/* Called with tlb_lock held */ | ||
121 | +static inline void tlb_flush_vtlb_page_locked(CPUArchState *env, int mmu_idx, | ||
122 | + target_ulong page) | ||
123 | { | ||
124 | int k; | ||
125 | + | ||
126 | + assert_cpu_is_self(ENV_GET_CPU(env)); | ||
127 | for (k = 0; k < CPU_VTLB_SIZE; k++) { | ||
128 | - tlb_flush_entry(&env->tlb_v_table[mmu_idx][k], page); | ||
129 | + tlb_flush_entry_locked(&env->tlb_v_table[mmu_idx][k], page); | ||
130 | } | ||
131 | } | ||
132 | |||
133 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_async_work(CPUState *cpu, run_on_cpu_data data) | ||
134 | |||
135 | addr &= TARGET_PAGE_MASK; | ||
136 | i = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
137 | + qemu_spin_lock(&env->tlb_lock); | ||
138 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
139 | - tlb_flush_entry(&env->tlb_table[mmu_idx][i], addr); | ||
140 | - tlb_flush_vtlb_page(env, mmu_idx, addr); | ||
141 | + tlb_flush_entry_locked(&env->tlb_table[mmu_idx][i], addr); | ||
142 | + tlb_flush_vtlb_page_locked(env, mmu_idx, addr); | ||
143 | } | ||
144 | + qemu_spin_unlock(&env->tlb_lock); | ||
145 | |||
146 | tb_flush_jmp_cache(cpu, addr); | ||
147 | } | ||
148 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_by_mmuidx_async_work(CPUState *cpu, | ||
149 | tlb_debug("page:%d addr:"TARGET_FMT_lx" mmu_idx:0x%lx\n", | ||
150 | page, addr, mmu_idx_bitmap); | ||
151 | |||
152 | + qemu_spin_lock(&env->tlb_lock); | ||
153 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
154 | if (test_bit(mmu_idx, &mmu_idx_bitmap)) { | ||
155 | - tlb_flush_entry(&env->tlb_table[mmu_idx][page], addr); | ||
156 | - tlb_flush_vtlb_page(env, mmu_idx, addr); | ||
157 | + tlb_flush_entry_locked(&env->tlb_table[mmu_idx][page], addr); | ||
158 | + tlb_flush_vtlb_page_locked(env, mmu_idx, addr); | ||
159 | } | ||
160 | } | ||
161 | + qemu_spin_unlock(&env->tlb_lock); | ||
162 | |||
163 | tb_flush_jmp_cache(cpu, addr); | ||
164 | } | ||
165 | @@ -XXX,XX +XXX,XX @@ void tlb_unprotect_code(ram_addr_t ram_addr) | ||
166 | * most usual is detecting writes to code regions which may invalidate | ||
167 | * generated code. | ||
168 | * | ||
169 | - * Because we want other vCPUs to respond to changes straight away we | ||
170 | - * update the te->addr_write field atomically. If the TLB entry has | ||
171 | - * been changed by the vCPU in the mean time we skip the update. | ||
172 | + * Other vCPUs might be reading their TLBs during guest execution, so we update | ||
173 | + * te->addr_write with atomic_set. We don't need to worry about this for | ||
174 | + * oversized guests as MTTCG is disabled for them. | ||
175 | * | ||
176 | - * As this function uses atomic accesses we also need to ensure | ||
177 | - * updates to tlb_entries follow the same access rules. We don't need | ||
178 | - * to worry about this for oversized guests as MTTCG is disabled for | ||
179 | - * them. | ||
180 | + * Called with tlb_lock held. | ||
181 | */ | ||
182 | - | ||
183 | -static void tlb_reset_dirty_range(CPUTLBEntry *tlb_entry, uintptr_t start, | ||
184 | - uintptr_t length) | ||
185 | +static void tlb_reset_dirty_range_locked(CPUTLBEntry *tlb_entry, | ||
186 | + uintptr_t start, uintptr_t length) | ||
187 | { | ||
188 | -#if TCG_OVERSIZED_GUEST | ||
189 | uintptr_t addr = tlb_entry->addr_write; | ||
190 | |||
191 | if ((addr & (TLB_INVALID_MASK | TLB_MMIO | TLB_NOTDIRTY)) == 0) { | ||
192 | addr &= TARGET_PAGE_MASK; | 79 | addr &= TARGET_PAGE_MASK; |
193 | addr += tlb_entry->addend; | 80 | addr += tlb_entry->addend; |
194 | if ((addr - start) < length) { | 81 | if ((addr - start) < length) { |
195 | +#if TCG_OVERSIZED_GUEST | 82 | -#if TARGET_LONG_BITS == 32 |
196 | tlb_entry->addr_write |= TLB_NOTDIRTY; | 83 | - uint32_t *ptr_write = (uint32_t *)&tlb_entry->addr_write; |
197 | - } | 84 | - ptr_write += HOST_BIG_ENDIAN; |
198 | - } | 85 | - qatomic_set(ptr_write, *ptr_write | TLB_NOTDIRTY); |
199 | #else | 86 | -#else |
200 | - /* paired with atomic_mb_set in tlb_set_page_with_attrs */ | 87 | qatomic_set(&tlb_entry->addr_write, |
201 | - uintptr_t orig_addr = atomic_mb_read(&tlb_entry->addr_write); | 88 | tlb_entry->addr_write | TLB_NOTDIRTY); |
202 | - uintptr_t addr = orig_addr; | 89 | -#endif |
203 | - | ||
204 | - if ((addr & (TLB_INVALID_MASK | TLB_MMIO | TLB_NOTDIRTY)) == 0) { | ||
205 | - addr &= TARGET_PAGE_MASK; | ||
206 | - addr += atomic_read(&tlb_entry->addend); | ||
207 | - if ((addr - start) < length) { | ||
208 | - uintptr_t notdirty_addr = orig_addr | TLB_NOTDIRTY; | ||
209 | - atomic_cmpxchg(&tlb_entry->addr_write, orig_addr, notdirty_addr); | ||
210 | + atomic_set(&tlb_entry->addr_write, | ||
211 | + tlb_entry->addr_write | TLB_NOTDIRTY); | ||
212 | +#endif | ||
213 | } | 90 | } |
214 | } | 91 | } |
215 | -#endif | ||
216 | } | 92 | } |
217 | 93 | diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc | |
218 | -/* For atomic correctness when running MTTCG we need to use the right | 94 | index XXXXXXX..XXXXXXX 100644 |
219 | - * primitives when copying entries */ | 95 | --- a/tcg/arm/tcg-target.c.inc |
220 | -static inline void copy_tlb_helper(CPUTLBEntry *d, CPUTLBEntry *s, | 96 | +++ b/tcg/arm/tcg-target.c.inc |
221 | - bool atomic_set) | 97 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
222 | +/* | 98 | * Add the tlb_table pointer, creating the CPUTLBEntry address in R1. |
223 | + * Called with tlb_lock held. | 99 | * Load the tlb comparator into R2 and the fast path addend into R1. |
224 | + * Called only from the vCPU context, i.e. the TLB's owner thread. | 100 | */ |
225 | + */ | 101 | - QEMU_BUILD_BUG_ON(HOST_BIG_ENDIAN); |
226 | +static inline void copy_tlb_helper_locked(CPUTLBEntry *d, const CPUTLBEntry *s) | 102 | if (cmp_off == 0) { |
227 | { | 103 | tcg_out_ld32_rwb(s, COND_AL, TCG_REG_R2, TCG_REG_R1, TCG_REG_R0); |
228 | -#if TCG_OVERSIZED_GUEST | 104 | } else { |
229 | *d = *s; | 105 | diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc |
230 | -#else | 106 | index XXXXXXX..XXXXXXX 100644 |
231 | - if (atomic_set) { | 107 | --- a/tcg/mips/tcg-target.c.inc |
232 | - d->addr_read = s->addr_read; | 108 | +++ b/tcg/mips/tcg-target.c.inc |
233 | - d->addr_code = s->addr_code; | 109 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, |
234 | - atomic_set(&d->addend, atomic_read(&s->addend)); | 110 | /* Add the tlb_table pointer, creating the CPUTLBEntry address. */ |
235 | - /* Pairs with flag setting in tlb_reset_dirty_range */ | 111 | tcg_out_opc_reg(s, ALIAS_PADD, TCG_TMP3, TCG_TMP3, TCG_TMP1); |
236 | - atomic_mb_set(&d->addr_write, atomic_read(&s->addr_write)); | 112 | |
237 | - } else { | 113 | - if (TCG_TARGET_REG_BITS == 32 || addr_type == TCG_TYPE_I32) { |
238 | - d->addr_read = s->addr_read; | 114 | - /* Load the (low half) tlb comparator. */ |
239 | - d->addr_write = atomic_read(&s->addr_write); | 115 | + /* Load the tlb comparator. */ |
240 | - d->addr_code = s->addr_code; | 116 | + if (TCG_TARGET_REG_BITS == 64 && addr_type == TCG_TYPE_I32) { |
241 | - d->addend = atomic_read(&s->addend); | 117 | tcg_out_ld(s, TCG_TYPE_I32, TCG_TMP0, TCG_TMP3, |
242 | - } | 118 | cmp_off + HOST_BIG_ENDIAN * 4); |
243 | -#endif | 119 | } else { |
244 | } | 120 | - tcg_out_ld(s, TCG_TYPE_I64, TCG_TMP0, TCG_TMP3, cmp_off); |
245 | 121 | + tcg_out_ld(s, TCG_TYPE_REG, TCG_TMP0, TCG_TMP3, cmp_off); | |
246 | /* This is a cross vCPU call (i.e. another vCPU resetting the flags of | ||
247 | - * the target vCPU). As such care needs to be taken that we don't | ||
248 | - * dangerously race with another vCPU update. The only thing actually | ||
249 | - * updated is the target TLB entry ->addr_write flags. | ||
250 | + * the target vCPU). | ||
251 | + * We must take tlb_lock to avoid racing with another vCPU update. The only | ||
252 | + * thing actually updated is the target TLB entry ->addr_write flags. | ||
253 | */ | ||
254 | void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length) | ||
255 | { | ||
256 | @@ -XXX,XX +XXX,XX @@ void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length) | ||
257 | int mmu_idx; | ||
258 | |||
259 | env = cpu->env_ptr; | ||
260 | + qemu_spin_lock(&env->tlb_lock); | ||
261 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
262 | unsigned int i; | ||
263 | |||
264 | for (i = 0; i < CPU_TLB_SIZE; i++) { | ||
265 | - tlb_reset_dirty_range(&env->tlb_table[mmu_idx][i], | ||
266 | - start1, length); | ||
267 | + tlb_reset_dirty_range_locked(&env->tlb_table[mmu_idx][i], start1, | ||
268 | + length); | ||
269 | } | 122 | } |
270 | 123 | ||
271 | for (i = 0; i < CPU_VTLB_SIZE; i++) { | 124 | - if (TCG_TARGET_REG_BITS == 64 || addr_type == TCG_TYPE_I32) { |
272 | - tlb_reset_dirty_range(&env->tlb_v_table[mmu_idx][i], | 125 | - /* Load the tlb addend for the fast path. */ |
273 | - start1, length); | 126 | - tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP3, TCG_TMP3, add_off); |
274 | + tlb_reset_dirty_range_locked(&env->tlb_v_table[mmu_idx][i], start1, | 127 | - } |
275 | + length); | 128 | + /* Load the tlb addend for the fast path. */ |
129 | + tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP3, TCG_TMP3, add_off); | ||
130 | |||
131 | /* | ||
132 | * Mask the page bits, keeping the alignment bits to compare against. | ||
133 | diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc | ||
134 | index XXXXXXX..XXXXXXX 100644 | ||
135 | --- a/tcg/ppc/tcg-target.c.inc | ||
136 | +++ b/tcg/ppc/tcg-target.c.inc | ||
137 | @@ -XXX,XX +XXX,XX @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h, | ||
138 | tcg_out32(s, AND | SAB(TCG_REG_TMP1, TCG_REG_TMP1, TCG_REG_R0)); | ||
139 | |||
140 | /* | ||
141 | - * Load the (low part) TLB comparator into TMP2. | ||
142 | + * Load the TLB comparator into TMP2. | ||
143 | * For 64-bit host, always load the entire 64-bit slot for simplicity. | ||
144 | * We will ignore the high bits with tcg_out_cmp(..., addr_type). | ||
145 | */ | ||
146 | - if (TCG_TARGET_REG_BITS == 64) { | ||
147 | - if (cmp_off == 0) { | ||
148 | - tcg_out32(s, LDUX | TAB(TCG_REG_TMP2, | ||
149 | - TCG_REG_TMP1, TCG_REG_TMP2)); | ||
150 | - } else { | ||
151 | - tcg_out32(s, ADD | TAB(TCG_REG_TMP1, | ||
152 | - TCG_REG_TMP1, TCG_REG_TMP2)); | ||
153 | - tcg_out_ld(s, TCG_TYPE_I64, TCG_REG_TMP2, | ||
154 | - TCG_REG_TMP1, cmp_off); | ||
155 | - } | ||
156 | - } else if (cmp_off == 0 && !HOST_BIG_ENDIAN) { | ||
157 | - tcg_out32(s, LWZUX | TAB(TCG_REG_TMP2, | ||
158 | - TCG_REG_TMP1, TCG_REG_TMP2)); | ||
159 | + if (cmp_off == 0) { | ||
160 | + tcg_out32(s, (TCG_TARGET_REG_BITS == 64 ? LDUX : LWZUX) | ||
161 | + | TAB(TCG_REG_TMP2, TCG_REG_TMP1, TCG_REG_TMP2)); | ||
162 | } else { | ||
163 | tcg_out32(s, ADD | TAB(TCG_REG_TMP1, TCG_REG_TMP1, TCG_REG_TMP2)); | ||
164 | - tcg_out_ld(s, TCG_TYPE_I32, TCG_REG_TMP2, TCG_REG_TMP1, | ||
165 | - cmp_off + 4 * HOST_BIG_ENDIAN); | ||
166 | + tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_TMP2, TCG_REG_TMP1, cmp_off); | ||
276 | } | 167 | } |
277 | } | 168 | |
278 | + qemu_spin_unlock(&env->tlb_lock); | 169 | /* |
279 | } | ||
280 | |||
281 | -static inline void tlb_set_dirty1(CPUTLBEntry *tlb_entry, target_ulong vaddr) | ||
282 | +/* Called with tlb_lock held */ | ||
283 | +static inline void tlb_set_dirty1_locked(CPUTLBEntry *tlb_entry, | ||
284 | + target_ulong vaddr) | ||
285 | { | ||
286 | if (tlb_entry->addr_write == (vaddr | TLB_NOTDIRTY)) { | ||
287 | tlb_entry->addr_write = vaddr; | ||
288 | @@ -XXX,XX +XXX,XX @@ void tlb_set_dirty(CPUState *cpu, target_ulong vaddr) | ||
289 | |||
290 | vaddr &= TARGET_PAGE_MASK; | ||
291 | i = (vaddr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
292 | + qemu_spin_lock(&env->tlb_lock); | ||
293 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
294 | - tlb_set_dirty1(&env->tlb_table[mmu_idx][i], vaddr); | ||
295 | + tlb_set_dirty1_locked(&env->tlb_table[mmu_idx][i], vaddr); | ||
296 | } | ||
297 | |||
298 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
299 | int k; | ||
300 | for (k = 0; k < CPU_VTLB_SIZE; k++) { | ||
301 | - tlb_set_dirty1(&env->tlb_v_table[mmu_idx][k], vaddr); | ||
302 | + tlb_set_dirty1_locked(&env->tlb_v_table[mmu_idx][k], vaddr); | ||
303 | } | ||
304 | } | ||
305 | + qemu_spin_unlock(&env->tlb_lock); | ||
306 | } | ||
307 | |||
308 | /* Our TLB does not support large pages, so remember the area covered by | ||
309 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr, | ||
310 | addend = (uintptr_t)memory_region_get_ram_ptr(section->mr) + xlat; | ||
311 | } | ||
312 | |||
313 | - /* Make sure there's no cached translation for the new page. */ | ||
314 | - tlb_flush_vtlb_page(env, mmu_idx, vaddr_page); | ||
315 | - | ||
316 | code_address = address; | ||
317 | iotlb = memory_region_section_get_iotlb(cpu, section, vaddr_page, | ||
318 | paddr_page, xlat, prot, &address); | ||
319 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr, | ||
320 | index = (vaddr_page >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
321 | te = &env->tlb_table[mmu_idx][index]; | ||
322 | |||
323 | + /* | ||
324 | + * Hold the TLB lock for the rest of the function. We could acquire/release | ||
325 | + * the lock several times in the function, but it is faster to amortize the | ||
326 | + * acquisition cost by acquiring it just once. Note that this leads to | ||
327 | + * a longer critical section, but this is not a concern since the TLB lock | ||
328 | + * is unlikely to be contended. | ||
329 | + */ | ||
330 | + qemu_spin_lock(&env->tlb_lock); | ||
331 | + | ||
332 | + /* Make sure there's no cached translation for the new page. */ | ||
333 | + tlb_flush_vtlb_page_locked(env, mmu_idx, vaddr_page); | ||
334 | + | ||
335 | /* | ||
336 | * Only evict the old entry to the victim tlb if it's for a | ||
337 | * different page; otherwise just overwrite the stale data. | ||
338 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr, | ||
339 | CPUTLBEntry *tv = &env->tlb_v_table[mmu_idx][vidx]; | ||
340 | |||
341 | /* Evict the old entry into the victim tlb. */ | ||
342 | - copy_tlb_helper(tv, te, true); | ||
343 | + copy_tlb_helper_locked(tv, te); | ||
344 | env->iotlb_v[mmu_idx][vidx] = env->iotlb[mmu_idx][index]; | ||
345 | } | ||
346 | |||
347 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr, | ||
348 | } | ||
349 | } | ||
350 | |||
351 | - /* Pairs with flag setting in tlb_reset_dirty_range */ | ||
352 | - copy_tlb_helper(te, &tn, true); | ||
353 | - /* atomic_mb_set(&te->addr_write, write_address); */ | ||
354 | + copy_tlb_helper_locked(te, &tn); | ||
355 | + qemu_spin_unlock(&env->tlb_lock); | ||
356 | } | ||
357 | |||
358 | /* Add a new TLB entry, but without specifying the memory | ||
359 | @@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index, | ||
360 | size_t elt_ofs, target_ulong page) | ||
361 | { | ||
362 | size_t vidx; | ||
363 | + | ||
364 | + assert_cpu_is_self(ENV_GET_CPU(env)); | ||
365 | for (vidx = 0; vidx < CPU_VTLB_SIZE; ++vidx) { | ||
366 | CPUTLBEntry *vtlb = &env->tlb_v_table[mmu_idx][vidx]; | ||
367 | target_ulong cmp = *(target_ulong *)((uintptr_t)vtlb + elt_ofs); | ||
368 | @@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index, | ||
369 | /* Found entry in victim tlb, swap tlb and iotlb. */ | ||
370 | CPUTLBEntry tmptlb, *tlb = &env->tlb_table[mmu_idx][index]; | ||
371 | |||
372 | - copy_tlb_helper(&tmptlb, tlb, false); | ||
373 | - copy_tlb_helper(tlb, vtlb, true); | ||
374 | - copy_tlb_helper(vtlb, &tmptlb, true); | ||
375 | + qemu_spin_lock(&env->tlb_lock); | ||
376 | + copy_tlb_helper_locked(&tmptlb, tlb); | ||
377 | + copy_tlb_helper_locked(tlb, vtlb); | ||
378 | + copy_tlb_helper_locked(vtlb, &tmptlb); | ||
379 | + qemu_spin_unlock(&env->tlb_lock); | ||
380 | |||
381 | CPUIOTLBEntry tmpio, *io = &env->iotlb[mmu_idx][index]; | ||
382 | CPUIOTLBEntry *vio = &env->iotlb_v[mmu_idx][vidx]; | ||
383 | -- | 170 | -- |
384 | 2.17.2 | 171 | 2.43.0 |
385 | 172 | ||
386 | 173 | diff view generated by jsdifflib |
1 | Reviewed-by: Emilio G. Cota <cota@braap.org> | 1 | For loongarch, mips, riscv and sparc, a zero register is |
---|---|---|---|
2 | available all the time. For aarch64, register index 31 | ||
3 | depends on context: sometimes it is the stack pointer, | ||
4 | and sometimes it is the zero register. | ||
5 | |||
6 | Introduce a new general-purpose constraint which maps 0 | ||
7 | to TCG_REG_ZERO, if defined. This differs from existing | ||
8 | constant constraints in that const_arg[*] is recorded as | ||
9 | false, indicating that the value is in a register. | ||
10 | |||
11 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 12 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
3 | --- | 13 | --- |
4 | target/arm/helper-a64.c | 259 +++++++++++++++++++++------------------- | 14 | include/tcg/tcg.h | 3 ++- |
5 | 1 file changed, 133 insertions(+), 126 deletions(-) | 15 | tcg/aarch64/tcg-target.h | 2 ++ |
16 | tcg/loongarch64/tcg-target.h | 2 ++ | ||
17 | tcg/mips/tcg-target.h | 2 ++ | ||
18 | tcg/riscv/tcg-target.h | 2 ++ | ||
19 | tcg/sparc64/tcg-target.h | 3 ++- | ||
20 | tcg/tcg.c | 29 ++++++++++++++++++++++------- | ||
21 | docs/devel/tcg-ops.rst | 4 +++- | ||
22 | 8 files changed, 37 insertions(+), 10 deletions(-) | ||
6 | 23 | ||
7 | diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c | 24 | diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h |
8 | index XXXXXXX..XXXXXXX 100644 | 25 | index XXXXXXX..XXXXXXX 100644 |
9 | --- a/target/arm/helper-a64.c | 26 | --- a/include/tcg/tcg.h |
10 | +++ b/target/arm/helper-a64.c | 27 | +++ b/include/tcg/tcg.h |
11 | @@ -XXX,XX +XXX,XX @@ | 28 | @@ -XXX,XX +XXX,XX @@ void tb_target_set_jmp_target(const TranslationBlock *, int, |
12 | #include "exec/exec-all.h" | 29 | |
13 | #include "exec/cpu_ldst.h" | 30 | void tcg_set_frame(TCGContext *s, TCGReg reg, intptr_t start, intptr_t size); |
14 | #include "qemu/int128.h" | 31 | |
15 | +#include "qemu/atomic128.h" | 32 | -#define TCG_CT_CONST 1 /* any constant of register size */ |
16 | #include "tcg.h" | 33 | +#define TCG_CT_CONST 1 /* any constant of register size */ |
17 | #include "fpu/softfloat.h" | 34 | +#define TCG_CT_REG_ZERO 2 /* zero, in TCG_REG_ZERO */ |
18 | #include <zlib.h> /* For crc32 */ | 35 | |
19 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(crc32c_64)(uint64_t acc, uint64_t val, uint32_t bytes) | 36 | typedef struct TCGArgConstraint { |
20 | return crc32c(acc, buf, bytes) ^ 0xffffffff; | 37 | unsigned ct : 16; |
21 | } | 38 | diff --git a/tcg/aarch64/tcg-target.h b/tcg/aarch64/tcg-target.h |
22 | 39 | index XXXXXXX..XXXXXXX 100644 | |
23 | -/* Returns 0 on success; 1 otherwise. */ | 40 | --- a/tcg/aarch64/tcg-target.h |
24 | -static uint64_t do_paired_cmpxchg64_le(CPUARMState *env, uint64_t addr, | 41 | +++ b/tcg/aarch64/tcg-target.h |
25 | - uint64_t new_lo, uint64_t new_hi, | 42 | @@ -XXX,XX +XXX,XX @@ typedef enum { |
26 | - bool parallel, uintptr_t ra) | 43 | TCG_AREG0 = TCG_REG_X19, |
27 | +uint64_t HELPER(paired_cmpxchg64_le)(CPUARMState *env, uint64_t addr, | 44 | } TCGReg; |
28 | + uint64_t new_lo, uint64_t new_hi) | 45 | |
29 | { | 46 | +#define TCG_REG_ZERO TCG_REG_XZR |
30 | - Int128 oldv, cmpv, newv; | ||
31 | + Int128 cmpv = int128_make128(env->exclusive_val, env->exclusive_high); | ||
32 | + Int128 newv = int128_make128(new_lo, new_hi); | ||
33 | + Int128 oldv; | ||
34 | + uintptr_t ra = GETPC(); | ||
35 | + uint64_t o0, o1; | ||
36 | bool success; | ||
37 | |||
38 | - cmpv = int128_make128(env->exclusive_val, env->exclusive_high); | ||
39 | - newv = int128_make128(new_lo, new_hi); | ||
40 | - | ||
41 | - if (parallel) { | ||
42 | -#ifndef CONFIG_ATOMIC128 | ||
43 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
44 | -#else | ||
45 | - int mem_idx = cpu_mmu_index(env, false); | ||
46 | - TCGMemOpIdx oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
47 | - oldv = helper_atomic_cmpxchgo_le_mmu(env, addr, cmpv, newv, oi, ra); | ||
48 | - success = int128_eq(oldv, cmpv); | ||
49 | -#endif | ||
50 | - } else { | ||
51 | - uint64_t o0, o1; | ||
52 | - | ||
53 | #ifdef CONFIG_USER_ONLY | ||
54 | - /* ??? Enforce alignment. */ | ||
55 | - uint64_t *haddr = g2h(addr); | ||
56 | + /* ??? Enforce alignment. */ | ||
57 | + uint64_t *haddr = g2h(addr); | ||
58 | |||
59 | - helper_retaddr = ra; | ||
60 | - o0 = ldq_le_p(haddr + 0); | ||
61 | - o1 = ldq_le_p(haddr + 1); | ||
62 | - oldv = int128_make128(o0, o1); | ||
63 | + helper_retaddr = ra; | ||
64 | + o0 = ldq_le_p(haddr + 0); | ||
65 | + o1 = ldq_le_p(haddr + 1); | ||
66 | + oldv = int128_make128(o0, o1); | ||
67 | |||
68 | - success = int128_eq(oldv, cmpv); | ||
69 | - if (success) { | ||
70 | - stq_le_p(haddr + 0, int128_getlo(newv)); | ||
71 | - stq_le_p(haddr + 1, int128_gethi(newv)); | ||
72 | - } | ||
73 | - helper_retaddr = 0; | ||
74 | -#else | ||
75 | - int mem_idx = cpu_mmu_index(env, false); | ||
76 | - TCGMemOpIdx oi0 = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
77 | - TCGMemOpIdx oi1 = make_memop_idx(MO_LEQ, mem_idx); | ||
78 | - | ||
79 | - o0 = helper_le_ldq_mmu(env, addr + 0, oi0, ra); | ||
80 | - o1 = helper_le_ldq_mmu(env, addr + 8, oi1, ra); | ||
81 | - oldv = int128_make128(o0, o1); | ||
82 | - | ||
83 | - success = int128_eq(oldv, cmpv); | ||
84 | - if (success) { | ||
85 | - helper_le_stq_mmu(env, addr + 0, int128_getlo(newv), oi1, ra); | ||
86 | - helper_le_stq_mmu(env, addr + 8, int128_gethi(newv), oi1, ra); | ||
87 | - } | ||
88 | -#endif | ||
89 | + success = int128_eq(oldv, cmpv); | ||
90 | + if (success) { | ||
91 | + stq_le_p(haddr + 0, int128_getlo(newv)); | ||
92 | + stq_le_p(haddr + 1, int128_gethi(newv)); | ||
93 | } | ||
94 | + helper_retaddr = 0; | ||
95 | +#else | ||
96 | + int mem_idx = cpu_mmu_index(env, false); | ||
97 | + TCGMemOpIdx oi0 = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
98 | + TCGMemOpIdx oi1 = make_memop_idx(MO_LEQ, mem_idx); | ||
99 | + | 47 | + |
100 | + o0 = helper_le_ldq_mmu(env, addr + 0, oi0, ra); | 48 | #define TCG_TARGET_NB_REGS 64 |
101 | + o1 = helper_le_ldq_mmu(env, addr + 8, oi1, ra); | 49 | |
102 | + oldv = int128_make128(o0, o1); | 50 | #endif /* AARCH64_TCG_TARGET_H */ |
51 | diff --git a/tcg/loongarch64/tcg-target.h b/tcg/loongarch64/tcg-target.h | ||
52 | index XXXXXXX..XXXXXXX 100644 | ||
53 | --- a/tcg/loongarch64/tcg-target.h | ||
54 | +++ b/tcg/loongarch64/tcg-target.h | ||
55 | @@ -XXX,XX +XXX,XX @@ typedef enum { | ||
56 | TCG_VEC_TMP0 = TCG_REG_V23, | ||
57 | } TCGReg; | ||
58 | |||
59 | +#define TCG_REG_ZERO TCG_REG_ZERO | ||
103 | + | 60 | + |
104 | + success = int128_eq(oldv, cmpv); | 61 | #endif /* LOONGARCH_TCG_TARGET_H */ |
105 | + if (success) { | 62 | diff --git a/tcg/mips/tcg-target.h b/tcg/mips/tcg-target.h |
106 | + helper_le_stq_mmu(env, addr + 0, int128_getlo(newv), oi1, ra); | 63 | index XXXXXXX..XXXXXXX 100644 |
107 | + helper_le_stq_mmu(env, addr + 8, int128_gethi(newv), oi1, ra); | 64 | --- a/tcg/mips/tcg-target.h |
108 | + } | 65 | +++ b/tcg/mips/tcg-target.h |
66 | @@ -XXX,XX +XXX,XX @@ typedef enum { | ||
67 | TCG_AREG0 = TCG_REG_S8, | ||
68 | } TCGReg; | ||
69 | |||
70 | +#define TCG_REG_ZERO TCG_REG_ZERO | ||
71 | + | ||
72 | #endif | ||
73 | diff --git a/tcg/riscv/tcg-target.h b/tcg/riscv/tcg-target.h | ||
74 | index XXXXXXX..XXXXXXX 100644 | ||
75 | --- a/tcg/riscv/tcg-target.h | ||
76 | +++ b/tcg/riscv/tcg-target.h | ||
77 | @@ -XXX,XX +XXX,XX @@ typedef enum { | ||
78 | TCG_REG_TMP2 = TCG_REG_T4, | ||
79 | } TCGReg; | ||
80 | |||
81 | +#define TCG_REG_ZERO TCG_REG_ZERO | ||
82 | + | ||
83 | #endif | ||
84 | diff --git a/tcg/sparc64/tcg-target.h b/tcg/sparc64/tcg-target.h | ||
85 | index XXXXXXX..XXXXXXX 100644 | ||
86 | --- a/tcg/sparc64/tcg-target.h | ||
87 | +++ b/tcg/sparc64/tcg-target.h | ||
88 | @@ -XXX,XX +XXX,XX @@ typedef enum { | ||
89 | TCG_REG_I7, | ||
90 | } TCGReg; | ||
91 | |||
92 | -#define TCG_AREG0 TCG_REG_I0 | ||
93 | +#define TCG_AREG0 TCG_REG_I0 | ||
94 | +#define TCG_REG_ZERO TCG_REG_G0 | ||
95 | |||
96 | #endif | ||
97 | diff --git a/tcg/tcg.c b/tcg/tcg.c | ||
98 | index XXXXXXX..XXXXXXX 100644 | ||
99 | --- a/tcg/tcg.c | ||
100 | +++ b/tcg/tcg.c | ||
101 | @@ -XXX,XX +XXX,XX @@ static void process_constraint_sets(void) | ||
102 | case 'i': | ||
103 | args_ct[i].ct |= TCG_CT_CONST; | ||
104 | break; | ||
105 | +#ifdef TCG_REG_ZERO | ||
106 | + case 'z': | ||
107 | + args_ct[i].ct |= TCG_CT_REG_ZERO; | ||
108 | + break; | ||
109 | +#endif | 109 | +#endif |
110 | 110 | ||
111 | return !success; | 111 | /* Include all of the target-specific constraints. */ |
112 | } | 112 | |
113 | 113 | @@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op) | |
114 | -uint64_t HELPER(paired_cmpxchg64_le)(CPUARMState *env, uint64_t addr, | 114 | arg_ct = &args_ct[i]; |
115 | - uint64_t new_lo, uint64_t new_hi) | 115 | ts = arg_temp(arg); |
116 | -{ | 116 | |
117 | - return do_paired_cmpxchg64_le(env, addr, new_lo, new_hi, false, GETPC()); | 117 | - if (ts->val_type == TEMP_VAL_CONST |
118 | -} | 118 | - && tcg_target_const_match(ts->val, arg_ct->ct, ts->type, |
119 | - | 119 | - op_cond, TCGOP_VECE(op))) { |
120 | uint64_t HELPER(paired_cmpxchg64_le_parallel)(CPUARMState *env, uint64_t addr, | 120 | - /* constant is OK for instruction */ |
121 | uint64_t new_lo, uint64_t new_hi) | 121 | - const_args[i] = 1; |
122 | -{ | 122 | - new_args[i] = ts->val; |
123 | - return do_paired_cmpxchg64_le(env, addr, new_lo, new_hi, true, GETPC()); | 123 | - continue; |
124 | -} | 124 | + if (ts->val_type == TEMP_VAL_CONST) { |
125 | - | 125 | +#ifdef TCG_REG_ZERO |
126 | -static uint64_t do_paired_cmpxchg64_be(CPUARMState *env, uint64_t addr, | 126 | + if (ts->val == 0 && (arg_ct->ct & TCG_CT_REG_ZERO)) { |
127 | - uint64_t new_lo, uint64_t new_hi, | 127 | + /* Hardware zero register: indicate register via non-const. */ |
128 | - bool parallel, uintptr_t ra) | 128 | + const_args[i] = 0; |
129 | { | 129 | + new_args[i] = TCG_REG_ZERO; |
130 | Int128 oldv, cmpv, newv; | 130 | + continue; |
131 | + uintptr_t ra = GETPC(); | 131 | + } |
132 | bool success; | ||
133 | + int mem_idx; | ||
134 | + TCGMemOpIdx oi; | ||
135 | |||
136 | - /* high and low need to be switched here because this is not actually a | ||
137 | - * 128bit store but two doublewords stored consecutively | ||
138 | - */ | ||
139 | - cmpv = int128_make128(env->exclusive_high, env->exclusive_val); | ||
140 | - newv = int128_make128(new_hi, new_lo); | ||
141 | - | ||
142 | - if (parallel) { | ||
143 | -#ifndef CONFIG_ATOMIC128 | ||
144 | + if (!HAVE_CMPXCHG128) { | ||
145 | cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
146 | -#else | ||
147 | - int mem_idx = cpu_mmu_index(env, false); | ||
148 | - TCGMemOpIdx oi = make_memop_idx(MO_BEQ | MO_ALIGN_16, mem_idx); | ||
149 | - oldv = helper_atomic_cmpxchgo_be_mmu(env, addr, cmpv, newv, oi, ra); | ||
150 | - success = int128_eq(oldv, cmpv); | ||
151 | -#endif | ||
152 | - } else { | ||
153 | - uint64_t o0, o1; | ||
154 | - | ||
155 | -#ifdef CONFIG_USER_ONLY | ||
156 | - /* ??? Enforce alignment. */ | ||
157 | - uint64_t *haddr = g2h(addr); | ||
158 | - | ||
159 | - helper_retaddr = ra; | ||
160 | - o1 = ldq_be_p(haddr + 0); | ||
161 | - o0 = ldq_be_p(haddr + 1); | ||
162 | - oldv = int128_make128(o0, o1); | ||
163 | - | ||
164 | - success = int128_eq(oldv, cmpv); | ||
165 | - if (success) { | ||
166 | - stq_be_p(haddr + 0, int128_gethi(newv)); | ||
167 | - stq_be_p(haddr + 1, int128_getlo(newv)); | ||
168 | - } | ||
169 | - helper_retaddr = 0; | ||
170 | -#else | ||
171 | - int mem_idx = cpu_mmu_index(env, false); | ||
172 | - TCGMemOpIdx oi0 = make_memop_idx(MO_BEQ | MO_ALIGN_16, mem_idx); | ||
173 | - TCGMemOpIdx oi1 = make_memop_idx(MO_BEQ, mem_idx); | ||
174 | - | ||
175 | - o1 = helper_be_ldq_mmu(env, addr + 0, oi0, ra); | ||
176 | - o0 = helper_be_ldq_mmu(env, addr + 8, oi1, ra); | ||
177 | - oldv = int128_make128(o0, o1); | ||
178 | - | ||
179 | - success = int128_eq(oldv, cmpv); | ||
180 | - if (success) { | ||
181 | - helper_be_stq_mmu(env, addr + 0, int128_gethi(newv), oi1, ra); | ||
182 | - helper_be_stq_mmu(env, addr + 8, int128_getlo(newv), oi1, ra); | ||
183 | - } | ||
184 | -#endif | ||
185 | } | ||
186 | |||
187 | + mem_idx = cpu_mmu_index(env, false); | ||
188 | + oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
189 | + | ||
190 | + cmpv = int128_make128(env->exclusive_val, env->exclusive_high); | ||
191 | + newv = int128_make128(new_lo, new_hi); | ||
192 | + oldv = helper_atomic_cmpxchgo_le_mmu(env, addr, cmpv, newv, oi, ra); | ||
193 | + | ||
194 | + success = int128_eq(oldv, cmpv); | ||
195 | return !success; | ||
196 | } | ||
197 | |||
198 | uint64_t HELPER(paired_cmpxchg64_be)(CPUARMState *env, uint64_t addr, | ||
199 | uint64_t new_lo, uint64_t new_hi) | ||
200 | { | ||
201 | - return do_paired_cmpxchg64_be(env, addr, new_lo, new_hi, false, GETPC()); | ||
202 | + /* | ||
203 | + * High and low need to be switched here because this is not actually a | ||
204 | + * 128bit store but two doublewords stored consecutively | ||
205 | + */ | ||
206 | + Int128 cmpv = int128_make128(env->exclusive_val, env->exclusive_high); | ||
207 | + Int128 newv = int128_make128(new_lo, new_hi); | ||
208 | + Int128 oldv; | ||
209 | + uintptr_t ra = GETPC(); | ||
210 | + uint64_t o0, o1; | ||
211 | + bool success; | ||
212 | + | ||
213 | +#ifdef CONFIG_USER_ONLY | ||
214 | + /* ??? Enforce alignment. */ | ||
215 | + uint64_t *haddr = g2h(addr); | ||
216 | + | ||
217 | + helper_retaddr = ra; | ||
218 | + o1 = ldq_be_p(haddr + 0); | ||
219 | + o0 = ldq_be_p(haddr + 1); | ||
220 | + oldv = int128_make128(o0, o1); | ||
221 | + | ||
222 | + success = int128_eq(oldv, cmpv); | ||
223 | + if (success) { | ||
224 | + stq_be_p(haddr + 0, int128_gethi(newv)); | ||
225 | + stq_be_p(haddr + 1, int128_getlo(newv)); | ||
226 | + } | ||
227 | + helper_retaddr = 0; | ||
228 | +#else | ||
229 | + int mem_idx = cpu_mmu_index(env, false); | ||
230 | + TCGMemOpIdx oi0 = make_memop_idx(MO_BEQ | MO_ALIGN_16, mem_idx); | ||
231 | + TCGMemOpIdx oi1 = make_memop_idx(MO_BEQ, mem_idx); | ||
232 | + | ||
233 | + o1 = helper_be_ldq_mmu(env, addr + 0, oi0, ra); | ||
234 | + o0 = helper_be_ldq_mmu(env, addr + 8, oi1, ra); | ||
235 | + oldv = int128_make128(o0, o1); | ||
236 | + | ||
237 | + success = int128_eq(oldv, cmpv); | ||
238 | + if (success) { | ||
239 | + helper_be_stq_mmu(env, addr + 0, int128_gethi(newv), oi1, ra); | ||
240 | + helper_be_stq_mmu(env, addr + 8, int128_getlo(newv), oi1, ra); | ||
241 | + } | ||
242 | +#endif | 132 | +#endif |
243 | + | 133 | + |
244 | + return !success; | 134 | + if (tcg_target_const_match(ts->val, arg_ct->ct, ts->type, |
245 | } | 135 | + op_cond, TCGOP_VECE(op))) { |
246 | 136 | + /* constant is OK for instruction */ | |
247 | uint64_t HELPER(paired_cmpxchg64_be_parallel)(CPUARMState *env, uint64_t addr, | 137 | + const_args[i] = 1; |
248 | - uint64_t new_lo, uint64_t new_hi) | 138 | + new_args[i] = ts->val; |
249 | + uint64_t new_lo, uint64_t new_hi) | 139 | + continue; |
250 | { | 140 | + } |
251 | - return do_paired_cmpxchg64_be(env, addr, new_lo, new_hi, true, GETPC()); | 141 | } |
252 | + Int128 oldv, cmpv, newv; | 142 | |
253 | + uintptr_t ra = GETPC(); | 143 | reg = ts->reg; |
254 | + bool success; | 144 | diff --git a/docs/devel/tcg-ops.rst b/docs/devel/tcg-ops.rst |
255 | + int mem_idx; | 145 | index XXXXXXX..XXXXXXX 100644 |
256 | + TCGMemOpIdx oi; | 146 | --- a/docs/devel/tcg-ops.rst |
257 | + | 147 | +++ b/docs/devel/tcg-ops.rst |
258 | + if (!HAVE_CMPXCHG128) { | 148 | @@ -XXX,XX +XXX,XX @@ operation uses a constant input constraint which does not allow all |
259 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 149 | constants, it must also accept registers in order to have a fallback. |
260 | + } | 150 | The constraint '``i``' is defined generically to accept any constant. |
261 | + | 151 | The constraint '``r``' is not defined generically, but is consistently |
262 | + mem_idx = cpu_mmu_index(env, false); | 152 | -used by each backend to indicate all registers. |
263 | + oi = make_memop_idx(MO_BEQ | MO_ALIGN_16, mem_idx); | 153 | +used by each backend to indicate all registers. If ``TCG_REG_ZERO`` |
264 | + | 154 | +is defined by the backend, the constraint '``z``' is defined generically |
265 | + /* | 155 | +to map constant 0 to the hardware zero register. |
266 | + * High and low need to be switched here because this is not actually a | 156 | |
267 | + * 128bit store but two doublewords stored consecutively | 157 | The movi_i32 and movi_i64 operations must accept any constants. |
268 | + */ | 158 | |
269 | + cmpv = int128_make128(env->exclusive_high, env->exclusive_val); | ||
270 | + newv = int128_make128(new_hi, new_lo); | ||
271 | + oldv = helper_atomic_cmpxchgo_be_mmu(env, addr, cmpv, newv, oi, ra); | ||
272 | + | ||
273 | + success = int128_eq(oldv, cmpv); | ||
274 | + return !success; | ||
275 | } | ||
276 | |||
277 | /* Writes back the old data into Rs. */ | ||
278 | void HELPER(casp_le_parallel)(CPUARMState *env, uint32_t rs, uint64_t addr, | ||
279 | uint64_t new_lo, uint64_t new_hi) | ||
280 | { | ||
281 | - uintptr_t ra = GETPC(); | ||
282 | -#ifndef CONFIG_ATOMIC128 | ||
283 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
284 | -#else | ||
285 | Int128 oldv, cmpv, newv; | ||
286 | + uintptr_t ra = GETPC(); | ||
287 | + int mem_idx; | ||
288 | + TCGMemOpIdx oi; | ||
289 | + | ||
290 | + if (!HAVE_CMPXCHG128) { | ||
291 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
292 | + } | ||
293 | + | ||
294 | + mem_idx = cpu_mmu_index(env, false); | ||
295 | + oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
296 | |||
297 | cmpv = int128_make128(env->xregs[rs], env->xregs[rs + 1]); | ||
298 | newv = int128_make128(new_lo, new_hi); | ||
299 | - | ||
300 | - int mem_idx = cpu_mmu_index(env, false); | ||
301 | - TCGMemOpIdx oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
302 | oldv = helper_atomic_cmpxchgo_le_mmu(env, addr, cmpv, newv, oi, ra); | ||
303 | |||
304 | env->xregs[rs] = int128_getlo(oldv); | ||
305 | env->xregs[rs + 1] = int128_gethi(oldv); | ||
306 | -#endif | ||
307 | } | ||
308 | |||
309 | void HELPER(casp_be_parallel)(CPUARMState *env, uint32_t rs, uint64_t addr, | ||
310 | uint64_t new_hi, uint64_t new_lo) | ||
311 | { | ||
312 | - uintptr_t ra = GETPC(); | ||
313 | -#ifndef CONFIG_ATOMIC128 | ||
314 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
315 | -#else | ||
316 | Int128 oldv, cmpv, newv; | ||
317 | + uintptr_t ra = GETPC(); | ||
318 | + int mem_idx; | ||
319 | + TCGMemOpIdx oi; | ||
320 | + | ||
321 | + if (!HAVE_CMPXCHG128) { | ||
322 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
323 | + } | ||
324 | + | ||
325 | + mem_idx = cpu_mmu_index(env, false); | ||
326 | + oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
327 | |||
328 | cmpv = int128_make128(env->xregs[rs + 1], env->xregs[rs]); | ||
329 | newv = int128_make128(new_lo, new_hi); | ||
330 | - | ||
331 | - int mem_idx = cpu_mmu_index(env, false); | ||
332 | - TCGMemOpIdx oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
333 | oldv = helper_atomic_cmpxchgo_be_mmu(env, addr, cmpv, newv, oi, ra); | ||
334 | |||
335 | env->xregs[rs + 1] = int128_getlo(oldv); | ||
336 | env->xregs[rs] = int128_gethi(oldv); | ||
337 | -#endif | ||
338 | } | ||
339 | |||
340 | /* | ||
341 | -- | 159 | -- |
342 | 2.17.2 | 160 | 2.43.0 |
343 | 161 | ||
344 | 162 | diff view generated by jsdifflib |
1 | Rather than test NOCHAIN before linking, do not emit the | 1 | Note that 'Z' is still used for addsub2. |
---|---|---|---|
2 | goto_tb opcode at all. We already do this for goto_ptr. | ||
3 | 2 | ||
4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
5 | --- | 4 | --- |
6 | accel/tcg/cpu-exec.c | 2 +- | 5 | tcg/aarch64/tcg-target-con-set.h | 12 ++++----- |
7 | tcg/tcg-op.c | 9 ++++++++- | 6 | tcg/aarch64/tcg-target.c.inc | 46 ++++++++++++++------------------ |
8 | 2 files changed, 9 insertions(+), 2 deletions(-) | 7 | 2 files changed, 26 insertions(+), 32 deletions(-) |
9 | 8 | ||
10 | diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c | 9 | diff --git a/tcg/aarch64/tcg-target-con-set.h b/tcg/aarch64/tcg-target-con-set.h |
11 | index XXXXXXX..XXXXXXX 100644 | 10 | index XXXXXXX..XXXXXXX 100644 |
12 | --- a/accel/tcg/cpu-exec.c | 11 | --- a/tcg/aarch64/tcg-target-con-set.h |
13 | +++ b/accel/tcg/cpu-exec.c | 12 | +++ b/tcg/aarch64/tcg-target-con-set.h |
14 | @@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_find(CPUState *cpu, | 13 | @@ -XXX,XX +XXX,XX @@ |
14 | */ | ||
15 | C_O0_I1(r) | ||
16 | C_O0_I2(r, rC) | ||
17 | -C_O0_I2(rZ, r) | ||
18 | +C_O0_I2(rz, r) | ||
19 | C_O0_I2(w, r) | ||
20 | -C_O0_I3(rZ, rZ, r) | ||
21 | +C_O0_I3(rz, rz, r) | ||
22 | C_O1_I1(r, r) | ||
23 | C_O1_I1(w, r) | ||
24 | C_O1_I1(w, w) | ||
25 | C_O1_I1(w, wr) | ||
26 | -C_O1_I2(r, 0, rZ) | ||
27 | +C_O1_I2(r, 0, rz) | ||
28 | C_O1_I2(r, r, r) | ||
29 | C_O1_I2(r, r, rA) | ||
30 | C_O1_I2(r, r, rAL) | ||
31 | C_O1_I2(r, r, rC) | ||
32 | C_O1_I2(r, r, ri) | ||
33 | C_O1_I2(r, r, rL) | ||
34 | -C_O1_I2(r, rZ, rZ) | ||
35 | +C_O1_I2(r, rz, rz) | ||
36 | C_O1_I2(w, 0, w) | ||
37 | C_O1_I2(w, w, w) | ||
38 | C_O1_I2(w, w, wN) | ||
39 | C_O1_I2(w, w, wO) | ||
40 | C_O1_I2(w, w, wZ) | ||
41 | C_O1_I3(w, w, w, w) | ||
42 | -C_O1_I4(r, r, rC, rZ, rZ) | ||
43 | +C_O1_I4(r, r, rC, rz, rz) | ||
44 | C_O2_I1(r, r, r) | ||
45 | -C_O2_I4(r, r, rZ, rZ, rA, rMZ) | ||
46 | +C_O2_I4(r, r, rz, rz, rA, rMZ) | ||
47 | diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/tcg/aarch64/tcg-target.c.inc | ||
50 | +++ b/tcg/aarch64/tcg-target.c.inc | ||
51 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
52 | TCGArg a2 = args[2]; | ||
53 | int c2 = const_args[2]; | ||
54 | |||
55 | - /* Some operands are defined with "rZ" constraint, a register or | ||
56 | - the zero register. These need not actually test args[I] == 0. */ | ||
57 | -#define REG0(I) (const_args[I] ? TCG_REG_XZR : (TCGReg)args[I]) | ||
58 | - | ||
59 | switch (opc) { | ||
60 | case INDEX_op_goto_ptr: | ||
61 | tcg_out_insn(s, 3207, BR, a0); | ||
62 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
63 | |||
64 | case INDEX_op_st8_i32: | ||
65 | case INDEX_op_st8_i64: | ||
66 | - tcg_out_ldst(s, I3312_STRB, REG0(0), a1, a2, 0); | ||
67 | + tcg_out_ldst(s, I3312_STRB, a0, a1, a2, 0); | ||
68 | break; | ||
69 | case INDEX_op_st16_i32: | ||
70 | case INDEX_op_st16_i64: | ||
71 | - tcg_out_ldst(s, I3312_STRH, REG0(0), a1, a2, 1); | ||
72 | + tcg_out_ldst(s, I3312_STRH, a0, a1, a2, 1); | ||
73 | break; | ||
74 | case INDEX_op_st_i32: | ||
75 | case INDEX_op_st32_i64: | ||
76 | - tcg_out_ldst(s, I3312_STRW, REG0(0), a1, a2, 2); | ||
77 | + tcg_out_ldst(s, I3312_STRW, a0, a1, a2, 2); | ||
78 | break; | ||
79 | case INDEX_op_st_i64: | ||
80 | - tcg_out_ldst(s, I3312_STRX, REG0(0), a1, a2, 3); | ||
81 | + tcg_out_ldst(s, I3312_STRX, a0, a1, a2, 3); | ||
82 | break; | ||
83 | |||
84 | case INDEX_op_add_i32: | ||
85 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
86 | /* FALLTHRU */ | ||
87 | case INDEX_op_movcond_i64: | ||
88 | tcg_out_cmp(s, ext, args[5], a1, a2, c2); | ||
89 | - tcg_out_insn(s, 3506, CSEL, ext, a0, REG0(3), REG0(4), args[5]); | ||
90 | + tcg_out_insn(s, 3506, CSEL, ext, a0, args[3], args[4], args[5]); | ||
91 | break; | ||
92 | |||
93 | case INDEX_op_qemu_ld_i32: | ||
94 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
95 | break; | ||
96 | case INDEX_op_qemu_st_i32: | ||
97 | case INDEX_op_qemu_st_i64: | ||
98 | - tcg_out_qemu_st(s, REG0(0), a1, a2, ext); | ||
99 | + tcg_out_qemu_st(s, a0, a1, a2, ext); | ||
100 | break; | ||
101 | case INDEX_op_qemu_ld_i128: | ||
102 | tcg_out_qemu_ldst_i128(s, a0, a1, a2, args[3], true); | ||
103 | break; | ||
104 | case INDEX_op_qemu_st_i128: | ||
105 | - tcg_out_qemu_ldst_i128(s, REG0(0), REG0(1), a2, args[3], false); | ||
106 | + tcg_out_qemu_ldst_i128(s, a0, a1, a2, args[3], false); | ||
107 | break; | ||
108 | |||
109 | case INDEX_op_bswap64_i64: | ||
110 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
111 | |||
112 | case INDEX_op_deposit_i64: | ||
113 | case INDEX_op_deposit_i32: | ||
114 | - tcg_out_dep(s, ext, a0, REG0(2), args[3], args[4]); | ||
115 | + tcg_out_dep(s, ext, a0, a2, args[3], args[4]); | ||
116 | break; | ||
117 | |||
118 | case INDEX_op_extract_i64: | ||
119 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
120 | |||
121 | case INDEX_op_extract2_i64: | ||
122 | case INDEX_op_extract2_i32: | ||
123 | - tcg_out_extr(s, ext, a0, REG0(2), REG0(1), args[3]); | ||
124 | + tcg_out_extr(s, ext, a0, a2, a1, args[3]); | ||
125 | break; | ||
126 | |||
127 | case INDEX_op_add2_i32: | ||
128 | - tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, REG0(2), REG0(3), | ||
129 | + tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, a2, args[3], | ||
130 | (int32_t)args[4], args[5], const_args[4], | ||
131 | const_args[5], false); | ||
132 | break; | ||
133 | case INDEX_op_add2_i64: | ||
134 | - tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, REG0(2), REG0(3), args[4], | ||
135 | + tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, a2, args[3], args[4], | ||
136 | args[5], const_args[4], const_args[5], false); | ||
137 | break; | ||
138 | case INDEX_op_sub2_i32: | ||
139 | - tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, REG0(2), REG0(3), | ||
140 | + tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, a2, args[3], | ||
141 | (int32_t)args[4], args[5], const_args[4], | ||
142 | const_args[5], true); | ||
143 | break; | ||
144 | case INDEX_op_sub2_i64: | ||
145 | - tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, REG0(2), REG0(3), args[4], | ||
146 | + tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, a2, args[3], args[4], | ||
147 | args[5], const_args[4], const_args[5], true); | ||
148 | break; | ||
149 | |||
150 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType ext, | ||
151 | default: | ||
152 | g_assert_not_reached(); | ||
15 | } | 153 | } |
16 | #endif | 154 | - |
17 | /* See if we can patch the calling TB. */ | 155 | -#undef REG0 |
18 | - if (last_tb && !qemu_loglevel_mask(CPU_LOG_TB_NOCHAIN)) { | ||
19 | + if (last_tb) { | ||
20 | tb_add_jump(last_tb, tb_exit, tb); | ||
21 | } | ||
22 | return tb; | ||
23 | diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c | ||
24 | index XXXXXXX..XXXXXXX 100644 | ||
25 | --- a/tcg/tcg-op.c | ||
26 | +++ b/tcg/tcg-op.c | ||
27 | @@ -XXX,XX +XXX,XX @@ void tcg_gen_exit_tb(TranslationBlock *tb, unsigned idx) | ||
28 | seen this numbered exit before, via tcg_gen_goto_tb. */ | ||
29 | tcg_debug_assert(tcg_ctx->goto_tb_issue_mask & (1 << idx)); | ||
30 | #endif | ||
31 | + /* When not chaining, exit without indicating a link. */ | ||
32 | + if (qemu_loglevel_mask(CPU_LOG_TB_NOCHAIN)) { | ||
33 | + val = 0; | ||
34 | + } | ||
35 | } else { | ||
36 | /* This is an exit via the exitreq label. */ | ||
37 | tcg_debug_assert(idx == TB_EXIT_REQUESTED); | ||
38 | @@ -XXX,XX +XXX,XX @@ void tcg_gen_goto_tb(unsigned idx) | ||
39 | tcg_debug_assert((tcg_ctx->goto_tb_issue_mask & (1 << idx)) == 0); | ||
40 | tcg_ctx->goto_tb_issue_mask |= 1 << idx; | ||
41 | #endif | ||
42 | - tcg_gen_op1i(INDEX_op_goto_tb, idx); | ||
43 | + /* When not chaining, we simply fall through to the "fallback" exit. */ | ||
44 | + if (!qemu_loglevel_mask(CPU_LOG_TB_NOCHAIN)) { | ||
45 | + tcg_gen_op1i(INDEX_op_goto_tb, idx); | ||
46 | + } | ||
47 | } | 156 | } |
48 | 157 | ||
49 | void tcg_gen_lookup_and_goto_ptr(void) | 158 | static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, |
159 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
160 | case INDEX_op_st16_i64: | ||
161 | case INDEX_op_st32_i64: | ||
162 | case INDEX_op_st_i64: | ||
163 | - return C_O0_I2(rZ, r); | ||
164 | + return C_O0_I2(rz, r); | ||
165 | |||
166 | case INDEX_op_add_i32: | ||
167 | case INDEX_op_add_i64: | ||
168 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
169 | |||
170 | case INDEX_op_movcond_i32: | ||
171 | case INDEX_op_movcond_i64: | ||
172 | - return C_O1_I4(r, r, rC, rZ, rZ); | ||
173 | + return C_O1_I4(r, r, rC, rz, rz); | ||
174 | |||
175 | case INDEX_op_qemu_ld_i32: | ||
176 | case INDEX_op_qemu_ld_i64: | ||
177 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
178 | return C_O2_I1(r, r, r); | ||
179 | case INDEX_op_qemu_st_i32: | ||
180 | case INDEX_op_qemu_st_i64: | ||
181 | - return C_O0_I2(rZ, r); | ||
182 | + return C_O0_I2(rz, r); | ||
183 | case INDEX_op_qemu_st_i128: | ||
184 | - return C_O0_I3(rZ, rZ, r); | ||
185 | + return C_O0_I3(rz, rz, r); | ||
186 | |||
187 | case INDEX_op_deposit_i32: | ||
188 | case INDEX_op_deposit_i64: | ||
189 | - return C_O1_I2(r, 0, rZ); | ||
190 | + return C_O1_I2(r, 0, rz); | ||
191 | |||
192 | case INDEX_op_extract2_i32: | ||
193 | case INDEX_op_extract2_i64: | ||
194 | - return C_O1_I2(r, rZ, rZ); | ||
195 | + return C_O1_I2(r, rz, rz); | ||
196 | |||
197 | case INDEX_op_add2_i32: | ||
198 | case INDEX_op_add2_i64: | ||
199 | case INDEX_op_sub2_i32: | ||
200 | case INDEX_op_sub2_i64: | ||
201 | - return C_O2_I4(r, r, rZ, rZ, rA, rMZ); | ||
202 | + return C_O2_I4(r, r, rz, rz, rA, rMZ); | ||
203 | |||
204 | case INDEX_op_add_vec: | ||
205 | case INDEX_op_sub_vec: | ||
50 | -- | 206 | -- |
51 | 2.17.2 | 207 | 2.43.0 |
52 | |||
53 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Replace target-specific 'Z' with generic 'z'. |
---|---|---|---|
2 | 2 | ||
3 | Paves the way for the addition of a per-TLB lock. | 3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | |||
5 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
6 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
7 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
8 | Message-Id: <20181009174557.16125-4-cota@braap.org> | ||
9 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
10 | --- | 5 | --- |
11 | include/exec/exec-all.h | 8 ++++++++ | 6 | tcg/loongarch64/tcg-target-con-set.h | 15 ++++++------- |
12 | accel/tcg/cputlb.c | 4 ++++ | 7 | tcg/loongarch64/tcg-target-con-str.h | 1 - |
13 | exec.c | 1 + | 8 | tcg/loongarch64/tcg-target.c.inc | 32 ++++++++++++---------------- |
14 | 3 files changed, 13 insertions(+) | 9 | 3 files changed, 21 insertions(+), 27 deletions(-) |
15 | 10 | ||
16 | diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h | 11 | diff --git a/tcg/loongarch64/tcg-target-con-set.h b/tcg/loongarch64/tcg-target-con-set.h |
17 | index XXXXXXX..XXXXXXX 100644 | 12 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/include/exec/exec-all.h | 13 | --- a/tcg/loongarch64/tcg-target-con-set.h |
19 | +++ b/include/exec/exec-all.h | 14 | +++ b/tcg/loongarch64/tcg-target-con-set.h |
20 | @@ -XXX,XX +XXX,XX @@ void cpu_address_space_init(CPUState *cpu, int asidx, | 15 | @@ -XXX,XX +XXX,XX @@ |
21 | 16 | * tcg-target-con-str.h; the constraint combination is inclusive or. | |
22 | #if !defined(CONFIG_USER_ONLY) && defined(CONFIG_TCG) | 17 | */ |
23 | /* cputlb.c */ | 18 | C_O0_I1(r) |
24 | +/** | 19 | -C_O0_I2(rZ, r) |
25 | + * tlb_init - initialize a CPU's TLB | 20 | -C_O0_I2(rZ, rZ) |
26 | + * @cpu: CPU whose TLB should be initialized | 21 | +C_O0_I2(rz, r) |
27 | + */ | 22 | +C_O0_I2(rz, rz) |
28 | +void tlb_init(CPUState *cpu); | 23 | C_O0_I2(w, r) |
29 | /** | 24 | C_O0_I3(r, r, r) |
30 | * tlb_flush_page: | 25 | C_O1_I1(r, r) |
31 | * @cpu: CPU whose TLB should be flushed | 26 | @@ -XXX,XX +XXX,XX @@ C_O1_I2(r, r, rI) |
32 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr, | 27 | C_O1_I2(r, r, rJ) |
33 | void probe_write(CPUArchState *env, target_ulong addr, int size, int mmu_idx, | 28 | C_O1_I2(r, r, rU) |
34 | uintptr_t retaddr); | 29 | C_O1_I2(r, r, rW) |
35 | #else | 30 | -C_O1_I2(r, r, rZ) |
36 | +static inline void tlb_init(CPUState *cpu) | 31 | -C_O1_I2(r, 0, rZ) |
37 | +{ | 32 | -C_O1_I2(r, rZ, ri) |
38 | +} | 33 | -C_O1_I2(r, rZ, rJ) |
39 | static inline void tlb_flush_page(CPUState *cpu, target_ulong addr) | 34 | -C_O1_I2(r, rZ, rZ) |
40 | { | 35 | +C_O1_I2(r, 0, rz) |
41 | } | 36 | +C_O1_I2(r, rz, ri) |
42 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | 37 | +C_O1_I2(r, rz, rJ) |
38 | +C_O1_I2(r, rz, rz) | ||
39 | C_O1_I2(w, w, w) | ||
40 | C_O1_I2(w, w, wM) | ||
41 | C_O1_I2(w, w, wA) | ||
42 | C_O1_I3(w, w, w, w) | ||
43 | -C_O1_I4(r, rZ, rJ, rZ, rZ) | ||
44 | +C_O1_I4(r, rz, rJ, rz, rz) | ||
45 | C_N2_I1(r, r, r) | ||
46 | diff --git a/tcg/loongarch64/tcg-target-con-str.h b/tcg/loongarch64/tcg-target-con-str.h | ||
43 | index XXXXXXX..XXXXXXX 100644 | 47 | index XXXXXXX..XXXXXXX 100644 |
44 | --- a/accel/tcg/cputlb.c | 48 | --- a/tcg/loongarch64/tcg-target-con-str.h |
45 | +++ b/accel/tcg/cputlb.c | 49 | +++ b/tcg/loongarch64/tcg-target-con-str.h |
46 | @@ -XXX,XX +XXX,XX @@ QEMU_BUILD_BUG_ON(sizeof(target_ulong) > sizeof(run_on_cpu_data)); | 50 | @@ -XXX,XX +XXX,XX @@ REGS('w', ALL_VECTOR_REGS) |
47 | QEMU_BUILD_BUG_ON(NB_MMU_MODES > 16); | 51 | CONST('I', TCG_CT_CONST_S12) |
48 | #define ALL_MMUIDX_BITS ((1 << NB_MMU_MODES) - 1) | 52 | CONST('J', TCG_CT_CONST_S32) |
49 | 53 | CONST('U', TCG_CT_CONST_U12) | |
50 | +void tlb_init(CPUState *cpu) | 54 | -CONST('Z', TCG_CT_CONST_ZERO) |
51 | +{ | 55 | CONST('C', TCG_CT_CONST_C12) |
52 | +} | 56 | CONST('W', TCG_CT_CONST_WSZ) |
53 | + | 57 | CONST('M', TCG_CT_CONST_VCMP) |
54 | /* flush_all_helper: run fn across all cpus | 58 | diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc |
55 | * | ||
56 | * If the wait flag is set then the src cpu's helper will be queued as | ||
57 | diff --git a/exec.c b/exec.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | 59 | index XXXXXXX..XXXXXXX 100644 |
59 | --- a/exec.c | 60 | --- a/tcg/loongarch64/tcg-target.c.inc |
60 | +++ b/exec.c | 61 | +++ b/tcg/loongarch64/tcg-target.c.inc |
61 | @@ -XXX,XX +XXX,XX @@ void cpu_exec_realizefn(CPUState *cpu, Error **errp) | 62 | @@ -XXX,XX +XXX,XX @@ static TCGReg tcg_target_call_oarg_reg(TCGCallReturnKind kind, int slot) |
62 | tcg_target_initialized = true; | 63 | |
63 | cc->tcg_initialize(); | 64 | #define TCG_GUEST_BASE_REG TCG_REG_S1 |
65 | |||
66 | -#define TCG_CT_CONST_ZERO 0x100 | ||
67 | -#define TCG_CT_CONST_S12 0x200 | ||
68 | -#define TCG_CT_CONST_S32 0x400 | ||
69 | -#define TCG_CT_CONST_U12 0x800 | ||
70 | -#define TCG_CT_CONST_C12 0x1000 | ||
71 | -#define TCG_CT_CONST_WSZ 0x2000 | ||
72 | -#define TCG_CT_CONST_VCMP 0x4000 | ||
73 | -#define TCG_CT_CONST_VADD 0x8000 | ||
74 | +#define TCG_CT_CONST_S12 0x100 | ||
75 | +#define TCG_CT_CONST_S32 0x200 | ||
76 | +#define TCG_CT_CONST_U12 0x400 | ||
77 | +#define TCG_CT_CONST_C12 0x800 | ||
78 | +#define TCG_CT_CONST_WSZ 0x1000 | ||
79 | +#define TCG_CT_CONST_VCMP 0x2000 | ||
80 | +#define TCG_CT_CONST_VADD 0x4000 | ||
81 | |||
82 | #define ALL_GENERAL_REGS MAKE_64BIT_MASK(0, 32) | ||
83 | #define ALL_VECTOR_REGS MAKE_64BIT_MASK(32, 32) | ||
84 | @@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, int ct, | ||
85 | if (ct & TCG_CT_CONST) { | ||
86 | return true; | ||
64 | } | 87 | } |
65 | + tlb_init(cpu); | 88 | - if ((ct & TCG_CT_CONST_ZERO) && val == 0) { |
66 | 89 | - return true; | |
67 | #ifndef CONFIG_USER_ONLY | 90 | - } |
68 | if (qdev_get_vmsd(DEVICE(cpu)) == NULL) { | 91 | if ((ct & TCG_CT_CONST_S12) && val == sextreg(val, 0, 12)) { |
92 | return true; | ||
93 | } | ||
94 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
95 | case INDEX_op_st_i64: | ||
96 | case INDEX_op_qemu_st_i32: | ||
97 | case INDEX_op_qemu_st_i64: | ||
98 | - return C_O0_I2(rZ, r); | ||
99 | + return C_O0_I2(rz, r); | ||
100 | |||
101 | case INDEX_op_qemu_ld_i128: | ||
102 | return C_N2_I1(r, r, r); | ||
103 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
104 | |||
105 | case INDEX_op_brcond_i32: | ||
106 | case INDEX_op_brcond_i64: | ||
107 | - return C_O0_I2(rZ, rZ); | ||
108 | + return C_O0_I2(rz, rz); | ||
109 | |||
110 | case INDEX_op_ext8s_i32: | ||
111 | case INDEX_op_ext8s_i64: | ||
112 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
113 | case INDEX_op_deposit_i32: | ||
114 | case INDEX_op_deposit_i64: | ||
115 | /* Must deposit into the same register as input */ | ||
116 | - return C_O1_I2(r, 0, rZ); | ||
117 | + return C_O1_I2(r, 0, rz); | ||
118 | |||
119 | case INDEX_op_sub_i32: | ||
120 | case INDEX_op_setcond_i32: | ||
121 | - return C_O1_I2(r, rZ, ri); | ||
122 | + return C_O1_I2(r, rz, ri); | ||
123 | case INDEX_op_sub_i64: | ||
124 | case INDEX_op_setcond_i64: | ||
125 | - return C_O1_I2(r, rZ, rJ); | ||
126 | + return C_O1_I2(r, rz, rJ); | ||
127 | |||
128 | case INDEX_op_mul_i32: | ||
129 | case INDEX_op_mul_i64: | ||
130 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
131 | case INDEX_op_rem_i64: | ||
132 | case INDEX_op_remu_i32: | ||
133 | case INDEX_op_remu_i64: | ||
134 | - return C_O1_I2(r, rZ, rZ); | ||
135 | + return C_O1_I2(r, rz, rz); | ||
136 | |||
137 | case INDEX_op_movcond_i32: | ||
138 | case INDEX_op_movcond_i64: | ||
139 | - return C_O1_I4(r, rZ, rJ, rZ, rZ); | ||
140 | + return C_O1_I4(r, rz, rJ, rz, rz); | ||
141 | |||
142 | case INDEX_op_ld_vec: | ||
143 | case INDEX_op_dupm_vec: | ||
69 | -- | 144 | -- |
70 | 2.17.2 | 145 | 2.43.0 |
71 | 146 | ||
72 | 147 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Replace target-specific 'Z' with generic 'z'. |
---|---|---|---|
2 | 2 | ||
3 | As far as I can tell tlb_flush does not need to be called | 3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | this early. tlb_flush is eventually called after the CPU | ||
5 | has been realized. | ||
6 | |||
7 | This change paves the way to the introduction of tlb_init, | ||
8 | which will be called from cpu_exec_realizefn. | ||
9 | |||
10 | Cc: Guan Xuetao <gxt@mprc.pku.edu.cn> | ||
11 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | ||
12 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
13 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
14 | Message-Id: <20181009174557.16125-3-cota@braap.org> | ||
15 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
16 | --- | 5 | --- |
17 | target/unicore32/cpu.c | 2 -- | 6 | tcg/mips/tcg-target-con-set.h | 26 ++++++++++----------- |
18 | 1 file changed, 2 deletions(-) | 7 | tcg/mips/tcg-target-con-str.h | 1 - |
8 | tcg/mips/tcg-target.c.inc | 44 ++++++++++++++--------------------- | ||
9 | 3 files changed, 31 insertions(+), 40 deletions(-) | ||
19 | 10 | ||
20 | diff --git a/target/unicore32/cpu.c b/target/unicore32/cpu.c | 11 | diff --git a/tcg/mips/tcg-target-con-set.h b/tcg/mips/tcg-target-con-set.h |
21 | index XXXXXXX..XXXXXXX 100644 | 12 | index XXXXXXX..XXXXXXX 100644 |
22 | --- a/target/unicore32/cpu.c | 13 | --- a/tcg/mips/tcg-target-con-set.h |
23 | +++ b/target/unicore32/cpu.c | 14 | +++ b/tcg/mips/tcg-target-con-set.h |
24 | @@ -XXX,XX +XXX,XX @@ static void uc32_cpu_initfn(Object *obj) | 15 | @@ -XXX,XX +XXX,XX @@ |
25 | env->uncached_asr = ASR_MODE_PRIV; | 16 | * tcg-target-con-str.h; the constraint combination is inclusive or. |
26 | env->regs[31] = 0x03000000; | 17 | */ |
27 | #endif | 18 | C_O0_I1(r) |
28 | - | 19 | -C_O0_I2(rZ, r) |
29 | - tlb_flush(cs); | 20 | -C_O0_I2(rZ, rZ) |
21 | -C_O0_I3(rZ, r, r) | ||
22 | -C_O0_I3(rZ, rZ, r) | ||
23 | -C_O0_I4(rZ, rZ, rZ, rZ) | ||
24 | -C_O0_I4(rZ, rZ, r, r) | ||
25 | +C_O0_I2(rz, r) | ||
26 | +C_O0_I2(rz, rz) | ||
27 | +C_O0_I3(rz, r, r) | ||
28 | +C_O0_I3(rz, rz, r) | ||
29 | +C_O0_I4(rz, rz, rz, rz) | ||
30 | +C_O0_I4(rz, rz, r, r) | ||
31 | C_O1_I1(r, r) | ||
32 | -C_O1_I2(r, 0, rZ) | ||
33 | +C_O1_I2(r, 0, rz) | ||
34 | C_O1_I2(r, r, r) | ||
35 | C_O1_I2(r, r, ri) | ||
36 | C_O1_I2(r, r, rI) | ||
37 | C_O1_I2(r, r, rIK) | ||
38 | C_O1_I2(r, r, rJ) | ||
39 | -C_O1_I2(r, r, rWZ) | ||
40 | -C_O1_I2(r, rZ, rN) | ||
41 | -C_O1_I2(r, rZ, rZ) | ||
42 | -C_O1_I4(r, rZ, rZ, rZ, 0) | ||
43 | -C_O1_I4(r, rZ, rZ, rZ, rZ) | ||
44 | +C_O1_I2(r, r, rzW) | ||
45 | +C_O1_I2(r, rz, rN) | ||
46 | +C_O1_I2(r, rz, rz) | ||
47 | +C_O1_I4(r, rz, rz, rz, 0) | ||
48 | +C_O1_I4(r, rz, rz, rz, rz) | ||
49 | C_O2_I1(r, r, r) | ||
50 | C_O2_I2(r, r, r, r) | ||
51 | -C_O2_I4(r, r, rZ, rZ, rN, rN) | ||
52 | +C_O2_I4(r, r, rz, rz, rN, rN) | ||
53 | diff --git a/tcg/mips/tcg-target-con-str.h b/tcg/mips/tcg-target-con-str.h | ||
54 | index XXXXXXX..XXXXXXX 100644 | ||
55 | --- a/tcg/mips/tcg-target-con-str.h | ||
56 | +++ b/tcg/mips/tcg-target-con-str.h | ||
57 | @@ -XXX,XX +XXX,XX @@ CONST('J', TCG_CT_CONST_S16) | ||
58 | CONST('K', TCG_CT_CONST_P2M1) | ||
59 | CONST('N', TCG_CT_CONST_N16) | ||
60 | CONST('W', TCG_CT_CONST_WSZ) | ||
61 | -CONST('Z', TCG_CT_CONST_ZERO) | ||
62 | diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc | ||
63 | index XXXXXXX..XXXXXXX 100644 | ||
64 | --- a/tcg/mips/tcg-target.c.inc | ||
65 | +++ b/tcg/mips/tcg-target.c.inc | ||
66 | @@ -XXX,XX +XXX,XX @@ static bool patch_reloc(tcg_insn_unit *code_ptr, int type, | ||
67 | g_assert_not_reached(); | ||
30 | } | 68 | } |
31 | 69 | ||
32 | static const VMStateDescription vmstate_uc32_cpu = { | 70 | -#define TCG_CT_CONST_ZERO 0x100 |
71 | -#define TCG_CT_CONST_U16 0x200 /* Unsigned 16-bit: 0 - 0xffff. */ | ||
72 | -#define TCG_CT_CONST_S16 0x400 /* Signed 16-bit: -32768 - 32767 */ | ||
73 | -#define TCG_CT_CONST_P2M1 0x800 /* Power of 2 minus 1. */ | ||
74 | -#define TCG_CT_CONST_N16 0x1000 /* "Negatable" 16-bit: -32767 - 32767 */ | ||
75 | -#define TCG_CT_CONST_WSZ 0x2000 /* word size */ | ||
76 | +#define TCG_CT_CONST_U16 0x100 /* Unsigned 16-bit: 0 - 0xffff. */ | ||
77 | +#define TCG_CT_CONST_S16 0x200 /* Signed 16-bit: -32768 - 32767 */ | ||
78 | +#define TCG_CT_CONST_P2M1 0x400 /* Power of 2 minus 1. */ | ||
79 | +#define TCG_CT_CONST_N16 0x800 /* "Negatable" 16-bit: -32767 - 32767 */ | ||
80 | +#define TCG_CT_CONST_WSZ 0x1000 /* word size */ | ||
81 | |||
82 | #define ALL_GENERAL_REGS 0xffffffffu | ||
83 | |||
84 | @@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, int ct, | ||
85 | { | ||
86 | if (ct & TCG_CT_CONST) { | ||
87 | return 1; | ||
88 | - } else if ((ct & TCG_CT_CONST_ZERO) && val == 0) { | ||
89 | - return 1; | ||
90 | } else if ((ct & TCG_CT_CONST_U16) && val == (uint16_t)val) { | ||
91 | return 1; | ||
92 | } else if ((ct & TCG_CT_CONST_S16) && val == (int16_t)val) { | ||
93 | @@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, TCGType type, | ||
94 | TCGArg a0, a1, a2; | ||
95 | int c2; | ||
96 | |||
97 | - /* | ||
98 | - * Note that many operands use the constraint set "rZ". | ||
99 | - * We make use of the fact that 0 is the ZERO register, | ||
100 | - * and hence such cases need not check for const_args. | ||
101 | - */ | ||
102 | a0 = args[0]; | ||
103 | a1 = args[1]; | ||
104 | a2 = args[2]; | ||
105 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
106 | case INDEX_op_st16_i64: | ||
107 | case INDEX_op_st32_i64: | ||
108 | case INDEX_op_st_i64: | ||
109 | - return C_O0_I2(rZ, r); | ||
110 | + return C_O0_I2(rz, r); | ||
111 | |||
112 | case INDEX_op_add_i32: | ||
113 | case INDEX_op_add_i64: | ||
114 | return C_O1_I2(r, r, rJ); | ||
115 | case INDEX_op_sub_i32: | ||
116 | case INDEX_op_sub_i64: | ||
117 | - return C_O1_I2(r, rZ, rN); | ||
118 | + return C_O1_I2(r, rz, rN); | ||
119 | case INDEX_op_mul_i32: | ||
120 | case INDEX_op_mulsh_i32: | ||
121 | case INDEX_op_muluh_i32: | ||
122 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
123 | case INDEX_op_remu_i64: | ||
124 | case INDEX_op_nor_i64: | ||
125 | case INDEX_op_setcond_i64: | ||
126 | - return C_O1_I2(r, rZ, rZ); | ||
127 | + return C_O1_I2(r, rz, rz); | ||
128 | case INDEX_op_muls2_i32: | ||
129 | case INDEX_op_mulu2_i32: | ||
130 | case INDEX_op_muls2_i64: | ||
131 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
132 | return C_O1_I2(r, r, ri); | ||
133 | case INDEX_op_clz_i32: | ||
134 | case INDEX_op_clz_i64: | ||
135 | - return C_O1_I2(r, r, rWZ); | ||
136 | + return C_O1_I2(r, r, rzW); | ||
137 | |||
138 | case INDEX_op_deposit_i32: | ||
139 | case INDEX_op_deposit_i64: | ||
140 | - return C_O1_I2(r, 0, rZ); | ||
141 | + return C_O1_I2(r, 0, rz); | ||
142 | case INDEX_op_brcond_i32: | ||
143 | case INDEX_op_brcond_i64: | ||
144 | - return C_O0_I2(rZ, rZ); | ||
145 | + return C_O0_I2(rz, rz); | ||
146 | case INDEX_op_movcond_i32: | ||
147 | case INDEX_op_movcond_i64: | ||
148 | return (use_mips32r6_instructions | ||
149 | - ? C_O1_I4(r, rZ, rZ, rZ, rZ) | ||
150 | - : C_O1_I4(r, rZ, rZ, rZ, 0)); | ||
151 | + ? C_O1_I4(r, rz, rz, rz, rz) | ||
152 | + : C_O1_I4(r, rz, rz, rz, 0)); | ||
153 | case INDEX_op_add2_i32: | ||
154 | case INDEX_op_sub2_i32: | ||
155 | - return C_O2_I4(r, r, rZ, rZ, rN, rN); | ||
156 | + return C_O2_I4(r, r, rz, rz, rN, rN); | ||
157 | case INDEX_op_setcond2_i32: | ||
158 | - return C_O1_I4(r, rZ, rZ, rZ, rZ); | ||
159 | + return C_O1_I4(r, rz, rz, rz, rz); | ||
160 | case INDEX_op_brcond2_i32: | ||
161 | - return C_O0_I4(rZ, rZ, rZ, rZ); | ||
162 | + return C_O0_I4(rz, rz, rz, rz); | ||
163 | |||
164 | case INDEX_op_qemu_ld_i32: | ||
165 | return C_O1_I1(r, r); | ||
166 | case INDEX_op_qemu_st_i32: | ||
167 | - return C_O0_I2(rZ, r); | ||
168 | + return C_O0_I2(rz, r); | ||
169 | case INDEX_op_qemu_ld_i64: | ||
170 | return TCG_TARGET_REG_BITS == 64 ? C_O1_I1(r, r) : C_O2_I1(r, r, r); | ||
171 | case INDEX_op_qemu_st_i64: | ||
172 | - return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(rZ, r) : C_O0_I3(rZ, rZ, r); | ||
173 | + return TCG_TARGET_REG_BITS == 64 ? C_O0_I2(rz, r) : C_O0_I3(rz, rz, r); | ||
174 | |||
175 | default: | ||
176 | return C_NotImplemented; | ||
33 | -- | 177 | -- |
34 | 2.17.2 | 178 | 2.43.0 |
35 | 179 | ||
36 | 180 | diff view generated by jsdifflib |
1 | From: "Emilio G. Cota" <cota@braap.org> | 1 | Replace target-specific 'Z' with generic 'z'. |
---|---|---|---|
2 | 2 | ||
3 | Consistently access u16.high with atomics to avoid | 3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> |
4 | undefined behaviour in MTTCG. | ||
5 | |||
6 | Note that icount_decr.u16.low is only used in icount mode, | ||
7 | so regular accesses to it are OK. | ||
8 | |||
9 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
10 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
11 | Message-Id: <20181010144853.13005-2-cota@braap.org> | ||
12 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
13 | --- | 5 | --- |
14 | accel/tcg/tcg-all.c | 2 +- | 6 | tcg/riscv/tcg-target-con-set.h | 10 +++++----- |
15 | accel/tcg/translate-all.c | 2 +- | 7 | tcg/riscv/tcg-target-con-str.h | 1 - |
16 | qom/cpu.c | 2 +- | 8 | tcg/riscv/tcg-target.c.inc | 28 ++++++++++++---------------- |
17 | 3 files changed, 3 insertions(+), 3 deletions(-) | 9 | 3 files changed, 17 insertions(+), 22 deletions(-) |
18 | 10 | ||
19 | diff --git a/accel/tcg/tcg-all.c b/accel/tcg/tcg-all.c | 11 | diff --git a/tcg/riscv/tcg-target-con-set.h b/tcg/riscv/tcg-target-con-set.h |
20 | index XXXXXXX..XXXXXXX 100644 | 12 | index XXXXXXX..XXXXXXX 100644 |
21 | --- a/accel/tcg/tcg-all.c | 13 | --- a/tcg/riscv/tcg-target-con-set.h |
22 | +++ b/accel/tcg/tcg-all.c | 14 | +++ b/tcg/riscv/tcg-target-con-set.h |
23 | @@ -XXX,XX +XXX,XX @@ static void tcg_handle_interrupt(CPUState *cpu, int mask) | 15 | @@ -XXX,XX +XXX,XX @@ |
24 | if (!qemu_cpu_is_self(cpu)) { | 16 | * tcg-target-con-str.h; the constraint combination is inclusive or. |
25 | qemu_cpu_kick(cpu); | 17 | */ |
26 | } else { | 18 | C_O0_I1(r) |
27 | - cpu->icount_decr.u16.high = -1; | 19 | -C_O0_I2(rZ, r) |
28 | + atomic_set(&cpu->icount_decr.u16.high, -1); | 20 | -C_O0_I2(rZ, rZ) |
29 | if (use_icount && | 21 | +C_O0_I2(rz, r) |
30 | !cpu->can_do_io | 22 | +C_O0_I2(rz, rz) |
31 | && (mask & ~old_mask) != 0) { | 23 | C_O1_I1(r, r) |
32 | diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c | 24 | C_O1_I2(r, r, ri) |
25 | C_O1_I2(r, r, rI) | ||
26 | C_O1_I2(r, r, rJ) | ||
27 | -C_O1_I2(r, rZ, rN) | ||
28 | -C_O1_I2(r, rZ, rZ) | ||
29 | +C_O1_I2(r, rz, rN) | ||
30 | +C_O1_I2(r, rz, rz) | ||
31 | C_N1_I2(r, r, rM) | ||
32 | C_O1_I4(r, r, rI, rM, rM) | ||
33 | -C_O2_I4(r, r, rZ, rZ, rM, rM) | ||
34 | +C_O2_I4(r, r, rz, rz, rM, rM) | ||
35 | C_O0_I2(v, r) | ||
36 | C_O1_I1(v, r) | ||
37 | C_O1_I1(v, v) | ||
38 | diff --git a/tcg/riscv/tcg-target-con-str.h b/tcg/riscv/tcg-target-con-str.h | ||
33 | index XXXXXXX..XXXXXXX 100644 | 39 | index XXXXXXX..XXXXXXX 100644 |
34 | --- a/accel/tcg/translate-all.c | 40 | --- a/tcg/riscv/tcg-target-con-str.h |
35 | +++ b/accel/tcg/translate-all.c | 41 | +++ b/tcg/riscv/tcg-target-con-str.h |
36 | @@ -XXX,XX +XXX,XX @@ void cpu_interrupt(CPUState *cpu, int mask) | 42 | @@ -XXX,XX +XXX,XX @@ CONST('K', TCG_CT_CONST_S5) |
37 | { | 43 | CONST('L', TCG_CT_CONST_CMP_VI) |
38 | g_assert(qemu_mutex_iothread_locked()); | 44 | CONST('N', TCG_CT_CONST_N12) |
39 | cpu->interrupt_request |= mask; | 45 | CONST('M', TCG_CT_CONST_M12) |
40 | - cpu->icount_decr.u16.high = -1; | 46 | -CONST('Z', TCG_CT_CONST_ZERO) |
41 | + atomic_set(&cpu->icount_decr.u16.high, -1); | 47 | diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc |
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/tcg/riscv/tcg-target.c.inc | ||
50 | +++ b/tcg/riscv/tcg-target.c.inc | ||
51 | @@ -XXX,XX +XXX,XX @@ static TCGReg tcg_target_call_oarg_reg(TCGCallReturnKind kind, int slot) | ||
52 | return TCG_REG_A0 + slot; | ||
42 | } | 53 | } |
43 | 54 | ||
44 | /* | 55 | -#define TCG_CT_CONST_ZERO 0x100 |
45 | diff --git a/qom/cpu.c b/qom/cpu.c | 56 | -#define TCG_CT_CONST_S12 0x200 |
46 | index XXXXXXX..XXXXXXX 100644 | 57 | -#define TCG_CT_CONST_N12 0x400 |
47 | --- a/qom/cpu.c | 58 | -#define TCG_CT_CONST_M12 0x800 |
48 | +++ b/qom/cpu.c | 59 | -#define TCG_CT_CONST_J12 0x1000 |
49 | @@ -XXX,XX +XXX,XX @@ static void cpu_common_reset(CPUState *cpu) | 60 | -#define TCG_CT_CONST_S5 0x2000 |
50 | cpu->mem_io_pc = 0; | 61 | -#define TCG_CT_CONST_CMP_VI 0x4000 |
51 | cpu->mem_io_vaddr = 0; | 62 | +#define TCG_CT_CONST_S12 0x100 |
52 | cpu->icount_extra = 0; | 63 | +#define TCG_CT_CONST_N12 0x200 |
53 | - cpu->icount_decr.u32 = 0; | 64 | +#define TCG_CT_CONST_M12 0x400 |
54 | + atomic_set(&cpu->icount_decr.u32, 0); | 65 | +#define TCG_CT_CONST_J12 0x800 |
55 | cpu->can_do_io = 1; | 66 | +#define TCG_CT_CONST_S5 0x1000 |
56 | cpu->exception_index = -1; | 67 | +#define TCG_CT_CONST_CMP_VI 0x2000 |
57 | cpu->crash_occurred = false; | 68 | |
69 | #define ALL_GENERAL_REGS MAKE_64BIT_MASK(0, 32) | ||
70 | #define ALL_VECTOR_REGS MAKE_64BIT_MASK(32, 32) | ||
71 | @@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, int ct, | ||
72 | if (ct & TCG_CT_CONST) { | ||
73 | return 1; | ||
74 | } | ||
75 | - if ((ct & TCG_CT_CONST_ZERO) && val == 0) { | ||
76 | - return 1; | ||
77 | - } | ||
78 | if (type >= TCG_TYPE_V64) { | ||
79 | /* Val is replicated by VECE; extract the highest element. */ | ||
80 | val >>= (-8 << vece) & 63; | ||
81 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
82 | case INDEX_op_st16_i64: | ||
83 | case INDEX_op_st32_i64: | ||
84 | case INDEX_op_st_i64: | ||
85 | - return C_O0_I2(rZ, r); | ||
86 | + return C_O0_I2(rz, r); | ||
87 | |||
88 | case INDEX_op_add_i32: | ||
89 | case INDEX_op_and_i32: | ||
90 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
91 | |||
92 | case INDEX_op_sub_i32: | ||
93 | case INDEX_op_sub_i64: | ||
94 | - return C_O1_I2(r, rZ, rN); | ||
95 | + return C_O1_I2(r, rz, rN); | ||
96 | |||
97 | case INDEX_op_mul_i32: | ||
98 | case INDEX_op_mulsh_i32: | ||
99 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
100 | case INDEX_op_divu_i64: | ||
101 | case INDEX_op_rem_i64: | ||
102 | case INDEX_op_remu_i64: | ||
103 | - return C_O1_I2(r, rZ, rZ); | ||
104 | + return C_O1_I2(r, rz, rz); | ||
105 | |||
106 | case INDEX_op_shl_i32: | ||
107 | case INDEX_op_shr_i32: | ||
108 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
109 | |||
110 | case INDEX_op_brcond_i32: | ||
111 | case INDEX_op_brcond_i64: | ||
112 | - return C_O0_I2(rZ, rZ); | ||
113 | + return C_O0_I2(rz, rz); | ||
114 | |||
115 | case INDEX_op_movcond_i32: | ||
116 | case INDEX_op_movcond_i64: | ||
117 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | ||
118 | case INDEX_op_add2_i64: | ||
119 | case INDEX_op_sub2_i32: | ||
120 | case INDEX_op_sub2_i64: | ||
121 | - return C_O2_I4(r, r, rZ, rZ, rM, rM); | ||
122 | + return C_O2_I4(r, r, rz, rz, rM, rM); | ||
123 | |||
124 | case INDEX_op_qemu_ld_i32: | ||
125 | case INDEX_op_qemu_ld_i64: | ||
126 | return C_O1_I1(r, r); | ||
127 | case INDEX_op_qemu_st_i32: | ||
128 | case INDEX_op_qemu_st_i64: | ||
129 | - return C_O0_I2(rZ, r); | ||
130 | + return C_O0_I2(rz, r); | ||
131 | |||
132 | case INDEX_op_st_vec: | ||
133 | return C_O0_I2(v, r); | ||
58 | -- | 134 | -- |
59 | 2.17.2 | 135 | 2.43.0 |
60 | 136 | ||
61 | 137 | diff view generated by jsdifflib |
1 | Reviewed-by: David Hildenbrand <david@redhat.com> | 1 | Replace target-specific 'Z' with generic 'z'. |
---|---|---|---|
2 | |||
3 | Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> | ||
2 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 4 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
3 | --- | 5 | --- |
4 | target/s390x/mem_helper.c | 92 +++++++++++++++++---------------------- | 6 | tcg/sparc64/tcg-target-con-set.h | 12 ++++++------ |
5 | 1 file changed, 41 insertions(+), 51 deletions(-) | 7 | tcg/sparc64/tcg-target-con-str.h | 1 - |
8 | tcg/sparc64/tcg-target.c.inc | 17 +++++++---------- | ||
9 | 3 files changed, 13 insertions(+), 17 deletions(-) | ||
6 | 10 | ||
7 | diff --git a/target/s390x/mem_helper.c b/target/s390x/mem_helper.c | 11 | diff --git a/tcg/sparc64/tcg-target-con-set.h b/tcg/sparc64/tcg-target-con-set.h |
8 | index XXXXXXX..XXXXXXX 100644 | 12 | index XXXXXXX..XXXXXXX 100644 |
9 | --- a/target/s390x/mem_helper.c | 13 | --- a/tcg/sparc64/tcg-target-con-set.h |
10 | +++ b/target/s390x/mem_helper.c | 14 | +++ b/tcg/sparc64/tcg-target-con-set.h |
11 | @@ -XXX,XX +XXX,XX @@ | 15 | @@ -XXX,XX +XXX,XX @@ |
12 | #include "exec/exec-all.h" | 16 | * tcg-target-con-str.h; the constraint combination is inclusive or. |
13 | #include "exec/cpu_ldst.h" | 17 | */ |
14 | #include "qemu/int128.h" | 18 | C_O0_I1(r) |
15 | +#include "qemu/atomic128.h" | 19 | -C_O0_I2(rZ, r) |
16 | 20 | -C_O0_I2(rZ, rJ) | |
17 | #if !defined(CONFIG_USER_ONLY) | 21 | +C_O0_I2(rz, r) |
18 | #include "hw/s390x/storage-keys.h" | 22 | +C_O0_I2(rz, rJ) |
19 | @@ -XXX,XX +XXX,XX @@ static void do_cdsg(CPUS390XState *env, uint64_t addr, | 23 | C_O1_I1(r, r) |
20 | bool fail; | 24 | C_O1_I2(r, r, r) |
21 | 25 | -C_O1_I2(r, rZ, rJ) | |
22 | if (parallel) { | 26 | -C_O1_I4(r, rZ, rJ, rI, 0) |
23 | -#ifndef CONFIG_ATOMIC128 | 27 | -C_O2_I2(r, r, rZ, rJ) |
24 | +#if !HAVE_CMPXCHG128 | 28 | -C_O2_I4(r, r, rZ, rZ, rJ, rJ) |
25 | cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 29 | +C_O1_I2(r, rz, rJ) |
26 | #else | 30 | +C_O1_I4(r, rz, rJ, rI, 0) |
27 | int mem_idx = cpu_mmu_index(env, false); | 31 | +C_O2_I2(r, r, rz, rJ) |
28 | @@ -XXX,XX +XXX,XX @@ void HELPER(cdsg_parallel)(CPUS390XState *env, uint64_t addr, | 32 | +C_O2_I4(r, r, rz, rz, rJ, rJ) |
29 | static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 33 | diff --git a/tcg/sparc64/tcg-target-con-str.h b/tcg/sparc64/tcg-target-con-str.h |
30 | uint64_t a2, bool parallel) | 34 | index XXXXXXX..XXXXXXX 100644 |
31 | { | 35 | --- a/tcg/sparc64/tcg-target-con-str.h |
32 | -#if !defined(CONFIG_USER_ONLY) || defined(CONFIG_ATOMIC128) | 36 | +++ b/tcg/sparc64/tcg-target-con-str.h |
33 | uint32_t mem_idx = cpu_mmu_index(env, false); | 37 | @@ -XXX,XX +XXX,XX @@ REGS('r', ALL_GENERAL_REGS) |
34 | -#endif | 38 | */ |
35 | uintptr_t ra = GETPC(); | 39 | CONST('I', TCG_CT_CONST_S11) |
36 | uint32_t fc = extract32(env->regs[0], 0, 8); | 40 | CONST('J', TCG_CT_CONST_S13) |
37 | uint32_t sc = extract32(env->regs[0], 8, 8); | 41 | -CONST('Z', TCG_CT_CONST_ZERO) |
38 | @@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 42 | diff --git a/tcg/sparc64/tcg-target.c.inc b/tcg/sparc64/tcg-target.c.inc |
39 | probe_write(env, a2, 0, mem_idx, ra); | 43 | index XXXXXXX..XXXXXXX 100644 |
40 | #endif | 44 | --- a/tcg/sparc64/tcg-target.c.inc |
41 | 45 | +++ b/tcg/sparc64/tcg-target.c.inc | |
42 | - /* Note that the compare-and-swap is atomic, and the store is atomic, but | 46 | @@ -XXX,XX +XXX,XX @@ static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = { |
43 | - the complete operation is not. Therefore we do not need to assert serial | 47 | |
44 | - context in order to implement this. That said, restart early if we can't | 48 | #define TCG_CT_CONST_S11 0x100 |
45 | - support either operation that is supposed to be atomic. */ | 49 | #define TCG_CT_CONST_S13 0x200 |
46 | + /* | 50 | -#define TCG_CT_CONST_ZERO 0x400 |
47 | + * Note that the compare-and-swap is atomic, and the store is atomic, | 51 | |
48 | + * but the complete operation is not. Therefore we do not need to | 52 | #define ALL_GENERAL_REGS MAKE_64BIT_MASK(0, 32) |
49 | + * assert serial context in order to implement this. That said, | 53 | |
50 | + * restart early if we can't support either operation that is supposed | 54 | @@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, int ct, |
51 | + * to be atomic. | 55 | val = (int32_t)val; |
52 | + */ | ||
53 | if (parallel) { | ||
54 | - int mask = 0; | ||
55 | -#if !defined(CONFIG_ATOMIC64) | ||
56 | - mask = -8; | ||
57 | -#elif !defined(CONFIG_ATOMIC128) | ||
58 | - mask = -16; | ||
59 | + uint32_t max = 2; | ||
60 | +#ifdef CONFIG_ATOMIC64 | ||
61 | + max = 3; | ||
62 | #endif | ||
63 | - if (((4 << fc) | (1 << sc)) & mask) { | ||
64 | + if ((HAVE_CMPXCHG128 ? 0 : fc + 2 > max) || | ||
65 | + (HAVE_ATOMIC128 ? 0 : sc > max)) { | ||
66 | cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
67 | } | ||
68 | } | 56 | } |
69 | @@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 57 | |
70 | Int128 cv = int128_make128(env->regs[r3 + 1], env->regs[r3]); | 58 | - if ((ct & TCG_CT_CONST_ZERO) && val == 0) { |
71 | Int128 ov; | 59 | - return 1; |
72 | 60 | - } else if ((ct & TCG_CT_CONST_S11) && check_fit_tl(val, 11)) { | |
73 | - if (parallel) { | 61 | + if ((ct & TCG_CT_CONST_S11) && check_fit_tl(val, 11)) { |
74 | -#ifdef CONFIG_ATOMIC128 | 62 | return 1; |
75 | - TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 63 | } else if ((ct & TCG_CT_CONST_S13) && check_fit_tl(val, 13)) { |
76 | - ov = helper_atomic_cmpxchgo_be_mmu(env, a1, cv, nv, oi, ra); | 64 | return 1; |
77 | - cc = !int128_eq(ov, cv); | 65 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) |
78 | -#else | 66 | case INDEX_op_st_i64: |
79 | - /* Note that we asserted !parallel above. */ | 67 | case INDEX_op_qemu_st_i32: |
80 | - g_assert_not_reached(); | 68 | case INDEX_op_qemu_st_i64: |
81 | -#endif | 69 | - return C_O0_I2(rZ, r); |
82 | - } else { | 70 | + return C_O0_I2(rz, r); |
83 | + if (!parallel) { | 71 | |
84 | uint64_t oh = cpu_ldq_data_ra(env, a1 + 0, ra); | 72 | case INDEX_op_add_i32: |
85 | uint64_t ol = cpu_ldq_data_ra(env, a1 + 8, ra); | 73 | case INDEX_op_add_i64: |
86 | 74 | @@ -XXX,XX +XXX,XX @@ tcg_target_op_def(TCGOpcode op, TCGType type, unsigned flags) | |
87 | @@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 75 | case INDEX_op_setcond_i64: |
88 | 76 | case INDEX_op_negsetcond_i32: | |
89 | cpu_stq_data_ra(env, a1 + 0, int128_gethi(nv), ra); | 77 | case INDEX_op_negsetcond_i64: |
90 | cpu_stq_data_ra(env, a1 + 8, int128_getlo(nv), ra); | 78 | - return C_O1_I2(r, rZ, rJ); |
91 | + } else if (HAVE_CMPXCHG128) { | 79 | + return C_O1_I2(r, rz, rJ); |
92 | + TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 80 | |
93 | + ov = helper_atomic_cmpxchgo_be_mmu(env, a1, cv, nv, oi, ra); | 81 | case INDEX_op_brcond_i32: |
94 | + cc = !int128_eq(ov, cv); | 82 | case INDEX_op_brcond_i64: |
95 | + } else { | 83 | - return C_O0_I2(rZ, rJ); |
96 | + /* Note that we asserted !parallel above. */ | 84 | + return C_O0_I2(rz, rJ); |
97 | + g_assert_not_reached(); | 85 | case INDEX_op_movcond_i32: |
98 | } | 86 | case INDEX_op_movcond_i64: |
99 | 87 | - return C_O1_I4(r, rZ, rJ, rI, 0); | |
100 | env->regs[r3 + 0] = int128_gethi(ov); | 88 | + return C_O1_I4(r, rz, rJ, rI, 0); |
101 | @@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1, | 89 | case INDEX_op_add2_i32: |
102 | cpu_stq_data_ra(env, a2, svh, ra); | 90 | case INDEX_op_add2_i64: |
103 | break; | 91 | case INDEX_op_sub2_i32: |
104 | case 4: | 92 | case INDEX_op_sub2_i64: |
105 | - if (parallel) { | 93 | - return C_O2_I4(r, r, rZ, rZ, rJ, rJ); |
106 | -#ifdef CONFIG_ATOMIC128 | 94 | + return C_O2_I4(r, r, rz, rz, rJ, rJ); |
107 | + if (!parallel) { | 95 | case INDEX_op_mulu2_i32: |
108 | + cpu_stq_data_ra(env, a2 + 0, svh, ra); | 96 | case INDEX_op_muls2_i32: |
109 | + cpu_stq_data_ra(env, a2 + 8, svl, ra); | 97 | - return C_O2_I2(r, r, rZ, rJ); |
110 | + } else if (HAVE_ATOMIC128) { | 98 | + return C_O2_I2(r, r, rz, rJ); |
111 | TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | 99 | case INDEX_op_muluh_i64: |
112 | Int128 sv = int128_make128(svl, svh); | 100 | return C_O1_I2(r, r, r); |
113 | helper_atomic_sto_be_mmu(env, a2, sv, oi, ra); | ||
114 | -#else | ||
115 | + } else { | ||
116 | /* Note that we asserted !parallel above. */ | ||
117 | g_assert_not_reached(); | ||
118 | -#endif | ||
119 | - } else { | ||
120 | - cpu_stq_data_ra(env, a2 + 0, svh, ra); | ||
121 | - cpu_stq_data_ra(env, a2 + 8, svl, ra); | ||
122 | } | ||
123 | break; | ||
124 | default: | ||
125 | @@ -XXX,XX +XXX,XX @@ static uint64_t do_lpq(CPUS390XState *env, uint64_t addr, bool parallel) | ||
126 | uintptr_t ra = GETPC(); | ||
127 | uint64_t hi, lo; | ||
128 | |||
129 | - if (parallel) { | ||
130 | -#ifndef CONFIG_ATOMIC128 | ||
131 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
132 | -#else | ||
133 | + if (!parallel) { | ||
134 | + check_alignment(env, addr, 16, ra); | ||
135 | + hi = cpu_ldq_data_ra(env, addr + 0, ra); | ||
136 | + lo = cpu_ldq_data_ra(env, addr + 8, ra); | ||
137 | + } else if (HAVE_ATOMIC128) { | ||
138 | int mem_idx = cpu_mmu_index(env, false); | ||
139 | TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
140 | Int128 v = helper_atomic_ldo_be_mmu(env, addr, oi, ra); | ||
141 | hi = int128_gethi(v); | ||
142 | lo = int128_getlo(v); | ||
143 | -#endif | ||
144 | } else { | ||
145 | - check_alignment(env, addr, 16, ra); | ||
146 | - | ||
147 | - hi = cpu_ldq_data_ra(env, addr + 0, ra); | ||
148 | - lo = cpu_ldq_data_ra(env, addr + 8, ra); | ||
149 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
150 | } | ||
151 | |||
152 | env->retxl = lo; | ||
153 | @@ -XXX,XX +XXX,XX @@ static void do_stpq(CPUS390XState *env, uint64_t addr, | ||
154 | { | ||
155 | uintptr_t ra = GETPC(); | ||
156 | |||
157 | - if (parallel) { | ||
158 | -#ifndef CONFIG_ATOMIC128 | ||
159 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
160 | -#else | ||
161 | - int mem_idx = cpu_mmu_index(env, false); | ||
162 | - TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
163 | - | ||
164 | - Int128 v = int128_make128(low, high); | ||
165 | - helper_atomic_sto_be_mmu(env, addr, v, oi, ra); | ||
166 | -#endif | ||
167 | - } else { | ||
168 | + if (!parallel) { | ||
169 | check_alignment(env, addr, 16, ra); | ||
170 | - | ||
171 | cpu_stq_data_ra(env, addr + 0, high, ra); | ||
172 | cpu_stq_data_ra(env, addr + 8, low, ra); | ||
173 | + } else if (HAVE_ATOMIC128) { | ||
174 | + int mem_idx = cpu_mmu_index(env, false); | ||
175 | + TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN_16, mem_idx); | ||
176 | + Int128 v = int128_make128(low, high); | ||
177 | + helper_atomic_sto_be_mmu(env, addr, v, oi, ra); | ||
178 | + } else { | ||
179 | + cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
180 | } | ||
181 | } | ||
182 | 101 | ||
183 | -- | 102 | -- |
184 | 2.17.2 | 103 | 2.43.0 |
185 | 104 | ||
186 | 105 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Fabiano Rosas <farosas@suse.de> | ||
1 | 2 | ||
3 | When complying with the alignment requested in the ELF and unmapping | ||
4 | the excess reservation, having align_end not aligned to the guest page | ||
5 | causes the unmap to be rejected by the alignment check at | ||
6 | target_munmap and later brk adjustments hit an EEXIST. | ||
7 | |||
8 | Fix by aligning the start of region to be unmapped. | ||
9 | |||
10 | Fixes: c81d1fafa6 ("linux-user: Honor elf alignment when placing images") | ||
11 | Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1913 | ||
12 | Signed-off-by: Fabiano Rosas <farosas@suse.de> | ||
13 | [rth: Align load_end as well.] | ||
14 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
15 | Message-ID: <20250213143558.10504-1-farosas@suse.de> | ||
16 | --- | ||
17 | linux-user/elfload.c | 4 ++-- | ||
18 | 1 file changed, 2 insertions(+), 2 deletions(-) | ||
19 | |||
20 | diff --git a/linux-user/elfload.c b/linux-user/elfload.c | ||
21 | index XXXXXXX..XXXXXXX 100644 | ||
22 | --- a/linux-user/elfload.c | ||
23 | +++ b/linux-user/elfload.c | ||
24 | @@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, const ImageSource *src, | ||
25 | |||
26 | if (align_size != reserve_size) { | ||
27 | abi_ulong align_addr = ROUND_UP(load_addr, align); | ||
28 | - abi_ulong align_end = align_addr + reserve_size; | ||
29 | - abi_ulong load_end = load_addr + align_size; | ||
30 | + abi_ulong align_end = TARGET_PAGE_ALIGN(align_addr + reserve_size); | ||
31 | + abi_ulong load_end = TARGET_PAGE_ALIGN(load_addr + align_size); | ||
32 | |||
33 | if (align_addr != load_addr) { | ||
34 | target_munmap(load_addr, align_addr - load_addr); | ||
35 | -- | ||
36 | 2.43.0 | diff view generated by jsdifflib |
1 | Reviewed-by: Emilio G. Cota <cota@braap.org> | 1 | From: Andreas Schwab <schwab@suse.de> |
---|---|---|---|
2 | Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com> | 2 | |
3 | SA_RESTORER and the associated sa_restorer field of struct sigaction are | ||
4 | an obsolete feature, not expected to be used by future architectures. | ||
5 | They are also absent on RISC-V, LoongArch, Hexagon and OpenRISC, but | ||
6 | defined due to their use of generic/signal.h. This leads to corrupted | ||
7 | data and out-of-bounds accesses. | ||
8 | |||
9 | Move the definition of TARGET_SA_RESTORER out of generic/signal.h into the | ||
10 | target_signal.h files that need it. Note that m68k has the sa_restorer | ||
11 | field, but does not use it and does not define SA_RESTORER. | ||
12 | |||
13 | Reported-by: Thomas Weißschuh <thomas@t-8ch.de> | ||
14 | Signed-off-by: Andreas Schwab <schwab@suse.de> | ||
15 | Reviewed-by: Thomas Weißschuh <thomas@t-8ch.de> | ||
16 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
3 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 17 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
18 | Message-ID: <mvmed060xc9.fsf@suse.de> | ||
4 | --- | 19 | --- |
5 | target/arm/helper-a64.c | 16 ++++------------ | 20 | linux-user/aarch64/target_signal.h | 2 ++ |
6 | target/arm/translate-a64.c | 38 ++++++++++++++++++++++---------------- | 21 | linux-user/arm/target_signal.h | 2 ++ |
7 | 2 files changed, 26 insertions(+), 28 deletions(-) | 22 | linux-user/generic/signal.h | 1 - |
23 | linux-user/i386/target_signal.h | 2 ++ | ||
24 | linux-user/m68k/target_signal.h | 1 + | ||
25 | linux-user/microblaze/target_signal.h | 2 ++ | ||
26 | linux-user/ppc/target_signal.h | 2 ++ | ||
27 | linux-user/s390x/target_signal.h | 2 ++ | ||
28 | linux-user/sh4/target_signal.h | 2 ++ | ||
29 | linux-user/x86_64/target_signal.h | 2 ++ | ||
30 | linux-user/xtensa/target_signal.h | 2 ++ | ||
31 | 11 files changed, 19 insertions(+), 1 deletion(-) | ||
8 | 32 | ||
9 | diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c | 33 | diff --git a/linux-user/aarch64/target_signal.h b/linux-user/aarch64/target_signal.h |
10 | index XXXXXXX..XXXXXXX 100644 | 34 | index XXXXXXX..XXXXXXX 100644 |
11 | --- a/target/arm/helper-a64.c | 35 | --- a/linux-user/aarch64/target_signal.h |
12 | +++ b/target/arm/helper-a64.c | 36 | +++ b/linux-user/aarch64/target_signal.h |
13 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(paired_cmpxchg64_le_parallel)(CPUARMState *env, uint64_t addr, | 37 | @@ -XXX,XX +XXX,XX @@ |
14 | int mem_idx; | 38 | |
15 | TCGMemOpIdx oi; | 39 | #include "../generic/signal.h" |
16 | 40 | ||
17 | - if (!HAVE_CMPXCHG128) { | 41 | +#define TARGET_SA_RESTORER 0x04000000 |
18 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | 42 | + |
19 | - } | 43 | #define TARGET_SEGV_MTEAERR 8 /* Asynchronous ARM MTE error */ |
20 | + assert(HAVE_CMPXCHG128); | 44 | #define TARGET_SEGV_MTESERR 9 /* Synchronous ARM MTE exception */ |
21 | 45 | ||
22 | mem_idx = cpu_mmu_index(env, false); | 46 | diff --git a/linux-user/arm/target_signal.h b/linux-user/arm/target_signal.h |
23 | oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
24 | @@ -XXX,XX +XXX,XX @@ uint64_t HELPER(paired_cmpxchg64_be_parallel)(CPUARMState *env, uint64_t addr, | ||
25 | int mem_idx; | ||
26 | TCGMemOpIdx oi; | ||
27 | |||
28 | - if (!HAVE_CMPXCHG128) { | ||
29 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
30 | - } | ||
31 | + assert(HAVE_CMPXCHG128); | ||
32 | |||
33 | mem_idx = cpu_mmu_index(env, false); | ||
34 | oi = make_memop_idx(MO_BEQ | MO_ALIGN_16, mem_idx); | ||
35 | @@ -XXX,XX +XXX,XX @@ void HELPER(casp_le_parallel)(CPUARMState *env, uint32_t rs, uint64_t addr, | ||
36 | int mem_idx; | ||
37 | TCGMemOpIdx oi; | ||
38 | |||
39 | - if (!HAVE_CMPXCHG128) { | ||
40 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
41 | - } | ||
42 | + assert(HAVE_CMPXCHG128); | ||
43 | |||
44 | mem_idx = cpu_mmu_index(env, false); | ||
45 | oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
46 | @@ -XXX,XX +XXX,XX @@ void HELPER(casp_be_parallel)(CPUARMState *env, uint32_t rs, uint64_t addr, | ||
47 | int mem_idx; | ||
48 | TCGMemOpIdx oi; | ||
49 | |||
50 | - if (!HAVE_CMPXCHG128) { | ||
51 | - cpu_loop_exit_atomic(ENV_GET_CPU(env), ra); | ||
52 | - } | ||
53 | + assert(HAVE_CMPXCHG128); | ||
54 | |||
55 | mem_idx = cpu_mmu_index(env, false); | ||
56 | oi = make_memop_idx(MO_LEQ | MO_ALIGN_16, mem_idx); | ||
57 | diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c | ||
58 | index XXXXXXX..XXXXXXX 100644 | 47 | index XXXXXXX..XXXXXXX 100644 |
59 | --- a/target/arm/translate-a64.c | 48 | --- a/linux-user/arm/target_signal.h |
60 | +++ b/target/arm/translate-a64.c | 49 | +++ b/linux-user/arm/target_signal.h |
61 | @@ -XXX,XX +XXX,XX @@ | 50 | @@ -XXX,XX +XXX,XX @@ |
62 | 51 | ||
63 | #include "trace-tcg.h" | 52 | #include "../generic/signal.h" |
64 | #include "translate-a64.h" | 53 | |
65 | +#include "qemu/atomic128.h" | 54 | +#define TARGET_SA_RESTORER 0x04000000 |
66 | 55 | + | |
67 | static TCGv_i64 cpu_X[32]; | 56 | #define TARGET_ARCH_HAS_SETUP_FRAME |
68 | static TCGv_i64 cpu_pc; | 57 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 |
69 | @@ -XXX,XX +XXX,XX @@ static void gen_store_exclusive(DisasContext *s, int rd, int rt, int rt2, | 58 | |
70 | get_mem_index(s), | 59 | diff --git a/linux-user/generic/signal.h b/linux-user/generic/signal.h |
71 | MO_64 | MO_ALIGN | s->be_data); | 60 | index XXXXXXX..XXXXXXX 100644 |
72 | tcg_gen_setcond_i64(TCG_COND_NE, tmp, tmp, cpu_exclusive_val); | 61 | --- a/linux-user/generic/signal.h |
73 | - } else if (s->be_data == MO_LE) { | 62 | +++ b/linux-user/generic/signal.h |
74 | - if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 63 | @@ -XXX,XX +XXX,XX @@ |
75 | + } else if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 64 | #define TARGET_SA_RESTART 0x10000000 |
76 | + if (!HAVE_CMPXCHG128) { | 65 | #define TARGET_SA_NODEFER 0x40000000 |
77 | + gen_helper_exit_atomic(cpu_env); | 66 | #define TARGET_SA_RESETHAND 0x80000000 |
78 | + s->base.is_jmp = DISAS_NORETURN; | 67 | -#define TARGET_SA_RESTORER 0x04000000 |
79 | + } else if (s->be_data == MO_LE) { | 68 | |
80 | gen_helper_paired_cmpxchg64_le_parallel(tmp, cpu_env, | 69 | #define TARGET_SIGHUP 1 |
81 | cpu_exclusive_addr, | 70 | #define TARGET_SIGINT 2 |
82 | cpu_reg(s, rt), | 71 | diff --git a/linux-user/i386/target_signal.h b/linux-user/i386/target_signal.h |
83 | cpu_reg(s, rt2)); | 72 | index XXXXXXX..XXXXXXX 100644 |
84 | } else { | 73 | --- a/linux-user/i386/target_signal.h |
85 | - gen_helper_paired_cmpxchg64_le(tmp, cpu_env, cpu_exclusive_addr, | 74 | +++ b/linux-user/i386/target_signal.h |
86 | - cpu_reg(s, rt), cpu_reg(s, rt2)); | 75 | @@ -XXX,XX +XXX,XX @@ |
87 | - } | 76 | |
88 | - } else { | 77 | #include "../generic/signal.h" |
89 | - if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 78 | |
90 | gen_helper_paired_cmpxchg64_be_parallel(tmp, cpu_env, | 79 | +#define TARGET_SA_RESTORER 0x04000000 |
91 | cpu_exclusive_addr, | 80 | + |
92 | cpu_reg(s, rt), | 81 | #define TARGET_ARCH_HAS_SETUP_FRAME |
93 | cpu_reg(s, rt2)); | 82 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 |
94 | - } else { | 83 | |
95 | - gen_helper_paired_cmpxchg64_be(tmp, cpu_env, cpu_exclusive_addr, | 84 | diff --git a/linux-user/m68k/target_signal.h b/linux-user/m68k/target_signal.h |
96 | - cpu_reg(s, rt), cpu_reg(s, rt2)); | 85 | index XXXXXXX..XXXXXXX 100644 |
97 | } | 86 | --- a/linux-user/m68k/target_signal.h |
98 | + } else if (s->be_data == MO_LE) { | 87 | +++ b/linux-user/m68k/target_signal.h |
99 | + gen_helper_paired_cmpxchg64_le(tmp, cpu_env, cpu_exclusive_addr, | 88 | @@ -XXX,XX +XXX,XX @@ |
100 | + cpu_reg(s, rt), cpu_reg(s, rt2)); | 89 | |
101 | + } else { | 90 | #include "../generic/signal.h" |
102 | + gen_helper_paired_cmpxchg64_be(tmp, cpu_env, cpu_exclusive_addr, | 91 | |
103 | + cpu_reg(s, rt), cpu_reg(s, rt2)); | 92 | +#define TARGET_ARCH_HAS_SA_RESTORER 1 |
104 | } | 93 | #define TARGET_ARCH_HAS_SETUP_FRAME |
105 | } else { | 94 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 |
106 | tcg_gen_atomic_cmpxchg_i64(tmp, cpu_exclusive_addr, cpu_exclusive_val, | 95 | |
107 | @@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt, | 96 | diff --git a/linux-user/microblaze/target_signal.h b/linux-user/microblaze/target_signal.h |
108 | } | 97 | index XXXXXXX..XXXXXXX 100644 |
109 | tcg_temp_free_i64(cmp); | 98 | --- a/linux-user/microblaze/target_signal.h |
110 | } else if (tb_cflags(s->base.tb) & CF_PARALLEL) { | 99 | +++ b/linux-user/microblaze/target_signal.h |
111 | - TCGv_i32 tcg_rs = tcg_const_i32(rs); | 100 | @@ -XXX,XX +XXX,XX @@ |
112 | - | 101 | |
113 | - if (s->be_data == MO_LE) { | 102 | #include "../generic/signal.h" |
114 | - gen_helper_casp_le_parallel(cpu_env, tcg_rs, addr, t1, t2); | 103 | |
115 | + if (HAVE_CMPXCHG128) { | 104 | +#define TARGET_SA_RESTORER 0x04000000 |
116 | + TCGv_i32 tcg_rs = tcg_const_i32(rs); | 105 | + |
117 | + if (s->be_data == MO_LE) { | 106 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 |
118 | + gen_helper_casp_le_parallel(cpu_env, tcg_rs, addr, t1, t2); | 107 | |
119 | + } else { | 108 | #endif /* MICROBLAZE_TARGET_SIGNAL_H */ |
120 | + gen_helper_casp_be_parallel(cpu_env, tcg_rs, addr, t1, t2); | 109 | diff --git a/linux-user/ppc/target_signal.h b/linux-user/ppc/target_signal.h |
121 | + } | 110 | index XXXXXXX..XXXXXXX 100644 |
122 | + tcg_temp_free_i32(tcg_rs); | 111 | --- a/linux-user/ppc/target_signal.h |
123 | } else { | 112 | +++ b/linux-user/ppc/target_signal.h |
124 | - gen_helper_casp_be_parallel(cpu_env, tcg_rs, addr, t1, t2); | 113 | @@ -XXX,XX +XXX,XX @@ |
125 | + gen_helper_exit_atomic(cpu_env); | 114 | |
126 | + s->base.is_jmp = DISAS_NORETURN; | 115 | #include "../generic/signal.h" |
127 | } | 116 | |
128 | - tcg_temp_free_i32(tcg_rs); | 117 | +#define TARGET_SA_RESTORER 0x04000000 |
129 | } else { | 118 | + |
130 | TCGv_i64 d1 = tcg_temp_new_i64(); | 119 | #if !defined(TARGET_PPC64) |
131 | TCGv_i64 d2 = tcg_temp_new_i64(); | 120 | #define TARGET_ARCH_HAS_SETUP_FRAME |
121 | #endif | ||
122 | diff --git a/linux-user/s390x/target_signal.h b/linux-user/s390x/target_signal.h | ||
123 | index XXXXXXX..XXXXXXX 100644 | ||
124 | --- a/linux-user/s390x/target_signal.h | ||
125 | +++ b/linux-user/s390x/target_signal.h | ||
126 | @@ -XXX,XX +XXX,XX @@ | ||
127 | |||
128 | #include "../generic/signal.h" | ||
129 | |||
130 | +#define TARGET_SA_RESTORER 0x04000000 | ||
131 | + | ||
132 | #define TARGET_ARCH_HAS_SETUP_FRAME | ||
133 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 | ||
134 | |||
135 | diff --git a/linux-user/sh4/target_signal.h b/linux-user/sh4/target_signal.h | ||
136 | index XXXXXXX..XXXXXXX 100644 | ||
137 | --- a/linux-user/sh4/target_signal.h | ||
138 | +++ b/linux-user/sh4/target_signal.h | ||
139 | @@ -XXX,XX +XXX,XX @@ | ||
140 | |||
141 | #include "../generic/signal.h" | ||
142 | |||
143 | +#define TARGET_SA_RESTORER 0x04000000 | ||
144 | + | ||
145 | #define TARGET_ARCH_HAS_SETUP_FRAME | ||
146 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 | ||
147 | |||
148 | diff --git a/linux-user/x86_64/target_signal.h b/linux-user/x86_64/target_signal.h | ||
149 | index XXXXXXX..XXXXXXX 100644 | ||
150 | --- a/linux-user/x86_64/target_signal.h | ||
151 | +++ b/linux-user/x86_64/target_signal.h | ||
152 | @@ -XXX,XX +XXX,XX @@ | ||
153 | |||
154 | #include "../generic/signal.h" | ||
155 | |||
156 | +#define TARGET_SA_RESTORER 0x04000000 | ||
157 | + | ||
158 | /* For x86_64, use of SA_RESTORER is mandatory. */ | ||
159 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 0 | ||
160 | |||
161 | diff --git a/linux-user/xtensa/target_signal.h b/linux-user/xtensa/target_signal.h | ||
162 | index XXXXXXX..XXXXXXX 100644 | ||
163 | --- a/linux-user/xtensa/target_signal.h | ||
164 | +++ b/linux-user/xtensa/target_signal.h | ||
165 | @@ -XXX,XX +XXX,XX @@ | ||
166 | |||
167 | #include "../generic/signal.h" | ||
168 | |||
169 | +#define TARGET_SA_RESTORER 0x04000000 | ||
170 | + | ||
171 | #define TARGET_ARCH_HAS_SIGTRAMP_PAGE 1 | ||
172 | |||
173 | #endif | ||
132 | -- | 174 | -- |
133 | 2.17.2 | 175 | 2.43.0 |
134 | 176 | ||
135 | 177 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Mikael Szreder <git@miszr.win> | ||
1 | 2 | ||
3 | A bug was introduced in commit 0bba7572d40d which causes the fdtox | ||
4 | and fqtox instructions to incorrectly select the destination registers. | ||
5 | More information and a test program can be found in issue #2802. | ||
6 | |||
7 | Fixes: 0bba7572d40d ("target/sparc: Perform DFPREG/QFPREG in decodetree") | ||
8 | Resolves: https://gitlab.com/qemu-project/qemu/-/issues/2802 | ||
9 | Signed-off-by: Mikael Szreder <git@miszr.win> | ||
10 | Acked-by: Artyom Tarasenko <atar4qemu@gmail.com> | ||
11 | [rth: Squash patches together, since the second fixes a typo in the first.] | ||
12 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
13 | Message-ID: <20250205090333.19626-3-git@miszr.win> | ||
14 | --- | ||
15 | target/sparc/insns.decode | 12 ++++++------ | ||
16 | 1 file changed, 6 insertions(+), 6 deletions(-) | ||
17 | |||
18 | diff --git a/target/sparc/insns.decode b/target/sparc/insns.decode | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/target/sparc/insns.decode | ||
21 | +++ b/target/sparc/insns.decode | ||
22 | @@ -XXX,XX +XXX,XX @@ FdMULq 10 ..... 110100 ..... 0 0110 1110 ..... @q_d_d | ||
23 | FNHADDs 10 ..... 110100 ..... 0 0111 0001 ..... @r_r_r | ||
24 | FNHADDd 10 ..... 110100 ..... 0 0111 0010 ..... @d_d_d | ||
25 | FNsMULd 10 ..... 110100 ..... 0 0111 1001 ..... @d_r_r | ||
26 | -FsTOx 10 ..... 110100 00000 0 1000 0001 ..... @r_r2 | ||
27 | -FdTOx 10 ..... 110100 00000 0 1000 0010 ..... @r_d2 | ||
28 | -FqTOx 10 ..... 110100 00000 0 1000 0011 ..... @r_q2 | ||
29 | -FxTOs 10 ..... 110100 00000 0 1000 0100 ..... @r_r2 | ||
30 | -FxTOd 10 ..... 110100 00000 0 1000 1000 ..... @d_r2 | ||
31 | -FxTOq 10 ..... 110100 00000 0 1000 1100 ..... @q_r2 | ||
32 | +FsTOx 10 ..... 110100 00000 0 1000 0001 ..... @d_r2 | ||
33 | +FdTOx 10 ..... 110100 00000 0 1000 0010 ..... @d_d2 | ||
34 | +FqTOx 10 ..... 110100 00000 0 1000 0011 ..... @d_q2 | ||
35 | +FxTOs 10 ..... 110100 00000 0 1000 0100 ..... @r_d2 | ||
36 | +FxTOd 10 ..... 110100 00000 0 1000 1000 ..... @d_d2 | ||
37 | +FxTOq 10 ..... 110100 00000 0 1000 1100 ..... @q_d2 | ||
38 | FiTOs 10 ..... 110100 00000 0 1100 0100 ..... @r_r2 | ||
39 | FdTOs 10 ..... 110100 00000 0 1100 0110 ..... @r_d2 | ||
40 | FqTOs 10 ..... 110100 00000 0 1100 0111 ..... @r_q2 | ||
41 | -- | ||
42 | 2.43.0 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Mikael Szreder <git@miszr.win> | ||
1 | 2 | ||
3 | The gdbstub implementation for the Sparc architecture would | ||
4 | incorrectly calculate the the floating point register offset. | ||
5 | This resulted in, for example, registers f32 and f34 to point to | ||
6 | the same value. | ||
7 | |||
8 | The issue was caused by the confusion between even register numbers | ||
9 | and even register indexes. For example, the register index of f32 is 64 | ||
10 | and f34 is 65. | ||
11 | |||
12 | Fixes: 30038fd81808 ("target-sparc: Change fpr representation to doubles.") | ||
13 | Signed-off-by: Mikael Szreder <git@miszr.win> | ||
14 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
15 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | ||
16 | Message-ID: <20250214070343.11501-1-git@miszr.win> | ||
17 | --- | ||
18 | target/sparc/gdbstub.c | 18 ++++++++++++++---- | ||
19 | 1 file changed, 14 insertions(+), 4 deletions(-) | ||
20 | |||
21 | diff --git a/target/sparc/gdbstub.c b/target/sparc/gdbstub.c | ||
22 | index XXXXXXX..XXXXXXX 100644 | ||
23 | --- a/target/sparc/gdbstub.c | ||
24 | +++ b/target/sparc/gdbstub.c | ||
25 | @@ -XXX,XX +XXX,XX @@ int sparc_cpu_gdb_read_register(CPUState *cs, GByteArray *mem_buf, int n) | ||
26 | } | ||
27 | } | ||
28 | if (n < 80) { | ||
29 | - /* f32-f62 (double width, even numbers only) */ | ||
30 | - return gdb_get_reg64(mem_buf, env->fpr[(n - 32) / 2].ll); | ||
31 | + /* f32-f62 (16 double width registers, even register numbers only) | ||
32 | + * n == 64: f32 : env->fpr[16] | ||
33 | + * n == 65: f34 : env->fpr[17] | ||
34 | + * etc... | ||
35 | + * n == 79: f62 : env->fpr[31] | ||
36 | + */ | ||
37 | + return gdb_get_reg64(mem_buf, env->fpr[(n - 64) + 16].ll); | ||
38 | } | ||
39 | switch (n) { | ||
40 | case 80: | ||
41 | @@ -XXX,XX +XXX,XX @@ int sparc_cpu_gdb_write_register(CPUState *cs, uint8_t *mem_buf, int n) | ||
42 | } | ||
43 | return 4; | ||
44 | } else if (n < 80) { | ||
45 | - /* f32-f62 (double width, even numbers only) */ | ||
46 | - env->fpr[(n - 32) / 2].ll = tmp; | ||
47 | + /* f32-f62 (16 double width registers, even register numbers only) | ||
48 | + * n == 64: f32 : env->fpr[16] | ||
49 | + * n == 65: f34 : env->fpr[17] | ||
50 | + * etc... | ||
51 | + * n == 79: f62 : env->fpr[31] | ||
52 | + */ | ||
53 | + env->fpr[(n - 64) + 16].ll = tmp; | ||
54 | } else { | ||
55 | switch (n) { | ||
56 | case 80: | ||
57 | -- | ||
58 | 2.43.0 | diff view generated by jsdifflib |
1 | Isolate the computation of an index from an address into a | 1 | From: Artyom Tarasenko <atar4qemu@gmail.com> |
---|---|---|---|
2 | helper before we change that function. | ||
3 | 2 | ||
4 | Reviewed-by: Alex Bennée <alex.bennee@linaro.org> | 3 | Fake access to |
4 | PCR Performance Control Register | ||
5 | and | ||
6 | PIC Performance Instrumentation Counter. | ||
7 | |||
8 | Ignore writes in privileged mode, and return 0 on reads. | ||
9 | |||
10 | This allows booting Tribblix, MilaX and v9os under Niagara target. | ||
11 | |||
12 | Signed-off-by: Artyom Tarasenko <atar4qemu@gmail.com> | ||
13 | Reviewed-by: Richard Henderson <richard.henderson@linaro.org> | ||
5 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> | 14 | Signed-off-by: Richard Henderson <richard.henderson@linaro.org> |
6 | [ cota: convert tlb_vaddr_to_host; use atomic_read on addr_write ] | 15 | Message-ID: <20250209211248.50383-1-atar4qemu@gmail.com> |
7 | Signed-off-by: Emilio G. Cota <cota@braap.org> | ||
8 | Message-Id: <20181009175129.17888-2-cota@braap.org> | ||
9 | --- | 16 | --- |
10 | accel/tcg/softmmu_template.h | 64 +++++++++++++++++--------------- | 17 | target/sparc/translate.c | 19 +++++++++++++++++++ |
11 | include/exec/cpu_ldst.h | 19 ++++++++-- | 18 | target/sparc/insns.decode | 7 ++++++- |
12 | include/exec/cpu_ldst_template.h | 25 +++++++------ | 19 | 2 files changed, 25 insertions(+), 1 deletion(-) |
13 | accel/tcg/cputlb.c | 60 ++++++++++++++---------------- | ||
14 | 4 files changed, 90 insertions(+), 78 deletions(-) | ||
15 | 20 | ||
16 | diff --git a/accel/tcg/softmmu_template.h b/accel/tcg/softmmu_template.h | 21 | diff --git a/target/sparc/translate.c b/target/sparc/translate.c |
17 | index XXXXXXX..XXXXXXX 100644 | 22 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/accel/tcg/softmmu_template.h | 23 | --- a/target/sparc/translate.c |
19 | +++ b/accel/tcg/softmmu_template.h | 24 | +++ b/target/sparc/translate.c |
20 | @@ -XXX,XX +XXX,XX @@ static inline DATA_TYPE glue(io_read, SUFFIX)(CPUArchState *env, | 25 | @@ -XXX,XX +XXX,XX @@ static TCGv do_rd_leon3_config(DisasContext *dc, TCGv dst) |
21 | WORD_TYPE helper_le_ld_name(CPUArchState *env, target_ulong addr, | 26 | |
22 | TCGMemOpIdx oi, uintptr_t retaddr) | 27 | TRANS(RDASR17, ASR17, do_rd_special, true, a->rd, do_rd_leon3_config) |
23 | { | 28 | |
24 | - unsigned mmu_idx = get_mmuidx(oi); | 29 | +static TCGv do_rdpic(DisasContext *dc, TCGv dst) |
25 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
26 | - target_ulong tlb_addr = env->tlb_table[mmu_idx][index].ADDR_READ; | ||
27 | + uintptr_t mmu_idx = get_mmuidx(oi); | ||
28 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
29 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
30 | + target_ulong tlb_addr = entry->ADDR_READ; | ||
31 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | ||
32 | uintptr_t haddr; | ||
33 | DATA_TYPE res; | ||
34 | @@ -XXX,XX +XXX,XX @@ WORD_TYPE helper_le_ld_name(CPUArchState *env, target_ulong addr, | ||
35 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, READ_ACCESS_TYPE, | ||
36 | mmu_idx, retaddr); | ||
37 | } | ||
38 | - tlb_addr = env->tlb_table[mmu_idx][index].ADDR_READ; | ||
39 | + tlb_addr = entry->ADDR_READ; | ||
40 | } | ||
41 | |||
42 | /* Handle an IO access. */ | ||
43 | @@ -XXX,XX +XXX,XX @@ WORD_TYPE helper_le_ld_name(CPUArchState *env, target_ulong addr, | ||
44 | return res; | ||
45 | } | ||
46 | |||
47 | - haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
48 | + haddr = addr + entry->addend; | ||
49 | #if DATA_SIZE == 1 | ||
50 | res = glue(glue(ld, LSUFFIX), _p)((uint8_t *)haddr); | ||
51 | #else | ||
52 | @@ -XXX,XX +XXX,XX @@ WORD_TYPE helper_le_ld_name(CPUArchState *env, target_ulong addr, | ||
53 | WORD_TYPE helper_be_ld_name(CPUArchState *env, target_ulong addr, | ||
54 | TCGMemOpIdx oi, uintptr_t retaddr) | ||
55 | { | ||
56 | - unsigned mmu_idx = get_mmuidx(oi); | ||
57 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
58 | - target_ulong tlb_addr = env->tlb_table[mmu_idx][index].ADDR_READ; | ||
59 | + uintptr_t mmu_idx = get_mmuidx(oi); | ||
60 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
61 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
62 | + target_ulong tlb_addr = entry->ADDR_READ; | ||
63 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | ||
64 | uintptr_t haddr; | ||
65 | DATA_TYPE res; | ||
66 | @@ -XXX,XX +XXX,XX @@ WORD_TYPE helper_be_ld_name(CPUArchState *env, target_ulong addr, | ||
67 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, READ_ACCESS_TYPE, | ||
68 | mmu_idx, retaddr); | ||
69 | } | ||
70 | - tlb_addr = env->tlb_table[mmu_idx][index].ADDR_READ; | ||
71 | + tlb_addr = entry->ADDR_READ; | ||
72 | } | ||
73 | |||
74 | /* Handle an IO access. */ | ||
75 | @@ -XXX,XX +XXX,XX @@ WORD_TYPE helper_be_ld_name(CPUArchState *env, target_ulong addr, | ||
76 | return res; | ||
77 | } | ||
78 | |||
79 | - haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
80 | + haddr = addr + entry->addend; | ||
81 | res = glue(glue(ld, LSUFFIX), _be_p)((uint8_t *)haddr); | ||
82 | return res; | ||
83 | } | ||
84 | @@ -XXX,XX +XXX,XX @@ static inline void glue(io_write, SUFFIX)(CPUArchState *env, | ||
85 | void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
86 | TCGMemOpIdx oi, uintptr_t retaddr) | ||
87 | { | ||
88 | - unsigned mmu_idx = get_mmuidx(oi); | ||
89 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
90 | - target_ulong tlb_addr = env->tlb_table[mmu_idx][index].addr_write; | ||
91 | + uintptr_t mmu_idx = get_mmuidx(oi); | ||
92 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
93 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
94 | + target_ulong tlb_addr = entry->addr_write; | ||
95 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | ||
96 | uintptr_t haddr; | ||
97 | |||
98 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
99 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, MMU_DATA_STORE, | ||
100 | mmu_idx, retaddr); | ||
101 | } | ||
102 | - tlb_addr = env->tlb_table[mmu_idx][index].addr_write & ~TLB_INVALID_MASK; | ||
103 | + tlb_addr = entry->addr_write & ~TLB_INVALID_MASK; | ||
104 | } | ||
105 | |||
106 | /* Handle an IO access. */ | ||
107 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
108 | if (DATA_SIZE > 1 | ||
109 | && unlikely((addr & ~TARGET_PAGE_MASK) + DATA_SIZE - 1 | ||
110 | >= TARGET_PAGE_SIZE)) { | ||
111 | - int i, index2; | ||
112 | - target_ulong page2, tlb_addr2; | ||
113 | + int i; | ||
114 | + target_ulong page2; | ||
115 | + CPUTLBEntry *entry2; | ||
116 | do_unaligned_access: | ||
117 | /* Ensure the second page is in the TLB. Note that the first page | ||
118 | is already guaranteed to be filled, and that the second page | ||
119 | cannot evict the first. */ | ||
120 | page2 = (addr + DATA_SIZE) & TARGET_PAGE_MASK; | ||
121 | - index2 = (page2 >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
122 | - tlb_addr2 = env->tlb_table[mmu_idx][index2].addr_write; | ||
123 | - if (!tlb_hit_page(tlb_addr2, page2) | ||
124 | + entry2 = tlb_entry(env, mmu_idx, page2); | ||
125 | + if (!tlb_hit_page(entry2->addr_write, page2) | ||
126 | && !VICTIM_TLB_HIT(addr_write, page2)) { | ||
127 | tlb_fill(ENV_GET_CPU(env), page2, DATA_SIZE, MMU_DATA_STORE, | ||
128 | mmu_idx, retaddr); | ||
129 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
130 | return; | ||
131 | } | ||
132 | |||
133 | - haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
134 | + haddr = addr + entry->addend; | ||
135 | #if DATA_SIZE == 1 | ||
136 | glue(glue(st, SUFFIX), _p)((uint8_t *)haddr, val); | ||
137 | #else | ||
138 | @@ -XXX,XX +XXX,XX @@ void helper_le_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
139 | void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
140 | TCGMemOpIdx oi, uintptr_t retaddr) | ||
141 | { | ||
142 | - unsigned mmu_idx = get_mmuidx(oi); | ||
143 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
144 | - target_ulong tlb_addr = env->tlb_table[mmu_idx][index].addr_write; | ||
145 | + uintptr_t mmu_idx = get_mmuidx(oi); | ||
146 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
147 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
148 | + target_ulong tlb_addr = entry->addr_write; | ||
149 | unsigned a_bits = get_alignment_bits(get_memop(oi)); | ||
150 | uintptr_t haddr; | ||
151 | |||
152 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
153 | tlb_fill(ENV_GET_CPU(env), addr, DATA_SIZE, MMU_DATA_STORE, | ||
154 | mmu_idx, retaddr); | ||
155 | } | ||
156 | - tlb_addr = env->tlb_table[mmu_idx][index].addr_write & ~TLB_INVALID_MASK; | ||
157 | + tlb_addr = entry->addr_write & ~TLB_INVALID_MASK; | ||
158 | } | ||
159 | |||
160 | /* Handle an IO access. */ | ||
161 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
162 | if (DATA_SIZE > 1 | ||
163 | && unlikely((addr & ~TARGET_PAGE_MASK) + DATA_SIZE - 1 | ||
164 | >= TARGET_PAGE_SIZE)) { | ||
165 | - int i, index2; | ||
166 | - target_ulong page2, tlb_addr2; | ||
167 | + int i; | ||
168 | + target_ulong page2; | ||
169 | + CPUTLBEntry *entry2; | ||
170 | do_unaligned_access: | ||
171 | /* Ensure the second page is in the TLB. Note that the first page | ||
172 | is already guaranteed to be filled, and that the second page | ||
173 | cannot evict the first. */ | ||
174 | page2 = (addr + DATA_SIZE) & TARGET_PAGE_MASK; | ||
175 | - index2 = (page2 >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
176 | - tlb_addr2 = env->tlb_table[mmu_idx][index2].addr_write; | ||
177 | - if (!tlb_hit_page(tlb_addr2, page2) | ||
178 | + entry2 = tlb_entry(env, mmu_idx, page2); | ||
179 | + if (!tlb_hit_page(entry2->addr_write, page2) | ||
180 | && !VICTIM_TLB_HIT(addr_write, page2)) { | ||
181 | tlb_fill(ENV_GET_CPU(env), page2, DATA_SIZE, MMU_DATA_STORE, | ||
182 | mmu_idx, retaddr); | ||
183 | @@ -XXX,XX +XXX,XX @@ void helper_be_st_name(CPUArchState *env, target_ulong addr, DATA_TYPE val, | ||
184 | return; | ||
185 | } | ||
186 | |||
187 | - haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
188 | + haddr = addr + entry->addend; | ||
189 | glue(glue(st, SUFFIX), _be_p)((uint8_t *)haddr, val); | ||
190 | } | ||
191 | #endif /* DATA_SIZE > 1 */ | ||
192 | diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h | ||
193 | index XXXXXXX..XXXXXXX 100644 | ||
194 | --- a/include/exec/cpu_ldst.h | ||
195 | +++ b/include/exec/cpu_ldst.h | ||
196 | @@ -XXX,XX +XXX,XX @@ extern __thread uintptr_t helper_retaddr; | ||
197 | /* The memory helpers for tcg-generated code need tcg_target_long etc. */ | ||
198 | #include "tcg.h" | ||
199 | |||
200 | +/* Find the TLB index corresponding to the mmu_idx + address pair. */ | ||
201 | +static inline uintptr_t tlb_index(CPUArchState *env, uintptr_t mmu_idx, | ||
202 | + target_ulong addr) | ||
203 | +{ | 30 | +{ |
204 | + return (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | 31 | + return tcg_constant_tl(0); |
205 | +} | 32 | +} |
206 | + | 33 | + |
207 | +/* Find the TLB entry corresponding to the mmu_idx + address pair. */ | 34 | +TRANS(RDPIC, HYPV, do_rd_special, supervisor(dc), a->rd, do_rdpic) |
208 | +static inline CPUTLBEntry *tlb_entry(CPUArchState *env, uintptr_t mmu_idx, | 35 | + |
209 | + target_ulong addr) | 36 | + |
37 | static TCGv do_rdccr(DisasContext *dc, TCGv dst) | ||
38 | { | ||
39 | gen_helper_rdccr(dst, tcg_env); | ||
40 | @@ -XXX,XX +XXX,XX @@ static void do_wrfprs(DisasContext *dc, TCGv src) | ||
41 | |||
42 | TRANS(WRFPRS, 64, do_wr_special, a, true, do_wrfprs) | ||
43 | |||
44 | +static bool do_priv_nop(DisasContext *dc, bool priv) | ||
210 | +{ | 45 | +{ |
211 | + return &env->tlb_table[mmu_idx][tlb_index(env, mmu_idx, addr)]; | 46 | + if (!priv) { |
47 | + return raise_priv(dc); | ||
48 | + } | ||
49 | + return advance_pc(dc); | ||
212 | +} | 50 | +} |
213 | + | 51 | + |
214 | #ifdef MMU_MODE0_SUFFIX | 52 | +TRANS(WRPCR, HYPV, do_priv_nop, supervisor(dc)) |
215 | #define CPU_MMU_INDEX 0 | 53 | +TRANS(WRPIC, HYPV, do_priv_nop, supervisor(dc)) |
216 | #define MEMSUFFIX MMU_MODE0_SUFFIX | 54 | + |
217 | @@ -XXX,XX +XXX,XX @@ static inline void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr, | 55 | static void do_wrgsr(DisasContext *dc, TCGv src) |
218 | #if defined(CONFIG_USER_ONLY) | 56 | { |
219 | return g2h(addr); | 57 | gen_trap_ifnofpu(dc); |
220 | #else | 58 | diff --git a/target/sparc/insns.decode b/target/sparc/insns.decode |
221 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
222 | - CPUTLBEntry *tlbentry = &env->tlb_table[mmu_idx][index]; | ||
223 | + CPUTLBEntry *tlbentry = tlb_entry(env, mmu_idx, addr); | ||
224 | abi_ptr tlb_addr; | ||
225 | uintptr_t haddr; | ||
226 | |||
227 | @@ -XXX,XX +XXX,XX @@ static inline void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr, | ||
228 | return NULL; | ||
229 | } | ||
230 | |||
231 | - haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
232 | + haddr = addr + tlbentry->addend; | ||
233 | return (void *)haddr; | ||
234 | #endif /* defined(CONFIG_USER_ONLY) */ | ||
235 | } | ||
236 | diff --git a/include/exec/cpu_ldst_template.h b/include/exec/cpu_ldst_template.h | ||
237 | index XXXXXXX..XXXXXXX 100644 | 59 | index XXXXXXX..XXXXXXX 100644 |
238 | --- a/include/exec/cpu_ldst_template.h | 60 | --- a/target/sparc/insns.decode |
239 | +++ b/include/exec/cpu_ldst_template.h | 61 | +++ b/target/sparc/insns.decode |
240 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_ld, USUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | 62 | @@ -XXX,XX +XXX,XX @@ CALL 01 i:s30 |
241 | target_ulong ptr, | 63 | RDTICK 10 rd:5 101000 00100 0 0000000000000 |
242 | uintptr_t retaddr) | 64 | RDPC 10 rd:5 101000 00101 0 0000000000000 |
243 | { | 65 | RDFPRS 10 rd:5 101000 00110 0 0000000000000 |
244 | - int page_index; | 66 | - RDASR17 10 rd:5 101000 10001 0 0000000000000 |
245 | + CPUTLBEntry *entry; | 67 | + { |
246 | RES_TYPE res; | 68 | + RDASR17 10 rd:5 101000 10001 0 0000000000000 |
247 | target_ulong addr; | 69 | + RDPIC 10 rd:5 101000 10001 0 0000000000000 |
248 | int mmu_idx; | 70 | + } |
249 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_ld, USUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | 71 | RDGSR 10 rd:5 101000 10011 0 0000000000000 |
250 | #endif | 72 | RDSOFTINT 10 rd:5 101000 10110 0 0000000000000 |
251 | 73 | RDTICK_CMPR 10 rd:5 101000 10111 0 0000000000000 | |
252 | addr = ptr; | 74 | @@ -XXX,XX +XXX,XX @@ CALL 01 i:s30 |
253 | - page_index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | 75 | WRCCR 10 00010 110000 ..... . ............. @n_r_ri |
254 | mmu_idx = CPU_MMU_INDEX; | 76 | WRASI 10 00011 110000 ..... . ............. @n_r_ri |
255 | - if (unlikely(env->tlb_table[mmu_idx][page_index].ADDR_READ != | 77 | WRFPRS 10 00110 110000 ..... . ............. @n_r_ri |
256 | + entry = tlb_entry(env, mmu_idx, addr); | 78 | + WRPCR 10 10000 110000 01000 0 0000000000000 |
257 | + if (unlikely(entry->ADDR_READ != | 79 | + WRPIC 10 10001 110000 01000 0 0000000000000 |
258 | (addr & (TARGET_PAGE_MASK | (DATA_SIZE - 1))))) { | 80 | { |
259 | oi = make_memop_idx(SHIFT, mmu_idx); | 81 | WRGSR 10 10011 110000 ..... . ............. @n_r_ri |
260 | res = glue(glue(helper_ret_ld, URETSUFFIX), MMUSUFFIX)(env, addr, | 82 | WRPOWERDOWN 10 10011 110000 ..... . ............. @n_r_ri |
261 | oi, retaddr); | ||
262 | } else { | ||
263 | - uintptr_t hostaddr = addr + env->tlb_table[mmu_idx][page_index].addend; | ||
264 | + uintptr_t hostaddr = addr + entry->addend; | ||
265 | res = glue(glue(ld, USUFFIX), _p)((uint8_t *)hostaddr); | ||
266 | } | ||
267 | return res; | ||
268 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_lds, SUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | ||
269 | target_ulong ptr, | ||
270 | uintptr_t retaddr) | ||
271 | { | ||
272 | - int res, page_index; | ||
273 | + CPUTLBEntry *entry; | ||
274 | + int res; | ||
275 | target_ulong addr; | ||
276 | int mmu_idx; | ||
277 | TCGMemOpIdx oi; | ||
278 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_lds, SUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | ||
279 | #endif | ||
280 | |||
281 | addr = ptr; | ||
282 | - page_index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
283 | mmu_idx = CPU_MMU_INDEX; | ||
284 | - if (unlikely(env->tlb_table[mmu_idx][page_index].ADDR_READ != | ||
285 | + entry = tlb_entry(env, mmu_idx, addr); | ||
286 | + if (unlikely(entry->ADDR_READ != | ||
287 | (addr & (TARGET_PAGE_MASK | (DATA_SIZE - 1))))) { | ||
288 | oi = make_memop_idx(SHIFT, mmu_idx); | ||
289 | res = (DATA_STYPE)glue(glue(helper_ret_ld, SRETSUFFIX), | ||
290 | MMUSUFFIX)(env, addr, oi, retaddr); | ||
291 | } else { | ||
292 | - uintptr_t hostaddr = addr + env->tlb_table[mmu_idx][page_index].addend; | ||
293 | + uintptr_t hostaddr = addr + entry->addend; | ||
294 | res = glue(glue(lds, SUFFIX), _p)((uint8_t *)hostaddr); | ||
295 | } | ||
296 | return res; | ||
297 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_st, SUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | ||
298 | target_ulong ptr, | ||
299 | RES_TYPE v, uintptr_t retaddr) | ||
300 | { | ||
301 | - int page_index; | ||
302 | + CPUTLBEntry *entry; | ||
303 | target_ulong addr; | ||
304 | int mmu_idx; | ||
305 | TCGMemOpIdx oi; | ||
306 | @@ -XXX,XX +XXX,XX @@ glue(glue(glue(cpu_st, SUFFIX), MEMSUFFIX), _ra)(CPUArchState *env, | ||
307 | #endif | ||
308 | |||
309 | addr = ptr; | ||
310 | - page_index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
311 | mmu_idx = CPU_MMU_INDEX; | ||
312 | - if (unlikely(env->tlb_table[mmu_idx][page_index].addr_write != | ||
313 | + entry = tlb_entry(env, mmu_idx, addr); | ||
314 | + if (unlikely(entry->addr_write != | ||
315 | (addr & (TARGET_PAGE_MASK | (DATA_SIZE - 1))))) { | ||
316 | oi = make_memop_idx(SHIFT, mmu_idx); | ||
317 | glue(glue(helper_ret_st, SUFFIX), MMUSUFFIX)(env, addr, v, oi, | ||
318 | retaddr); | ||
319 | } else { | ||
320 | - uintptr_t hostaddr = addr + env->tlb_table[mmu_idx][page_index].addend; | ||
321 | + uintptr_t hostaddr = addr + entry->addend; | ||
322 | glue(glue(st, SUFFIX), _p)((uint8_t *)hostaddr, v); | ||
323 | } | ||
324 | } | ||
325 | diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c | ||
326 | index XXXXXXX..XXXXXXX 100644 | ||
327 | --- a/accel/tcg/cputlb.c | ||
328 | +++ b/accel/tcg/cputlb.c | ||
329 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_async_work(CPUState *cpu, run_on_cpu_data data) | ||
330 | { | ||
331 | CPUArchState *env = cpu->env_ptr; | ||
332 | target_ulong addr = (target_ulong) data.target_ptr; | ||
333 | - int i; | ||
334 | int mmu_idx; | ||
335 | |||
336 | assert_cpu_is_self(cpu); | ||
337 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_async_work(CPUState *cpu, run_on_cpu_data data) | ||
338 | } | ||
339 | |||
340 | addr &= TARGET_PAGE_MASK; | ||
341 | - i = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
342 | qemu_spin_lock(&env->tlb_lock); | ||
343 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
344 | - tlb_flush_entry_locked(&env->tlb_table[mmu_idx][i], addr); | ||
345 | + tlb_flush_entry_locked(tlb_entry(env, mmu_idx, addr), addr); | ||
346 | tlb_flush_vtlb_page_locked(env, mmu_idx, addr); | ||
347 | } | ||
348 | qemu_spin_unlock(&env->tlb_lock); | ||
349 | @@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_by_mmuidx_async_work(CPUState *cpu, | ||
350 | target_ulong addr_and_mmuidx = (target_ulong) data.target_ptr; | ||
351 | target_ulong addr = addr_and_mmuidx & TARGET_PAGE_MASK; | ||
352 | unsigned long mmu_idx_bitmap = addr_and_mmuidx & ALL_MMUIDX_BITS; | ||
353 | - int page = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
354 | int mmu_idx; | ||
355 | |||
356 | assert_cpu_is_self(cpu); | ||
357 | |||
358 | - tlb_debug("page:%d addr:"TARGET_FMT_lx" mmu_idx:0x%lx\n", | ||
359 | - page, addr, mmu_idx_bitmap); | ||
360 | + tlb_debug("flush page addr:"TARGET_FMT_lx" mmu_idx:0x%lx\n", | ||
361 | + addr, mmu_idx_bitmap); | ||
362 | |||
363 | qemu_spin_lock(&env->tlb_lock); | ||
364 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
365 | if (test_bit(mmu_idx, &mmu_idx_bitmap)) { | ||
366 | - tlb_flush_entry_locked(&env->tlb_table[mmu_idx][page], addr); | ||
367 | + tlb_flush_entry_locked(tlb_entry(env, mmu_idx, addr), addr); | ||
368 | tlb_flush_vtlb_page_locked(env, mmu_idx, addr); | ||
369 | } | ||
370 | } | ||
371 | @@ -XXX,XX +XXX,XX @@ static inline void tlb_set_dirty1_locked(CPUTLBEntry *tlb_entry, | ||
372 | void tlb_set_dirty(CPUState *cpu, target_ulong vaddr) | ||
373 | { | ||
374 | CPUArchState *env = cpu->env_ptr; | ||
375 | - int i; | ||
376 | int mmu_idx; | ||
377 | |||
378 | assert_cpu_is_self(cpu); | ||
379 | |||
380 | vaddr &= TARGET_PAGE_MASK; | ||
381 | - i = (vaddr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
382 | qemu_spin_lock(&env->tlb_lock); | ||
383 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
384 | - tlb_set_dirty1_locked(&env->tlb_table[mmu_idx][i], vaddr); | ||
385 | + tlb_set_dirty1_locked(tlb_entry(env, mmu_idx, vaddr), vaddr); | ||
386 | } | ||
387 | |||
388 | for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) { | ||
389 | @@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr, | ||
390 | iotlb = memory_region_section_get_iotlb(cpu, section, vaddr_page, | ||
391 | paddr_page, xlat, prot, &address); | ||
392 | |||
393 | - index = (vaddr_page >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
394 | - te = &env->tlb_table[mmu_idx][index]; | ||
395 | + index = tlb_index(env, mmu_idx, vaddr_page); | ||
396 | + te = tlb_entry(env, mmu_idx, vaddr_page); | ||
397 | |||
398 | /* | ||
399 | * Hold the TLB lock for the rest of the function. We could acquire/release | ||
400 | @@ -XXX,XX +XXX,XX @@ static uint64_t io_readx(CPUArchState *env, CPUIOTLBEntry *iotlbentry, | ||
401 | * repeat the MMU check here. This tlb_fill() call might | ||
402 | * longjump out if this access should cause a guest exception. | ||
403 | */ | ||
404 | - int index; | ||
405 | + CPUTLBEntry *entry; | ||
406 | target_ulong tlb_addr; | ||
407 | |||
408 | tlb_fill(cpu, addr, size, MMU_DATA_LOAD, mmu_idx, retaddr); | ||
409 | |||
410 | - index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
411 | - tlb_addr = env->tlb_table[mmu_idx][index].addr_read; | ||
412 | + entry = tlb_entry(env, mmu_idx, addr); | ||
413 | + tlb_addr = entry->addr_read; | ||
414 | if (!(tlb_addr & ~(TARGET_PAGE_MASK | TLB_RECHECK))) { | ||
415 | /* RAM access */ | ||
416 | - uintptr_t haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
417 | + uintptr_t haddr = addr + entry->addend; | ||
418 | |||
419 | return ldn_p((void *)haddr, size); | ||
420 | } | ||
421 | @@ -XXX,XX +XXX,XX @@ static void io_writex(CPUArchState *env, CPUIOTLBEntry *iotlbentry, | ||
422 | * repeat the MMU check here. This tlb_fill() call might | ||
423 | * longjump out if this access should cause a guest exception. | ||
424 | */ | ||
425 | - int index; | ||
426 | + CPUTLBEntry *entry; | ||
427 | target_ulong tlb_addr; | ||
428 | |||
429 | tlb_fill(cpu, addr, size, MMU_DATA_STORE, mmu_idx, retaddr); | ||
430 | |||
431 | - index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
432 | - tlb_addr = env->tlb_table[mmu_idx][index].addr_write; | ||
433 | + entry = tlb_entry(env, mmu_idx, addr); | ||
434 | + tlb_addr = entry->addr_write; | ||
435 | if (!(tlb_addr & ~(TARGET_PAGE_MASK | TLB_RECHECK))) { | ||
436 | /* RAM access */ | ||
437 | - uintptr_t haddr = addr + env->tlb_table[mmu_idx][index].addend; | ||
438 | + uintptr_t haddr = addr + entry->addend; | ||
439 | |||
440 | stn_p((void *)haddr, size, val); | ||
441 | return; | ||
442 | @@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index, | ||
443 | */ | ||
444 | tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr) | ||
445 | { | ||
446 | - int mmu_idx, index; | ||
447 | + uintptr_t mmu_idx = cpu_mmu_index(env, true); | ||
448 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
449 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
450 | void *p; | ||
451 | |||
452 | - index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
453 | - mmu_idx = cpu_mmu_index(env, true); | ||
454 | - if (unlikely(!tlb_hit(env->tlb_table[mmu_idx][index].addr_code, addr))) { | ||
455 | + if (unlikely(!tlb_hit(entry->addr_code, addr))) { | ||
456 | if (!VICTIM_TLB_HIT(addr_code, addr)) { | ||
457 | tlb_fill(ENV_GET_CPU(env), addr, 0, MMU_INST_FETCH, mmu_idx, 0); | ||
458 | } | ||
459 | - assert(tlb_hit(env->tlb_table[mmu_idx][index].addr_code, addr)); | ||
460 | + assert(tlb_hit(entry->addr_code, addr)); | ||
461 | } | ||
462 | |||
463 | - if (unlikely(env->tlb_table[mmu_idx][index].addr_code & | ||
464 | - (TLB_RECHECK | TLB_MMIO))) { | ||
465 | + if (unlikely(entry->addr_code & (TLB_RECHECK | TLB_MMIO))) { | ||
466 | /* | ||
467 | * Return -1 if we can't translate and execute from an entire | ||
468 | * page of RAM here, which will cause us to execute by loading | ||
469 | @@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr) | ||
470 | return -1; | ||
471 | } | ||
472 | |||
473 | - p = (void *)((uintptr_t)addr + env->tlb_table[mmu_idx][index].addend); | ||
474 | + p = (void *)((uintptr_t)addr + entry->addend); | ||
475 | return qemu_ram_addr_from_host_nofail(p); | ||
476 | } | ||
477 | |||
478 | @@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr) | ||
479 | void probe_write(CPUArchState *env, target_ulong addr, int size, int mmu_idx, | ||
480 | uintptr_t retaddr) | ||
481 | { | ||
482 | - int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
483 | - target_ulong tlb_addr = env->tlb_table[mmu_idx][index].addr_write; | ||
484 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
485 | + CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr); | ||
486 | |||
487 | - if (!tlb_hit(tlb_addr, addr)) { | ||
488 | + if (!tlb_hit(entry->addr_write, addr)) { | ||
489 | /* TLB entry is for a different page */ | ||
490 | if (!VICTIM_TLB_HIT(addr_write, addr)) { | ||
491 | tlb_fill(ENV_GET_CPU(env), addr, size, MMU_DATA_STORE, | ||
492 | @@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr, | ||
493 | NotDirtyInfo *ndi) | ||
494 | { | ||
495 | size_t mmu_idx = get_mmuidx(oi); | ||
496 | - size_t index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1); | ||
497 | - CPUTLBEntry *tlbe = &env->tlb_table[mmu_idx][index]; | ||
498 | + uintptr_t index = tlb_index(env, mmu_idx, addr); | ||
499 | + CPUTLBEntry *tlbe = tlb_entry(env, mmu_idx, addr); | ||
500 | target_ulong tlb_addr = tlbe->addr_write; | ||
501 | TCGMemOp mop = get_memop(oi); | ||
502 | int a_bits = get_alignment_bits(mop); | ||
503 | -- | 83 | -- |
504 | 2.17.2 | 84 | 2.43.0 |
505 | |||
506 | diff view generated by jsdifflib |