1
The following changes since commit 13d5f87cc3b94bfccc501142df4a7b12fee3a6e7:
1
TCG patch queue, plus one target/sh4 patch that
2
Yoshinori Sato asked me to process.
2
3
3
Merge remote-tracking branch 'remotes/rth-gitlab/tags/pull-axp-20210628' into staging (2021-06-29 10:02:42 +0100)
4
5
r~
6
7
8
The following changes since commit efbf38d73e5dcc4d5f8b98c6e7a12be1f3b91745:
9
10
Merge tag 'for-upstream' of git://repo.or.cz/qemu/kevin into staging (2022-10-03 15:06:07 -0400)
4
11
5
are available in the Git repository at:
12
are available in the Git repository at:
6
13
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20210629
14
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20221004
8
15
9
for you to fetch changes up to c86bd2dc4c1d37653c27293b2dacee6bb46bb995:
16
for you to fetch changes up to ab419fd8a035a65942de4e63effcd55ccbf1a9fe:
10
17
11
tcg/riscv: Remove MO_BSWAP handling (2021-06-29 10:04:57 -0700)
18
target/sh4: Fix TB_FLAG_UNALIGN (2022-10-04 12:33:05 -0700)
12
19
13
----------------------------------------------------------------
20
----------------------------------------------------------------
14
TranslatorOps conversion for target/avr
21
Cache CPUClass for use in hot code paths.
15
TranslatorOps conversion for target/cris
22
Add CPUTLBEntryFull, probe_access_full, tlb_set_page_full.
16
TranslatorOps conversion for target/nios2
23
Add generic support for TARGET_TB_PCREL.
17
Simple vector operations on TCGv_i32
24
tcg/ppc: Optimize 26-bit jumps using STQ for POWER 2.07
18
Host signal fixes for *BSD
25
target/sh4: Fix TB_FLAG_UNALIGN
19
Improvements to tcg bswap operations
20
26
21
----------------------------------------------------------------
27
----------------------------------------------------------------
22
LIU Zhiwei (5):
28
Alex Bennée (3):
23
tcg: Add tcg_gen_vec_add{sub}16_i32
29
cpu: cache CPUClass in CPUState for hot code paths
24
tcg: Add tcg_gen_vec_add{sub}8_i32
30
hw/core/cpu-sysemu: used cached class in cpu_asidx_from_attrs
25
tcg: Add tcg_gen_vec_shl{shr}{sar}16i_i32
31
cputlb: used cached CPUClass in our hot-paths
26
tcg: Add tcg_gen_vec_shl{shr}{sar}8i_i32
27
tcg: Implement tcg_gen_vec_add{sub}32_tl
28
32
29
Richard Henderson (57):
33
Leandro Lupori (1):
30
target/nios2: Replace DISAS_TB_JUMP with DISAS_NORETURN
34
tcg/ppc: Optimize 26-bit jumps
31
target/nios2: Use global cpu_env
32
target/nios2: Use global cpu_R
33
target/nios2: Add DisasContextBase to DisasContext
34
target/nios2: Convert to TranslatorOps
35
target/nios2: Remove assignment to env in handle_instruction
36
target/nios2: Clean up goto in handle_instruction
37
target/nios2: Inline handle_instruction
38
target/nios2: Use pc_next for pc + 4
39
target/avr: Add DisasContextBase to DisasContext
40
target/avr: Change ctx to DisasContext* in gen_intermediate_code
41
target/avr: Convert to TranslatorOps
42
target/cris: Add DisasContextBase to DisasContext
43
target/cris: Remove DISAS_SWI
44
target/cris: Replace DISAS_TB_JUMP with DISAS_NORETURN
45
target/cris: Mark exceptions as DISAS_NORETURN
46
target/cris: Fix use_goto_tb
47
target/cris: Convert to TranslatorOps
48
target/cris: Mark helper_raise_exception noreturn
49
target/cris: Mark static arrays const
50
target/cris: Fold unhandled X_FLAG changes into cpustate_changed
51
target/cris: Set cpustate_changed for rfe/rfn
52
target/cris: Add DISAS_UPDATE_NEXT
53
target/cris: Add DISAS_DBRANCH
54
target/cris: Use tcg_gen_lookup_and_goto_ptr
55
target/cris: Improve JMP_INDIRECT
56
target/cris: Remove dc->flagx_known
57
target/cris: Do not exit tb for X_FLAG changes
58
tcg: Add flags argument to bswap opcodes
59
tcg/i386: Support bswap flags
60
tcg/aarch64: Merge tcg_out_rev{16,32,64}
61
tcg/aarch64: Support bswap flags
62
tcg/arm: Support bswap flags
63
tcg/ppc: Split out tcg_out_ext{8,16,32}s
64
tcg/ppc: Split out tcg_out_sari{32,64}
65
tcg/ppc: Split out tcg_out_bswap16
66
tcg/ppc: Split out tcg_out_bswap32
67
tcg/ppc: Split out tcg_out_bswap64
68
tcg/ppc: Support bswap flags
69
tcg/ppc: Use power10 byte-reverse instructions
70
tcg/s390: Support bswap flags
71
tcg/mips: Support bswap flags in tcg_out_bswap16
72
tcg/mips: Support bswap flags in tcg_out_bswap32
73
tcg/tci: Support bswap flags
74
tcg: Handle new bswap flags during optimize
75
tcg: Add flags argument to tcg_gen_bswap16_*, tcg_gen_bswap32_i64
76
tcg: Make use of bswap flags in tcg_gen_qemu_ld_*
77
tcg: Make use of bswap flags in tcg_gen_qemu_st_*
78
target/arm: Improve REV32
79
target/arm: Improve vector REV
80
target/arm: Improve REVSH
81
target/i386: Improve bswap translation
82
target/sh4: Improve swap.b translation
83
target/mips: Fix gen_mxu_s32ldd_s32lddr
84
tcg/arm: Unset TCG_TARGET_HAS_MEMORY_BSWAP
85
tcg/aarch64: Unset TCG_TARGET_HAS_MEMORY_BSWAP
86
tcg/riscv: Remove MO_BSWAP handling
87
35
88
Warner Losh (1):
36
Richard Henderson (16):
89
tcg: Use correct trap number for page faults on *BSD systems
37
accel/tcg: Rename CPUIOTLBEntry to CPUTLBEntryFull
38
accel/tcg: Drop addr member from SavedIOTLB
39
accel/tcg: Suppress auto-invalidate in probe_access_internal
40
accel/tcg: Introduce probe_access_full
41
accel/tcg: Introduce tlb_set_page_full
42
include/exec: Introduce TARGET_PAGE_ENTRY_EXTRA
43
accel/tcg: Remove PageDesc code_bitmap
44
accel/tcg: Use bool for page_find_alloc
45
accel/tcg: Use DisasContextBase in plugin_gen_tb_start
46
accel/tcg: Do not align tb->page_addr[0]
47
accel/tcg: Inline tb_flush_jmp_cache
48
include/hw/core: Create struct CPUJumpCache
49
hw/core: Add CPUClass.get_pc
50
accel/tcg: Introduce tb_pc and log_pc
51
accel/tcg: Introduce TARGET_TB_PCREL
52
target/sh4: Fix TB_FLAG_UNALIGN
90
53
91
include/tcg/tcg-op-gvec.h | 43 ++++
54
accel/tcg/internal.h | 10 ++
92
include/tcg/tcg-op.h | 8 +-
55
accel/tcg/tb-hash.h | 1 +
93
include/tcg/tcg-opc.h | 10 +-
56
accel/tcg/tb-jmp-cache.h | 65 ++++++++
94
include/tcg/tcg.h | 12 +
57
include/exec/cpu-common.h | 1 +
95
target/cris/helper.h | 2 +-
58
include/exec/cpu-defs.h | 48 ++++--
96
tcg/aarch64/tcg-target.h | 2 +-
59
include/exec/exec-all.h | 75 ++++++++-
97
tcg/arm/tcg-target.h | 2 +-
60
include/exec/plugin-gen.h | 7 +-
98
accel/tcg/user-exec.c | 20 +-
61
include/hw/core/cpu.h | 28 ++--
99
target/arm/translate-a64.c | 21 +-
62
include/qemu/typedefs.h | 2 +
100
target/arm/translate.c | 4 +-
63
include/tcg/tcg.h | 2 +-
101
target/avr/translate.c | 284 ++++++++++++----------
64
target/sh4/cpu.h | 56 ++++---
102
target/cris/translate.c | 515 ++++++++++++++++++++--------------------
65
accel/stubs/tcg-stub.c | 4 +
103
target/i386/tcg/translate.c | 14 +-
66
accel/tcg/cpu-exec.c | 80 +++++-----
104
target/mips/tcg/mxu_translate.c | 6 +-
67
accel/tcg/cputlb.c | 259 ++++++++++++++++++--------------
105
target/nios2/translate.c | 318 ++++++++++++-------------
68
accel/tcg/plugin-gen.c | 22 +--
106
target/s390x/translate.c | 4 +-
69
accel/tcg/translate-all.c | 214 ++++++++++++--------------
107
target/sh4/translate.c | 3 +-
70
accel/tcg/translator.c | 2 +-
108
tcg/optimize.c | 56 ++++-
71
cpu.c | 9 +-
109
tcg/tcg-op-gvec.c | 122 ++++++++++
72
hw/core/cpu-common.c | 3 +-
110
tcg/tcg-op.c | 143 +++++++----
73
hw/core/cpu-sysemu.c | 5 +-
111
tcg/tcg.c | 28 +++
74
linux-user/sh4/signal.c | 6 +-
112
tcg/tci.c | 3 +-
75
plugins/core.c | 2 +-
113
target/cris/translate_v10.c.inc | 17 +-
76
target/alpha/cpu.c | 9 ++
114
tcg/aarch64/tcg-target.c.inc | 125 ++++------
77
target/arm/cpu.c | 17 ++-
115
tcg/arm/tcg-target.c.inc | 295 ++++++++++-------------
78
target/arm/mte_helper.c | 14 +-
116
tcg/i386/tcg-target.c.inc | 20 +-
79
target/arm/sve_helper.c | 4 +-
117
tcg/mips/tcg-target.c.inc | 102 ++++----
80
target/arm/translate-a64.c | 2 +-
118
tcg/ppc/tcg-target.c.inc | 230 ++++++++++++------
81
target/avr/cpu.c | 10 +-
119
tcg/riscv/tcg-target.c.inc | 64 ++---
82
target/cris/cpu.c | 8 +
120
tcg/s390/tcg-target.c.inc | 34 ++-
83
target/hexagon/cpu.c | 10 +-
121
tcg/tci/tcg-target.c.inc | 23 +-
84
target/hppa/cpu.c | 12 +-
122
tcg/README | 22 +-
85
target/i386/cpu.c | 9 ++
123
32 files changed, 1458 insertions(+), 1094 deletions(-)
86
target/i386/tcg/tcg-cpu.c | 2 +-
87
target/loongarch/cpu.c | 11 +-
88
target/m68k/cpu.c | 8 +
89
target/microblaze/cpu.c | 10 +-
90
target/mips/cpu.c | 8 +
91
target/mips/tcg/exception.c | 2 +-
92
target/mips/tcg/sysemu/special_helper.c | 2 +-
93
target/nios2/cpu.c | 9 ++
94
target/openrisc/cpu.c | 10 +-
95
target/ppc/cpu_init.c | 8 +
96
target/riscv/cpu.c | 17 ++-
97
target/rx/cpu.c | 10 +-
98
target/s390x/cpu.c | 8 +
99
target/s390x/tcg/mem_helper.c | 4 -
100
target/sh4/cpu.c | 18 ++-
101
target/sh4/helper.c | 6 +-
102
target/sh4/translate.c | 90 +++++------
103
target/sparc/cpu.c | 10 +-
104
target/tricore/cpu.c | 11 +-
105
target/xtensa/cpu.c | 8 +
106
tcg/tcg.c | 8 +-
107
trace/control-target.c | 2 +-
108
tcg/ppc/tcg-target.c.inc | 119 +++++++++++----
109
55 files changed, 915 insertions(+), 462 deletions(-)
110
create mode 100644 accel/tcg/tb-jmp-cache.h
124
111
diff view generated by jsdifflib
1
We have pre-computed the next instruction address into
1
From: Alex Bennée <alex.bennee@linaro.org>
2
dc->base.pc_next, so we might as well use it.
3
2
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
The class cast checkers are quite expensive and always on (unlike the
5
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
4
dynamic case who's checks are gated by CONFIG_QOM_CAST_DEBUG). To
5
avoid the overhead of repeatedly checking something which should never
6
change we cache the CPUClass reference for use in the hot code paths.
7
8
Signed-off-by: Alex Bennée <alex.bennee@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-Id: <20220811151413.3350684-3-alex.bennee@linaro.org>
11
Signed-off-by: Cédric Le Goater <clg@kaod.org>
12
Message-Id: <20220923084803.498337-3-clg@kaod.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
14
---
8
target/nios2/translate.c | 12 ++++++------
15
include/hw/core/cpu.h | 9 +++++++++
9
1 file changed, 6 insertions(+), 6 deletions(-)
16
cpu.c | 9 ++++-----
17
2 files changed, 13 insertions(+), 5 deletions(-)
10
18
11
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
19
diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
12
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
13
--- a/target/nios2/translate.c
21
--- a/include/hw/core/cpu.h
14
+++ b/target/nios2/translate.c
22
+++ b/include/hw/core/cpu.h
15
@@ -XXX,XX +XXX,XX @@ static void jmpi(DisasContext *dc, uint32_t code, uint32_t flags)
23
@@ -XXX,XX +XXX,XX @@ typedef int (*WriteCoreDumpFunction)(const void *buf, size_t size,
16
24
*/
17
static void call(DisasContext *dc, uint32_t code, uint32_t flags)
25
#define CPU(obj) ((CPUState *)(obj))
26
27
+/*
28
+ * The class checkers bring in CPU_GET_CLASS() which is potentially
29
+ * expensive given the eventual call to
30
+ * object_class_dynamic_cast_assert(). Because of this the CPUState
31
+ * has a cached value for the class in cs->cc which is set up in
32
+ * cpu_exec_realizefn() for use in hot code paths.
33
+ */
34
typedef struct CPUClass CPUClass;
35
DECLARE_CLASS_CHECKERS(CPUClass, CPU,
36
TYPE_CPU)
37
@@ -XXX,XX +XXX,XX @@ struct qemu_work_item;
38
struct CPUState {
39
/*< private >*/
40
DeviceState parent_obj;
41
+ /* cache to avoid expensive CPU_GET_CLASS */
42
+ CPUClass *cc;
43
/*< public >*/
44
45
int nr_cores;
46
diff --git a/cpu.c b/cpu.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/cpu.c
49
+++ b/cpu.c
50
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_cpu_common = {
51
52
void cpu_exec_realizefn(CPUState *cpu, Error **errp)
18
{
53
{
19
- tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
54
-#ifndef CONFIG_USER_ONLY
20
+ tcg_gen_movi_tl(cpu_R[R_RA], dc->base.pc_next);
55
- CPUClass *cc = CPU_GET_CLASS(cpu);
21
jmpi(dc, code, flags);
56
-#endif
22
}
57
+ /* cache the cpu class for the hotpath */
23
58
+ cpu->cc = CPU_GET_CLASS(cpu);
24
@@ -XXX,XX +XXX,XX @@ static void br(DisasContext *dc, uint32_t code, uint32_t flags)
59
25
{
60
cpu_list_add(cpu);
26
I_TYPE(instr, code);
61
if (!accel_cpu_realizefn(cpu, errp)) {
27
62
@@ -XXX,XX +XXX,XX @@ void cpu_exec_realizefn(CPUState *cpu, Error **errp)
28
- gen_goto_tb(dc, 0, dc->pc + 4 + (instr.imm16.s & -4));
63
if (qdev_get_vmsd(DEVICE(cpu)) == NULL) {
29
+ gen_goto_tb(dc, 0, dc->base.pc_next + (instr.imm16.s & -4));
64
vmstate_register(NULL, cpu->cpu_index, &vmstate_cpu_common, cpu);
30
dc->base.is_jmp = DISAS_NORETURN;
31
}
32
33
@@ -XXX,XX +XXX,XX @@ static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
34
35
TCGLabel *l1 = gen_new_label();
36
tcg_gen_brcond_tl(flags, cpu_R[instr.a], cpu_R[instr.b], l1);
37
- gen_goto_tb(dc, 0, dc->pc + 4);
38
+ gen_goto_tb(dc, 0, dc->base.pc_next);
39
gen_set_label(l1);
40
- gen_goto_tb(dc, 1, dc->pc + 4 + (instr.imm16.s & -4));
41
+ gen_goto_tb(dc, 1, dc->base.pc_next + (instr.imm16.s & -4));
42
dc->base.is_jmp = DISAS_NORETURN;
43
}
44
45
@@ -XXX,XX +XXX,XX @@ static void nextpc(DisasContext *dc, uint32_t code, uint32_t flags)
46
R_TYPE(instr, code);
47
48
if (likely(instr.c != R_ZERO)) {
49
- tcg_gen_movi_tl(cpu_R[instr.c], dc->pc + 4);
50
+ tcg_gen_movi_tl(cpu_R[instr.c], dc->base.pc_next);
51
}
65
}
52
}
66
- if (cc->sysemu_ops->legacy_vmsd != NULL) {
53
67
- vmstate_register(NULL, cpu->cpu_index, cc->sysemu_ops->legacy_vmsd, cpu);
54
@@ -XXX,XX +XXX,XX @@ static void callr(DisasContext *dc, uint32_t code, uint32_t flags)
68
+ if (cpu->cc->sysemu_ops->legacy_vmsd != NULL) {
55
R_TYPE(instr, code);
69
+ vmstate_register(NULL, cpu->cpu_index, cpu->cc->sysemu_ops->legacy_vmsd, cpu);
56
70
}
57
tcg_gen_mov_tl(cpu_R[R_PC], load_gpr(dc, instr.a));
71
#endif /* CONFIG_USER_ONLY */
58
- tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
59
+ tcg_gen_movi_tl(cpu_R[R_RA], dc->base.pc_next);
60
61
dc->base.is_jmp = DISAS_JUMP;
62
}
72
}
63
--
73
--
64
2.25.1
74
2.34.1
65
75
66
76
diff view generated by jsdifflib
1
TCG_TARGET_HAS_MEMORY_BSWAP is already unset for this backend,
1
From: Alex Bennée <alex.bennee@linaro.org>
2
which means that MO_BSWAP be handled by the middle-end and
3
will never be seen by the backend. Thus the indexes used with
4
qemu_{ld,st}_helpers will always be zero.
5
2
6
Tidy the comments and asserts in tcg_out_qemu_{ld,st}_direct.
3
This is a heavily used function so lets avoid the cost of
7
It is not that we do not handle bswap "yet", but never will.
4
CPU_GET_CLASS. On the romulus-bmc run it has a modest effect:
8
5
9
Acked-by: Alistair Francis <alistair.francis@wdc.com>
6
Before: 36.812 s ± 0.506 s
7
After: 35.912 s ± 0.168 s
8
9
Signed-off-by: Alex Bennée <alex.bennee@linaro.org>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-Id: <20220811151413.3350684-4-alex.bennee@linaro.org>
12
Signed-off-by: Cédric Le Goater <clg@kaod.org>
13
Message-Id: <20220923084803.498337-4-clg@kaod.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
14
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
15
---
12
tcg/riscv/tcg-target.c.inc | 64 ++++++++++++++++++++------------------
16
hw/core/cpu-sysemu.c | 5 ++---
13
1 file changed, 33 insertions(+), 31 deletions(-)
17
1 file changed, 2 insertions(+), 3 deletions(-)
14
18
15
diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc
19
diff --git a/hw/core/cpu-sysemu.c b/hw/core/cpu-sysemu.c
16
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
17
--- a/tcg/riscv/tcg-target.c.inc
21
--- a/hw/core/cpu-sysemu.c
18
+++ b/tcg/riscv/tcg-target.c.inc
22
+++ b/hw/core/cpu-sysemu.c
19
@@ -XXX,XX +XXX,XX @@ static void tcg_out_mb(TCGContext *s, TCGArg a0)
23
@@ -XXX,XX +XXX,XX @@ hwaddr cpu_get_phys_page_debug(CPUState *cpu, vaddr addr)
20
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
24
21
* TCGMemOpIdx oi, uintptr_t ra)
25
int cpu_asidx_from_attrs(CPUState *cpu, MemTxAttrs attrs)
22
*/
23
-static void * const qemu_ld_helpers[16] = {
24
- [MO_UB] = helper_ret_ldub_mmu,
25
- [MO_SB] = helper_ret_ldsb_mmu,
26
- [MO_LEUW] = helper_le_lduw_mmu,
27
- [MO_LESW] = helper_le_ldsw_mmu,
28
- [MO_LEUL] = helper_le_ldul_mmu,
29
+static void * const qemu_ld_helpers[8] = {
30
+ [MO_UB] = helper_ret_ldub_mmu,
31
+ [MO_SB] = helper_ret_ldsb_mmu,
32
+#ifdef HOST_WORDS_BIGENDIAN
33
+ [MO_UW] = helper_be_lduw_mmu,
34
+ [MO_SW] = helper_be_ldsw_mmu,
35
+ [MO_UL] = helper_be_ldul_mmu,
36
#if TCG_TARGET_REG_BITS == 64
37
- [MO_LESL] = helper_le_ldsl_mmu,
38
+ [MO_SL] = helper_be_ldsl_mmu,
39
#endif
40
- [MO_LEQ] = helper_le_ldq_mmu,
41
- [MO_BEUW] = helper_be_lduw_mmu,
42
- [MO_BESW] = helper_be_ldsw_mmu,
43
- [MO_BEUL] = helper_be_ldul_mmu,
44
+ [MO_Q] = helper_be_ldq_mmu,
45
+#else
46
+ [MO_UW] = helper_le_lduw_mmu,
47
+ [MO_SW] = helper_le_ldsw_mmu,
48
+ [MO_UL] = helper_le_ldul_mmu,
49
#if TCG_TARGET_REG_BITS == 64
50
- [MO_BESL] = helper_be_ldsl_mmu,
51
+ [MO_SL] = helper_le_ldsl_mmu,
52
+#endif
53
+ [MO_Q] = helper_le_ldq_mmu,
54
#endif
55
- [MO_BEQ] = helper_be_ldq_mmu,
56
};
57
58
/* helper signature: helper_ret_st_mmu(CPUState *env, target_ulong addr,
59
* uintxx_t val, TCGMemOpIdx oi,
60
* uintptr_t ra)
61
*/
62
-static void * const qemu_st_helpers[16] = {
63
- [MO_UB] = helper_ret_stb_mmu,
64
- [MO_LEUW] = helper_le_stw_mmu,
65
- [MO_LEUL] = helper_le_stl_mmu,
66
- [MO_LEQ] = helper_le_stq_mmu,
67
- [MO_BEUW] = helper_be_stw_mmu,
68
- [MO_BEUL] = helper_be_stl_mmu,
69
- [MO_BEQ] = helper_be_stq_mmu,
70
+static void * const qemu_st_helpers[4] = {
71
+ [MO_8] = helper_ret_stb_mmu,
72
+#ifdef HOST_WORDS_BIGENDIAN
73
+ [MO_16] = helper_be_stw_mmu,
74
+ [MO_32] = helper_be_stl_mmu,
75
+ [MO_64] = helper_be_stq_mmu,
76
+#else
77
+ [MO_16] = helper_le_stw_mmu,
78
+ [MO_32] = helper_le_stl_mmu,
79
+ [MO_64] = helper_le_stq_mmu,
80
+#endif
81
};
82
83
/* We don't support oversize guests */
84
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
85
tcg_out_movi(s, TCG_TYPE_PTR, a2, oi);
86
tcg_out_movi(s, TCG_TYPE_PTR, a3, (tcg_target_long)l->raddr);
87
88
- tcg_out_call(s, qemu_ld_helpers[opc & (MO_BSWAP | MO_SSIZE)]);
89
+ tcg_out_call(s, qemu_ld_helpers[opc & MO_SSIZE]);
90
tcg_out_mov(s, (opc & MO_SIZE) == MO_64, l->datalo_reg, a0);
91
92
tcg_out_goto(s, l->raddr);
93
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
94
tcg_out_movi(s, TCG_TYPE_PTR, a3, oi);
95
tcg_out_movi(s, TCG_TYPE_PTR, a4, (tcg_target_long)l->raddr);
96
97
- tcg_out_call(s, qemu_st_helpers[opc & (MO_BSWAP | MO_SSIZE)]);
98
+ tcg_out_call(s, qemu_st_helpers[opc & MO_SIZE]);
99
100
tcg_out_goto(s, l->raddr);
101
return true;
102
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
103
static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
104
TCGReg base, MemOp opc, bool is_64)
105
{
26
{
106
- const MemOp bswap = opc & MO_BSWAP;
27
- CPUClass *cc = CPU_GET_CLASS(cpu);
107
-
28
int ret = 0;
108
- /* We don't yet handle byteswapping, assert */
29
109
- g_assert(!bswap);
30
- if (cc->sysemu_ops->asidx_from_attrs) {
110
+ /* Byte swapping is left to middle-end expansion. */
31
- ret = cc->sysemu_ops->asidx_from_attrs(cpu, attrs);
111
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
32
+ if (cpu->cc->sysemu_ops->asidx_from_attrs) {
112
33
+ ret = cpu->cc->sysemu_ops->asidx_from_attrs(cpu, attrs);
113
switch (opc & (MO_SSIZE)) {
34
assert(ret < cpu->num_ases && ret >= 0);
114
case MO_UB:
35
}
115
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
36
return ret;
116
static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
117
TCGReg base, MemOp opc)
118
{
119
- const MemOp bswap = opc & MO_BSWAP;
120
-
121
- /* We don't yet handle byteswapping, assert */
122
- g_assert(!bswap);
123
+ /* Byte swapping is left to middle-end expansion. */
124
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
125
126
switch (opc & (MO_SSIZE)) {
127
case MO_8:
128
--
37
--
129
2.25.1
38
2.34.1
130
39
131
40
diff view generated by jsdifflib
1
Merge tcg_out_bswap16 and tcg_out_bswap16s. Use the flags
1
From: Alex Bennée <alex.bennee@linaro.org>
2
in the internal uses for loads and stores.
3
2
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
3
Before: 35.912 s ± 0.168 s
4
After: 35.565 s ± 0.087 s
5
6
Signed-off-by: Alex Bennée <alex.bennee@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-Id: <20220811151413.3350684-5-alex.bennee@linaro.org>
9
Signed-off-by: Cédric Le Goater <clg@kaod.org>
10
Message-Id: <20220923084803.498337-5-clg@kaod.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
12
---
7
tcg/mips/tcg-target.c.inc | 63 +++++++++++++++++++--------------------
13
accel/tcg/cputlb.c | 15 ++++++---------
8
1 file changed, 30 insertions(+), 33 deletions(-)
14
1 file changed, 6 insertions(+), 9 deletions(-)
9
15
10
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
16
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
11
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/mips/tcg-target.c.inc
18
--- a/accel/tcg/cputlb.c
13
+++ b/tcg/mips/tcg-target.c.inc
19
+++ b/accel/tcg/cputlb.c
14
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi(TCGContext *s, TCGType type,
20
@@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr,
15
}
21
static void tlb_fill(CPUState *cpu, target_ulong addr, int size,
22
MMUAccessType access_type, int mmu_idx, uintptr_t retaddr)
23
{
24
- CPUClass *cc = CPU_GET_CLASS(cpu);
25
bool ok;
26
27
/*
28
* This is not a probe, so only valid return is success; failure
29
* should result in exception + longjmp to the cpu loop.
30
*/
31
- ok = cc->tcg_ops->tlb_fill(cpu, addr, size,
32
- access_type, mmu_idx, false, retaddr);
33
+ ok = cpu->cc->tcg_ops->tlb_fill(cpu, addr, size,
34
+ access_type, mmu_idx, false, retaddr);
35
assert(ok);
16
}
36
}
17
37
18
-static inline void tcg_out_bswap16(TCGContext *s, TCGReg ret, TCGReg arg)
38
@@ -XXX,XX +XXX,XX @@ static inline void cpu_unaligned_access(CPUState *cpu, vaddr addr,
19
+static void tcg_out_bswap16(TCGContext *s, TCGReg ret, TCGReg arg, int flags)
39
MMUAccessType access_type,
40
int mmu_idx, uintptr_t retaddr)
20
{
41
{
21
+ /* ret and arg can't be register tmp0 */
42
- CPUClass *cc = CPU_GET_CLASS(cpu);
22
+ tcg_debug_assert(ret != TCG_TMP0);
23
+ tcg_debug_assert(arg != TCG_TMP0);
24
+
25
+ /* With arg = abcd: */
26
if (use_mips32r2_instructions) {
27
- tcg_out_opc_reg(s, OPC_WSBH, ret, 0, arg);
28
- } else {
29
- /* ret and arg can't be register at */
30
- if (ret == TCG_TMP0 || arg == TCG_TMP0) {
31
- tcg_abort();
32
+ tcg_out_opc_reg(s, OPC_WSBH, ret, 0, arg); /* badc */
33
+ if (flags & TCG_BSWAP_OS) {
34
+ tcg_out_opc_reg(s, OPC_SEH, ret, 0, ret); /* ssdc */
35
+ } else if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
36
+ tcg_out_opc_imm(s, OPC_ANDI, ret, ret, 0xffff); /* 00dc */
37
}
38
-
43
-
39
- tcg_out_opc_sa(s, OPC_SRL, TCG_TMP0, arg, 8);
44
- cc->tcg_ops->do_unaligned_access(cpu, addr, access_type, mmu_idx, retaddr);
40
- tcg_out_opc_sa(s, OPC_SLL, ret, arg, 8);
45
+ cpu->cc->tcg_ops->do_unaligned_access(cpu, addr, access_type,
41
- tcg_out_opc_imm(s, OPC_ANDI, ret, ret, 0xff00);
46
+ mmu_idx, retaddr);
42
- tcg_out_opc_reg(s, OPC_OR, ret, ret, TCG_TMP0);
43
+ return;
44
}
45
-}
46
47
-static inline void tcg_out_bswap16s(TCGContext *s, TCGReg ret, TCGReg arg)
48
-{
49
- if (use_mips32r2_instructions) {
50
- tcg_out_opc_reg(s, OPC_WSBH, ret, 0, arg);
51
- tcg_out_opc_reg(s, OPC_SEH, ret, 0, ret);
52
+ tcg_out_opc_sa(s, OPC_SRL, TCG_TMP0, arg, 8); /* 0abc */
53
+ if (!(flags & TCG_BSWAP_IZ)) {
54
+ tcg_out_opc_imm(s, OPC_ANDI, TCG_TMP0, TCG_TMP0, 0x00ff); /* 000c */
55
+ }
56
+ if (flags & TCG_BSWAP_OS) {
57
+ tcg_out_opc_sa(s, OPC_SLL, ret, arg, 24); /* d000 */
58
+ tcg_out_opc_sa(s, OPC_SRA, ret, ret, 16); /* ssd0 */
59
} else {
60
- /* ret and arg can't be register at */
61
- if (ret == TCG_TMP0 || arg == TCG_TMP0) {
62
- tcg_abort();
63
+ tcg_out_opc_sa(s, OPC_SLL, ret, arg, 8); /* bcd0 */
64
+ if (flags & TCG_BSWAP_OZ) {
65
+ tcg_out_opc_imm(s, OPC_ANDI, ret, ret, 0xff00); /* 00d0 */
66
}
67
-
68
- tcg_out_opc_sa(s, OPC_SRL, TCG_TMP0, arg, 8);
69
- tcg_out_opc_sa(s, OPC_SLL, ret, arg, 24);
70
- tcg_out_opc_sa(s, OPC_SRA, ret, ret, 16);
71
- tcg_out_opc_reg(s, OPC_OR, ret, ret, TCG_TMP0);
72
}
73
+ tcg_out_opc_reg(s, OPC_OR, ret, ret, TCG_TMP0); /* ssdc */
74
}
47
}
75
48
76
static void tcg_out_bswap_subr(TCGContext *s, const tcg_insn_unit *sub)
49
static inline void cpu_transaction_failed(CPUState *cpu, hwaddr physaddr,
77
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
50
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
78
break;
51
if (!tlb_hit_page(tlb_addr, page_addr)) {
79
case MO_UW | MO_BSWAP:
52
if (!victim_tlb_hit(env, mmu_idx, index, elt_ofs, page_addr)) {
80
tcg_out_opc_imm(s, OPC_LHU, TCG_TMP1, base, 0);
53
CPUState *cs = env_cpu(env);
81
- tcg_out_bswap16(s, lo, TCG_TMP1);
54
- CPUClass *cc = CPU_GET_CLASS(cs);
82
+ tcg_out_bswap16(s, lo, TCG_TMP1, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
55
83
break;
56
- if (!cc->tcg_ops->tlb_fill(cs, addr, fault_size, access_type,
84
case MO_UW:
57
- mmu_idx, nonfault, retaddr)) {
85
tcg_out_opc_imm(s, OPC_LHU, lo, base, 0);
58
+ if (!cs->cc->tcg_ops->tlb_fill(cs, addr, fault_size, access_type,
86
break;
59
+ mmu_idx, nonfault, retaddr)) {
87
case MO_SW | MO_BSWAP:
60
/* Non-faulting page table read failed. */
88
tcg_out_opc_imm(s, OPC_LHU, TCG_TMP1, base, 0);
61
*phost = NULL;
89
- tcg_out_bswap16s(s, lo, TCG_TMP1);
62
return TLB_INVALID_MASK;
90
+ tcg_out_bswap16(s, lo, TCG_TMP1, TCG_BSWAP_IZ | TCG_BSWAP_OS);
91
break;
92
case MO_SW:
93
tcg_out_opc_imm(s, OPC_LH, lo, base, 0);
94
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
95
break;
96
97
case MO_16 | MO_BSWAP:
98
- tcg_out_opc_imm(s, OPC_ANDI, TCG_TMP1, lo, 0xffff);
99
- tcg_out_bswap16(s, TCG_TMP1, TCG_TMP1);
100
+ tcg_out_bswap16(s, TCG_TMP1, lo, 0);
101
lo = TCG_TMP1;
102
/* FALLTHRU */
103
case MO_16:
104
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
105
case INDEX_op_not_i64:
106
i1 = OPC_NOR;
107
goto do_unary;
108
- case INDEX_op_bswap16_i32:
109
- case INDEX_op_bswap16_i64:
110
- i1 = OPC_WSBH;
111
- goto do_unary;
112
case INDEX_op_ext8s_i32:
113
case INDEX_op_ext8s_i64:
114
i1 = OPC_SEB;
115
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
116
tcg_out_opc_reg(s, i1, a0, TCG_REG_ZERO, a1);
117
break;
118
119
+ case INDEX_op_bswap16_i32:
120
+ case INDEX_op_bswap16_i64:
121
+ tcg_out_bswap16(s, a0, a1, a2);
122
+ break;
123
case INDEX_op_bswap32_i32:
124
tcg_out_bswap32(s, a0, a1);
125
break;
126
--
63
--
127
2.25.1
64
2.34.1
128
65
129
66
diff view generated by jsdifflib
1
For the sf version, we are performing two 32-bit bswaps
1
This structure will shortly contain more than just
2
in either half of the register. This is equivalent to
2
data for accessing MMIO. Rename the 'addr' member
3
performing one 64-bit bswap followed by a rotate.
3
to 'xlat_section' to more clearly indicate its purpose.
4
4
5
For the non-sf version, we can remove TCG_BSWAP_IZ
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
and the preceding zero-extension.
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
9
---
12
target/arm/translate-a64.c | 17 ++++-------------
10
include/exec/cpu-defs.h | 22 ++++----
13
1 file changed, 4 insertions(+), 13 deletions(-)
11
accel/tcg/cputlb.c | 102 +++++++++++++++++++------------------
12
target/arm/mte_helper.c | 14 ++---
13
target/arm/sve_helper.c | 4 +-
14
target/arm/translate-a64.c | 2 +-
15
5 files changed, 73 insertions(+), 71 deletions(-)
14
16
17
diff --git a/include/exec/cpu-defs.h b/include/exec/cpu-defs.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/exec/cpu-defs.h
20
+++ b/include/exec/cpu-defs.h
21
@@ -XXX,XX +XXX,XX @@ typedef uint64_t target_ulong;
22
# endif
23
# endif
24
25
+/* Minimalized TLB entry for use by TCG fast path. */
26
typedef struct CPUTLBEntry {
27
/* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
28
bit TARGET_PAGE_BITS-1..4 : Nonzero for accesses that should not
29
@@ -XXX,XX +XXX,XX @@ typedef struct CPUTLBEntry {
30
31
QEMU_BUILD_BUG_ON(sizeof(CPUTLBEntry) != (1 << CPU_TLB_ENTRY_BITS));
32
33
-/* The IOTLB is not accessed directly inline by generated TCG code,
34
- * so the CPUIOTLBEntry layout is not as critical as that of the
35
- * CPUTLBEntry. (This is also why we don't want to combine the two
36
- * structs into one.)
37
+/*
38
+ * The full TLB entry, which is not accessed by generated TCG code,
39
+ * so the layout is not as critical as that of CPUTLBEntry. This is
40
+ * also why we don't want to combine the two structs.
41
*/
42
-typedef struct CPUIOTLBEntry {
43
+typedef struct CPUTLBEntryFull {
44
/*
45
- * @addr contains:
46
+ * @xlat_section contains:
47
* - in the lower TARGET_PAGE_BITS, a physical section number
48
* - with the lower TARGET_PAGE_BITS masked off, an offset which
49
* must be added to the virtual address to obtain:
50
@@ -XXX,XX +XXX,XX @@ typedef struct CPUIOTLBEntry {
51
* number is PHYS_SECTION_NOTDIRTY or PHYS_SECTION_ROM)
52
* + the offset within the target MemoryRegion (otherwise)
53
*/
54
- hwaddr addr;
55
+ hwaddr xlat_section;
56
MemTxAttrs attrs;
57
-} CPUIOTLBEntry;
58
+} CPUTLBEntryFull;
59
60
/*
61
* Data elements that are per MMU mode, minus the bits accessed by
62
@@ -XXX,XX +XXX,XX @@ typedef struct CPUTLBDesc {
63
size_t vindex;
64
/* The tlb victim table, in two parts. */
65
CPUTLBEntry vtable[CPU_VTLB_SIZE];
66
- CPUIOTLBEntry viotlb[CPU_VTLB_SIZE];
67
- /* The iotlb. */
68
- CPUIOTLBEntry *iotlb;
69
+ CPUTLBEntryFull vfulltlb[CPU_VTLB_SIZE];
70
+ CPUTLBEntryFull *fulltlb;
71
} CPUTLBDesc;
72
73
/*
74
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
75
index XXXXXXX..XXXXXXX 100644
76
--- a/accel/tcg/cputlb.c
77
+++ b/accel/tcg/cputlb.c
78
@@ -XXX,XX +XXX,XX @@ static void tlb_mmu_resize_locked(CPUTLBDesc *desc, CPUTLBDescFast *fast,
79
}
80
81
g_free(fast->table);
82
- g_free(desc->iotlb);
83
+ g_free(desc->fulltlb);
84
85
tlb_window_reset(desc, now, 0);
86
/* desc->n_used_entries is cleared by the caller */
87
fast->mask = (new_size - 1) << CPU_TLB_ENTRY_BITS;
88
fast->table = g_try_new(CPUTLBEntry, new_size);
89
- desc->iotlb = g_try_new(CPUIOTLBEntry, new_size);
90
+ desc->fulltlb = g_try_new(CPUTLBEntryFull, new_size);
91
92
/*
93
* If the allocations fail, try smaller sizes. We just freed some
94
@@ -XXX,XX +XXX,XX @@ static void tlb_mmu_resize_locked(CPUTLBDesc *desc, CPUTLBDescFast *fast,
95
* allocations to fail though, so we progressively reduce the allocation
96
* size, aborting if we cannot even allocate the smallest TLB we support.
97
*/
98
- while (fast->table == NULL || desc->iotlb == NULL) {
99
+ while (fast->table == NULL || desc->fulltlb == NULL) {
100
if (new_size == (1 << CPU_TLB_DYN_MIN_BITS)) {
101
error_report("%s: %s", __func__, strerror(errno));
102
abort();
103
@@ -XXX,XX +XXX,XX @@ static void tlb_mmu_resize_locked(CPUTLBDesc *desc, CPUTLBDescFast *fast,
104
fast->mask = (new_size - 1) << CPU_TLB_ENTRY_BITS;
105
106
g_free(fast->table);
107
- g_free(desc->iotlb);
108
+ g_free(desc->fulltlb);
109
fast->table = g_try_new(CPUTLBEntry, new_size);
110
- desc->iotlb = g_try_new(CPUIOTLBEntry, new_size);
111
+ desc->fulltlb = g_try_new(CPUTLBEntryFull, new_size);
112
}
113
}
114
115
@@ -XXX,XX +XXX,XX @@ static void tlb_mmu_init(CPUTLBDesc *desc, CPUTLBDescFast *fast, int64_t now)
116
desc->n_used_entries = 0;
117
fast->mask = (n_entries - 1) << CPU_TLB_ENTRY_BITS;
118
fast->table = g_new(CPUTLBEntry, n_entries);
119
- desc->iotlb = g_new(CPUIOTLBEntry, n_entries);
120
+ desc->fulltlb = g_new(CPUTLBEntryFull, n_entries);
121
tlb_mmu_flush_locked(desc, fast);
122
}
123
124
@@ -XXX,XX +XXX,XX @@ void tlb_destroy(CPUState *cpu)
125
CPUTLBDescFast *fast = &env_tlb(env)->f[i];
126
127
g_free(fast->table);
128
- g_free(desc->iotlb);
129
+ g_free(desc->fulltlb);
130
}
131
}
132
133
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
134
135
/* Evict the old entry into the victim tlb. */
136
copy_tlb_helper_locked(tv, te);
137
- desc->viotlb[vidx] = desc->iotlb[index];
138
+ desc->vfulltlb[vidx] = desc->fulltlb[index];
139
tlb_n_used_entries_dec(env, mmu_idx);
140
}
141
142
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
143
* subtract here is that of the page base, and not the same as the
144
* vaddr we add back in io_readx()/io_writex()/get_page_addr_code().
145
*/
146
- desc->iotlb[index].addr = iotlb - vaddr_page;
147
- desc->iotlb[index].attrs = attrs;
148
+ desc->fulltlb[index].xlat_section = iotlb - vaddr_page;
149
+ desc->fulltlb[index].attrs = attrs;
150
151
/* Now calculate the new entry */
152
tn.addend = addend - vaddr_page;
153
@@ -XXX,XX +XXX,XX @@ static inline void cpu_transaction_failed(CPUState *cpu, hwaddr physaddr,
154
}
155
}
156
157
-static uint64_t io_readx(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
158
+static uint64_t io_readx(CPUArchState *env, CPUTLBEntryFull *full,
159
int mmu_idx, target_ulong addr, uintptr_t retaddr,
160
MMUAccessType access_type, MemOp op)
161
{
162
@@ -XXX,XX +XXX,XX @@ static uint64_t io_readx(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
163
bool locked = false;
164
MemTxResult r;
165
166
- section = iotlb_to_section(cpu, iotlbentry->addr, iotlbentry->attrs);
167
+ section = iotlb_to_section(cpu, full->xlat_section, full->attrs);
168
mr = section->mr;
169
- mr_offset = (iotlbentry->addr & TARGET_PAGE_MASK) + addr;
170
+ mr_offset = (full->xlat_section & TARGET_PAGE_MASK) + addr;
171
cpu->mem_io_pc = retaddr;
172
if (!cpu->can_do_io) {
173
cpu_io_recompile(cpu, retaddr);
174
@@ -XXX,XX +XXX,XX @@ static uint64_t io_readx(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
175
qemu_mutex_lock_iothread();
176
locked = true;
177
}
178
- r = memory_region_dispatch_read(mr, mr_offset, &val, op, iotlbentry->attrs);
179
+ r = memory_region_dispatch_read(mr, mr_offset, &val, op, full->attrs);
180
if (r != MEMTX_OK) {
181
hwaddr physaddr = mr_offset +
182
section->offset_within_address_space -
183
section->offset_within_region;
184
185
cpu_transaction_failed(cpu, physaddr, addr, memop_size(op), access_type,
186
- mmu_idx, iotlbentry->attrs, r, retaddr);
187
+ mmu_idx, full->attrs, r, retaddr);
188
}
189
if (locked) {
190
qemu_mutex_unlock_iothread();
191
@@ -XXX,XX +XXX,XX @@ static uint64_t io_readx(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
192
}
193
194
/*
195
- * Save a potentially trashed IOTLB entry for later lookup by plugin.
196
- * This is read by tlb_plugin_lookup if the iotlb entry doesn't match
197
+ * Save a potentially trashed CPUTLBEntryFull for later lookup by plugin.
198
+ * This is read by tlb_plugin_lookup if the fulltlb entry doesn't match
199
* because of the side effect of io_writex changing memory layout.
200
*/
201
static void save_iotlb_data(CPUState *cs, hwaddr addr,
202
@@ -XXX,XX +XXX,XX @@ static void save_iotlb_data(CPUState *cs, hwaddr addr,
203
#endif
204
}
205
206
-static void io_writex(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
207
+static void io_writex(CPUArchState *env, CPUTLBEntryFull *full,
208
int mmu_idx, uint64_t val, target_ulong addr,
209
uintptr_t retaddr, MemOp op)
210
{
211
@@ -XXX,XX +XXX,XX @@ static void io_writex(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
212
bool locked = false;
213
MemTxResult r;
214
215
- section = iotlb_to_section(cpu, iotlbentry->addr, iotlbentry->attrs);
216
+ section = iotlb_to_section(cpu, full->xlat_section, full->attrs);
217
mr = section->mr;
218
- mr_offset = (iotlbentry->addr & TARGET_PAGE_MASK) + addr;
219
+ mr_offset = (full->xlat_section & TARGET_PAGE_MASK) + addr;
220
if (!cpu->can_do_io) {
221
cpu_io_recompile(cpu, retaddr);
222
}
223
@@ -XXX,XX +XXX,XX @@ static void io_writex(CPUArchState *env, CPUIOTLBEntry *iotlbentry,
224
* The memory_region_dispatch may trigger a flush/resize
225
* so for plugins we save the iotlb_data just in case.
226
*/
227
- save_iotlb_data(cpu, iotlbentry->addr, section, mr_offset);
228
+ save_iotlb_data(cpu, full->xlat_section, section, mr_offset);
229
230
if (!qemu_mutex_iothread_locked()) {
231
qemu_mutex_lock_iothread();
232
locked = true;
233
}
234
- r = memory_region_dispatch_write(mr, mr_offset, val, op, iotlbentry->attrs);
235
+ r = memory_region_dispatch_write(mr, mr_offset, val, op, full->attrs);
236
if (r != MEMTX_OK) {
237
hwaddr physaddr = mr_offset +
238
section->offset_within_address_space -
239
section->offset_within_region;
240
241
cpu_transaction_failed(cpu, physaddr, addr, memop_size(op),
242
- MMU_DATA_STORE, mmu_idx, iotlbentry->attrs, r,
243
+ MMU_DATA_STORE, mmu_idx, full->attrs, r,
244
retaddr);
245
}
246
if (locked) {
247
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
248
copy_tlb_helper_locked(vtlb, &tmptlb);
249
qemu_spin_unlock(&env_tlb(env)->c.lock);
250
251
- CPUIOTLBEntry tmpio, *io = &env_tlb(env)->d[mmu_idx].iotlb[index];
252
- CPUIOTLBEntry *vio = &env_tlb(env)->d[mmu_idx].viotlb[vidx];
253
- tmpio = *io; *io = *vio; *vio = tmpio;
254
+ CPUTLBEntryFull *f1 = &env_tlb(env)->d[mmu_idx].fulltlb[index];
255
+ CPUTLBEntryFull *f2 = &env_tlb(env)->d[mmu_idx].vfulltlb[vidx];
256
+ CPUTLBEntryFull tmpf;
257
+ tmpf = *f1; *f1 = *f2; *f2 = tmpf;
258
return true;
259
}
260
}
261
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
262
(ADDR) & TARGET_PAGE_MASK)
263
264
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
265
- CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
266
+ CPUTLBEntryFull *full, uintptr_t retaddr)
267
{
268
- ram_addr_t ram_addr = mem_vaddr + iotlbentry->addr;
269
+ ram_addr_t ram_addr = mem_vaddr + full->xlat_section;
270
271
trace_memory_notdirty_write_access(mem_vaddr, ram_addr, size);
272
273
@@ -XXX,XX +XXX,XX @@ int probe_access_flags(CPUArchState *env, target_ulong addr,
274
/* Handle clean RAM pages. */
275
if (unlikely(flags & TLB_NOTDIRTY)) {
276
uintptr_t index = tlb_index(env, mmu_idx, addr);
277
- CPUIOTLBEntry *iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
278
+ CPUTLBEntryFull *full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
279
280
- notdirty_write(env_cpu(env), addr, 1, iotlbentry, retaddr);
281
+ notdirty_write(env_cpu(env), addr, 1, full, retaddr);
282
flags &= ~TLB_NOTDIRTY;
283
}
284
285
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
286
287
if (unlikely(flags & (TLB_NOTDIRTY | TLB_WATCHPOINT))) {
288
uintptr_t index = tlb_index(env, mmu_idx, addr);
289
- CPUIOTLBEntry *iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
290
+ CPUTLBEntryFull *full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
291
292
/* Handle watchpoints. */
293
if (flags & TLB_WATCHPOINT) {
294
int wp_access = (access_type == MMU_DATA_STORE
295
? BP_MEM_WRITE : BP_MEM_READ);
296
cpu_check_watchpoint(env_cpu(env), addr, size,
297
- iotlbentry->attrs, wp_access, retaddr);
298
+ full->attrs, wp_access, retaddr);
299
}
300
301
/* Handle clean RAM pages. */
302
if (flags & TLB_NOTDIRTY) {
303
- notdirty_write(env_cpu(env), addr, 1, iotlbentry, retaddr);
304
+ notdirty_write(env_cpu(env), addr, 1, full, retaddr);
305
}
306
}
307
308
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
309
* should have just filled the TLB. The one corner case is io_writex
310
* which can cause TLB flushes and potential resizing of the TLBs
311
* losing the information we need. In those cases we need to recover
312
- * data from a copy of the iotlbentry. As long as this always occurs
313
+ * data from a copy of the CPUTLBEntryFull. As long as this always occurs
314
* from the same thread (which a mem callback will be) this is safe.
315
*/
316
317
@@ -XXX,XX +XXX,XX @@ bool tlb_plugin_lookup(CPUState *cpu, target_ulong addr, int mmu_idx,
318
if (likely(tlb_hit(tlb_addr, addr))) {
319
/* We must have an iotlb entry for MMIO */
320
if (tlb_addr & TLB_MMIO) {
321
- CPUIOTLBEntry *iotlbentry;
322
- iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
323
+ CPUTLBEntryFull *full;
324
+ full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
325
data->is_io = true;
326
- data->v.io.section = iotlb_to_section(cpu, iotlbentry->addr, iotlbentry->attrs);
327
- data->v.io.offset = (iotlbentry->addr & TARGET_PAGE_MASK) + addr;
328
+ data->v.io.section =
329
+ iotlb_to_section(cpu, full->xlat_section, full->attrs);
330
+ data->v.io.offset = (full->xlat_section & TARGET_PAGE_MASK) + addr;
331
} else {
332
data->is_io = false;
333
data->v.ram.hostaddr = (void *)((uintptr_t)addr + tlbe->addend);
334
@@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr,
335
336
if (unlikely(tlb_addr & TLB_NOTDIRTY)) {
337
notdirty_write(env_cpu(env), addr, size,
338
- &env_tlb(env)->d[mmu_idx].iotlb[index], retaddr);
339
+ &env_tlb(env)->d[mmu_idx].fulltlb[index], retaddr);
340
}
341
342
return hostaddr;
343
@@ -XXX,XX +XXX,XX @@ load_helper(CPUArchState *env, target_ulong addr, MemOpIdx oi,
344
345
/* Handle anything that isn't just a straight memory access. */
346
if (unlikely(tlb_addr & ~TARGET_PAGE_MASK)) {
347
- CPUIOTLBEntry *iotlbentry;
348
+ CPUTLBEntryFull *full;
349
bool need_swap;
350
351
/* For anything that is unaligned, recurse through full_load. */
352
@@ -XXX,XX +XXX,XX @@ load_helper(CPUArchState *env, target_ulong addr, MemOpIdx oi,
353
goto do_unaligned_access;
354
}
355
356
- iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
357
+ full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
358
359
/* Handle watchpoints. */
360
if (unlikely(tlb_addr & TLB_WATCHPOINT)) {
361
/* On watchpoint hit, this will longjmp out. */
362
cpu_check_watchpoint(env_cpu(env), addr, size,
363
- iotlbentry->attrs, BP_MEM_READ, retaddr);
364
+ full->attrs, BP_MEM_READ, retaddr);
365
}
366
367
need_swap = size > 1 && (tlb_addr & TLB_BSWAP);
368
369
/* Handle I/O access. */
370
if (likely(tlb_addr & TLB_MMIO)) {
371
- return io_readx(env, iotlbentry, mmu_idx, addr, retaddr,
372
+ return io_readx(env, full, mmu_idx, addr, retaddr,
373
access_type, op ^ (need_swap * MO_BSWAP));
374
}
375
376
@@ -XXX,XX +XXX,XX @@ store_helper_unaligned(CPUArchState *env, target_ulong addr, uint64_t val,
377
*/
378
if (unlikely(tlb_addr & TLB_WATCHPOINT)) {
379
cpu_check_watchpoint(env_cpu(env), addr, size - size2,
380
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs,
381
+ env_tlb(env)->d[mmu_idx].fulltlb[index].attrs,
382
BP_MEM_WRITE, retaddr);
383
}
384
if (unlikely(tlb_addr2 & TLB_WATCHPOINT)) {
385
cpu_check_watchpoint(env_cpu(env), page2, size2,
386
- env_tlb(env)->d[mmu_idx].iotlb[index2].attrs,
387
+ env_tlb(env)->d[mmu_idx].fulltlb[index2].attrs,
388
BP_MEM_WRITE, retaddr);
389
}
390
391
@@ -XXX,XX +XXX,XX @@ store_helper(CPUArchState *env, target_ulong addr, uint64_t val,
392
393
/* Handle anything that isn't just a straight memory access. */
394
if (unlikely(tlb_addr & ~TARGET_PAGE_MASK)) {
395
- CPUIOTLBEntry *iotlbentry;
396
+ CPUTLBEntryFull *full;
397
bool need_swap;
398
399
/* For anything that is unaligned, recurse through byte stores. */
400
@@ -XXX,XX +XXX,XX @@ store_helper(CPUArchState *env, target_ulong addr, uint64_t val,
401
goto do_unaligned_access;
402
}
403
404
- iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
405
+ full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
406
407
/* Handle watchpoints. */
408
if (unlikely(tlb_addr & TLB_WATCHPOINT)) {
409
/* On watchpoint hit, this will longjmp out. */
410
cpu_check_watchpoint(env_cpu(env), addr, size,
411
- iotlbentry->attrs, BP_MEM_WRITE, retaddr);
412
+ full->attrs, BP_MEM_WRITE, retaddr);
413
}
414
415
need_swap = size > 1 && (tlb_addr & TLB_BSWAP);
416
417
/* Handle I/O access. */
418
if (tlb_addr & TLB_MMIO) {
419
- io_writex(env, iotlbentry, mmu_idx, val, addr, retaddr,
420
+ io_writex(env, full, mmu_idx, val, addr, retaddr,
421
op ^ (need_swap * MO_BSWAP));
422
return;
423
}
424
@@ -XXX,XX +XXX,XX @@ store_helper(CPUArchState *env, target_ulong addr, uint64_t val,
425
426
/* Handle clean RAM pages. */
427
if (tlb_addr & TLB_NOTDIRTY) {
428
- notdirty_write(env_cpu(env), addr, size, iotlbentry, retaddr);
429
+ notdirty_write(env_cpu(env), addr, size, full, retaddr);
430
}
431
432
haddr = (void *)((uintptr_t)addr + entry->addend);
433
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
434
index XXXXXXX..XXXXXXX 100644
435
--- a/target/arm/mte_helper.c
436
+++ b/target/arm/mte_helper.c
437
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
438
return tags + index;
439
#else
440
uintptr_t index;
441
- CPUIOTLBEntry *iotlbentry;
442
+ CPUTLBEntryFull *full;
443
int in_page, flags;
444
ram_addr_t ptr_ra;
445
hwaddr ptr_paddr, tag_paddr, xlat;
446
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
447
assert(!(flags & TLB_INVALID_MASK));
448
449
/*
450
- * Find the iotlbentry for ptr. This *must* be present in the TLB
451
+ * Find the CPUTLBEntryFull for ptr. This *must* be present in the TLB
452
* because we just found the mapping.
453
* TODO: Perhaps there should be a cputlb helper that returns a
454
* matching tlb entry + iotlb entry.
455
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
456
g_assert(tlb_hit(comparator, ptr));
457
}
458
# endif
459
- iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
460
+ full = &env_tlb(env)->d[ptr_mmu_idx].fulltlb[index];
461
462
/* If the virtual page MemAttr != Tagged, access unchecked. */
463
- if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
464
+ if (!arm_tlb_mte_tagged(&full->attrs)) {
465
return NULL;
466
}
467
468
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
469
int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
470
assert(ra != 0);
471
cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
472
- iotlbentry->attrs, wp, ra);
473
+ full->attrs, wp, ra);
474
}
475
476
/*
477
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
478
tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
479
480
/* Look up the address in tag space. */
481
- tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
482
+ tag_asi = full->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
483
tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
484
mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
485
tag_access == MMU_DATA_STORE,
486
- iotlbentry->attrs);
487
+ full->attrs);
488
489
/*
490
* Note that @mr will never be NULL. If there is nothing in the address
491
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
492
index XXXXXXX..XXXXXXX 100644
493
--- a/target/arm/sve_helper.c
494
+++ b/target/arm/sve_helper.c
495
@@ -XXX,XX +XXX,XX @@ bool sve_probe_page(SVEHostPage *info, bool nofault, CPUARMState *env,
496
g_assert(tlb_hit(comparator, addr));
497
# endif
498
499
- CPUIOTLBEntry *iotlbentry = &env_tlb(env)->d[mmu_idx].iotlb[index];
500
- info->attrs = iotlbentry->attrs;
501
+ CPUTLBEntryFull *full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
502
+ info->attrs = full->attrs;
503
}
504
#endif
505
15
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
506
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
16
index XXXXXXX..XXXXXXX 100644
507
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate-a64.c
508
--- a/target/arm/translate-a64.c
18
+++ b/target/arm/translate-a64.c
509
+++ b/target/arm/translate-a64.c
19
@@ -XXX,XX +XXX,XX @@ static void handle_rev32(DisasContext *s, unsigned int sf,
510
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
20
unsigned int rn, unsigned int rd)
511
* table entry even for that case.
21
{
512
*/
22
TCGv_i64 tcg_rd = cpu_reg(s, rd);
513
return (tlb_hit(entry->addr_code, addr) &&
23
+ TCGv_i64 tcg_rn = cpu_reg(s, rn);
514
- arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
24
515
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].fulltlb[index].attrs));
25
if (sf) {
516
#endif
26
- TCGv_i64 tcg_tmp = tcg_temp_new_i64();
517
}
27
- TCGv_i64 tcg_rn = read_cpu_reg(s, rn, sf);
28
-
29
- /* bswap32_i64 requires zero high word */
30
- tcg_gen_ext32u_i64(tcg_tmp, tcg_rn);
31
- tcg_gen_bswap32_i64(tcg_rd, tcg_tmp, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
32
- tcg_gen_shri_i64(tcg_tmp, tcg_rn, 32);
33
- tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
34
- tcg_gen_concat32_i64(tcg_rd, tcg_rd, tcg_tmp);
35
-
36
- tcg_temp_free_i64(tcg_tmp);
37
+ tcg_gen_bswap64_i64(tcg_rd, tcg_rn);
38
+ tcg_gen_rotri_i64(tcg_rd, tcg_rd, 32);
39
} else {
40
- tcg_gen_ext32u_i64(tcg_rd, cpu_reg(s, rn));
41
- tcg_gen_bswap32_i64(tcg_rd, tcg_rd, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
42
+ tcg_gen_bswap32_i64(tcg_rd, tcg_rn, TCG_BSWAP_OZ);
43
}
44
}
45
518
46
--
519
--
47
2.25.1
520
2.34.1
48
521
49
522
diff view generated by jsdifflib
1
The existing interpreter zero-extends, ignoring high bits.
1
This field is only written, not read; remove it.
2
Simply add a separate sign-extension opcode if required.
3
Ensure that the interpreter supports ext16s when bswap16 is enabled.
4
2
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
7
---
8
tcg/tci.c | 3 ++-
8
include/hw/core/cpu.h | 1 -
9
tcg/tci/tcg-target.c.inc | 23 ++++++++++++++++++++---
9
accel/tcg/cputlb.c | 7 +++----
10
2 files changed, 22 insertions(+), 4 deletions(-)
10
2 files changed, 3 insertions(+), 5 deletions(-)
11
11
12
diff --git a/tcg/tci.c b/tcg/tci.c
12
diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
13
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/tci.c
14
--- a/include/hw/core/cpu.h
15
+++ b/tcg/tci.c
15
+++ b/include/hw/core/cpu.h
16
@@ -XXX,XX +XXX,XX @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState *env,
16
@@ -XXX,XX +XXX,XX @@ struct CPUWatchpoint {
17
regs[r0] = (int8_t)regs[r1];
17
* the memory regions get moved around by io_writex.
18
break;
18
*/
19
typedef struct SavedIOTLB {
20
- hwaddr addr;
21
MemoryRegionSection *section;
22
hwaddr mr_offset;
23
} SavedIOTLB;
24
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
25
index XXXXXXX..XXXXXXX 100644
26
--- a/accel/tcg/cputlb.c
27
+++ b/accel/tcg/cputlb.c
28
@@ -XXX,XX +XXX,XX @@ static uint64_t io_readx(CPUArchState *env, CPUTLBEntryFull *full,
29
* This is read by tlb_plugin_lookup if the fulltlb entry doesn't match
30
* because of the side effect of io_writex changing memory layout.
31
*/
32
-static void save_iotlb_data(CPUState *cs, hwaddr addr,
33
- MemoryRegionSection *section, hwaddr mr_offset)
34
+static void save_iotlb_data(CPUState *cs, MemoryRegionSection *section,
35
+ hwaddr mr_offset)
36
{
37
#ifdef CONFIG_PLUGIN
38
SavedIOTLB *saved = &cs->saved_iotlb;
39
- saved->addr = addr;
40
saved->section = section;
41
saved->mr_offset = mr_offset;
19
#endif
42
#endif
20
-#if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64
43
@@ -XXX,XX +XXX,XX @@ static void io_writex(CPUArchState *env, CPUTLBEntryFull *full,
21
+#if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64 || \
44
* The memory_region_dispatch may trigger a flush/resize
22
+ TCG_TARGET_HAS_bswap16_i32 || TCG_TARGET_HAS_bswap16_i64
45
* so for plugins we save the iotlb_data just in case.
23
CASE_32_64(ext16s)
46
*/
24
tci_args_rr(insn, &r0, &r1);
47
- save_iotlb_data(cpu, full->xlat_section, section, mr_offset);
25
regs[r0] = (int16_t)regs[r1];
48
+ save_iotlb_data(cpu, section, mr_offset);
26
diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc
49
27
index XXXXXXX..XXXXXXX 100644
50
if (!qemu_mutex_iothread_locked()) {
28
--- a/tcg/tci/tcg-target.c.inc
51
qemu_mutex_lock_iothread();
29
+++ b/tcg/tci/tcg-target.c.inc
30
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
31
const TCGArg args[TCG_MAX_OP_ARGS],
32
const int const_args[TCG_MAX_OP_ARGS])
33
{
34
+ TCGOpcode exts;
35
+
36
switch (opc) {
37
case INDEX_op_exit_tb:
38
tcg_out_op_p(s, opc, (void *)args[0]);
39
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
40
CASE_64(ext32u) /* Optional (TCG_TARGET_HAS_ext32u_i64). */
41
CASE_64(ext_i32)
42
CASE_64(extu_i32)
43
- CASE_32_64(bswap16) /* Optional (TCG_TARGET_HAS_bswap16_*). */
44
- CASE_32_64(bswap32) /* Optional (TCG_TARGET_HAS_bswap32_*). */
45
- CASE_64(bswap64) /* Optional (TCG_TARGET_HAS_bswap64_i64). */
46
CASE_32_64(ctpop) /* Optional (TCG_TARGET_HAS_ctpop_*). */
47
+ case INDEX_op_bswap32_i32: /* Optional (TCG_TARGET_HAS_bswap32_i32). */
48
+ case INDEX_op_bswap64_i64: /* Optional (TCG_TARGET_HAS_bswap64_i64). */
49
tcg_out_op_rr(s, opc, args[0], args[1]);
50
break;
51
52
+ case INDEX_op_bswap16_i32: /* Optional (TCG_TARGET_HAS_bswap16_i32). */
53
+ exts = INDEX_op_ext16s_i32;
54
+ goto do_bswap;
55
+ case INDEX_op_bswap16_i64: /* Optional (TCG_TARGET_HAS_bswap16_i64). */
56
+ exts = INDEX_op_ext16s_i64;
57
+ goto do_bswap;
58
+ case INDEX_op_bswap32_i64: /* Optional (TCG_TARGET_HAS_bswap32_i64). */
59
+ exts = INDEX_op_ext32s_i64;
60
+ do_bswap:
61
+ /* The base tci bswaps zero-extend, and ignore high bits. */
62
+ tcg_out_op_rr(s, opc, args[0], args[1]);
63
+ if (args[2] & TCG_BSWAP_OS) {
64
+ tcg_out_op_rr(s, exts, args[0], args[0]);
65
+ }
66
+ break;
67
+
68
CASE_32_64(add2)
69
CASE_32_64(sub2)
70
tcg_out_op_rrrrrr(s, opc, args[0], args[1], args[2],
71
--
52
--
72
2.25.1
53
2.34.1
73
54
74
55
diff view generated by jsdifflib
1
The memory bswap support in the aarch64 backend merely dates from
1
When PAGE_WRITE_INV is set when calling tlb_set_page,
2
a time when it was required. There is nothing special about the
2
we immediately set TLB_INVALID_MASK in order to force
3
backend support that could not have been provided by the middle-end
3
tlb_fill to be called on the next lookup. Here in
4
even prior to the introduction of the bswap flags.
4
probe_access_internal, we have just called tlb_fill
5
and eliminated true misses, thus the lookup must be valid.
5
6
7
This allows us to remove a warning comment from s390x.
8
There doesn't seem to be a reason to change the code though.
9
10
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
11
Reviewed-by: David Hildenbrand <david@redhat.com>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
14
---
9
tcg/aarch64/tcg-target.h | 2 +-
15
accel/tcg/cputlb.c | 10 +++++++++-
10
tcg/aarch64/tcg-target.c.inc | 87 +++++++++++++-----------------------
16
target/s390x/tcg/mem_helper.c | 4 ----
11
2 files changed, 32 insertions(+), 57 deletions(-)
17
2 files changed, 9 insertions(+), 5 deletions(-)
12
18
13
diff --git a/tcg/aarch64/tcg-target.h b/tcg/aarch64/tcg-target.h
19
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
14
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
15
--- a/tcg/aarch64/tcg-target.h
21
--- a/accel/tcg/cputlb.c
16
+++ b/tcg/aarch64/tcg-target.h
22
+++ b/accel/tcg/cputlb.c
17
@@ -XXX,XX +XXX,XX @@ typedef enum {
23
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
18
#define TCG_TARGET_HAS_cmpsel_vec 0
24
}
19
25
tlb_addr = tlb_read_ofs(entry, elt_ofs);
20
#define TCG_TARGET_DEFAULT_MO (0)
26
21
-#define TCG_TARGET_HAS_MEMORY_BSWAP 1
27
+ flags = TLB_FLAGS_MASK;
22
+#define TCG_TARGET_HAS_MEMORY_BSWAP 0
28
page_addr = addr & TARGET_PAGE_MASK;
23
29
if (!tlb_hit_page(tlb_addr, page_addr)) {
24
void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
30
if (!victim_tlb_hit(env, mmu_idx, index, elt_ofs, page_addr)) {
25
31
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
26
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
32
33
/* TLB resize via tlb_fill may have moved the entry. */
34
entry = tlb_entry(env, mmu_idx, addr);
35
+
36
+ /*
37
+ * With PAGE_WRITE_INV, we set TLB_INVALID_MASK immediately,
38
+ * to force the next access through tlb_fill. We've just
39
+ * called tlb_fill, so we know that this entry *is* valid.
40
+ */
41
+ flags &= ~TLB_INVALID_MASK;
42
}
43
tlb_addr = tlb_read_ofs(entry, elt_ofs);
44
}
45
- flags = tlb_addr & TLB_FLAGS_MASK;
46
+ flags &= tlb_addr;
47
48
/* Fold all "mmio-like" bits into TLB_MMIO. This is not RAM. */
49
if (unlikely(flags & ~(TLB_WATCHPOINT | TLB_NOTDIRTY))) {
50
diff --git a/target/s390x/tcg/mem_helper.c b/target/s390x/tcg/mem_helper.c
27
index XXXXXXX..XXXXXXX 100644
51
index XXXXXXX..XXXXXXX 100644
28
--- a/tcg/aarch64/tcg-target.c.inc
52
--- a/target/s390x/tcg/mem_helper.c
29
+++ b/tcg/aarch64/tcg-target.c.inc
53
+++ b/target/s390x/tcg/mem_helper.c
30
@@ -XXX,XX +XXX,XX @@ static void tcg_out_cltz(TCGContext *s, TCGType ext, TCGReg d,
54
@@ -XXX,XX +XXX,XX @@ static int s390_probe_access(CPUArchState *env, target_ulong addr, int size,
31
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
55
#else
32
* TCGMemOpIdx oi, uintptr_t ra)
56
int flags;
33
*/
57
34
-static void * const qemu_ld_helpers[16] = {
58
- /*
35
- [MO_UB] = helper_ret_ldub_mmu,
59
- * For !CONFIG_USER_ONLY, we cannot rely on TLB_INVALID_MASK or haddr==NULL
36
- [MO_LEUW] = helper_le_lduw_mmu,
60
- * to detect if there was an exception during tlb_fill().
37
- [MO_LEUL] = helper_le_ldul_mmu,
61
- */
38
- [MO_LEQ] = helper_le_ldq_mmu,
62
env->tlb_fill_exc = 0;
39
- [MO_BEUW] = helper_be_lduw_mmu,
63
flags = probe_access_flags(env, addr, access_type, mmu_idx, nonfault, phost,
40
- [MO_BEUL] = helper_be_ldul_mmu,
64
ra);
41
- [MO_BEQ] = helper_be_ldq_mmu,
42
+static void * const qemu_ld_helpers[4] = {
43
+ [MO_8] = helper_ret_ldub_mmu,
44
+#ifdef HOST_WORDS_BIGENDIAN
45
+ [MO_16] = helper_be_lduw_mmu,
46
+ [MO_32] = helper_be_ldul_mmu,
47
+ [MO_64] = helper_be_ldq_mmu,
48
+#else
49
+ [MO_16] = helper_le_lduw_mmu,
50
+ [MO_32] = helper_le_ldul_mmu,
51
+ [MO_64] = helper_le_ldq_mmu,
52
+#endif
53
};
54
55
/* helper signature: helper_ret_st_mmu(CPUState *env, target_ulong addr,
56
* uintxx_t val, TCGMemOpIdx oi,
57
* uintptr_t ra)
58
*/
59
-static void * const qemu_st_helpers[16] = {
60
- [MO_UB] = helper_ret_stb_mmu,
61
- [MO_LEUW] = helper_le_stw_mmu,
62
- [MO_LEUL] = helper_le_stl_mmu,
63
- [MO_LEQ] = helper_le_stq_mmu,
64
- [MO_BEUW] = helper_be_stw_mmu,
65
- [MO_BEUL] = helper_be_stl_mmu,
66
- [MO_BEQ] = helper_be_stq_mmu,
67
+static void * const qemu_st_helpers[4] = {
68
+ [MO_8] = helper_ret_stb_mmu,
69
+#ifdef HOST_WORDS_BIGENDIAN
70
+ [MO_16] = helper_be_stw_mmu,
71
+ [MO_32] = helper_be_stl_mmu,
72
+ [MO_64] = helper_be_stq_mmu,
73
+#else
74
+ [MO_16] = helper_le_stw_mmu,
75
+ [MO_32] = helper_le_stl_mmu,
76
+ [MO_64] = helper_le_stq_mmu,
77
+#endif
78
};
79
80
static inline void tcg_out_adr(TCGContext *s, TCGReg rd, const void *target)
81
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
82
tcg_out_mov(s, TARGET_LONG_BITS == 64, TCG_REG_X1, lb->addrlo_reg);
83
tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_X2, oi);
84
tcg_out_adr(s, TCG_REG_X3, lb->raddr);
85
- tcg_out_call(s, qemu_ld_helpers[opc & (MO_BSWAP | MO_SIZE)]);
86
+ tcg_out_call(s, qemu_ld_helpers[opc & MO_SIZE]);
87
if (opc & MO_SIGN) {
88
tcg_out_sxt(s, lb->type, size, lb->datalo_reg, TCG_REG_X0);
89
} else {
90
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
91
tcg_out_mov(s, size == MO_64, TCG_REG_X2, lb->datalo_reg);
92
tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_X3, oi);
93
tcg_out_adr(s, TCG_REG_X4, lb->raddr);
94
- tcg_out_call(s, qemu_st_helpers[opc & (MO_BSWAP | MO_SIZE)]);
95
+ tcg_out_call(s, qemu_st_helpers[opc & MO_SIZE]);
96
tcg_out_goto(s, lb->raddr);
97
return true;
98
}
99
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
100
TCGReg data_r, TCGReg addr_r,
101
TCGType otype, TCGReg off_r)
102
{
103
- const MemOp bswap = memop & MO_BSWAP;
104
+ /* Byte swapping is left to middle-end expansion. */
105
+ tcg_debug_assert((memop & MO_BSWAP) == 0);
106
107
switch (memop & MO_SSIZE) {
108
case MO_UB:
109
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
110
break;
111
case MO_UW:
112
tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
113
- if (bswap) {
114
- tcg_out_rev(s, TCG_TYPE_I32, MO_16, data_r, data_r);
115
- }
116
break;
117
case MO_SW:
118
- if (bswap) {
119
- tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
120
- tcg_out_rev(s, TCG_TYPE_I32, MO_16, data_r, data_r);
121
- tcg_out_sxt(s, ext, MO_16, data_r, data_r);
122
- } else {
123
- tcg_out_ldst_r(s, (ext ? I3312_LDRSHX : I3312_LDRSHW),
124
- data_r, addr_r, otype, off_r);
125
- }
126
+ tcg_out_ldst_r(s, (ext ? I3312_LDRSHX : I3312_LDRSHW),
127
+ data_r, addr_r, otype, off_r);
128
break;
129
case MO_UL:
130
tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
131
- if (bswap) {
132
- tcg_out_rev(s, TCG_TYPE_I32, MO_32, data_r, data_r);
133
- }
134
break;
135
case MO_SL:
136
- if (bswap) {
137
- tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
138
- tcg_out_rev(s, TCG_TYPE_I32, MO_32, data_r, data_r);
139
- tcg_out_sxt(s, TCG_TYPE_I64, MO_32, data_r, data_r);
140
- } else {
141
- tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, otype, off_r);
142
- }
143
+ tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, otype, off_r);
144
break;
145
case MO_Q:
146
tcg_out_ldst_r(s, I3312_LDRX, data_r, addr_r, otype, off_r);
147
- if (bswap) {
148
- tcg_out_rev(s, TCG_TYPE_I64, MO_64, data_r, data_r);
149
- }
150
break;
151
default:
152
tcg_abort();
153
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp memop,
154
TCGReg data_r, TCGReg addr_r,
155
TCGType otype, TCGReg off_r)
156
{
157
- const MemOp bswap = memop & MO_BSWAP;
158
+ /* Byte swapping is left to middle-end expansion. */
159
+ tcg_debug_assert((memop & MO_BSWAP) == 0);
160
161
switch (memop & MO_SIZE) {
162
case MO_8:
163
tcg_out_ldst_r(s, I3312_STRB, data_r, addr_r, otype, off_r);
164
break;
165
case MO_16:
166
- if (bswap && data_r != TCG_REG_XZR) {
167
- tcg_out_rev(s, TCG_TYPE_I32, MO_16, TCG_REG_TMP, data_r);
168
- data_r = TCG_REG_TMP;
169
- }
170
tcg_out_ldst_r(s, I3312_STRH, data_r, addr_r, otype, off_r);
171
break;
172
case MO_32:
173
- if (bswap && data_r != TCG_REG_XZR) {
174
- tcg_out_rev(s, TCG_TYPE_I32, MO_32, TCG_REG_TMP, data_r);
175
- data_r = TCG_REG_TMP;
176
- }
177
tcg_out_ldst_r(s, I3312_STRW, data_r, addr_r, otype, off_r);
178
break;
179
case MO_64:
180
- if (bswap && data_r != TCG_REG_XZR) {
181
- tcg_out_rev(s, TCG_TYPE_I64, MO_64, TCG_REG_TMP, data_r);
182
- data_r = TCG_REG_TMP;
183
- }
184
tcg_out_ldst_r(s, I3312_STRX, data_r, addr_r, otype, off_r);
185
break;
186
default:
187
--
65
--
188
2.25.1
66
2.34.1
189
67
190
68
diff view generated by jsdifflib
1
Migrate the is_jmp, tb and singlestep_enabled fields from
1
Add an interface to return the CPUTLBEntryFull struct
2
DisasContext into the base. Use pc_first instead of tb->pc.
2
that goes with the lookup. The result is not intended
3
Increment pc_next prior to decode, leaving the address of
3
to be valid across multiple lookups, so the user must
4
the current insn in dc->pc.
4
use the results immediately.
5
5
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
10
---
9
target/nios2/translate.c | 70 +++++++++++++++++++++-------------------
11
include/exec/exec-all.h | 15 +++++++++++++
10
1 file changed, 36 insertions(+), 34 deletions(-)
12
include/qemu/typedefs.h | 1 +
13
accel/tcg/cputlb.c | 47 +++++++++++++++++++++++++----------------
14
3 files changed, 45 insertions(+), 18 deletions(-)
11
15
12
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
16
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
13
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
14
--- a/target/nios2/translate.c
18
--- a/include/exec/exec-all.h
15
+++ b/target/nios2/translate.c
19
+++ b/include/exec/exec-all.h
16
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ int probe_access_flags(CPUArchState *env, target_ulong addr,
21
MMUAccessType access_type, int mmu_idx,
22
bool nonfault, void **phost, uintptr_t retaddr);
23
24
+#ifndef CONFIG_USER_ONLY
25
+/**
26
+ * probe_access_full:
27
+ * Like probe_access_flags, except also return into @pfull.
28
+ *
29
+ * The CPUTLBEntryFull structure returned via @pfull is transient
30
+ * and must be consumed or copied immediately, before any further
31
+ * access or changes to TLB @mmu_idx.
32
+ */
33
+int probe_access_full(CPUArchState *env, target_ulong addr,
34
+ MMUAccessType access_type, int mmu_idx,
35
+ bool nonfault, void **phost,
36
+ CPUTLBEntryFull **pfull, uintptr_t retaddr);
37
+#endif
38
+
39
#define CODE_GEN_ALIGN 16 /* must be >= of the size of a icache line */
40
41
/* Estimated block size for TB allocation. */
42
diff --git a/include/qemu/typedefs.h b/include/qemu/typedefs.h
43
index XXXXXXX..XXXXXXX 100644
44
--- a/include/qemu/typedefs.h
45
+++ b/include/qemu/typedefs.h
46
@@ -XXX,XX +XXX,XX @@ typedef struct ConfidentialGuestSupport ConfidentialGuestSupport;
47
typedef struct CPUAddressSpace CPUAddressSpace;
48
typedef struct CPUArchState CPUArchState;
49
typedef struct CPUState CPUState;
50
+typedef struct CPUTLBEntryFull CPUTLBEntryFull;
51
typedef struct DeviceListener DeviceListener;
52
typedef struct DeviceState DeviceState;
53
typedef struct DirtyBitmapSnapshot DirtyBitmapSnapshot;
54
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
55
index XXXXXXX..XXXXXXX 100644
56
--- a/accel/tcg/cputlb.c
57
+++ b/accel/tcg/cputlb.c
58
@@ -XXX,XX +XXX,XX @@ static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
59
static int probe_access_internal(CPUArchState *env, target_ulong addr,
60
int fault_size, MMUAccessType access_type,
61
int mmu_idx, bool nonfault,
62
- void **phost, uintptr_t retaddr)
63
+ void **phost, CPUTLBEntryFull **pfull,
64
+ uintptr_t retaddr)
65
{
66
uintptr_t index = tlb_index(env, mmu_idx, addr);
67
CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr);
68
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
69
mmu_idx, nonfault, retaddr)) {
70
/* Non-faulting page table read failed. */
71
*phost = NULL;
72
+ *pfull = NULL;
73
return TLB_INVALID_MASK;
74
}
75
76
/* TLB resize via tlb_fill may have moved the entry. */
77
+ index = tlb_index(env, mmu_idx, addr);
78
entry = tlb_entry(env, mmu_idx, addr);
79
80
/*
81
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
17
}
82
}
18
83
flags &= tlb_addr;
19
typedef struct DisasContext {
84
20
+ DisasContextBase base;
85
+ *pfull = &env_tlb(env)->d[mmu_idx].fulltlb[index];
21
TCGv_i32 zero;
86
+
22
- int is_jmp;
87
/* Fold all "mmio-like" bits into TLB_MMIO. This is not RAM. */
23
target_ulong pc;
88
if (unlikely(flags & ~(TLB_WATCHPOINT | TLB_NOTDIRTY))) {
24
- TranslationBlock *tb;
89
*phost = NULL;
25
int mem_idx;
90
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
26
- bool singlestep_enabled;
91
return flags;
27
} DisasContext;
28
29
static TCGv cpu_R[NUM_CORE_REGS];
30
@@ -XXX,XX +XXX,XX @@ static void t_gen_helper_raise_exception(DisasContext *dc,
31
tcg_gen_movi_tl(cpu_R[R_PC], dc->pc);
32
gen_helper_raise_exception(cpu_env, tmp);
33
tcg_temp_free_i32(tmp);
34
- dc->is_jmp = DISAS_NORETURN;
35
+ dc->base.is_jmp = DISAS_NORETURN;
36
}
92
}
37
93
38
static bool use_goto_tb(DisasContext *dc, uint32_t dest)
94
-int probe_access_flags(CPUArchState *env, target_ulong addr,
95
- MMUAccessType access_type, int mmu_idx,
96
- bool nonfault, void **phost, uintptr_t retaddr)
97
+int probe_access_full(CPUArchState *env, target_ulong addr,
98
+ MMUAccessType access_type, int mmu_idx,
99
+ bool nonfault, void **phost, CPUTLBEntryFull **pfull,
100
+ uintptr_t retaddr)
39
{
101
{
40
- if (unlikely(dc->singlestep_enabled)) {
102
- int flags;
41
+ if (unlikely(dc->base.singlestep_enabled)) {
103
-
42
return false;
104
- flags = probe_access_internal(env, addr, 0, access_type, mmu_idx,
105
- nonfault, phost, retaddr);
106
+ int flags = probe_access_internal(env, addr, 0, access_type, mmu_idx,
107
+ nonfault, phost, pfull, retaddr);
108
109
/* Handle clean RAM pages. */
110
if (unlikely(flags & TLB_NOTDIRTY)) {
111
- uintptr_t index = tlb_index(env, mmu_idx, addr);
112
- CPUTLBEntryFull *full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
113
-
114
- notdirty_write(env_cpu(env), addr, 1, full, retaddr);
115
+ notdirty_write(env_cpu(env), addr, 1, *pfull, retaddr);
116
flags &= ~TLB_NOTDIRTY;
43
}
117
}
44
118
45
#ifndef CONFIG_USER_ONLY
119
return flags;
46
- return (dc->tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
120
}
47
+ return (dc->base.pc_first & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
121
48
#else
122
+int probe_access_flags(CPUArchState *env, target_ulong addr,
49
return true;
123
+ MMUAccessType access_type, int mmu_idx,
50
#endif
124
+ bool nonfault, void **phost, uintptr_t retaddr)
51
@@ -XXX,XX +XXX,XX @@ static bool use_goto_tb(DisasContext *dc, uint32_t dest)
125
+{
52
126
+ CPUTLBEntryFull *full;
53
static void gen_goto_tb(DisasContext *dc, int n, uint32_t dest)
127
+
128
+ return probe_access_full(env, addr, access_type, mmu_idx,
129
+ nonfault, phost, &full, retaddr);
130
+}
131
+
132
void *probe_access(CPUArchState *env, target_ulong addr, int size,
133
MMUAccessType access_type, int mmu_idx, uintptr_t retaddr)
54
{
134
{
55
- TranslationBlock *tb = dc->tb;
135
+ CPUTLBEntryFull *full;
56
+ const TranslationBlock *tb = dc->base.tb;
136
void *host;
57
137
int flags;
58
if (use_goto_tb(dc, dest)) {
138
59
tcg_gen_goto_tb(n);
139
g_assert(-(addr | TARGET_PAGE_MASK) >= size);
60
@@ -XXX,XX +XXX,XX @@ static void gen_excp(DisasContext *dc, uint32_t code, uint32_t flags)
140
61
141
flags = probe_access_internal(env, addr, size, access_type, mmu_idx,
62
static void gen_check_supervisor(DisasContext *dc)
142
- false, &host, retaddr);
143
+ false, &host, &full, retaddr);
144
145
/* Per the interface, size == 0 merely faults the access. */
146
if (size == 0) {
147
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
148
}
149
150
if (unlikely(flags & (TLB_NOTDIRTY | TLB_WATCHPOINT))) {
151
- uintptr_t index = tlb_index(env, mmu_idx, addr);
152
- CPUTLBEntryFull *full = &env_tlb(env)->d[mmu_idx].fulltlb[index];
153
-
154
/* Handle watchpoints. */
155
if (flags & TLB_WATCHPOINT) {
156
int wp_access = (access_type == MMU_DATA_STORE
157
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
158
void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
159
MMUAccessType access_type, int mmu_idx)
63
{
160
{
64
- if (dc->tb->flags & CR_STATUS_U) {
161
+ CPUTLBEntryFull *full;
65
+ if (dc->base.tb->flags & CR_STATUS_U) {
162
void *host;
66
/* CPU in user mode, privileged instruction called, stop. */
163
int flags;
67
t_gen_helper_raise_exception(dc, EXCP_SUPERI);
164
68
}
165
flags = probe_access_internal(env, addr, 0, access_type,
69
@@ -XXX,XX +XXX,XX @@ static void jmpi(DisasContext *dc, uint32_t code, uint32_t flags)
166
- mmu_idx, true, &host, 0);
167
+ mmu_idx, true, &host, &full, 0);
168
169
/* No combination of flags are expected by the caller. */
170
return flags ? NULL : host;
171
@@ -XXX,XX +XXX,XX @@ void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
172
tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
173
void **hostp)
70
{
174
{
71
J_TYPE(instr, code);
175
+ CPUTLBEntryFull *full;
72
gen_goto_tb(dc, 0, (dc->pc & 0xF0000000) | (instr.imm26 << 2));
176
void *p;
73
- dc->is_jmp = DISAS_NORETURN;
177
74
+ dc->base.is_jmp = DISAS_NORETURN;
178
(void)probe_access_internal(env, addr, 1, MMU_INST_FETCH,
75
}
179
- cpu_mmu_index(env, true), false, &p, 0);
76
180
+ cpu_mmu_index(env, true), false, &p, &full, 0);
77
static void call(DisasContext *dc, uint32_t code, uint32_t flags)
181
if (p == NULL) {
78
@@ -XXX,XX +XXX,XX @@ static void br(DisasContext *dc, uint32_t code, uint32_t flags)
182
return -1;
79
I_TYPE(instr, code);
80
81
gen_goto_tb(dc, 0, dc->pc + 4 + (instr.imm16.s & -4));
82
- dc->is_jmp = DISAS_NORETURN;
83
+ dc->base.is_jmp = DISAS_NORETURN;
84
}
85
86
static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
87
@@ -XXX,XX +XXX,XX @@ static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
88
gen_goto_tb(dc, 0, dc->pc + 4);
89
gen_set_label(l1);
90
gen_goto_tb(dc, 1, dc->pc + 4 + (instr.imm16.s & -4));
91
- dc->is_jmp = DISAS_NORETURN;
92
+ dc->base.is_jmp = DISAS_NORETURN;
93
}
94
95
/* Comparison instructions */
96
@@ -XXX,XX +XXX,XX @@ static void eret(DisasContext *dc, uint32_t code, uint32_t flags)
97
tcg_gen_mov_tl(cpu_R[CR_STATUS], cpu_R[CR_ESTATUS]);
98
tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_EA]);
99
100
- dc->is_jmp = DISAS_JUMP;
101
+ dc->base.is_jmp = DISAS_JUMP;
102
}
103
104
/* PC <- ra */
105
@@ -XXX,XX +XXX,XX @@ static void ret(DisasContext *dc, uint32_t code, uint32_t flags)
106
{
107
tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_RA]);
108
109
- dc->is_jmp = DISAS_JUMP;
110
+ dc->base.is_jmp = DISAS_JUMP;
111
}
112
113
/* PC <- ba */
114
@@ -XXX,XX +XXX,XX @@ static void bret(DisasContext *dc, uint32_t code, uint32_t flags)
115
{
116
tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_BA]);
117
118
- dc->is_jmp = DISAS_JUMP;
119
+ dc->base.is_jmp = DISAS_JUMP;
120
}
121
122
/* PC <- rA */
123
@@ -XXX,XX +XXX,XX @@ static void jmp(DisasContext *dc, uint32_t code, uint32_t flags)
124
125
tcg_gen_mov_tl(cpu_R[R_PC], load_gpr(dc, instr.a));
126
127
- dc->is_jmp = DISAS_JUMP;
128
+ dc->base.is_jmp = DISAS_JUMP;
129
}
130
131
/* rC <- PC + 4 */
132
@@ -XXX,XX +XXX,XX @@ static void callr(DisasContext *dc, uint32_t code, uint32_t flags)
133
tcg_gen_mov_tl(cpu_R[R_PC], load_gpr(dc, instr.a));
134
tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
135
136
- dc->is_jmp = DISAS_JUMP;
137
+ dc->base.is_jmp = DISAS_JUMP;
138
}
139
140
/* rC <- ctlN */
141
@@ -XXX,XX +XXX,XX @@ static void wrctl(DisasContext *dc, uint32_t code, uint32_t flags)
142
/* If interrupts were enabled using WRCTL, trigger them. */
143
#if !defined(CONFIG_USER_ONLY)
144
if ((instr.imm5 + CR_BASE) == CR_STATUS) {
145
- if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
146
+ if (tb_cflags(dc->base.tb) & CF_USE_ICOUNT) {
147
gen_io_start();
148
}
149
gen_helper_check_interrupts(cpu_env);
150
- dc->is_jmp = DISAS_UPDATE;
151
+ dc->base.is_jmp = DISAS_UPDATE;
152
}
153
#endif
154
}
155
@@ -XXX,XX +XXX,XX @@ static void gen_exception(DisasContext *dc, uint32_t excp)
156
tcg_gen_movi_tl(cpu_R[R_PC], dc->pc);
157
gen_helper_raise_exception(cpu_env, tmp);
158
tcg_temp_free_i32(tmp);
159
- dc->is_jmp = DISAS_NORETURN;
160
+ dc->base.is_jmp = DISAS_NORETURN;
161
}
162
163
/* generate intermediate code for basic block 'tb'. */
164
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
165
int num_insns;
166
167
/* Initialize DC */
168
- dc->is_jmp = DISAS_NEXT;
169
- dc->pc = tb->pc;
170
- dc->tb = tb;
171
+
172
+ dc->base.tb = tb;
173
+ dc->base.singlestep_enabled = cs->singlestep_enabled;
174
+ dc->base.is_jmp = DISAS_NEXT;
175
+ dc->base.pc_first = tb->pc;
176
+ dc->base.pc_next = tb->pc;
177
+
178
dc->mem_idx = cpu_mmu_index(env, false);
179
- dc->singlestep_enabled = cs->singlestep_enabled;
180
181
/* Set up instruction counts */
182
num_insns = 0;
183
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
184
185
gen_tb_start(tb);
186
do {
187
- tcg_gen_insn_start(dc->pc);
188
+ tcg_gen_insn_start(dc->base.pc_next);
189
num_insns++;
190
191
- if (unlikely(cpu_breakpoint_test(cs, dc->pc, BP_ANY))) {
192
+ if (unlikely(cpu_breakpoint_test(cs, dc->base.pc_next, BP_ANY))) {
193
gen_exception(dc, EXCP_DEBUG);
194
/* The address covered by the breakpoint must be included in
195
[tb->pc, tb->pc + tb->size) in order to for it to be
196
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
197
gen_io_start();
198
}
199
200
+ dc->pc = dc->base.pc_next;
201
+ dc->base.pc_next += 4;
202
+
203
/* Decode an instruction */
204
handle_instruction(dc, env);
205
206
- dc->pc += 4;
207
-
208
/* Translation stops when a conditional branch is encountered.
209
* Otherwise the subsequent code could get translated several times.
210
* Also stop translation when a page boundary is reached. This
211
* ensures prefetch aborts occur at the right place. */
212
- } while (!dc->is_jmp &&
213
+ } while (!dc->base.is_jmp &&
214
!tcg_op_buf_full() &&
215
num_insns < max_insns);
216
217
/* Indicate where the next block should start */
218
- switch (dc->is_jmp) {
219
+ switch (dc->base.is_jmp) {
220
case DISAS_NEXT:
221
case DISAS_UPDATE:
222
/* Save the current PC back into the CPU register */
223
- tcg_gen_movi_tl(cpu_R[R_PC], dc->pc);
224
+ tcg_gen_movi_tl(cpu_R[R_PC], dc->base.pc_next);
225
tcg_gen_exit_tb(NULL, 0);
226
break;
227
228
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
229
gen_tb_end(tb, num_insns);
230
231
/* Mark instruction starts for the final generated instruction */
232
- tb->size = dc->pc - tb->pc;
233
+ tb->size = dc->base.pc_next - dc->base.pc_first;
234
tb->icount = num_insns;
235
236
#ifdef DEBUG_DISAS
237
if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
238
- && qemu_log_in_addr_range(tb->pc)) {
239
+ && qemu_log_in_addr_range(dc->base.pc_first)) {
240
FILE *logfile = qemu_log_lock();
241
- qemu_log("IN: %s\n", lookup_symbol(tb->pc));
242
- log_target_disas(cs, tb->pc, dc->pc - tb->pc);
243
+ qemu_log("IN: %s\n", lookup_symbol(dc->base.pc_first));
244
+ log_target_disas(cs, tb->pc, tb->size);
245
qemu_log("\n");
246
qemu_log_unlock(logfile);
247
}
183
}
248
--
184
--
249
2.25.1
185
2.34.1
250
186
251
187
diff view generated by jsdifflib
1
The new bswap flags can implement the semantics exactly.
1
Now that we have collected all of the page data into
2
CPUTLBEntryFull, provide an interface to record that
3
all in one go, instead of using 4 arguments. This interface
4
allows CPUTLBEntryFull to be extended without having to
5
change the number of arguments.
2
6
7
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
11
---
7
target/arm/translate.c | 4 +---
12
include/exec/cpu-defs.h | 14 +++++++++++
8
1 file changed, 1 insertion(+), 3 deletions(-)
13
include/exec/exec-all.h | 22 ++++++++++++++++++
14
accel/tcg/cputlb.c | 51 ++++++++++++++++++++++++++---------------
15
3 files changed, 69 insertions(+), 18 deletions(-)
9
16
10
diff --git a/target/arm/translate.c b/target/arm/translate.c
17
diff --git a/include/exec/cpu-defs.h b/include/exec/cpu-defs.h
11
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
12
--- a/target/arm/translate.c
19
--- a/include/exec/cpu-defs.h
13
+++ b/target/arm/translate.c
20
+++ b/include/exec/cpu-defs.h
14
@@ -XXX,XX +XXX,XX @@ void gen_rev16(TCGv_i32 dest, TCGv_i32 var)
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUTLBEntryFull {
15
/* Byteswap low halfword and sign extend. */
22
* + the offset within the target MemoryRegion (otherwise)
16
static void gen_revsh(TCGv_i32 dest, TCGv_i32 var)
23
*/
24
hwaddr xlat_section;
25
+
26
+ /*
27
+ * @phys_addr contains the physical address in the address space
28
+ * given by cpu_asidx_from_attrs(cpu, @attrs).
29
+ */
30
+ hwaddr phys_addr;
31
+
32
+ /* @attrs contains the memory transaction attributes for the page. */
33
MemTxAttrs attrs;
34
+
35
+ /* @prot contains the complete protections for the page. */
36
+ uint8_t prot;
37
+
38
+ /* @lg_page_size contains the log2 of the page size. */
39
+ uint8_t lg_page_size;
40
} CPUTLBEntryFull;
41
42
/*
43
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
44
index XXXXXXX..XXXXXXX 100644
45
--- a/include/exec/exec-all.h
46
+++ b/include/exec/exec-all.h
47
@@ -XXX,XX +XXX,XX @@ void tlb_flush_range_by_mmuidx_all_cpus_synced(CPUState *cpu,
48
uint16_t idxmap,
49
unsigned bits);
50
51
+/**
52
+ * tlb_set_page_full:
53
+ * @cpu: CPU context
54
+ * @mmu_idx: mmu index of the tlb to modify
55
+ * @vaddr: virtual address of the entry to add
56
+ * @full: the details of the tlb entry
57
+ *
58
+ * Add an entry to @cpu tlb index @mmu_idx. All of the fields of
59
+ * @full must be filled, except for xlat_section, and constitute
60
+ * the complete description of the translated page.
61
+ *
62
+ * This is generally called by the target tlb_fill function after
63
+ * having performed a successful page table walk to find the physical
64
+ * address and attributes for the translation.
65
+ *
66
+ * At most one entry for a given virtual address is permitted. Only a
67
+ * single TARGET_PAGE_SIZE region is mapped; @full->lg_page_size is only
68
+ * used by tlb_flush_page.
69
+ */
70
+void tlb_set_page_full(CPUState *cpu, int mmu_idx, target_ulong vaddr,
71
+ CPUTLBEntryFull *full);
72
+
73
/**
74
* tlb_set_page_with_attrs:
75
* @cpu: CPU to add this TLB entry for
76
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
77
index XXXXXXX..XXXXXXX 100644
78
--- a/accel/tcg/cputlb.c
79
+++ b/accel/tcg/cputlb.c
80
@@ -XXX,XX +XXX,XX @@ static void tlb_add_large_page(CPUArchState *env, int mmu_idx,
81
env_tlb(env)->d[mmu_idx].large_page_mask = lp_mask;
82
}
83
84
-/* Add a new TLB entry. At most one entry for a given virtual address
85
+/*
86
+ * Add a new TLB entry. At most one entry for a given virtual address
87
* is permitted. Only a single TARGET_PAGE_SIZE region is mapped, the
88
* supplied size is only used by tlb_flush_page.
89
*
90
* Called from TCG-generated code, which is under an RCU read-side
91
* critical section.
92
*/
93
-void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
94
- hwaddr paddr, MemTxAttrs attrs, int prot,
95
- int mmu_idx, target_ulong size)
96
+void tlb_set_page_full(CPUState *cpu, int mmu_idx,
97
+ target_ulong vaddr, CPUTLBEntryFull *full)
17
{
98
{
18
- tcg_gen_ext16u_i32(var, var);
99
CPUArchState *env = cpu->env_ptr;
19
- tcg_gen_bswap16_i32(var, var, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
100
CPUTLB *tlb = env_tlb(env);
20
- tcg_gen_ext16s_i32(dest, var);
101
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
21
+ tcg_gen_bswap16_i32(var, var, TCG_BSWAP_OS);
102
CPUTLBEntry *te, tn;
103
hwaddr iotlb, xlat, sz, paddr_page;
104
target_ulong vaddr_page;
105
- int asidx = cpu_asidx_from_attrs(cpu, attrs);
106
- int wp_flags;
107
+ int asidx, wp_flags, prot;
108
bool is_ram, is_romd;
109
110
assert_cpu_is_self(cpu);
111
112
- if (size <= TARGET_PAGE_SIZE) {
113
+ if (full->lg_page_size <= TARGET_PAGE_BITS) {
114
sz = TARGET_PAGE_SIZE;
115
} else {
116
- tlb_add_large_page(env, mmu_idx, vaddr, size);
117
- sz = size;
118
+ sz = (hwaddr)1 << full->lg_page_size;
119
+ tlb_add_large_page(env, mmu_idx, vaddr, sz);
120
}
121
vaddr_page = vaddr & TARGET_PAGE_MASK;
122
- paddr_page = paddr & TARGET_PAGE_MASK;
123
+ paddr_page = full->phys_addr & TARGET_PAGE_MASK;
124
125
+ prot = full->prot;
126
+ asidx = cpu_asidx_from_attrs(cpu, full->attrs);
127
section = address_space_translate_for_iotlb(cpu, asidx, paddr_page,
128
- &xlat, &sz, attrs, &prot);
129
+ &xlat, &sz, full->attrs, &prot);
130
assert(sz >= TARGET_PAGE_SIZE);
131
132
tlb_debug("vaddr=" TARGET_FMT_lx " paddr=0x" TARGET_FMT_plx
133
" prot=%x idx=%d\n",
134
- vaddr, paddr, prot, mmu_idx);
135
+ vaddr, full->phys_addr, prot, mmu_idx);
136
137
address = vaddr_page;
138
- if (size < TARGET_PAGE_SIZE) {
139
+ if (full->lg_page_size < TARGET_PAGE_BITS) {
140
/* Repeat the MMU check and TLB fill on every access. */
141
address |= TLB_INVALID_MASK;
142
}
143
- if (attrs.byte_swap) {
144
+ if (full->attrs.byte_swap) {
145
address |= TLB_BSWAP;
146
}
147
148
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
149
* subtract here is that of the page base, and not the same as the
150
* vaddr we add back in io_readx()/io_writex()/get_page_addr_code().
151
*/
152
+ desc->fulltlb[index] = *full;
153
desc->fulltlb[index].xlat_section = iotlb - vaddr_page;
154
- desc->fulltlb[index].attrs = attrs;
155
+ desc->fulltlb[index].phys_addr = paddr_page;
156
+ desc->fulltlb[index].prot = prot;
157
158
/* Now calculate the new entry */
159
tn.addend = addend - vaddr_page;
160
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
161
qemu_spin_unlock(&tlb->c.lock);
22
}
162
}
23
163
24
/* Dual 16-bit add. Result placed in t0 and t1 is marked as dead.
164
-/* Add a new TLB entry, but without specifying the memory
165
- * transaction attributes to be used.
166
- */
167
+void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
168
+ hwaddr paddr, MemTxAttrs attrs, int prot,
169
+ int mmu_idx, target_ulong size)
170
+{
171
+ CPUTLBEntryFull full = {
172
+ .phys_addr = paddr,
173
+ .attrs = attrs,
174
+ .prot = prot,
175
+ .lg_page_size = ctz64(size)
176
+ };
177
+
178
+ assert(is_power_of_2(size));
179
+ tlb_set_page_full(cpu, mmu_idx, vaddr, &full);
180
+}
181
+
182
void tlb_set_page(CPUState *cpu, target_ulong vaddr,
183
hwaddr paddr, int prot,
184
int mmu_idx, target_ulong size)
25
--
185
--
26
2.25.1
186
2.34.1
27
187
28
188
diff view generated by jsdifflib
1
Migrate the bstate, tb and singlestep_enabled fields
1
Allow the target to cache items from the guest page tables.
2
from DisasContext into the base.
3
2
4
Tested-by: Michael Rolnik <mrolnik@gmail.com>
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
7
---
9
target/avr/translate.c | 58 +++++++++++++++++++++---------------------
8
include/exec/cpu-defs.h | 9 +++++++++
10
1 file changed, 29 insertions(+), 29 deletions(-)
9
1 file changed, 9 insertions(+)
11
10
12
diff --git a/target/avr/translate.c b/target/avr/translate.c
11
diff --git a/include/exec/cpu-defs.h b/include/exec/cpu-defs.h
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/target/avr/translate.c
13
--- a/include/exec/cpu-defs.h
15
+++ b/target/avr/translate.c
14
+++ b/include/exec/cpu-defs.h
16
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext DisasContext;
15
@@ -XXX,XX +XXX,XX @@ typedef struct CPUTLBEntryFull {
17
16
18
/* This is the state at translation time. */
17
/* @lg_page_size contains the log2 of the page size. */
19
struct DisasContext {
18
uint8_t lg_page_size;
20
- TranslationBlock *tb;
19
+
21
+ DisasContextBase base;
20
+ /*
22
21
+ * Allow target-specific additions to this structure.
23
CPUAVRState *env;
22
+ * This may be used to cache items from the guest cpu
24
CPUState *cs;
23
+ * page tables for later use by the implementation.
25
@@ -XXX,XX +XXX,XX @@ struct DisasContext {
24
+ */
26
25
+#ifdef TARGET_PAGE_ENTRY_EXTRA
27
/* Routine used to access memory */
26
+ TARGET_PAGE_ENTRY_EXTRA
28
int memidx;
27
+#endif
29
- int bstate;
28
} CPUTLBEntryFull;
30
- int singlestep;
31
32
/*
33
* some AVR instructions can make the following instruction to be skipped
34
@@ -XXX,XX +XXX,XX @@ static bool avr_have_feature(DisasContext *ctx, int feature)
35
{
36
if (!avr_feature(ctx->env, feature)) {
37
gen_helper_unsupported(cpu_env);
38
- ctx->bstate = DISAS_NORETURN;
39
+ ctx->base.is_jmp = DISAS_NORETURN;
40
return false;
41
}
42
return true;
43
@@ -XXX,XX +XXX,XX @@ static void gen_jmp_ez(DisasContext *ctx)
44
{
45
tcg_gen_deposit_tl(cpu_pc, cpu_r[30], cpu_r[31], 8, 8);
46
tcg_gen_or_tl(cpu_pc, cpu_pc, cpu_eind);
47
- ctx->bstate = DISAS_LOOKUP;
48
+ ctx->base.is_jmp = DISAS_LOOKUP;
49
}
50
51
static void gen_jmp_z(DisasContext *ctx)
52
{
53
tcg_gen_deposit_tl(cpu_pc, cpu_r[30], cpu_r[31], 8, 8);
54
- ctx->bstate = DISAS_LOOKUP;
55
+ ctx->base.is_jmp = DISAS_LOOKUP;
56
}
57
58
static void gen_push_ret(DisasContext *ctx, int ret)
59
@@ -XXX,XX +XXX,XX @@ static void gen_pop_ret(DisasContext *ctx, TCGv ret)
60
61
static void gen_goto_tb(DisasContext *ctx, int n, target_ulong dest)
62
{
63
- TranslationBlock *tb = ctx->tb;
64
+ const TranslationBlock *tb = ctx->base.tb;
65
66
- if (ctx->singlestep == 0) {
67
+ if (!ctx->base.singlestep_enabled) {
68
tcg_gen_goto_tb(n);
69
tcg_gen_movi_i32(cpu_pc, dest);
70
tcg_gen_exit_tb(tb, n);
71
@@ -XXX,XX +XXX,XX @@ static void gen_goto_tb(DisasContext *ctx, int n, target_ulong dest)
72
gen_helper_debug(cpu_env);
73
tcg_gen_exit_tb(NULL, 0);
74
}
75
- ctx->bstate = DISAS_NORETURN;
76
+ ctx->base.is_jmp = DISAS_NORETURN;
77
}
78
29
79
/*
30
/*
80
@@ -XXX,XX +XXX,XX @@ static bool trans_RET(DisasContext *ctx, arg_RET *a)
81
{
82
gen_pop_ret(ctx, cpu_pc);
83
84
- ctx->bstate = DISAS_LOOKUP;
85
+ ctx->base.is_jmp = DISAS_LOOKUP;
86
return true;
87
}
88
89
@@ -XXX,XX +XXX,XX @@ static bool trans_RETI(DisasContext *ctx, arg_RETI *a)
90
tcg_gen_movi_tl(cpu_If, 1);
91
92
/* Need to return to main loop to re-evaluate interrupts. */
93
- ctx->bstate = DISAS_EXIT;
94
+ ctx->base.is_jmp = DISAS_EXIT;
95
return true;
96
}
97
98
@@ -XXX,XX +XXX,XX @@ static bool trans_BRBC(DisasContext *ctx, arg_BRBC *a)
99
gen_goto_tb(ctx, 0, ctx->npc + a->imm);
100
gen_set_label(not_taken);
101
102
- ctx->bstate = DISAS_CHAIN;
103
+ ctx->base.is_jmp = DISAS_CHAIN;
104
return true;
105
}
106
107
@@ -XXX,XX +XXX,XX @@ static bool trans_BRBS(DisasContext *ctx, arg_BRBS *a)
108
gen_goto_tb(ctx, 0, ctx->npc + a->imm);
109
gen_set_label(not_taken);
110
111
- ctx->bstate = DISAS_CHAIN;
112
+ ctx->base.is_jmp = DISAS_CHAIN;
113
return true;
114
}
115
116
@@ -XXX,XX +XXX,XX @@ static TCGv gen_get_zaddr(void)
117
*/
118
static void gen_data_store(DisasContext *ctx, TCGv data, TCGv addr)
119
{
120
- if (ctx->tb->flags & TB_FLAGS_FULL_ACCESS) {
121
+ if (ctx->base.tb->flags & TB_FLAGS_FULL_ACCESS) {
122
gen_helper_fullwr(cpu_env, data, addr);
123
} else {
124
tcg_gen_qemu_st8(data, addr, MMU_DATA_IDX); /* mem[addr] = data */
125
@@ -XXX,XX +XXX,XX @@ static void gen_data_store(DisasContext *ctx, TCGv data, TCGv addr)
126
127
static void gen_data_load(DisasContext *ctx, TCGv data, TCGv addr)
128
{
129
- if (ctx->tb->flags & TB_FLAGS_FULL_ACCESS) {
130
+ if (ctx->base.tb->flags & TB_FLAGS_FULL_ACCESS) {
131
gen_helper_fullrd(data, cpu_env, addr);
132
} else {
133
tcg_gen_qemu_ld8u(data, addr, MMU_DATA_IDX); /* data = mem[addr] */
134
@@ -XXX,XX +XXX,XX @@ static bool trans_BREAK(DisasContext *ctx, arg_BREAK *a)
135
#ifdef BREAKPOINT_ON_BREAK
136
tcg_gen_movi_tl(cpu_pc, ctx->npc - 1);
137
gen_helper_debug(cpu_env);
138
- ctx->bstate = DISAS_EXIT;
139
+ ctx->base.is_jmp = DISAS_EXIT;
140
#else
141
/* NOP */
142
#endif
143
@@ -XXX,XX +XXX,XX @@ static bool trans_NOP(DisasContext *ctx, arg_NOP *a)
144
static bool trans_SLEEP(DisasContext *ctx, arg_SLEEP *a)
145
{
146
gen_helper_sleep(cpu_env);
147
- ctx->bstate = DISAS_NORETURN;
148
+ ctx->base.is_jmp = DISAS_NORETURN;
149
return true;
150
}
151
152
@@ -XXX,XX +XXX,XX @@ static void translate(DisasContext *ctx)
153
154
if (!decode_insn(ctx, opcode)) {
155
gen_helper_unsupported(cpu_env);
156
- ctx->bstate = DISAS_NORETURN;
157
+ ctx->base.is_jmp = DISAS_NORETURN;
158
}
159
}
160
161
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
162
{
163
CPUAVRState *env = cs->env_ptr;
164
DisasContext ctx = {
165
- .tb = tb,
166
+ .base.tb = tb,
167
+ .base.is_jmp = DISAS_NEXT,
168
+ .base.pc_first = tb->pc,
169
+ .base.pc_next = tb->pc,
170
+ .base.singlestep_enabled = cs->singlestep_enabled,
171
.cs = cs,
172
.env = env,
173
.memidx = 0,
174
- .bstate = DISAS_NEXT,
175
.skip_cond = TCG_COND_NEVER,
176
- .singlestep = cs->singlestep_enabled,
177
};
178
target_ulong pc_start = tb->pc / 2;
179
int num_insns = 0;
180
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
181
*/
182
max_insns = 1;
183
}
184
- if (ctx.singlestep) {
185
+ if (ctx.base.singlestep_enabled) {
186
max_insns = 1;
187
}
188
189
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
190
* b main - sets breakpoint at address 0x00000100 (code)
191
* b *0x100 - sets breakpoint at address 0x00800100 (data)
192
*/
193
- if (unlikely(!ctx.singlestep &&
194
+ if (unlikely(!ctx.base.singlestep_enabled &&
195
(cpu_breakpoint_test(cs, OFFSET_CODE + ctx.npc * 2, BP_ANY) ||
196
cpu_breakpoint_test(cs, OFFSET_DATA + ctx.npc * 2, BP_ANY)))) {
197
canonicalize_skip(&ctx);
198
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
199
if (skip_label) {
200
canonicalize_skip(&ctx);
201
gen_set_label(skip_label);
202
- if (ctx.bstate == DISAS_NORETURN) {
203
- ctx.bstate = DISAS_CHAIN;
204
+ if (ctx.base.is_jmp == DISAS_NORETURN) {
205
+ ctx.base.is_jmp = DISAS_CHAIN;
206
}
207
}
208
- } while (ctx.bstate == DISAS_NEXT
209
+ } while (ctx.base.is_jmp == DISAS_NEXT
210
&& num_insns < max_insns
211
&& (ctx.npc - pc_start) * 2 < TARGET_PAGE_SIZE - 4
212
&& !tcg_op_buf_full());
213
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
214
215
bool nonconst_skip = canonicalize_skip(&ctx);
216
217
- switch (ctx.bstate) {
218
+ switch (ctx.base.is_jmp) {
219
case DISAS_NORETURN:
220
assert(!nonconst_skip);
221
break;
222
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
223
tcg_gen_movi_tl(cpu_pc, ctx.npc);
224
/* fall through */
225
case DISAS_LOOKUP:
226
- if (!ctx.singlestep) {
227
+ if (!ctx.base.singlestep_enabled) {
228
tcg_gen_lookup_and_goto_ptr();
229
break;
230
}
231
/* fall through */
232
case DISAS_EXIT:
233
- if (ctx.singlestep) {
234
+ if (ctx.base.singlestep_enabled) {
235
gen_helper_debug(cpu_env);
236
} else {
237
tcg_gen_exit_tb(NULL, 0);
238
--
31
--
239
2.25.1
32
2.34.1
240
33
241
34
diff view generated by jsdifflib
1
Combine the three bswap16 routines, and differentiate via the flags.
1
This bitmap is created and discarded immediately.
2
Use the correct flags combination from the load/store routines, and
2
We gain nothing by its existence.
3
pass along the constant parameter from tcg_out_op.
4
3
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-Id: <20220822232338.1727934-2-richard.henderson@linaro.org>
7
---
7
---
8
tcg/arm/tcg-target.c.inc | 101 ++++++++++++++++++++++++---------------
8
accel/tcg/translate-all.c | 78 ++-------------------------------------
9
1 file changed, 63 insertions(+), 38 deletions(-)
9
1 file changed, 4 insertions(+), 74 deletions(-)
10
10
11
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
11
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
12
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/arm/tcg-target.c.inc
13
--- a/accel/tcg/translate-all.c
14
+++ b/tcg/arm/tcg-target.c.inc
14
+++ b/accel/tcg/translate-all.c
15
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_ext16u(TCGContext *s, int cond,
15
@@ -XXX,XX +XXX,XX @@
16
#define assert_memory_lock() tcg_debug_assert(have_mmap_lock())
17
#endif
18
19
-#define SMC_BITMAP_USE_THRESHOLD 10
20
-
21
typedef struct PageDesc {
22
/* list of TBs intersecting this ram page */
23
uintptr_t first_tb;
24
-#ifdef CONFIG_SOFTMMU
25
- /* in order to optimize self modifying code, we count the number
26
- of lookups we do to a given page to use a bitmap */
27
- unsigned long *code_bitmap;
28
- unsigned int code_write_count;
29
-#else
30
+#ifdef CONFIG_USER_ONLY
31
unsigned long flags;
32
void *target_data;
33
#endif
34
-#ifndef CONFIG_USER_ONLY
35
+#ifdef CONFIG_SOFTMMU
36
QemuSpin lock;
37
#endif
38
} PageDesc;
39
@@ -XXX,XX +XXX,XX @@ void tb_htable_init(void)
40
qht_init(&tb_ctx.htable, tb_cmp, CODE_GEN_HTABLE_SIZE, mode);
41
}
42
43
-/* call with @p->lock held */
44
-static inline void invalidate_page_bitmap(PageDesc *p)
45
-{
46
- assert_page_locked(p);
47
-#ifdef CONFIG_SOFTMMU
48
- g_free(p->code_bitmap);
49
- p->code_bitmap = NULL;
50
- p->code_write_count = 0;
51
-#endif
52
-}
53
-
54
/* Set to NULL all the 'first_tb' fields in all PageDescs. */
55
static void page_flush_tb_1(int level, void **lp)
56
{
57
@@ -XXX,XX +XXX,XX @@ static void page_flush_tb_1(int level, void **lp)
58
for (i = 0; i < V_L2_SIZE; ++i) {
59
page_lock(&pd[i]);
60
pd[i].first_tb = (uintptr_t)NULL;
61
- invalidate_page_bitmap(pd + i);
62
page_unlock(&pd[i]);
63
}
64
} else {
65
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
66
if (rm_from_page_list) {
67
p = page_find(tb->page_addr[0] >> TARGET_PAGE_BITS);
68
tb_page_remove(p, tb);
69
- invalidate_page_bitmap(p);
70
if (tb->page_addr[1] != -1) {
71
p = page_find(tb->page_addr[1] >> TARGET_PAGE_BITS);
72
tb_page_remove(p, tb);
73
- invalidate_page_bitmap(p);
74
}
75
}
76
77
@@ -XXX,XX +XXX,XX @@ void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr)
16
}
78
}
17
}
79
}
18
80
19
-static inline void tcg_out_bswap16s(TCGContext *s, int cond, int rd, int rn)
81
-#ifdef CONFIG_SOFTMMU
20
+static void tcg_out_bswap16(TCGContext *s, int cond, int rd, int rn, int flags)
82
-/* call with @p->lock held */
21
{
83
-static void build_page_bitmap(PageDesc *p)
22
if (use_armv6_instructions) {
84
-{
23
- /* revsh */
85
- int n, tb_start, tb_end;
24
- tcg_out32(s, 0x06ff0fb0 | (cond << 28) | (rd << 12) | rn);
86
- TranslationBlock *tb;
25
- } else {
87
-
26
- tcg_out_dat_reg(s, cond, ARITH_MOV,
88
- assert_page_locked(p);
27
- TCG_REG_TMP, 0, rn, SHIFT_IMM_LSL(24));
89
- p->code_bitmap = bitmap_new(TARGET_PAGE_SIZE);
28
- tcg_out_dat_reg(s, cond, ARITH_MOV,
90
-
29
- TCG_REG_TMP, 0, TCG_REG_TMP, SHIFT_IMM_ASR(16));
91
- PAGE_FOR_EACH_TB(p, tb, n) {
30
- tcg_out_dat_reg(s, cond, ARITH_ORR,
92
- /* NOTE: this is subtle as a TB may span two physical pages */
31
- rd, TCG_REG_TMP, rn, SHIFT_IMM_LSR(8));
93
- if (n == 0) {
94
- /* NOTE: tb_end may be after the end of the page, but
95
- it is not a problem */
96
- tb_start = tb->pc & ~TARGET_PAGE_MASK;
97
- tb_end = tb_start + tb->size;
98
- if (tb_end > TARGET_PAGE_SIZE) {
99
- tb_end = TARGET_PAGE_SIZE;
100
- }
101
- } else {
102
- tb_start = 0;
103
- tb_end = ((tb->pc + tb->size) & ~TARGET_PAGE_MASK);
104
- }
105
- bitmap_set(p->code_bitmap, tb_start, tb_end - tb_start);
32
- }
106
- }
33
-}
107
-}
34
+ if (flags & TCG_BSWAP_OS) {
108
-#endif
35
+ /* revsh */
109
-
36
+ tcg_out32(s, 0x06ff0fb0 | (cond << 28) | (rd << 12) | rn);
110
/* add the tb in the target page and protect it if necessary
37
+ return;
111
*
38
+ }
112
* Called with mmap_lock held for user-mode emulation.
39
113
@@ -XXX,XX +XXX,XX @@ static inline void tb_page_add(PageDesc *p, TranslationBlock *tb,
40
-static inline void tcg_out_bswap16(TCGContext *s, int cond, int rd, int rn)
114
page_already_protected = p->first_tb != (uintptr_t)NULL;
41
-{
115
#endif
42
- if (use_armv6_instructions) {
116
p->first_tb = (uintptr_t)tb | n;
43
/* rev16 */
117
- invalidate_page_bitmap(p);
44
tcg_out32(s, 0x06bf0fb0 | (cond << 28) | (rd << 12) | rn);
118
119
#if defined(CONFIG_USER_ONLY)
120
/* translator_loop() must have made all TB pages non-writable */
121
@@ -XXX,XX +XXX,XX @@ tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
122
/* remove TB from the page(s) if we couldn't insert it */
123
if (unlikely(existing_tb)) {
124
tb_page_remove(p, tb);
125
- invalidate_page_bitmap(p);
126
if (p2) {
127
tb_page_remove(p2, tb);
128
- invalidate_page_bitmap(p2);
129
}
130
tb = existing_tb;
131
}
132
@@ -XXX,XX +XXX,XX @@ tb_invalidate_phys_page_range__locked(struct page_collection *pages,
133
#if !defined(CONFIG_USER_ONLY)
134
/* if no code remaining, no need to continue to use slow writes */
135
if (!p->first_tb) {
136
- invalidate_page_bitmap(p);
137
tlb_unprotect_code(start);
138
}
139
#endif
140
@@ -XXX,XX +XXX,XX @@ void tb_invalidate_phys_page_fast(struct page_collection *pages,
141
}
142
143
assert_page_locked(p);
144
- if (!p->code_bitmap &&
145
- ++p->code_write_count >= SMC_BITMAP_USE_THRESHOLD) {
146
- build_page_bitmap(p);
147
- }
148
- if (p->code_bitmap) {
149
- unsigned int nr;
150
- unsigned long b;
151
-
152
- nr = start & ~TARGET_PAGE_MASK;
153
- b = p->code_bitmap[BIT_WORD(nr)] >> (nr & (BITS_PER_LONG - 1));
154
- if (b & ((1 << len) - 1)) {
155
- goto do_invalidate;
156
- }
45
- } else {
157
- } else {
46
- tcg_out_dat_reg(s, cond, ARITH_MOV,
158
- do_invalidate:
47
- TCG_REG_TMP, 0, rn, SHIFT_IMM_LSL(24));
159
- tb_invalidate_phys_page_range__locked(pages, p, start, start + len,
48
- tcg_out_dat_reg(s, cond, ARITH_MOV,
160
- retaddr);
49
- TCG_REG_TMP, 0, TCG_REG_TMP, SHIFT_IMM_LSR(16));
161
- }
50
- tcg_out_dat_reg(s, cond, ARITH_ORR,
162
+ tb_invalidate_phys_page_range__locked(pages, p, start, start + len,
51
- rd, TCG_REG_TMP, rn, SHIFT_IMM_LSR(8));
163
+ retaddr);
52
+ if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
53
+ /* uxth */
54
+ tcg_out32(s, 0x06ff0070 | (cond << 28) | (rd << 12) | rd);
55
+ }
56
+ return;
57
}
58
-}
59
60
-/* swap the two low bytes assuming that the two high input bytes and the
61
- two high output bit can hold any value. */
62
-static inline void tcg_out_bswap16st(TCGContext *s, int cond, int rd, int rn)
63
-{
64
- if (use_armv6_instructions) {
65
- /* rev16 */
66
- tcg_out32(s, 0x06bf0fb0 | (cond << 28) | (rd << 12) | rn);
67
- } else {
68
+ if (flags == 0) {
69
+ /*
70
+ * For stores, no input or output extension:
71
+ * rn = xxAB
72
+ * lsr tmp, rn, #8 tmp = 0xxA
73
+ * and tmp, tmp, #0xff tmp = 000A
74
+ * orr rd, tmp, rn, lsl #8 rd = xABA
75
+ */
76
tcg_out_dat_reg(s, cond, ARITH_MOV,
77
TCG_REG_TMP, 0, rn, SHIFT_IMM_LSR(8));
78
tcg_out_dat_imm(s, cond, ARITH_AND, TCG_REG_TMP, TCG_REG_TMP, 0xff);
79
tcg_out_dat_reg(s, cond, ARITH_ORR,
80
rd, TCG_REG_TMP, rn, SHIFT_IMM_LSL(8));
81
+ return;
82
}
83
+
84
+ /*
85
+ * Byte swap, leaving the result at the top of the register.
86
+ * We will then shift down, zero or sign-extending.
87
+ */
88
+ if (flags & TCG_BSWAP_IZ) {
89
+ /*
90
+ * rn = 00AB
91
+ * ror tmp, rn, #8 tmp = B00A
92
+ * orr tmp, tmp, tmp, lsl #16 tmp = BA00
93
+ */
94
+ tcg_out_dat_reg(s, cond, ARITH_MOV,
95
+ TCG_REG_TMP, 0, rn, SHIFT_IMM_ROR(8));
96
+ tcg_out_dat_reg(s, cond, ARITH_ORR,
97
+ TCG_REG_TMP, TCG_REG_TMP, TCG_REG_TMP,
98
+ SHIFT_IMM_LSL(16));
99
+ } else {
100
+ /*
101
+ * rn = xxAB
102
+ * and tmp, rn, #0xff00 tmp = 00A0
103
+ * lsl tmp, tmp, #8 tmp = 0A00
104
+ * orr tmp, tmp, rn, lsl #24 tmp = BA00
105
+ */
106
+ tcg_out_dat_rI(s, cond, ARITH_AND, TCG_REG_TMP, rn, 0xff00, 1);
107
+ tcg_out_dat_reg(s, cond, ARITH_MOV,
108
+ TCG_REG_TMP, 0, TCG_REG_TMP, SHIFT_IMM_LSL(8));
109
+ tcg_out_dat_reg(s, cond, ARITH_ORR,
110
+ TCG_REG_TMP, TCG_REG_TMP, rn, SHIFT_IMM_LSL(24));
111
+ }
112
+ tcg_out_dat_reg(s, cond, ARITH_MOV, rd, 0, TCG_REG_TMP,
113
+ (flags & TCG_BSWAP_OS
114
+ ? SHIFT_IMM_ASR(8) : SHIFT_IMM_LSR(8)));
115
}
164
}
116
165
#else
117
static inline void tcg_out_bswap32(TCGContext *s, int cond, int rd, int rn)
166
/* Called with mmap_lock held. If pc is not 0 then it indicates the
118
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
119
case MO_UW:
120
tcg_out_ld16u_r(s, COND_AL, datalo, addrlo, addend);
121
if (bswap) {
122
- tcg_out_bswap16(s, COND_AL, datalo, datalo);
123
+ tcg_out_bswap16(s, COND_AL, datalo, datalo,
124
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
125
}
126
break;
127
case MO_SW:
128
if (bswap) {
129
tcg_out_ld16u_r(s, COND_AL, datalo, addrlo, addend);
130
- tcg_out_bswap16s(s, COND_AL, datalo, datalo);
131
+ tcg_out_bswap16(s, COND_AL, datalo, datalo,
132
+ TCG_BSWAP_IZ | TCG_BSWAP_OS);
133
} else {
134
tcg_out_ld16s_r(s, COND_AL, datalo, addrlo, addend);
135
}
136
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc,
137
case MO_UW:
138
tcg_out_ld16u_8(s, COND_AL, datalo, addrlo, 0);
139
if (bswap) {
140
- tcg_out_bswap16(s, COND_AL, datalo, datalo);
141
+ tcg_out_bswap16(s, COND_AL, datalo, datalo,
142
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
143
}
144
break;
145
case MO_SW:
146
if (bswap) {
147
tcg_out_ld16u_8(s, COND_AL, datalo, addrlo, 0);
148
- tcg_out_bswap16s(s, COND_AL, datalo, datalo);
149
+ tcg_out_bswap16(s, COND_AL, datalo, datalo,
150
+ TCG_BSWAP_IZ | TCG_BSWAP_OS);
151
} else {
152
tcg_out_ld16s_8(s, COND_AL, datalo, addrlo, 0);
153
}
154
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_st_index(TCGContext *s, int cond, MemOp opc,
155
break;
156
case MO_16:
157
if (bswap) {
158
- tcg_out_bswap16st(s, cond, TCG_REG_R0, datalo);
159
+ tcg_out_bswap16(s, cond, TCG_REG_R0, datalo, 0);
160
tcg_out_st16_r(s, cond, TCG_REG_R0, addrlo, addend);
161
} else {
162
tcg_out_st16_r(s, cond, datalo, addrlo, addend);
163
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc,
164
break;
165
case MO_16:
166
if (bswap) {
167
- tcg_out_bswap16st(s, COND_AL, TCG_REG_R0, datalo);
168
+ tcg_out_bswap16(s, COND_AL, TCG_REG_R0, datalo, 0);
169
tcg_out_st16_8(s, COND_AL, TCG_REG_R0, addrlo, 0);
170
} else {
171
tcg_out_st16_8(s, COND_AL, datalo, addrlo, 0);
172
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
173
break;
174
175
case INDEX_op_bswap16_i32:
176
- tcg_out_bswap16(s, COND_AL, args[0], args[1]);
177
+ tcg_out_bswap16(s, COND_AL, args[0], args[1], args[2]);
178
break;
179
case INDEX_op_bswap32_i32:
180
tcg_out_bswap32(s, COND_AL, args[0], args[1]);
181
--
167
--
182
2.25.1
168
2.34.1
183
169
184
170
diff view generated by jsdifflib
1
For INDEX_op_bswap32_i32, pass 0 for flags: input not zero-extended,
1
Bool is more appropriate type for the alloc parameter.
2
output does not need extension within the host 64-bit register.
3
2
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
6
---
7
tcg/ppc/tcg-target.c.inc | 22 ++++++++++++++++------
7
accel/tcg/translate-all.c | 14 +++++++-------
8
1 file changed, 16 insertions(+), 6 deletions(-)
8
1 file changed, 7 insertions(+), 7 deletions(-)
9
9
10
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
10
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
11
index XXXXXXX..XXXXXXX 100644
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/ppc/tcg-target.c.inc
12
--- a/accel/tcg/translate-all.c
13
+++ b/tcg/ppc/tcg-target.c.inc
13
+++ b/accel/tcg/translate-all.c
14
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_sari64(TCGContext *s, TCGReg dst, TCGReg src, int c)
14
@@ -XXX,XX +XXX,XX @@ void page_init(void)
15
tcg_out32(s, SRADI | RA(dst) | RS(src) | SH(c & 0x1f) | ((c >> 4) & 2));
15
#endif
16
}
16
}
17
17
18
-static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src)
18
-static PageDesc *page_find_alloc(tb_page_addr_t index, int alloc)
19
+static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src, int flags)
19
+static PageDesc *page_find_alloc(tb_page_addr_t index, bool alloc)
20
{
20
{
21
TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
21
PageDesc *pd;
22
22
void **lp;
23
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src)
23
@@ -XXX,XX +XXX,XX @@ static PageDesc *page_find_alloc(tb_page_addr_t index, int alloc)
24
/* tmp = dep(tmp, rol32(src, 8), 0x0000ff00) = 000000dc */
24
25
tcg_out_rlw(s, RLWIMI, tmp, src, 8, 16, 23);
25
static inline PageDesc *page_find(tb_page_addr_t index)
26
26
{
27
- tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
27
- return page_find_alloc(index, 0);
28
+ if (flags & TCG_BSWAP_OS) {
28
+ return page_find_alloc(index, false);
29
+ tcg_out_ext16s(s, dst, tmp);
30
+ } else {
31
+ tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
32
+ }
33
}
29
}
34
30
35
-static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src)
31
static void page_lock_pair(PageDesc **ret_p1, tb_page_addr_t phys1,
36
+static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src, int flags)
32
- PageDesc **ret_p2, tb_page_addr_t phys2, int alloc);
33
+ PageDesc **ret_p2, tb_page_addr_t phys2, bool alloc);
34
35
/* In user-mode page locks aren't used; mmap_lock is enough */
36
#ifdef CONFIG_USER_ONLY
37
@@ -XXX,XX +XXX,XX @@ static inline void page_unlock(PageDesc *pd)
38
/* lock the page(s) of a TB in the correct acquisition order */
39
static inline void page_lock_tb(const TranslationBlock *tb)
37
{
40
{
38
TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
41
- page_lock_pair(NULL, tb->page_addr[0], NULL, tb->page_addr[1], 0);
39
42
+ page_lock_pair(NULL, tb->page_addr[0], NULL, tb->page_addr[1], false);
40
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src)
41
/* tmp = dep(tmp, rol32(src, 24), 0x0000ff00) = 0000dcba */
42
tcg_out_rlw(s, RLWIMI, tmp, src, 24, 16, 23);
43
44
- tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
45
+ if (flags & TCG_BSWAP_OS) {
46
+ tcg_out_ext32s(s, dst, tmp);
47
+ } else {
48
+ tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
49
+ }
50
}
43
}
51
44
52
static void tcg_out_bswap64(TCGContext *s, TCGReg dst, TCGReg src)
45
static inline void page_unlock_tb(const TranslationBlock *tb)
53
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
46
@@ -XXX,XX +XXX,XX @@ void page_collection_unlock(struct page_collection *set)
54
47
#endif /* !CONFIG_USER_ONLY */
55
case INDEX_op_bswap16_i32:
48
56
case INDEX_op_bswap16_i64:
49
static void page_lock_pair(PageDesc **ret_p1, tb_page_addr_t phys1,
57
- tcg_out_bswap16(s, args[0], args[1]);
50
- PageDesc **ret_p2, tb_page_addr_t phys2, int alloc)
58
+ tcg_out_bswap16(s, args[0], args[1], args[2]);
51
+ PageDesc **ret_p2, tb_page_addr_t phys2, bool alloc)
59
break;
52
{
60
case INDEX_op_bswap32_i32:
53
PageDesc *p1, *p2;
61
+ tcg_out_bswap32(s, args[0], args[1], 0);
54
tb_page_addr_t page1;
62
+ break;
55
@@ -XXX,XX +XXX,XX @@ tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
63
case INDEX_op_bswap32_i64:
56
* Note that inserting into the hash table first isn't an option, since
64
- tcg_out_bswap32(s, args[0], args[1]);
57
* we can only insert TBs that are fully initialized.
65
+ tcg_out_bswap32(s, args[0], args[1], args[2]);
58
*/
66
break;
59
- page_lock_pair(&p, phys_pc, &p2, phys_page2, 1);
67
case INDEX_op_bswap64_i64:
60
+ page_lock_pair(&p, phys_pc, &p2, phys_page2, true);
68
tcg_out_bswap64(s, args[0], args[1]);
61
tb_page_add(p, tb, 0, phys_pc & TARGET_PAGE_MASK);
62
if (p2) {
63
tb_page_add(p2, tb, 1, phys_page2);
64
@@ -XXX,XX +XXX,XX @@ void page_set_flags(target_ulong start, target_ulong end, int flags)
65
for (addr = start, len = end - start;
66
len != 0;
67
len -= TARGET_PAGE_SIZE, addr += TARGET_PAGE_SIZE) {
68
- PageDesc *p = page_find_alloc(addr >> TARGET_PAGE_BITS, 1);
69
+ PageDesc *p = page_find_alloc(addr >> TARGET_PAGE_BITS, true);
70
71
/* If the write protection bit is set, then we invalidate
72
the code inside. */
69
--
73
--
70
2.25.1
74
2.34.1
71
75
72
76
diff view generated by jsdifflib
1
The only semantic of DISAS_TB_JUMP is that we've done goto_tb,
1
Use the pc coming from db->pc_first rather than the TB.
2
which is the same as DISAS_NORETURN -- we've exited the tb.
3
2
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Use the cached host_addr rather than re-computing for the
4
first page. We still need a separate lookup for the second
5
page because it won't be computed for DisasContextBase until
6
the translator actually performs a read from the page.
7
8
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
10
---
7
target/nios2/translate.c | 8 +++-----
11
include/exec/plugin-gen.h | 7 ++++---
8
1 file changed, 3 insertions(+), 5 deletions(-)
12
accel/tcg/plugin-gen.c | 22 +++++++++++-----------
13
accel/tcg/translator.c | 2 +-
14
3 files changed, 16 insertions(+), 15 deletions(-)
9
15
10
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
16
diff --git a/include/exec/plugin-gen.h b/include/exec/plugin-gen.h
11
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
12
--- a/target/nios2/translate.c
18
--- a/include/exec/plugin-gen.h
13
+++ b/target/nios2/translate.c
19
+++ b/include/exec/plugin-gen.h
14
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ struct DisasContextBase;
15
/* is_jmp field values */
21
16
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
22
#ifdef CONFIG_PLUGIN
17
#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
23
18
-#define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */
24
-bool plugin_gen_tb_start(CPUState *cpu, const TranslationBlock *tb, bool supress);
19
25
+bool plugin_gen_tb_start(CPUState *cpu, const struct DisasContextBase *db,
20
#define INSTRUCTION_FLG(func, flags) { (func), (flags) }
26
+ bool supress);
21
#define INSTRUCTION(func) \
27
void plugin_gen_tb_end(CPUState *cpu);
22
@@ -XXX,XX +XXX,XX @@ static void jmpi(DisasContext *dc, uint32_t code, uint32_t flags)
28
void plugin_gen_insn_start(CPUState *cpu, const struct DisasContextBase *db);
29
void plugin_gen_insn_end(void);
30
@@ -XXX,XX +XXX,XX @@ static inline void plugin_insn_append(abi_ptr pc, const void *from, size_t size)
31
32
#else /* !CONFIG_PLUGIN */
33
34
-static inline
35
-bool plugin_gen_tb_start(CPUState *cpu, const TranslationBlock *tb, bool supress)
36
+static inline bool
37
+plugin_gen_tb_start(CPUState *cpu, const struct DisasContextBase *db, bool sup)
23
{
38
{
24
J_TYPE(instr, code);
39
return false;
25
gen_goto_tb(dc, 0, (dc->pc & 0xF0000000) | (instr.imm26 << 2));
26
- dc->is_jmp = DISAS_TB_JUMP;
27
+ dc->is_jmp = DISAS_NORETURN;
28
}
40
}
29
41
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
30
static void call(DisasContext *dc, uint32_t code, uint32_t flags)
42
index XXXXXXX..XXXXXXX 100644
31
@@ -XXX,XX +XXX,XX @@ static void br(DisasContext *dc, uint32_t code, uint32_t flags)
43
--- a/accel/tcg/plugin-gen.c
32
I_TYPE(instr, code);
44
+++ b/accel/tcg/plugin-gen.c
33
45
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(const struct qemu_plugin_tb *plugin_tb)
34
gen_goto_tb(dc, 0, dc->pc + 4 + (instr.imm16.s & -4));
46
pr_ops();
35
- dc->is_jmp = DISAS_TB_JUMP;
36
+ dc->is_jmp = DISAS_NORETURN;
37
}
47
}
38
48
39
static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
49
-bool plugin_gen_tb_start(CPUState *cpu, const TranslationBlock *tb, bool mem_only)
40
@@ -XXX,XX +XXX,XX @@ static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
50
+bool plugin_gen_tb_start(CPUState *cpu, const DisasContextBase *db,
41
gen_goto_tb(dc, 0, dc->pc + 4);
51
+ bool mem_only)
42
gen_set_label(l1);
52
{
43
gen_goto_tb(dc, 1, dc->pc + 4 + (instr.imm16.s & -4));
53
bool ret = false;
44
- dc->is_jmp = DISAS_TB_JUMP;
54
45
+ dc->is_jmp = DISAS_NORETURN;
55
@@ -XXX,XX +XXX,XX @@ bool plugin_gen_tb_start(CPUState *cpu, const TranslationBlock *tb, bool mem_onl
56
57
ret = true;
58
59
- ptb->vaddr = tb->pc;
60
+ ptb->vaddr = db->pc_first;
61
ptb->vaddr2 = -1;
62
- get_page_addr_code_hostp(cpu->env_ptr, tb->pc, &ptb->haddr1);
63
+ ptb->haddr1 = db->host_addr[0];
64
ptb->haddr2 = NULL;
65
ptb->mem_only = mem_only;
66
67
@@ -XXX,XX +XXX,XX @@ void plugin_gen_insn_start(CPUState *cpu, const DisasContextBase *db)
68
* Note that we skip this when haddr1 == NULL, e.g. when we're
69
* fetching instructions from a region not backed by RAM.
70
*/
71
- if (likely(ptb->haddr1 != NULL && ptb->vaddr2 == -1) &&
72
- unlikely((db->pc_next & TARGET_PAGE_MASK) !=
73
- (db->pc_first & TARGET_PAGE_MASK))) {
74
- get_page_addr_code_hostp(cpu->env_ptr, db->pc_next,
75
- &ptb->haddr2);
76
- ptb->vaddr2 = db->pc_next;
77
- }
78
- if (likely(ptb->vaddr2 == -1)) {
79
+ if (ptb->haddr1 == NULL) {
80
+ pinsn->haddr = NULL;
81
+ } else if (is_same_page(db, db->pc_next)) {
82
pinsn->haddr = ptb->haddr1 + pinsn->vaddr - ptb->vaddr;
83
} else {
84
+ if (ptb->vaddr2 == -1) {
85
+ ptb->vaddr2 = TARGET_PAGE_ALIGN(db->pc_first);
86
+ get_page_addr_code_hostp(cpu->env_ptr, ptb->vaddr2, &ptb->haddr2);
87
+ }
88
pinsn->haddr = ptb->haddr2 + pinsn->vaddr - ptb->vaddr2;
89
}
46
}
90
}
47
91
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
48
/* Comparison instructions */
92
index XXXXXXX..XXXXXXX 100644
49
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
93
--- a/accel/tcg/translator.c
50
break;
94
+++ b/accel/tcg/translator.c
51
95
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
52
case DISAS_NORETURN:
96
ops->tb_start(db, cpu);
53
- case DISAS_TB_JUMP:
97
tcg_debug_assert(db->is_jmp == DISAS_NEXT); /* no early exit */
54
/* nothing more to generate */
98
55
break;
99
- plugin_enabled = plugin_gen_tb_start(cpu, tb, cflags & CF_MEMI_ONLY);
56
}
100
+ plugin_enabled = plugin_gen_tb_start(cpu, db, cflags & CF_MEMI_ONLY);
101
102
while (true) {
103
db->num_insns++;
57
--
104
--
58
2.25.1
105
2.34.1
59
106
60
107
diff view generated by jsdifflib
Deleted patch
1
We do not need to copy this into DisasContext.
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
6
target/nios2/translate.c | 10 ++++------
7
1 file changed, 4 insertions(+), 6 deletions(-)
8
9
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
10
index XXXXXXX..XXXXXXX 100644
11
--- a/target/nios2/translate.c
12
+++ b/target/nios2/translate.c
13
@@ -XXX,XX +XXX,XX @@
14
}
15
16
typedef struct DisasContext {
17
- TCGv_ptr cpu_env;
18
TCGv *cpu_R;
19
TCGv_i32 zero;
20
int is_jmp;
21
@@ -XXX,XX +XXX,XX @@ static void t_gen_helper_raise_exception(DisasContext *dc,
22
TCGv_i32 tmp = tcg_const_i32(index);
23
24
tcg_gen_movi_tl(dc->cpu_R[R_PC], dc->pc);
25
- gen_helper_raise_exception(dc->cpu_env, tmp);
26
+ gen_helper_raise_exception(cpu_env, tmp);
27
tcg_temp_free_i32(tmp);
28
dc->is_jmp = DISAS_NORETURN;
29
}
30
@@ -XXX,XX +XXX,XX @@ static void rdctl(DisasContext *dc, uint32_t code, uint32_t flags)
31
tcg_gen_mov_tl(dc->cpu_R[instr.c], dc->cpu_R[instr.imm5 + CR_BASE]);
32
#ifdef DEBUG_MMU
33
TCGv_i32 tmp = tcg_const_i32(instr.imm5 + CR_BASE);
34
- gen_helper_mmu_read_debug(dc->cpu_R[instr.c], dc->cpu_env, tmp);
35
+ gen_helper_mmu_read_debug(dc->cpu_R[instr.c], cpu_env, tmp);
36
tcg_temp_free_i32(tmp);
37
#endif
38
}
39
@@ -XXX,XX +XXX,XX @@ static void wrctl(DisasContext *dc, uint32_t code, uint32_t flags)
40
{
41
#if !defined(CONFIG_USER_ONLY)
42
TCGv_i32 tmp = tcg_const_i32(instr.imm5 + CR_BASE);
43
- gen_helper_mmu_write(dc->cpu_env, tmp, load_gpr(dc, instr.a));
44
+ gen_helper_mmu_write(cpu_env, tmp, load_gpr(dc, instr.a));
45
tcg_temp_free_i32(tmp);
46
#endif
47
break;
48
@@ -XXX,XX +XXX,XX @@ static void wrctl(DisasContext *dc, uint32_t code, uint32_t flags)
49
if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
50
gen_io_start();
51
}
52
- gen_helper_check_interrupts(dc->cpu_env);
53
+ gen_helper_check_interrupts(cpu_env);
54
dc->is_jmp = DISAS_UPDATE;
55
}
56
#endif
57
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
58
int num_insns;
59
60
/* Initialize DC */
61
- dc->cpu_env = cpu_env;
62
dc->cpu_R = cpu_R;
63
dc->is_jmp = DISAS_NEXT;
64
dc->pc = tb->pc;
65
--
66
2.25.1
67
68
diff view generated by jsdifflib
1
Move handle_instruction into nios2_tr_translate_insn
1
Let tb->page_addr[0] contain the address of the first byte of the
2
as the only caller.
2
translated block, rather than the address of the page containing the
3
start of the translated block. We need to recover this value anyway
4
at various points, and it is easier to discard a page offset when it
5
is not needed, which happens naturally via the existing find_page shift.
3
6
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
9
---
7
target/nios2/translate.c | 66 +++++++++++++++++++---------------------
10
accel/tcg/cpu-exec.c | 16 ++++++++--------
8
1 file changed, 31 insertions(+), 35 deletions(-)
11
accel/tcg/cputlb.c | 3 ++-
12
accel/tcg/translate-all.c | 9 +++++----
13
3 files changed, 15 insertions(+), 13 deletions(-)
9
14
10
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
15
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
11
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
12
--- a/target/nios2/translate.c
17
--- a/accel/tcg/cpu-exec.c
13
+++ b/target/nios2/translate.c
18
+++ b/accel/tcg/cpu-exec.c
14
@@ -XXX,XX +XXX,XX @@ illegal_op:
19
@@ -XXX,XX +XXX,XX @@ struct tb_desc {
15
t_gen_helper_raise_exception(dc, EXCP_ILLEGAL);
20
target_ulong pc;
21
target_ulong cs_base;
22
CPUArchState *env;
23
- tb_page_addr_t phys_page1;
24
+ tb_page_addr_t page_addr0;
25
uint32_t flags;
26
uint32_t cflags;
27
uint32_t trace_vcpu_dstate;
28
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
29
const struct tb_desc *desc = d;
30
31
if (tb->pc == desc->pc &&
32
- tb->page_addr[0] == desc->phys_page1 &&
33
+ tb->page_addr[0] == desc->page_addr0 &&
34
tb->cs_base == desc->cs_base &&
35
tb->flags == desc->flags &&
36
tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
37
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
38
if (tb->page_addr[1] == -1) {
39
return true;
40
} else {
41
- tb_page_addr_t phys_page2;
42
- target_ulong virt_page2;
43
+ tb_page_addr_t phys_page1;
44
+ target_ulong virt_page1;
45
46
/*
47
* We know that the first page matched, and an otherwise valid TB
48
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
49
* is different for the new TB. Therefore any exception raised
50
* here by the faulting lookup is not premature.
51
*/
52
- virt_page2 = TARGET_PAGE_ALIGN(desc->pc);
53
- phys_page2 = get_page_addr_code(desc->env, virt_page2);
54
- if (tb->page_addr[1] == phys_page2) {
55
+ virt_page1 = TARGET_PAGE_ALIGN(desc->pc);
56
+ phys_page1 = get_page_addr_code(desc->env, virt_page1);
57
+ if (tb->page_addr[1] == phys_page1) {
58
return true;
59
}
60
}
61
@@ -XXX,XX +XXX,XX @@ static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
62
if (phys_pc == -1) {
63
return NULL;
64
}
65
- desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
66
+ desc.page_addr0 = phys_pc;
67
h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
68
return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
16
}
69
}
17
70
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
18
-static void handle_instruction(DisasContext *dc, CPUNios2State *env)
71
index XXXXXXX..XXXXXXX 100644
19
-{
72
--- a/accel/tcg/cputlb.c
20
- uint32_t code;
73
+++ b/accel/tcg/cputlb.c
21
- uint8_t op;
74
@@ -XXX,XX +XXX,XX @@ void tlb_flush_page_bits_by_mmuidx_all_cpus_synced(CPUState *src_cpu,
22
- const Nios2Instruction *instr;
75
can be detected */
23
-
76
void tlb_protect_code(ram_addr_t ram_addr)
24
-#if defined(CONFIG_USER_ONLY)
25
- /* FIXME: Is this needed ? */
26
- if (dc->pc >= 0x1000 && dc->pc < 0x2000) {
27
- t_gen_helper_raise_exception(dc, 0xaa);
28
- return;
29
- }
30
-#endif
31
-
32
- code = cpu_ldl_code(env, dc->pc);
33
- op = get_opcode(code);
34
-
35
- if (unlikely(op >= ARRAY_SIZE(i_type_instructions))) {
36
- t_gen_helper_raise_exception(dc, EXCP_ILLEGAL);
37
- return;
38
- }
39
-
40
- dc->zero = NULL;
41
-
42
- instr = &i_type_instructions[op];
43
- instr->handler(dc, code, instr->flags);
44
-
45
- if (dc->zero) {
46
- tcg_temp_free(dc->zero);
47
- }
48
-}
49
-
50
static const char * const regnames[] = {
51
"zero", "at", "r2", "r3",
52
"r4", "r5", "r6", "r7",
53
@@ -XXX,XX +XXX,XX @@ static void nios2_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
54
{
77
{
55
DisasContext *dc = container_of(dcbase, DisasContext, base);
78
- cpu_physical_memory_test_and_clear_dirty(ram_addr, TARGET_PAGE_SIZE,
56
CPUNios2State *env = cs->env_ptr;
79
+ cpu_physical_memory_test_and_clear_dirty(ram_addr & TARGET_PAGE_MASK,
57
+ const Nios2Instruction *instr;
80
+ TARGET_PAGE_SIZE,
58
+ uint32_t code, pc;
81
DIRTY_MEMORY_CODE);
59
+ uint8_t op;
60
61
- dc->pc = dc->base.pc_next;
62
- dc->base.pc_next += 4;
63
+ pc = dc->base.pc_next;
64
+ dc->pc = pc;
65
+ dc->base.pc_next = pc + 4;
66
67
/* Decode an instruction */
68
- handle_instruction(dc, env);
69
+
70
+#if defined(CONFIG_USER_ONLY)
71
+ /* FIXME: Is this needed ? */
72
+ if (pc >= 0x1000 && pc < 0x2000) {
73
+ t_gen_helper_raise_exception(dc, 0xaa);
74
+ return;
75
+ }
76
+#endif
77
+
78
+ code = cpu_ldl_code(env, pc);
79
+ op = get_opcode(code);
80
+
81
+ if (unlikely(op >= ARRAY_SIZE(i_type_instructions))) {
82
+ t_gen_helper_raise_exception(dc, EXCP_ILLEGAL);
83
+ return;
84
+ }
85
+
86
+ dc->zero = NULL;
87
+
88
+ instr = &i_type_instructions[op];
89
+ instr->handler(dc, code, instr->flags);
90
+
91
+ if (dc->zero) {
92
+ tcg_temp_free(dc->zero);
93
+ }
94
}
82
}
95
83
96
static void nios2_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
84
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
85
index XXXXXXX..XXXXXXX 100644
86
--- a/accel/tcg/translate-all.c
87
+++ b/accel/tcg/translate-all.c
88
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
89
qemu_spin_unlock(&tb->jmp_lock);
90
91
/* remove the TB from the hash list */
92
- phys_pc = tb->page_addr[0] + (tb->pc & ~TARGET_PAGE_MASK);
93
+ phys_pc = tb->page_addr[0];
94
h = tb_hash_func(phys_pc, tb->pc, tb->flags, orig_cflags,
95
tb->trace_vcpu_dstate);
96
if (!qht_remove(&tb_ctx.htable, tb, h)) {
97
@@ -XXX,XX +XXX,XX @@ tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
98
* we can only insert TBs that are fully initialized.
99
*/
100
page_lock_pair(&p, phys_pc, &p2, phys_page2, true);
101
- tb_page_add(p, tb, 0, phys_pc & TARGET_PAGE_MASK);
102
+ tb_page_add(p, tb, 0, phys_pc);
103
if (p2) {
104
tb_page_add(p2, tb, 1, phys_page2);
105
} else {
106
@@ -XXX,XX +XXX,XX @@ tb_invalidate_phys_page_range__locked(struct page_collection *pages,
107
if (n == 0) {
108
/* NOTE: tb_end may be after the end of the page, but
109
it is not a problem */
110
- tb_start = tb->page_addr[0] + (tb->pc & ~TARGET_PAGE_MASK);
111
+ tb_start = tb->page_addr[0];
112
tb_end = tb_start + tb->size;
113
} else {
114
tb_start = tb->page_addr[1];
115
- tb_end = tb_start + ((tb->pc + tb->size) & ~TARGET_PAGE_MASK);
116
+ tb_end = tb_start + ((tb->page_addr[0] + tb->size)
117
+ & ~TARGET_PAGE_MASK);
118
}
119
if (!(tb_end <= start || tb_start >= end)) {
120
#ifdef TARGET_HAS_PRECISE_SMC
97
--
121
--
98
2.25.1
122
2.34.1
99
123
100
124
diff view generated by jsdifflib
1
We do not need to copy this into DisasContext.
1
This function has two users, who use it incompatibly.
2
In tlb_flush_page_by_mmuidx_async_0, when flushing a
3
single page, we need to flush exactly two pages.
4
In tlb_flush_range_by_mmuidx_async_0, when flushing a
5
range of pages, we need to flush N+1 pages.
2
6
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
This avoids double-flushing of jmp cache pages in a range.
8
9
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
11
---
6
target/nios2/translate.c | 73 +++++++++++++++++++---------------------
12
accel/tcg/cputlb.c | 25 ++++++++++++++-----------
7
1 file changed, 34 insertions(+), 39 deletions(-)
13
1 file changed, 14 insertions(+), 11 deletions(-)
8
14
9
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
15
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
10
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
11
--- a/target/nios2/translate.c
17
--- a/accel/tcg/cputlb.c
12
+++ b/target/nios2/translate.c
18
+++ b/accel/tcg/cputlb.c
13
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ static void tb_jmp_cache_clear_page(CPUState *cpu, target_ulong page_addr)
14
}
15
16
typedef struct DisasContext {
17
- TCGv *cpu_R;
18
TCGv_i32 zero;
19
int is_jmp;
20
target_ulong pc;
21
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
22
bool singlestep_enabled;
23
} DisasContext;
24
25
+static TCGv cpu_R[NUM_CORE_REGS];
26
+
27
typedef struct Nios2Instruction {
28
void (*handler)(DisasContext *dc, uint32_t code, uint32_t flags);
29
uint32_t flags;
30
@@ -XXX,XX +XXX,XX @@ static TCGv load_zero(DisasContext *dc)
31
static TCGv load_gpr(DisasContext *dc, uint8_t reg)
32
{
33
if (likely(reg != R_ZERO)) {
34
- return dc->cpu_R[reg];
35
+ return cpu_R[reg];
36
} else {
37
return load_zero(dc);
38
}
39
@@ -XXX,XX +XXX,XX @@ static void t_gen_helper_raise_exception(DisasContext *dc,
40
{
41
TCGv_i32 tmp = tcg_const_i32(index);
42
43
- tcg_gen_movi_tl(dc->cpu_R[R_PC], dc->pc);
44
+ tcg_gen_movi_tl(cpu_R[R_PC], dc->pc);
45
gen_helper_raise_exception(cpu_env, tmp);
46
tcg_temp_free_i32(tmp);
47
dc->is_jmp = DISAS_NORETURN;
48
@@ -XXX,XX +XXX,XX @@ static void gen_goto_tb(DisasContext *dc, int n, uint32_t dest)
49
50
if (use_goto_tb(dc, dest)) {
51
tcg_gen_goto_tb(n);
52
- tcg_gen_movi_tl(dc->cpu_R[R_PC], dest);
53
+ tcg_gen_movi_tl(cpu_R[R_PC], dest);
54
tcg_gen_exit_tb(tb, n);
55
} else {
56
- tcg_gen_movi_tl(dc->cpu_R[R_PC], dest);
57
+ tcg_gen_movi_tl(cpu_R[R_PC], dest);
58
tcg_gen_exit_tb(NULL, 0);
59
}
20
}
60
}
21
}
61
@@ -XXX,XX +XXX,XX @@ static void jmpi(DisasContext *dc, uint32_t code, uint32_t flags)
22
62
23
-static void tb_flush_jmp_cache(CPUState *cpu, target_ulong addr)
63
static void call(DisasContext *dc, uint32_t code, uint32_t flags)
24
-{
64
{
25
- /* Discard jump cache entries for any tb which might potentially
65
- tcg_gen_movi_tl(dc->cpu_R[R_RA], dc->pc + 4);
26
- overlap the flushed page. */
66
+ tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
27
- tb_jmp_cache_clear_page(cpu, addr - TARGET_PAGE_SIZE);
67
jmpi(dc, code, flags);
28
- tb_jmp_cache_clear_page(cpu, addr);
29
-}
30
-
31
/**
32
* tlb_mmu_resize_locked() - perform TLB resize bookkeeping; resize if necessary
33
* @desc: The CPUTLBDesc portion of the TLB
34
@@ -XXX,XX +XXX,XX @@ static void tlb_flush_page_by_mmuidx_async_0(CPUState *cpu,
35
}
36
qemu_spin_unlock(&env_tlb(env)->c.lock);
37
38
- tb_flush_jmp_cache(cpu, addr);
39
+ /*
40
+ * Discard jump cache entries for any tb which might potentially
41
+ * overlap the flushed page, which includes the previous.
42
+ */
43
+ tb_jmp_cache_clear_page(cpu, addr - TARGET_PAGE_SIZE);
44
+ tb_jmp_cache_clear_page(cpu, addr);
68
}
45
}
69
46
70
@@ -XXX,XX +XXX,XX @@ static void gen_ldx(DisasContext *dc, uint32_t code, uint32_t flags)
47
/**
71
* the Nios2 CPU.
48
@@ -XXX,XX +XXX,XX @@ static void tlb_flush_range_by_mmuidx_async_0(CPUState *cpu,
72
*/
49
return;
73
if (likely(instr.b != R_ZERO)) {
74
- data = dc->cpu_R[instr.b];
75
+ data = cpu_R[instr.b];
76
} else {
77
data = tcg_temp_new();
78
}
50
}
79
@@ -XXX,XX +XXX,XX @@ static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
51
80
I_TYPE(instr, code);
52
- for (target_ulong i = 0; i < d.len; i += TARGET_PAGE_SIZE) {
81
53
- tb_flush_jmp_cache(cpu, d.addr + i);
82
TCGLabel *l1 = gen_new_label();
54
+ /*
83
- tcg_gen_brcond_tl(flags, dc->cpu_R[instr.a], dc->cpu_R[instr.b], l1);
55
+ * Discard jump cache entries for any tb which might potentially
84
+ tcg_gen_brcond_tl(flags, cpu_R[instr.a], cpu_R[instr.b], l1);
56
+ * overlap the flushed pages, which includes the previous.
85
gen_goto_tb(dc, 0, dc->pc + 4);
57
+ */
86
gen_set_label(l1);
58
+ d.addr -= TARGET_PAGE_SIZE;
87
gen_goto_tb(dc, 1, dc->pc + 4 + (instr.imm16.s & -4));
59
+ for (target_ulong i = 0, n = d.len / TARGET_PAGE_SIZE + 1; i < n; i++) {
88
@@ -XXX,XX +XXX,XX @@ static void gen_bxx(DisasContext *dc, uint32_t code, uint32_t flags)
60
+ tb_jmp_cache_clear_page(cpu, d.addr);
89
static void (fname)(DisasContext *dc, uint32_t code, uint32_t flags) \
61
+ d.addr += TARGET_PAGE_SIZE;
90
{ \
91
I_TYPE(instr, (code)); \
92
- tcg_gen_setcondi_tl(flags, (dc)->cpu_R[instr.b], (dc)->cpu_R[instr.a], \
93
- (op3)); \
94
+ tcg_gen_setcondi_tl(flags, cpu_R[instr.b], cpu_R[instr.a], (op3)); \
95
}
96
97
gen_i_cmpxx(gen_cmpxxsi, instr.imm16.s)
98
@@ -XXX,XX +XXX,XX @@ static void (fname)(DisasContext *dc, uint32_t code, uint32_t flags) \
99
if (unlikely(instr.b == R_ZERO)) { /* Store to R_ZERO is ignored */ \
100
return; \
101
} else if (instr.a == R_ZERO) { /* MOVxI optimizations */ \
102
- tcg_gen_movi_tl(dc->cpu_R[instr.b], (resimm) ? (op3) : 0); \
103
+ tcg_gen_movi_tl(cpu_R[instr.b], (resimm) ? (op3) : 0); \
104
} else { \
105
- tcg_gen_##insn##_tl((dc)->cpu_R[instr.b], (dc)->cpu_R[instr.a], \
106
- (op3)); \
107
+ tcg_gen_##insn##_tl(cpu_R[instr.b], cpu_R[instr.a], (op3)); \
108
} \
109
}
110
111
@@ -XXX,XX +XXX,XX @@ static const Nios2Instruction i_type_instructions[] = {
112
*/
113
static void eret(DisasContext *dc, uint32_t code, uint32_t flags)
114
{
115
- tcg_gen_mov_tl(dc->cpu_R[CR_STATUS], dc->cpu_R[CR_ESTATUS]);
116
- tcg_gen_mov_tl(dc->cpu_R[R_PC], dc->cpu_R[R_EA]);
117
+ tcg_gen_mov_tl(cpu_R[CR_STATUS], cpu_R[CR_ESTATUS]);
118
+ tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_EA]);
119
120
dc->is_jmp = DISAS_JUMP;
121
}
122
@@ -XXX,XX +XXX,XX @@ static void eret(DisasContext *dc, uint32_t code, uint32_t flags)
123
/* PC <- ra */
124
static void ret(DisasContext *dc, uint32_t code, uint32_t flags)
125
{
126
- tcg_gen_mov_tl(dc->cpu_R[R_PC], dc->cpu_R[R_RA]);
127
+ tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_RA]);
128
129
dc->is_jmp = DISAS_JUMP;
130
}
131
@@ -XXX,XX +XXX,XX @@ static void ret(DisasContext *dc, uint32_t code, uint32_t flags)
132
/* PC <- ba */
133
static void bret(DisasContext *dc, uint32_t code, uint32_t flags)
134
{
135
- tcg_gen_mov_tl(dc->cpu_R[R_PC], dc->cpu_R[R_BA]);
136
+ tcg_gen_mov_tl(cpu_R[R_PC], cpu_R[R_BA]);
137
138
dc->is_jmp = DISAS_JUMP;
139
}
140
@@ -XXX,XX +XXX,XX @@ static void jmp(DisasContext *dc, uint32_t code, uint32_t flags)
141
{
142
R_TYPE(instr, code);
143
144
- tcg_gen_mov_tl(dc->cpu_R[R_PC], load_gpr(dc, instr.a));
145
+ tcg_gen_mov_tl(cpu_R[R_PC], load_gpr(dc, instr.a));
146
147
dc->is_jmp = DISAS_JUMP;
148
}
149
@@ -XXX,XX +XXX,XX @@ static void nextpc(DisasContext *dc, uint32_t code, uint32_t flags)
150
R_TYPE(instr, code);
151
152
if (likely(instr.c != R_ZERO)) {
153
- tcg_gen_movi_tl(dc->cpu_R[instr.c], dc->pc + 4);
154
+ tcg_gen_movi_tl(cpu_R[instr.c], dc->pc + 4);
155
}
62
}
156
}
63
}
157
64
158
@@ -XXX,XX +XXX,XX @@ static void callr(DisasContext *dc, uint32_t code, uint32_t flags)
159
{
160
R_TYPE(instr, code);
161
162
- tcg_gen_mov_tl(dc->cpu_R[R_PC], load_gpr(dc, instr.a));
163
- tcg_gen_movi_tl(dc->cpu_R[R_RA], dc->pc + 4);
164
+ tcg_gen_mov_tl(cpu_R[R_PC], load_gpr(dc, instr.a));
165
+ tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
166
167
dc->is_jmp = DISAS_JUMP;
168
}
169
@@ -XXX,XX +XXX,XX @@ static void rdctl(DisasContext *dc, uint32_t code, uint32_t flags)
170
{
171
#if !defined(CONFIG_USER_ONLY)
172
if (likely(instr.c != R_ZERO)) {
173
- tcg_gen_mov_tl(dc->cpu_R[instr.c], dc->cpu_R[instr.imm5 + CR_BASE]);
174
+ tcg_gen_mov_tl(cpu_R[instr.c], cpu_R[instr.imm5 + CR_BASE]);
175
#ifdef DEBUG_MMU
176
TCGv_i32 tmp = tcg_const_i32(instr.imm5 + CR_BASE);
177
- gen_helper_mmu_read_debug(dc->cpu_R[instr.c], cpu_env, tmp);
178
+ gen_helper_mmu_read_debug(cpu_R[instr.c], cpu_env, tmp);
179
tcg_temp_free_i32(tmp);
180
#endif
181
}
182
@@ -XXX,XX +XXX,XX @@ static void rdctl(DisasContext *dc, uint32_t code, uint32_t flags)
183
184
default:
185
if (likely(instr.c != R_ZERO)) {
186
- tcg_gen_mov_tl(dc->cpu_R[instr.c], dc->cpu_R[instr.imm5 + CR_BASE]);
187
+ tcg_gen_mov_tl(cpu_R[instr.c], cpu_R[instr.imm5 + CR_BASE]);
188
}
189
break;
190
}
191
@@ -XXX,XX +XXX,XX @@ static void wrctl(DisasContext *dc, uint32_t code, uint32_t flags)
192
}
193
194
default:
195
- tcg_gen_mov_tl(dc->cpu_R[instr.imm5 + CR_BASE], load_gpr(dc, instr.a));
196
+ tcg_gen_mov_tl(cpu_R[instr.imm5 + CR_BASE], load_gpr(dc, instr.a));
197
break;
198
}
199
200
@@ -XXX,XX +XXX,XX @@ static void gen_cmpxx(DisasContext *dc, uint32_t code, uint32_t flags)
201
{
202
R_TYPE(instr, code);
203
if (likely(instr.c != R_ZERO)) {
204
- tcg_gen_setcond_tl(flags, dc->cpu_R[instr.c], dc->cpu_R[instr.a],
205
- dc->cpu_R[instr.b]);
206
+ tcg_gen_setcond_tl(flags, cpu_R[instr.c], cpu_R[instr.a],
207
+ cpu_R[instr.b]);
208
}
209
}
210
211
@@ -XXX,XX +XXX,XX @@ static void (fname)(DisasContext *dc, uint32_t code, uint32_t flags) \
212
{ \
213
R_TYPE(instr, (code)); \
214
if (likely(instr.c != R_ZERO)) { \
215
- tcg_gen_##insn((dc)->cpu_R[instr.c], load_gpr((dc), instr.a), \
216
- (op3)); \
217
+ tcg_gen_##insn(cpu_R[instr.c], load_gpr((dc), instr.a), (op3)); \
218
} \
219
}
220
221
@@ -XXX,XX +XXX,XX @@ static void (fname)(DisasContext *dc, uint32_t code, uint32_t flags) \
222
R_TYPE(instr, (code)); \
223
if (likely(instr.c != R_ZERO)) { \
224
TCGv t0 = tcg_temp_new(); \
225
- tcg_gen_##insn(t0, dc->cpu_R[instr.c], \
226
- load_gpr(dc, instr.a), load_gpr(dc, instr.b)); \
227
+ tcg_gen_##insn(t0, cpu_R[instr.c], \
228
+ load_gpr(dc, instr.a), load_gpr(dc, instr.b)); \
229
tcg_temp_free(t0); \
230
} \
231
}
232
@@ -XXX,XX +XXX,XX @@ static void (fname)(DisasContext *dc, uint32_t code, uint32_t flags) \
233
if (likely(instr.c != R_ZERO)) { \
234
TCGv t0 = tcg_temp_new(); \
235
tcg_gen_andi_tl(t0, load_gpr((dc), instr.b), 31); \
236
- tcg_gen_##insn((dc)->cpu_R[instr.c], load_gpr((dc), instr.a), t0); \
237
+ tcg_gen_##insn(cpu_R[instr.c], load_gpr((dc), instr.a), t0); \
238
tcg_temp_free(t0); \
239
} \
240
}
241
@@ -XXX,XX +XXX,XX @@ static void divs(DisasContext *dc, uint32_t code, uint32_t flags)
242
tcg_gen_or_tl(t2, t2, t3);
243
tcg_gen_movi_tl(t3, 0);
244
tcg_gen_movcond_tl(TCG_COND_NE, t1, t2, t3, t2, t1);
245
- tcg_gen_div_tl(dc->cpu_R[instr.c], t0, t1);
246
- tcg_gen_ext32s_tl(dc->cpu_R[instr.c], dc->cpu_R[instr.c]);
247
+ tcg_gen_div_tl(cpu_R[instr.c], t0, t1);
248
+ tcg_gen_ext32s_tl(cpu_R[instr.c], cpu_R[instr.c]);
249
250
tcg_temp_free(t3);
251
tcg_temp_free(t2);
252
@@ -XXX,XX +XXX,XX @@ static void divu(DisasContext *dc, uint32_t code, uint32_t flags)
253
tcg_gen_ext32u_tl(t0, load_gpr(dc, instr.a));
254
tcg_gen_ext32u_tl(t1, load_gpr(dc, instr.b));
255
tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1);
256
- tcg_gen_divu_tl(dc->cpu_R[instr.c], t0, t1);
257
- tcg_gen_ext32s_tl(dc->cpu_R[instr.c], dc->cpu_R[instr.c]);
258
+ tcg_gen_divu_tl(cpu_R[instr.c], t0, t1);
259
+ tcg_gen_ext32s_tl(cpu_R[instr.c], cpu_R[instr.c]);
260
261
tcg_temp_free(t3);
262
tcg_temp_free(t2);
263
@@ -XXX,XX +XXX,XX @@ static const char * const regnames[] = {
264
"rpc"
265
};
266
267
-static TCGv cpu_R[NUM_CORE_REGS];
268
-
269
#include "exec/gen-icount.h"
270
271
static void gen_exception(DisasContext *dc, uint32_t excp)
272
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
273
int num_insns;
274
275
/* Initialize DC */
276
- dc->cpu_R = cpu_R;
277
dc->is_jmp = DISAS_NEXT;
278
dc->pc = tb->pc;
279
dc->tb = tb;
280
--
65
--
281
2.25.1
66
2.34.1
282
67
283
68
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
4
target/nios2/translate.c | 128 ++++++++++++++++++++-------------------
5
1 file changed, 65 insertions(+), 63 deletions(-)
6
1
7
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
8
index XXXXXXX..XXXXXXX 100644
9
--- a/target/nios2/translate.c
10
+++ b/target/nios2/translate.c
11
@@ -XXX,XX +XXX,XX @@ static void gen_exception(DisasContext *dc, uint32_t excp)
12
}
13
14
/* generate intermediate code for basic block 'tb'. */
15
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
16
+static void nios2_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
17
{
18
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
19
CPUNios2State *env = cs->env_ptr;
20
- DisasContext dc1, *dc = &dc1;
21
- int num_insns;
22
-
23
- /* Initialize DC */
24
-
25
- dc->base.tb = tb;
26
- dc->base.singlestep_enabled = cs->singlestep_enabled;
27
- dc->base.is_jmp = DISAS_NEXT;
28
- dc->base.pc_first = tb->pc;
29
- dc->base.pc_next = tb->pc;
30
+ int page_insns;
31
32
dc->mem_idx = cpu_mmu_index(env, false);
33
34
- /* Set up instruction counts */
35
- num_insns = 0;
36
- if (max_insns > 1) {
37
- int page_insns = (TARGET_PAGE_SIZE - (tb->pc & ~TARGET_PAGE_MASK)) / 4;
38
- if (max_insns > page_insns) {
39
- max_insns = page_insns;
40
- }
41
- }
42
+ /* Bound the number of insns to execute to those left on the page. */
43
+ page_insns = -(dc->base.pc_first | TARGET_PAGE_MASK) / 4;
44
+ dc->base.max_insns = MIN(page_insns, dc->base.max_insns);
45
+}
46
47
- gen_tb_start(tb);
48
- do {
49
- tcg_gen_insn_start(dc->base.pc_next);
50
- num_insns++;
51
+static void nios2_tr_tb_start(DisasContextBase *db, CPUState *cs)
52
+{
53
+}
54
55
- if (unlikely(cpu_breakpoint_test(cs, dc->base.pc_next, BP_ANY))) {
56
- gen_exception(dc, EXCP_DEBUG);
57
- /* The address covered by the breakpoint must be included in
58
- [tb->pc, tb->pc + tb->size) in order to for it to be
59
- properly cleared -- thus we increment the PC here so that
60
- the logic setting tb->size below does the right thing. */
61
- dc->pc += 4;
62
- break;
63
- }
64
+static void nios2_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
65
+{
66
+ tcg_gen_insn_start(dcbase->pc_next);
67
+}
68
69
- if (num_insns == max_insns && (tb_cflags(tb) & CF_LAST_IO)) {
70
- gen_io_start();
71
- }
72
+static bool nios2_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cs,
73
+ const CPUBreakpoint *bp)
74
+{
75
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
76
77
- dc->pc = dc->base.pc_next;
78
- dc->base.pc_next += 4;
79
+ gen_exception(dc, EXCP_DEBUG);
80
+ /*
81
+ * The address covered by the breakpoint must be included in
82
+ * [tb->pc, tb->pc + tb->size) in order to for it to be
83
+ * properly cleared -- thus we increment the PC here so that
84
+ * the logic setting tb->size below does the right thing.
85
+ */
86
+ dc->base.pc_next += 4;
87
+ return true;
88
+}
89
90
- /* Decode an instruction */
91
- handle_instruction(dc, env);
92
+static void nios2_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
93
+{
94
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
95
+ CPUNios2State *env = cs->env_ptr;
96
97
- /* Translation stops when a conditional branch is encountered.
98
- * Otherwise the subsequent code could get translated several times.
99
- * Also stop translation when a page boundary is reached. This
100
- * ensures prefetch aborts occur at the right place. */
101
- } while (!dc->base.is_jmp &&
102
- !tcg_op_buf_full() &&
103
- num_insns < max_insns);
104
+ dc->pc = dc->base.pc_next;
105
+ dc->base.pc_next += 4;
106
+
107
+ /* Decode an instruction */
108
+ handle_instruction(dc, env);
109
+}
110
+
111
+static void nios2_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
112
+{
113
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
114
115
/* Indicate where the next block should start */
116
switch (dc->base.is_jmp) {
117
- case DISAS_NEXT:
118
+ case DISAS_TOO_MANY:
119
case DISAS_UPDATE:
120
/* Save the current PC back into the CPU register */
121
tcg_gen_movi_tl(cpu_R[R_PC], dc->base.pc_next);
122
tcg_gen_exit_tb(NULL, 0);
123
break;
124
125
- default:
126
case DISAS_JUMP:
127
/* The jump will already have updated the PC register */
128
tcg_gen_exit_tb(NULL, 0);
129
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
130
case DISAS_NORETURN:
131
/* nothing more to generate */
132
break;
133
+
134
+ default:
135
+ g_assert_not_reached();
136
}
137
+}
138
139
- /* End off the block */
140
- gen_tb_end(tb, num_insns);
141
+static void nios2_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
142
+{
143
+ qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
144
+ log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size);
145
+}
146
147
- /* Mark instruction starts for the final generated instruction */
148
- tb->size = dc->base.pc_next - dc->base.pc_first;
149
- tb->icount = num_insns;
150
+static const TranslatorOps nios2_tr_ops = {
151
+ .init_disas_context = nios2_tr_init_disas_context,
152
+ .tb_start = nios2_tr_tb_start,
153
+ .insn_start = nios2_tr_insn_start,
154
+ .breakpoint_check = nios2_tr_breakpoint_check,
155
+ .translate_insn = nios2_tr_translate_insn,
156
+ .tb_stop = nios2_tr_tb_stop,
157
+ .disas_log = nios2_tr_disas_log,
158
+};
159
160
-#ifdef DEBUG_DISAS
161
- if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
162
- && qemu_log_in_addr_range(dc->base.pc_first)) {
163
- FILE *logfile = qemu_log_lock();
164
- qemu_log("IN: %s\n", lookup_symbol(dc->base.pc_first));
165
- log_target_disas(cs, tb->pc, tb->size);
166
- qemu_log("\n");
167
- qemu_log_unlock(logfile);
168
- }
169
-#endif
170
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
171
+{
172
+ DisasContext dc;
173
+ translator_loop(&nios2_tr_ops, &dc.base, cs, tb, max_insns);
174
}
175
176
void nios2_cpu_dump_state(CPUState *cs, FILE *f, int flags)
177
--
178
2.25.1
179
180
diff view generated by jsdifflib
Deleted patch
1
Direct assignments to env during translation do not work.
2
1
3
As it happens, the only way we can get here is if env->pc
4
is already set to dc->pc. We will trap on the first insn
5
we execute anywhere on the page.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
10
target/nios2/translate.c | 3 ++-
11
1 file changed, 2 insertions(+), 1 deletion(-)
12
13
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/nios2/translate.c
16
+++ b/target/nios2/translate.c
17
@@ -XXX,XX +XXX,XX @@ static void handle_instruction(DisasContext *dc, CPUNios2State *env)
18
uint32_t code;
19
uint8_t op;
20
const Nios2Instruction *instr;
21
+
22
#if defined(CONFIG_USER_ONLY)
23
/* FIXME: Is this needed ? */
24
if (dc->pc >= 0x1000 && dc->pc < 0x2000) {
25
- env->regs[R_PC] = dc->pc;
26
t_gen_helper_raise_exception(dc, 0xaa);
27
return;
28
}
29
#endif
30
+
31
code = cpu_ldl_code(env, dc->pc);
32
op = get_opcode(code);
33
34
--
35
2.25.1
36
37
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
4
target/nios2/translate.c | 8 ++------
5
1 file changed, 2 insertions(+), 6 deletions(-)
6
1
7
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
8
index XXXXXXX..XXXXXXX 100644
9
--- a/target/nios2/translate.c
10
+++ b/target/nios2/translate.c
11
@@ -XXX,XX +XXX,XX @@ static void handle_instruction(DisasContext *dc, CPUNios2State *env)
12
op = get_opcode(code);
13
14
if (unlikely(op >= ARRAY_SIZE(i_type_instructions))) {
15
- goto illegal_op;
16
+ t_gen_helper_raise_exception(dc, EXCP_ILLEGAL);
17
+ return;
18
}
19
20
dc->zero = NULL;
21
@@ -XXX,XX +XXX,XX @@ static void handle_instruction(DisasContext *dc, CPUNios2State *env)
22
if (dc->zero) {
23
tcg_temp_free(dc->zero);
24
}
25
-
26
- return;
27
-
28
-illegal_op:
29
- t_gen_helper_raise_exception(dc, EXCP_ILLEGAL);
30
}
31
32
static const char * const regnames[] = {
33
--
34
2.25.1
35
36
diff view generated by jsdifflib
Deleted patch
1
Prepare for receiving it as a pointer input.
2
1
3
Tested-by: Michael Rolnik <mrolnik@gmail.com>
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/avr/translate.c | 84 +++++++++++++++++++++---------------------
9
1 file changed, 43 insertions(+), 41 deletions(-)
10
11
diff --git a/target/avr/translate.c b/target/avr/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/avr/translate.c
14
+++ b/target/avr/translate.c
15
@@ -XXX,XX +XXX,XX @@ struct DisasContext {
16
* used in the following manner (sketch)
17
*
18
* TCGLabel *skip_label = NULL;
19
- * if (ctx.skip_cond != TCG_COND_NEVER) {
20
+ * if (ctx->skip_cond != TCG_COND_NEVER) {
21
* skip_label = gen_new_label();
22
* tcg_gen_brcond_tl(skip_cond, skip_var0, skip_var1, skip_label);
23
* }
24
@@ -XXX,XX +XXX,XX @@ struct DisasContext {
25
* free_skip_var0 = false;
26
* }
27
*
28
- * translate(&ctx);
29
+ * translate(ctx);
30
*
31
* if (skip_label) {
32
* gen_set_label(skip_label);
33
@@ -XXX,XX +XXX,XX @@ static bool canonicalize_skip(DisasContext *ctx)
34
void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
35
{
36
CPUAVRState *env = cs->env_ptr;
37
- DisasContext ctx = {
38
+ DisasContext ctx1 = {
39
.base.tb = tb,
40
.base.is_jmp = DISAS_NEXT,
41
.base.pc_first = tb->pc,
42
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
43
.memidx = 0,
44
.skip_cond = TCG_COND_NEVER,
45
};
46
+ DisasContext *ctx = &ctx1;
47
target_ulong pc_start = tb->pc / 2;
48
int num_insns = 0;
49
50
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
51
*/
52
max_insns = 1;
53
}
54
- if (ctx.base.singlestep_enabled) {
55
+ if (ctx->base.singlestep_enabled) {
56
max_insns = 1;
57
}
58
59
gen_tb_start(tb);
60
61
- ctx.npc = pc_start;
62
+ ctx->npc = pc_start;
63
if (tb->flags & TB_FLAGS_SKIP) {
64
- ctx.skip_cond = TCG_COND_ALWAYS;
65
- ctx.skip_var0 = cpu_skip;
66
+ ctx->skip_cond = TCG_COND_ALWAYS;
67
+ ctx->skip_var0 = cpu_skip;
68
}
69
70
do {
71
TCGLabel *skip_label = NULL;
72
73
/* translate current instruction */
74
- tcg_gen_insn_start(ctx.npc);
75
+ tcg_gen_insn_start(ctx->npc);
76
num_insns++;
77
78
/*
79
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
80
* b main - sets breakpoint at address 0x00000100 (code)
81
* b *0x100 - sets breakpoint at address 0x00800100 (data)
82
*/
83
- if (unlikely(!ctx.base.singlestep_enabled &&
84
- (cpu_breakpoint_test(cs, OFFSET_CODE + ctx.npc * 2, BP_ANY) ||
85
- cpu_breakpoint_test(cs, OFFSET_DATA + ctx.npc * 2, BP_ANY)))) {
86
- canonicalize_skip(&ctx);
87
- tcg_gen_movi_tl(cpu_pc, ctx.npc);
88
+ if (unlikely(!ctx->base.singlestep_enabled &&
89
+ (cpu_breakpoint_test(cs, OFFSET_CODE + ctx->npc * 2, BP_ANY) ||
90
+ cpu_breakpoint_test(cs, OFFSET_DATA + ctx->npc * 2, BP_ANY)))) {
91
+ canonicalize_skip(ctx);
92
+ tcg_gen_movi_tl(cpu_pc, ctx->npc);
93
gen_helper_debug(cpu_env);
94
goto done_generating;
95
}
96
97
/* Conditionally skip the next instruction, if indicated. */
98
- if (ctx.skip_cond != TCG_COND_NEVER) {
99
+ if (ctx->skip_cond != TCG_COND_NEVER) {
100
skip_label = gen_new_label();
101
- if (ctx.skip_var0 == cpu_skip) {
102
+ if (ctx->skip_var0 == cpu_skip) {
103
/*
104
* Copy cpu_skip so that we may zero it before the branch.
105
* This ensures that cpu_skip is non-zero after the label
106
* if and only if the skipped insn itself sets a skip.
107
*/
108
- ctx.free_skip_var0 = true;
109
- ctx.skip_var0 = tcg_temp_new();
110
- tcg_gen_mov_tl(ctx.skip_var0, cpu_skip);
111
+ ctx->free_skip_var0 = true;
112
+ ctx->skip_var0 = tcg_temp_new();
113
+ tcg_gen_mov_tl(ctx->skip_var0, cpu_skip);
114
tcg_gen_movi_tl(cpu_skip, 0);
115
}
116
- if (ctx.skip_var1 == NULL) {
117
- tcg_gen_brcondi_tl(ctx.skip_cond, ctx.skip_var0, 0, skip_label);
118
+ if (ctx->skip_var1 == NULL) {
119
+ tcg_gen_brcondi_tl(ctx->skip_cond, ctx->skip_var0,
120
+ 0, skip_label);
121
} else {
122
- tcg_gen_brcond_tl(ctx.skip_cond, ctx.skip_var0,
123
- ctx.skip_var1, skip_label);
124
- ctx.skip_var1 = NULL;
125
+ tcg_gen_brcond_tl(ctx->skip_cond, ctx->skip_var0,
126
+ ctx->skip_var1, skip_label);
127
+ ctx->skip_var1 = NULL;
128
}
129
- if (ctx.free_skip_var0) {
130
- tcg_temp_free(ctx.skip_var0);
131
- ctx.free_skip_var0 = false;
132
+ if (ctx->free_skip_var0) {
133
+ tcg_temp_free(ctx->skip_var0);
134
+ ctx->free_skip_var0 = false;
135
}
136
- ctx.skip_cond = TCG_COND_NEVER;
137
- ctx.skip_var0 = NULL;
138
+ ctx->skip_cond = TCG_COND_NEVER;
139
+ ctx->skip_var0 = NULL;
140
}
141
142
- translate(&ctx);
143
+ translate(ctx);
144
145
if (skip_label) {
146
- canonicalize_skip(&ctx);
147
+ canonicalize_skip(ctx);
148
gen_set_label(skip_label);
149
- if (ctx.base.is_jmp == DISAS_NORETURN) {
150
- ctx.base.is_jmp = DISAS_CHAIN;
151
+ if (ctx->base.is_jmp == DISAS_NORETURN) {
152
+ ctx->base.is_jmp = DISAS_CHAIN;
153
}
154
}
155
- } while (ctx.base.is_jmp == DISAS_NEXT
156
+ } while (ctx->base.is_jmp == DISAS_NEXT
157
&& num_insns < max_insns
158
- && (ctx.npc - pc_start) * 2 < TARGET_PAGE_SIZE - 4
159
+ && (ctx->npc - pc_start) * 2 < TARGET_PAGE_SIZE - 4
160
&& !tcg_op_buf_full());
161
162
if (tb->cflags & CF_LAST_IO) {
163
gen_io_end();
164
}
165
166
- bool nonconst_skip = canonicalize_skip(&ctx);
167
+ bool nonconst_skip = canonicalize_skip(ctx);
168
169
- switch (ctx.base.is_jmp) {
170
+ switch (ctx->base.is_jmp) {
171
case DISAS_NORETURN:
172
assert(!nonconst_skip);
173
break;
174
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
175
case DISAS_CHAIN:
176
if (!nonconst_skip) {
177
/* Note gen_goto_tb checks singlestep. */
178
- gen_goto_tb(&ctx, 1, ctx.npc);
179
+ gen_goto_tb(ctx, 1, ctx->npc);
180
break;
181
}
182
- tcg_gen_movi_tl(cpu_pc, ctx.npc);
183
+ tcg_gen_movi_tl(cpu_pc, ctx->npc);
184
/* fall through */
185
case DISAS_LOOKUP:
186
- if (!ctx.base.singlestep_enabled) {
187
+ if (!ctx->base.singlestep_enabled) {
188
tcg_gen_lookup_and_goto_ptr();
189
break;
190
}
191
/* fall through */
192
case DISAS_EXIT:
193
- if (ctx.base.singlestep_enabled) {
194
+ if (ctx->base.singlestep_enabled) {
195
gen_helper_debug(cpu_env);
196
} else {
197
tcg_gen_exit_tb(NULL, 0);
198
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
199
done_generating:
200
gen_tb_end(tb, num_insns);
201
202
- tb->size = (ctx.npc - pc_start) * 2;
203
+ tb->size = (ctx->npc - pc_start) * 2;
204
tb->icount = num_insns;
205
206
#ifdef DEBUG_DISAS
207
--
208
2.25.1
209
210
diff view generated by jsdifflib
1
Merge tcg_out_bswap32 and tcg_out_bswap32s.
1
Wrap the bare TranslationBlock pointer into a structure.
2
Use the flags in the internal uses for loads and stores.
3
2
4
For mips32r2 bswap32 with zero-extension, standardize on
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
WSBH+ROTR+DEXT. This is the same number of insns as the
6
previous DSBH+DSHD+DSRL but fits in better with the flags check.
7
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
6
---
11
tcg/mips/tcg-target.c.inc | 39 ++++++++++++++++-----------------------
7
accel/tcg/tb-hash.h | 1 +
12
1 file changed, 16 insertions(+), 23 deletions(-)
8
accel/tcg/tb-jmp-cache.h | 24 ++++++++++++++++++++++++
9
include/exec/cpu-common.h | 1 +
10
include/hw/core/cpu.h | 15 +--------------
11
include/qemu/typedefs.h | 1 +
12
accel/stubs/tcg-stub.c | 4 ++++
13
accel/tcg/cpu-exec.c | 10 +++++++---
14
accel/tcg/cputlb.c | 9 +++++----
15
accel/tcg/translate-all.c | 28 +++++++++++++++++++++++++---
16
hw/core/cpu-common.c | 3 +--
17
plugins/core.c | 2 +-
18
trace/control-target.c | 2 +-
19
12 files changed, 72 insertions(+), 28 deletions(-)
20
create mode 100644 accel/tcg/tb-jmp-cache.h
13
21
14
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
22
diff --git a/accel/tcg/tb-hash.h b/accel/tcg/tb-hash.h
15
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
16
--- a/tcg/mips/tcg-target.c.inc
24
--- a/accel/tcg/tb-hash.h
17
+++ b/tcg/mips/tcg-target.c.inc
25
+++ b/accel/tcg/tb-hash.h
18
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap_subr(TCGContext *s, const tcg_insn_unit *sub)
26
@@ -XXX,XX +XXX,XX @@
19
tcg_debug_assert(ok);
27
#include "exec/cpu-defs.h"
20
}
28
#include "exec/exec-all.h"
21
29
#include "qemu/xxhash.h"
22
-static void tcg_out_bswap32(TCGContext *s, TCGReg ret, TCGReg arg)
30
+#include "tb-jmp-cache.h"
23
+static void tcg_out_bswap32(TCGContext *s, TCGReg ret, TCGReg arg, int flags)
31
24
{
32
#ifdef CONFIG_SOFTMMU
25
if (use_mips32r2_instructions) {
33
26
tcg_out_opc_reg(s, OPC_WSBH, ret, 0, arg);
34
diff --git a/accel/tcg/tb-jmp-cache.h b/accel/tcg/tb-jmp-cache.h
27
tcg_out_opc_sa(s, OPC_ROTR, ret, ret, 16);
35
new file mode 100644
28
+ if (flags & TCG_BSWAP_OZ) {
36
index XXXXXXX..XXXXXXX
29
+ tcg_out_opc_bf(s, OPC_DEXT, ret, ret, 31, 0);
37
--- /dev/null
30
+ }
38
+++ b/accel/tcg/tb-jmp-cache.h
31
} else {
39
@@ -XXX,XX +XXX,XX @@
32
- tcg_out_bswap_subr(s, bswap32_addr);
40
+/*
33
- /* delay slot -- never omit the insn, like tcg_out_mov might. */
41
+ * The per-CPU TranslationBlock jump cache.
34
- tcg_out_opc_reg(s, OPC_OR, TCG_TMP0, arg, TCG_REG_ZERO);
42
+ *
35
- tcg_out_mov(s, TCG_TYPE_I32, ret, TCG_TMP3);
43
+ * Copyright (c) 2003 Fabrice Bellard
44
+ *
45
+ * SPDX-License-Identifier: GPL-2.0-or-later
46
+ */
47
+
48
+#ifndef ACCEL_TCG_TB_JMP_CACHE_H
49
+#define ACCEL_TCG_TB_JMP_CACHE_H
50
+
51
+#define TB_JMP_CACHE_BITS 12
52
+#define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
53
+
54
+/*
55
+ * Accessed in parallel; all accesses to 'tb' must be atomic.
56
+ */
57
+struct CPUJumpCache {
58
+ struct {
59
+ TranslationBlock *tb;
60
+ } array[TB_JMP_CACHE_SIZE];
61
+};
62
+
63
+#endif /* ACCEL_TCG_TB_JMP_CACHE_H */
64
diff --git a/include/exec/cpu-common.h b/include/exec/cpu-common.h
65
index XXXXXXX..XXXXXXX 100644
66
--- a/include/exec/cpu-common.h
67
+++ b/include/exec/cpu-common.h
68
@@ -XXX,XX +XXX,XX @@ void cpu_list_unlock(void);
69
unsigned int cpu_list_generation_id_get(void);
70
71
void tcg_flush_softmmu_tlb(CPUState *cs);
72
+void tcg_flush_jmp_cache(CPUState *cs);
73
74
void tcg_iommu_init_notifier_list(CPUState *cpu);
75
void tcg_iommu_free_notifier_list(CPUState *cpu);
76
diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
77
index XXXXXXX..XXXXXXX 100644
78
--- a/include/hw/core/cpu.h
79
+++ b/include/hw/core/cpu.h
80
@@ -XXX,XX +XXX,XX @@ struct kvm_run;
81
struct hax_vcpu_state;
82
struct hvf_vcpu_state;
83
84
-#define TB_JMP_CACHE_BITS 12
85
-#define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
86
-
87
/* work queue */
88
89
/* The union type allows passing of 64 bit target pointers on 32 bit
90
@@ -XXX,XX +XXX,XX @@ struct CPUState {
91
CPUArchState *env_ptr;
92
IcountDecr *icount_decr_ptr;
93
94
- /* Accessed in parallel; all accesses must be atomic */
95
- TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE];
96
+ CPUJumpCache *tb_jmp_cache;
97
98
struct GDBRegisterState *gdb_regs;
99
int gdb_num_regs;
100
@@ -XXX,XX +XXX,XX @@ extern CPUTailQ cpus;
101
102
extern __thread CPUState *current_cpu;
103
104
-static inline void cpu_tb_jmp_cache_clear(CPUState *cpu)
105
-{
106
- unsigned int i;
107
-
108
- for (i = 0; i < TB_JMP_CACHE_SIZE; i++) {
109
- qatomic_set(&cpu->tb_jmp_cache[i], NULL);
36
- }
110
- }
37
-}
111
-}
38
-
112
-
39
-static void tcg_out_bswap32u(TCGContext *s, TCGReg ret, TCGReg arg)
113
/**
40
-{
114
* qemu_tcg_mttcg_enabled:
41
- if (use_mips32r2_instructions) {
115
* Check whether we are running MultiThread TCG or not.
42
- tcg_out_opc_reg(s, OPC_DSBH, ret, 0, arg);
116
diff --git a/include/qemu/typedefs.h b/include/qemu/typedefs.h
43
- tcg_out_opc_reg(s, OPC_DSHD, ret, 0, ret);
117
index XXXXXXX..XXXXXXX 100644
44
- tcg_out_dsrl(s, ret, ret, 32);
118
--- a/include/qemu/typedefs.h
45
- } else {
119
+++ b/include/qemu/typedefs.h
46
- tcg_out_bswap_subr(s, bswap32u_addr);
120
@@ -XXX,XX +XXX,XX @@ typedef struct CoMutex CoMutex;
47
+ if (flags & TCG_BSWAP_OZ) {
121
typedef struct ConfidentialGuestSupport ConfidentialGuestSupport;
48
+ tcg_out_bswap_subr(s, bswap32u_addr);
122
typedef struct CPUAddressSpace CPUAddressSpace;
49
+ } else {
123
typedef struct CPUArchState CPUArchState;
50
+ tcg_out_bswap_subr(s, bswap32_addr);
124
+typedef struct CPUJumpCache CPUJumpCache;
125
typedef struct CPUState CPUState;
126
typedef struct CPUTLBEntryFull CPUTLBEntryFull;
127
typedef struct DeviceListener DeviceListener;
128
diff --git a/accel/stubs/tcg-stub.c b/accel/stubs/tcg-stub.c
129
index XXXXXXX..XXXXXXX 100644
130
--- a/accel/stubs/tcg-stub.c
131
+++ b/accel/stubs/tcg-stub.c
132
@@ -XXX,XX +XXX,XX @@ void tlb_set_dirty(CPUState *cpu, target_ulong vaddr)
133
{
134
}
135
136
+void tcg_flush_jmp_cache(CPUState *cpu)
137
+{
138
+}
139
+
140
int probe_access_flags(CPUArchState *env, target_ulong addr,
141
MMUAccessType access_type, int mmu_idx,
142
bool nonfault, void **phost, uintptr_t retaddr)
143
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
144
index XXXXXXX..XXXXXXX 100644
145
--- a/accel/tcg/cpu-exec.c
146
+++ b/accel/tcg/cpu-exec.c
147
@@ -XXX,XX +XXX,XX @@
148
#include "sysemu/replay.h"
149
#include "sysemu/tcg.h"
150
#include "exec/helper-proto.h"
151
+#include "tb-jmp-cache.h"
152
#include "tb-hash.h"
153
#include "tb-context.h"
154
#include "internal.h"
155
@@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
156
tcg_debug_assert(!(cflags & CF_INVALID));
157
158
hash = tb_jmp_cache_hash_func(pc);
159
- tb = qatomic_rcu_read(&cpu->tb_jmp_cache[hash]);
160
+ tb = qatomic_rcu_read(&cpu->tb_jmp_cache->array[hash].tb);
161
162
if (likely(tb &&
163
tb->pc == pc &&
164
@@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
165
if (tb == NULL) {
166
return NULL;
167
}
168
- qatomic_set(&cpu->tb_jmp_cache[hash], tb);
169
+ qatomic_set(&cpu->tb_jmp_cache->array[hash].tb, tb);
170
return tb;
171
}
172
173
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
174
175
tb = tb_lookup(cpu, pc, cs_base, flags, cflags);
176
if (tb == NULL) {
177
+ uint32_t h;
178
+
179
mmap_lock();
180
tb = tb_gen_code(cpu, pc, cs_base, flags, cflags);
181
mmap_unlock();
182
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
183
* We add the TB in the virtual pc hash table
184
* for the fast lookup
185
*/
186
- qatomic_set(&cpu->tb_jmp_cache[tb_jmp_cache_hash_func(pc)], tb);
187
+ h = tb_jmp_cache_hash_func(pc);
188
+ qatomic_set(&cpu->tb_jmp_cache->array[h].tb, tb);
189
}
190
191
#ifndef CONFIG_USER_ONLY
192
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
193
index XXXXXXX..XXXXXXX 100644
194
--- a/accel/tcg/cputlb.c
195
+++ b/accel/tcg/cputlb.c
196
@@ -XXX,XX +XXX,XX @@ static void tlb_window_reset(CPUTLBDesc *desc, int64_t ns,
197
198
static void tb_jmp_cache_clear_page(CPUState *cpu, target_ulong page_addr)
199
{
200
- unsigned int i, i0 = tb_jmp_cache_hash_page(page_addr);
201
+ int i, i0 = tb_jmp_cache_hash_page(page_addr);
202
+ CPUJumpCache *jc = cpu->tb_jmp_cache;
203
204
for (i = 0; i < TB_JMP_PAGE_SIZE; i++) {
205
- qatomic_set(&cpu->tb_jmp_cache[i0 + i], NULL);
206
+ qatomic_set(&jc->array[i0 + i].tb, NULL);
207
}
208
}
209
210
@@ -XXX,XX +XXX,XX @@ static void tlb_flush_by_mmuidx_async_work(CPUState *cpu, run_on_cpu_data data)
211
212
qemu_spin_unlock(&env_tlb(env)->c.lock);
213
214
- cpu_tb_jmp_cache_clear(cpu);
215
+ tcg_flush_jmp_cache(cpu);
216
217
if (to_clean == ALL_MMUIDX_BITS) {
218
qatomic_set(&env_tlb(env)->c.full_flush_count,
219
@@ -XXX,XX +XXX,XX @@ static void tlb_flush_range_by_mmuidx_async_0(CPUState *cpu,
220
* longer to clear each entry individually than it will to clear it all.
221
*/
222
if (d.len >= (TARGET_PAGE_SIZE * TB_JMP_CACHE_SIZE)) {
223
- cpu_tb_jmp_cache_clear(cpu);
224
+ tcg_flush_jmp_cache(cpu);
225
return;
226
}
227
228
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
229
index XXXXXXX..XXXXXXX 100644
230
--- a/accel/tcg/translate-all.c
231
+++ b/accel/tcg/translate-all.c
232
@@ -XXX,XX +XXX,XX @@
233
#include "sysemu/tcg.h"
234
#include "qapi/error.h"
235
#include "hw/core/tcg-cpu-ops.h"
236
+#include "tb-jmp-cache.h"
237
#include "tb-hash.h"
238
#include "tb-context.h"
239
#include "internal.h"
240
@@ -XXX,XX +XXX,XX @@ static void do_tb_flush(CPUState *cpu, run_on_cpu_data tb_flush_count)
241
}
242
243
CPU_FOREACH(cpu) {
244
- cpu_tb_jmp_cache_clear(cpu);
245
+ tcg_flush_jmp_cache(cpu);
246
}
247
248
qht_reset_size(&tb_ctx.htable, CODE_GEN_HTABLE_SIZE);
249
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
250
/* remove the TB from the hash list */
251
h = tb_jmp_cache_hash_func(tb->pc);
252
CPU_FOREACH(cpu) {
253
- if (qatomic_read(&cpu->tb_jmp_cache[h]) == tb) {
254
- qatomic_set(&cpu->tb_jmp_cache[h], NULL);
255
+ CPUJumpCache *jc = cpu->tb_jmp_cache;
256
+ if (qatomic_read(&jc->array[h].tb) == tb) {
257
+ qatomic_set(&jc->array[h].tb, NULL);
258
}
259
}
260
261
@@ -XXX,XX +XXX,XX @@ int page_unprotect(target_ulong address, uintptr_t pc)
262
}
263
#endif /* CONFIG_USER_ONLY */
264
265
+/*
266
+ * Called by generic code at e.g. cpu reset after cpu creation,
267
+ * therefore we must be prepared to allocate the jump cache.
268
+ */
269
+void tcg_flush_jmp_cache(CPUState *cpu)
270
+{
271
+ CPUJumpCache *jc = cpu->tb_jmp_cache;
272
+
273
+ if (likely(jc)) {
274
+ for (int i = 0; i < TB_JMP_CACHE_SIZE; i++) {
275
+ qatomic_set(&jc->array[i].tb, NULL);
51
+ }
276
+ }
52
/* delay slot -- never omit the insn, like tcg_out_mov might. */
277
+ } else {
53
tcg_out_opc_reg(s, OPC_OR, TCG_TMP0, arg, TCG_REG_ZERO);
278
+ /* This should happen once during realize, and thus never race. */
54
tcg_out_mov(s, TCG_TYPE_I32, ret, TCG_TMP3);
279
+ jc = g_new0(CPUJumpCache, 1);
55
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
280
+ jc = qatomic_xchg(&cpu->tb_jmp_cache, jc);
56
if (TCG_TARGET_REG_BITS == 64 && is_64) {
281
+ assert(jc == NULL);
57
if (use_mips32r2_instructions) {
282
+ }
58
tcg_out_opc_imm(s, OPC_LWU, lo, base, 0);
283
+}
59
- tcg_out_bswap32u(s, lo, lo);
284
+
60
+ tcg_out_bswap32(s, lo, lo, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
285
/* This is a wrapper for common code that can not use CONFIG_SOFTMMU */
61
} else {
286
void tcg_flush_softmmu_tlb(CPUState *cs)
62
tcg_out_bswap_subr(s, bswap32u_addr);
287
{
63
/* delay slot */
288
diff --git a/hw/core/cpu-common.c b/hw/core/cpu-common.c
64
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
289
index XXXXXXX..XXXXXXX 100644
65
case MO_SL | MO_BSWAP:
290
--- a/hw/core/cpu-common.c
66
if (use_mips32r2_instructions) {
291
+++ b/hw/core/cpu-common.c
67
tcg_out_opc_imm(s, OPC_LW, lo, base, 0);
292
@@ -XXX,XX +XXX,XX @@ static void cpu_common_reset(DeviceState *dev)
68
- tcg_out_bswap32(s, lo, lo);
293
cpu->cflags_next_tb = -1;
69
+ tcg_out_bswap32(s, lo, lo, 0);
294
70
} else {
295
if (tcg_enabled()) {
71
tcg_out_bswap_subr(s, bswap32_addr);
296
- cpu_tb_jmp_cache_clear(cpu);
72
/* delay slot */
297
-
73
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
298
+ tcg_flush_jmp_cache(cpu);
74
break;
299
tcg_flush_softmmu_tlb(cpu);
75
300
}
76
case MO_32 | MO_BSWAP:
301
}
77
- tcg_out_bswap32(s, TCG_TMP3, lo);
302
diff --git a/plugins/core.c b/plugins/core.c
78
+ tcg_out_bswap32(s, TCG_TMP3, lo, 0);
303
index XXXXXXX..XXXXXXX 100644
79
lo = TCG_TMP3;
304
--- a/plugins/core.c
80
/* FALLTHRU */
305
+++ b/plugins/core.c
81
case MO_32:
306
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_ctx *plugin_id_to_ctx_locked(qemu_plugin_id_t id)
82
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
307
static void plugin_cpu_update__async(CPUState *cpu, run_on_cpu_data data)
83
tcg_out_opc_imm(s, OPC_SW, TCG_TMP0, base, 0);
308
{
84
tcg_out_opc_imm(s, OPC_SW, TCG_TMP1, base, 4);
309
bitmap_copy(cpu->plugin_mask, &data.host_ulong, QEMU_PLUGIN_EV_MAX);
85
} else {
310
- cpu_tb_jmp_cache_clear(cpu);
86
- tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? lo : hi);
311
+ tcg_flush_jmp_cache(cpu);
87
+ tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? lo : hi, 0);
312
}
88
tcg_out_opc_imm(s, OPC_SW, TCG_TMP3, base, 0);
313
89
- tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? hi : lo);
314
static void plugin_cpu_update__locked(gpointer k, gpointer v, gpointer udata)
90
+ tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? hi : lo, 0);
315
diff --git a/trace/control-target.c b/trace/control-target.c
91
tcg_out_opc_imm(s, OPC_SW, TCG_TMP3, base, 4);
316
index XXXXXXX..XXXXXXX 100644
92
}
317
--- a/trace/control-target.c
93
break;
318
+++ b/trace/control-target.c
94
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
319
@@ -XXX,XX +XXX,XX @@ static void trace_event_synchronize_vcpu_state_dynamic(
95
tcg_out_bswap16(s, a0, a1, a2);
320
{
96
break;
321
bitmap_copy(vcpu->trace_dstate, vcpu->trace_dstate_delayed,
97
case INDEX_op_bswap32_i32:
322
CPU_TRACE_DSTATE_MAX_EVENTS);
98
- tcg_out_bswap32(s, a0, a1);
323
- cpu_tb_jmp_cache_clear(vcpu);
99
+ tcg_out_bswap32(s, a0, a1, 0);
324
+ tcg_flush_jmp_cache(vcpu);
100
break;
325
}
101
case INDEX_op_bswap32_i64:
326
102
- tcg_out_bswap32u(s, a0, a1);
327
void trace_event_set_vcpu_state_dynamic(CPUState *vcpu,
103
+ tcg_out_bswap32(s, a0, a1, a2);
104
break;
105
case INDEX_op_bswap64_i64:
106
tcg_out_bswap64(s, a0, a1);
107
--
328
--
108
2.25.1
329
2.34.1
109
330
110
331
diff view generated by jsdifflib
1
Now that the middle-end can replicate the same tricks as tcg/arm
1
Populate this new method for all targets. Always match
2
used for optimizing bswap for signed loads and for stores, do not
2
the result that would be given by cpu_get_tb_cpu_state,
3
pretend to have these memory ops in the backend.
3
as we will want these values to correspond in the logs.
4
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Taylor Simpson <tsimpson@quicinc.com>
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
7
Reviewed-by: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk> (target/sparc)
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
9
---
8
tcg/arm/tcg-target.h | 2 +-
10
Cc: Eduardo Habkost <eduardo@habkost.net> (supporter:Machine core)
9
tcg/arm/tcg-target.c.inc | 214 ++++++++++++++-------------------------
11
Cc: Marcel Apfelbaum <marcel.apfelbaum@gmail.com> (supporter:Machine core)
10
2 files changed, 77 insertions(+), 139 deletions(-)
12
Cc: "Philippe Mathieu-Daudé" <f4bug@amsat.org> (reviewer:Machine core)
13
Cc: Yanan Wang <wangyanan55@huawei.com> (reviewer:Machine core)
14
Cc: Michael Rolnik <mrolnik@gmail.com> (maintainer:AVR TCG CPUs)
15
Cc: "Edgar E. Iglesias" <edgar.iglesias@gmail.com> (maintainer:CRIS TCG CPUs)
16
Cc: Taylor Simpson <tsimpson@quicinc.com> (supporter:Hexagon TCG CPUs)
17
Cc: Song Gao <gaosong@loongson.cn> (maintainer:LoongArch TCG CPUs)
18
Cc: Xiaojuan Yang <yangxiaojuan@loongson.cn> (maintainer:LoongArch TCG CPUs)
19
Cc: Laurent Vivier <laurent@vivier.eu> (maintainer:M68K TCG CPUs)
20
Cc: Jiaxun Yang <jiaxun.yang@flygoat.com> (reviewer:MIPS TCG CPUs)
21
Cc: Aleksandar Rikalo <aleksandar.rikalo@syrmia.com> (reviewer:MIPS TCG CPUs)
22
Cc: Chris Wulff <crwulff@gmail.com> (maintainer:NiosII TCG CPUs)
23
Cc: Marek Vasut <marex@denx.de> (maintainer:NiosII TCG CPUs)
24
Cc: Stafford Horne <shorne@gmail.com> (odd fixer:OpenRISC TCG CPUs)
25
Cc: Yoshinori Sato <ysato@users.sourceforge.jp> (reviewer:RENESAS RX CPUs)
26
Cc: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk> (maintainer:SPARC TCG CPUs)
27
Cc: Bastian Koppelmann <kbastian@mail.uni-paderborn.de> (maintainer:TriCore TCG CPUs)
28
Cc: Max Filippov <jcmvbkbc@gmail.com> (maintainer:Xtensa TCG CPUs)
29
Cc: qemu-arm@nongnu.org (open list:ARM TCG CPUs)
30
Cc: qemu-ppc@nongnu.org (open list:PowerPC TCG CPUs)
31
Cc: qemu-riscv@nongnu.org (open list:RISC-V TCG CPUs)
32
Cc: qemu-s390x@nongnu.org (open list:S390 TCG CPUs)
33
---
34
include/hw/core/cpu.h | 3 +++
35
target/alpha/cpu.c | 9 +++++++++
36
target/arm/cpu.c | 13 +++++++++++++
37
target/avr/cpu.c | 8 ++++++++
38
target/cris/cpu.c | 8 ++++++++
39
target/hexagon/cpu.c | 8 ++++++++
40
target/hppa/cpu.c | 8 ++++++++
41
target/i386/cpu.c | 9 +++++++++
42
target/loongarch/cpu.c | 9 +++++++++
43
target/m68k/cpu.c | 8 ++++++++
44
target/microblaze/cpu.c | 8 ++++++++
45
target/mips/cpu.c | 8 ++++++++
46
target/nios2/cpu.c | 9 +++++++++
47
target/openrisc/cpu.c | 8 ++++++++
48
target/ppc/cpu_init.c | 8 ++++++++
49
target/riscv/cpu.c | 13 +++++++++++++
50
target/rx/cpu.c | 8 ++++++++
51
target/s390x/cpu.c | 8 ++++++++
52
target/sh4/cpu.c | 8 ++++++++
53
target/sparc/cpu.c | 8 ++++++++
54
target/tricore/cpu.c | 9 +++++++++
55
target/xtensa/cpu.c | 8 ++++++++
56
22 files changed, 186 insertions(+)
11
57
12
diff --git a/tcg/arm/tcg-target.h b/tcg/arm/tcg-target.h
58
diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
13
index XXXXXXX..XXXXXXX 100644
59
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/arm/tcg-target.h
60
--- a/include/hw/core/cpu.h
15
+++ b/tcg/arm/tcg-target.h
61
+++ b/include/hw/core/cpu.h
16
@@ -XXX,XX +XXX,XX @@ extern bool use_neon_instructions;
62
@@ -XXX,XX +XXX,XX @@ struct SysemuCPUOps;
17
#define TCG_TARGET_HAS_cmpsel_vec 0
63
* If the target behaviour here is anything other than "set
18
64
* the PC register to the value passed in" then the target must
19
#define TCG_TARGET_DEFAULT_MO (0)
65
* also implement the synchronize_from_tb hook.
20
-#define TCG_TARGET_HAS_MEMORY_BSWAP 1
66
+ * @get_pc: Callback for getting the Program Counter register.
21
+#define TCG_TARGET_HAS_MEMORY_BSWAP 0
67
+ * As above, with the semantics of the target architecture.
22
68
* @gdb_read_register: Callback for letting GDB read a register.
23
/* not defined -- call should be eliminated at compile time */
69
* @gdb_write_register: Callback for letting GDB write a register.
24
void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
70
* @gdb_adjust_breakpoint: Callback for adjusting the address of a
25
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
71
@@ -XXX,XX +XXX,XX @@ struct CPUClass {
26
index XXXXXXX..XXXXXXX 100644
72
void (*dump_state)(CPUState *cpu, FILE *, int flags);
27
--- a/tcg/arm/tcg-target.c.inc
73
int64_t (*get_arch_id)(CPUState *cpu);
28
+++ b/tcg/arm/tcg-target.c.inc
74
void (*set_pc)(CPUState *cpu, vaddr value);
29
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vldst(TCGContext *s, ARMInsn insn,
75
+ vaddr (*get_pc)(CPUState *cpu);
30
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
76
int (*gdb_read_register)(CPUState *cpu, GByteArray *buf, int reg);
31
* int mmu_idx, uintptr_t ra)
77
int (*gdb_write_register)(CPUState *cpu, uint8_t *buf, int reg);
32
*/
78
vaddr (*gdb_adjust_breakpoint)(CPUState *cpu, vaddr addr);
33
-static void * const qemu_ld_helpers[16] = {
79
diff --git a/target/alpha/cpu.c b/target/alpha/cpu.c
34
+static void * const qemu_ld_helpers[8] = {
80
index XXXXXXX..XXXXXXX 100644
35
[MO_UB] = helper_ret_ldub_mmu,
81
--- a/target/alpha/cpu.c
36
[MO_SB] = helper_ret_ldsb_mmu,
82
+++ b/target/alpha/cpu.c
37
-
83
@@ -XXX,XX +XXX,XX @@ static void alpha_cpu_set_pc(CPUState *cs, vaddr value)
38
- [MO_LEUW] = helper_le_lduw_mmu,
84
cpu->env.pc = value;
39
- [MO_LEUL] = helper_le_ldul_mmu,
85
}
40
- [MO_LEQ] = helper_le_ldq_mmu,
86
41
- [MO_LESW] = helper_le_ldsw_mmu,
87
+static vaddr alpha_cpu_get_pc(CPUState *cs)
42
- [MO_LESL] = helper_le_ldul_mmu,
88
+{
43
-
89
+ AlphaCPU *cpu = ALPHA_CPU(cs);
44
- [MO_BEUW] = helper_be_lduw_mmu,
90
+
45
- [MO_BEUL] = helper_be_ldul_mmu,
91
+ return cpu->env.pc;
46
- [MO_BEQ] = helper_be_ldq_mmu,
92
+}
47
- [MO_BESW] = helper_be_ldsw_mmu,
93
+
48
- [MO_BESL] = helper_be_ldul_mmu,
94
+
49
+#ifdef HOST_WORDS_BIGENDIAN
95
static bool alpha_cpu_has_work(CPUState *cs)
50
+ [MO_UW] = helper_be_lduw_mmu,
96
{
51
+ [MO_UL] = helper_be_ldul_mmu,
97
/* Here we are checking to see if the CPU should wake up from HALT.
52
+ [MO_Q] = helper_be_ldq_mmu,
98
@@ -XXX,XX +XXX,XX @@ static void alpha_cpu_class_init(ObjectClass *oc, void *data)
53
+ [MO_SW] = helper_be_ldsw_mmu,
99
cc->has_work = alpha_cpu_has_work;
54
+ [MO_SL] = helper_be_ldul_mmu,
100
cc->dump_state = alpha_cpu_dump_state;
55
+#else
101
cc->set_pc = alpha_cpu_set_pc;
56
+ [MO_UW] = helper_le_lduw_mmu,
102
+ cc->get_pc = alpha_cpu_get_pc;
57
+ [MO_UL] = helper_le_ldul_mmu,
103
cc->gdb_read_register = alpha_cpu_gdb_read_register;
58
+ [MO_Q] = helper_le_ldq_mmu,
104
cc->gdb_write_register = alpha_cpu_gdb_write_register;
59
+ [MO_SW] = helper_le_ldsw_mmu,
105
#ifndef CONFIG_USER_ONLY
60
+ [MO_SL] = helper_le_ldul_mmu,
106
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
61
+#endif
107
index XXXXXXX..XXXXXXX 100644
62
};
108
--- a/target/arm/cpu.c
63
109
+++ b/target/arm/cpu.c
64
/* helper signature: helper_ret_st_mmu(CPUState *env, target_ulong addr,
110
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_set_pc(CPUState *cs, vaddr value)
65
* uintxx_t val, int mmu_idx, uintptr_t ra)
66
*/
67
-static void * const qemu_st_helpers[16] = {
68
- [MO_UB] = helper_ret_stb_mmu,
69
- [MO_LEUW] = helper_le_stw_mmu,
70
- [MO_LEUL] = helper_le_stl_mmu,
71
- [MO_LEQ] = helper_le_stq_mmu,
72
- [MO_BEUW] = helper_be_stw_mmu,
73
- [MO_BEUL] = helper_be_stl_mmu,
74
- [MO_BEQ] = helper_be_stq_mmu,
75
+static void * const qemu_st_helpers[4] = {
76
+ [MO_8] = helper_ret_stb_mmu,
77
+#ifdef HOST_WORDS_BIGENDIAN
78
+ [MO_16] = helper_be_stw_mmu,
79
+ [MO_32] = helper_be_stl_mmu,
80
+ [MO_64] = helper_be_stq_mmu,
81
+#else
82
+ [MO_16] = helper_le_stw_mmu,
83
+ [MO_32] = helper_le_stl_mmu,
84
+ [MO_64] = helper_le_stq_mmu,
85
+#endif
86
};
87
88
/* Helper routines for marshalling helper function arguments into
89
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
90
icache usage. For pre-armv6, use the signed helpers since we do
91
not have a single insn sign-extend. */
92
if (use_armv6_instructions) {
93
- func = qemu_ld_helpers[opc & (MO_BSWAP | MO_SIZE)];
94
+ func = qemu_ld_helpers[opc & MO_SIZE];
95
} else {
96
- func = qemu_ld_helpers[opc & (MO_BSWAP | MO_SSIZE)];
97
+ func = qemu_ld_helpers[opc & MO_SSIZE];
98
if (opc & MO_SIGN) {
99
opc = MO_UL;
100
}
101
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
102
argreg = tcg_out_arg_reg32(s, argreg, TCG_REG_R14);
103
104
/* Tail-call to the helper, which will return to the fast path. */
105
- tcg_out_goto(s, COND_AL, qemu_st_helpers[opc & (MO_BSWAP | MO_SIZE)]);
106
+ tcg_out_goto(s, COND_AL, qemu_st_helpers[opc & MO_SIZE]);
107
return true;
108
}
109
#endif /* SOFTMMU */
110
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
111
TCGReg datalo, TCGReg datahi,
112
TCGReg addrlo, TCGReg addend)
113
{
114
- MemOp bswap = opc & MO_BSWAP;
115
+ /* Byte swapping is left to middle-end expansion. */
116
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
117
118
switch (opc & MO_SSIZE) {
119
case MO_UB:
120
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
121
break;
122
case MO_UW:
123
tcg_out_ld16u_r(s, COND_AL, datalo, addrlo, addend);
124
- if (bswap) {
125
- tcg_out_bswap16(s, COND_AL, datalo, datalo,
126
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
127
- }
128
break;
129
case MO_SW:
130
- if (bswap) {
131
- tcg_out_ld16u_r(s, COND_AL, datalo, addrlo, addend);
132
- tcg_out_bswap16(s, COND_AL, datalo, datalo,
133
- TCG_BSWAP_IZ | TCG_BSWAP_OS);
134
- } else {
135
- tcg_out_ld16s_r(s, COND_AL, datalo, addrlo, addend);
136
- }
137
+ tcg_out_ld16s_r(s, COND_AL, datalo, addrlo, addend);
138
break;
139
case MO_UL:
140
- default:
141
tcg_out_ld32_r(s, COND_AL, datalo, addrlo, addend);
142
- if (bswap) {
143
- tcg_out_bswap32(s, COND_AL, datalo, datalo);
144
- }
145
break;
146
case MO_Q:
147
- {
148
- TCGReg dl = (bswap ? datahi : datalo);
149
- TCGReg dh = (bswap ? datalo : datahi);
150
-
151
- /* Avoid ldrd for user-only emulation, to handle unaligned. */
152
- if (USING_SOFTMMU && use_armv6_instructions
153
- && (dl & 1) == 0 && dh == dl + 1) {
154
- tcg_out_ldrd_r(s, COND_AL, dl, addrlo, addend);
155
- } else if (dl != addend) {
156
- tcg_out_ld32_rwb(s, COND_AL, dl, addend, addrlo);
157
- tcg_out_ld32_12(s, COND_AL, dh, addend, 4);
158
- } else {
159
- tcg_out_dat_reg(s, COND_AL, ARITH_ADD, TCG_REG_TMP,
160
- addend, addrlo, SHIFT_IMM_LSL(0));
161
- tcg_out_ld32_12(s, COND_AL, dl, TCG_REG_TMP, 0);
162
- tcg_out_ld32_12(s, COND_AL, dh, TCG_REG_TMP, 4);
163
- }
164
- if (bswap) {
165
- tcg_out_bswap32(s, COND_AL, dl, dl);
166
- tcg_out_bswap32(s, COND_AL, dh, dh);
167
- }
168
+ /* Avoid ldrd for user-only emulation, to handle unaligned. */
169
+ if (USING_SOFTMMU && use_armv6_instructions
170
+ && (datalo & 1) == 0 && datahi == datalo + 1) {
171
+ tcg_out_ldrd_r(s, COND_AL, datalo, addrlo, addend);
172
+ } else if (datalo != addend) {
173
+ tcg_out_ld32_rwb(s, COND_AL, datalo, addend, addrlo);
174
+ tcg_out_ld32_12(s, COND_AL, datahi, addend, 4);
175
+ } else {
176
+ tcg_out_dat_reg(s, COND_AL, ARITH_ADD, TCG_REG_TMP,
177
+ addend, addrlo, SHIFT_IMM_LSL(0));
178
+ tcg_out_ld32_12(s, COND_AL, datalo, TCG_REG_TMP, 0);
179
+ tcg_out_ld32_12(s, COND_AL, datahi, TCG_REG_TMP, 4);
180
}
181
break;
182
+ default:
183
+ g_assert_not_reached();
184
}
111
}
185
}
112
}
186
113
187
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc,
114
+static vaddr arm_cpu_get_pc(CPUState *cs)
188
TCGReg datalo, TCGReg datahi,
115
+{
189
TCGReg addrlo)
116
+ ARMCPU *cpu = ARM_CPU(cs);
190
{
117
+ CPUARMState *env = &cpu->env;
191
- MemOp bswap = opc & MO_BSWAP;
118
+
192
+ /* Byte swapping is left to middle-end expansion. */
119
+ if (is_a64(env)) {
193
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
120
+ return env->pc;
194
121
+ } else {
195
switch (opc & MO_SSIZE) {
122
+ return env->regs[15];
196
case MO_UB:
123
+ }
197
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc,
124
+}
198
break;
125
+
199
case MO_UW:
126
#ifdef CONFIG_TCG
200
tcg_out_ld16u_8(s, COND_AL, datalo, addrlo, 0);
127
void arm_cpu_synchronize_from_tb(CPUState *cs,
201
- if (bswap) {
128
const TranslationBlock *tb)
202
- tcg_out_bswap16(s, COND_AL, datalo, datalo,
129
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
203
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
130
cc->has_work = arm_cpu_has_work;
204
- }
131
cc->dump_state = arm_cpu_dump_state;
205
break;
132
cc->set_pc = arm_cpu_set_pc;
206
case MO_SW:
133
+ cc->get_pc = arm_cpu_get_pc;
207
- if (bswap) {
134
cc->gdb_read_register = arm_cpu_gdb_read_register;
208
- tcg_out_ld16u_8(s, COND_AL, datalo, addrlo, 0);
135
cc->gdb_write_register = arm_cpu_gdb_write_register;
209
- tcg_out_bswap16(s, COND_AL, datalo, datalo,
136
#ifndef CONFIG_USER_ONLY
210
- TCG_BSWAP_IZ | TCG_BSWAP_OS);
137
diff --git a/target/avr/cpu.c b/target/avr/cpu.c
211
- } else {
138
index XXXXXXX..XXXXXXX 100644
212
- tcg_out_ld16s_8(s, COND_AL, datalo, addrlo, 0);
139
--- a/target/avr/cpu.c
213
- }
140
+++ b/target/avr/cpu.c
214
+ tcg_out_ld16s_8(s, COND_AL, datalo, addrlo, 0);
141
@@ -XXX,XX +XXX,XX @@ static void avr_cpu_set_pc(CPUState *cs, vaddr value)
215
break;
142
cpu->env.pc_w = value / 2; /* internally PC points to words */
216
case MO_UL:
143
}
217
- default:
144
218
tcg_out_ld32_12(s, COND_AL, datalo, addrlo, 0);
145
+static vaddr avr_cpu_get_pc(CPUState *cs)
219
- if (bswap) {
146
+{
220
- tcg_out_bswap32(s, COND_AL, datalo, datalo);
147
+ AVRCPU *cpu = AVR_CPU(cs);
221
- }
148
+
222
break;
149
+ return cpu->env.pc_w * 2;
223
case MO_Q:
150
+}
224
- {
151
+
225
- TCGReg dl = (bswap ? datahi : datalo);
152
static bool avr_cpu_has_work(CPUState *cs)
226
- TCGReg dh = (bswap ? datalo : datahi);
153
{
227
-
154
AVRCPU *cpu = AVR_CPU(cs);
228
- /* Avoid ldrd for user-only emulation, to handle unaligned. */
155
@@ -XXX,XX +XXX,XX @@ static void avr_cpu_class_init(ObjectClass *oc, void *data)
229
- if (USING_SOFTMMU && use_armv6_instructions
156
cc->has_work = avr_cpu_has_work;
230
- && (dl & 1) == 0 && dh == dl + 1) {
157
cc->dump_state = avr_cpu_dump_state;
231
- tcg_out_ldrd_8(s, COND_AL, dl, addrlo, 0);
158
cc->set_pc = avr_cpu_set_pc;
232
- } else if (dl == addrlo) {
159
+ cc->get_pc = avr_cpu_get_pc;
233
- tcg_out_ld32_12(s, COND_AL, dh, addrlo, bswap ? 0 : 4);
160
dc->vmsd = &vms_avr_cpu;
234
- tcg_out_ld32_12(s, COND_AL, dl, addrlo, bswap ? 4 : 0);
161
cc->sysemu_ops = &avr_sysemu_ops;
235
- } else {
162
cc->disas_set_info = avr_cpu_disas_set_info;
236
- tcg_out_ld32_12(s, COND_AL, dl, addrlo, bswap ? 4 : 0);
163
diff --git a/target/cris/cpu.c b/target/cris/cpu.c
237
- tcg_out_ld32_12(s, COND_AL, dh, addrlo, bswap ? 0 : 4);
164
index XXXXXXX..XXXXXXX 100644
238
- }
165
--- a/target/cris/cpu.c
239
- if (bswap) {
166
+++ b/target/cris/cpu.c
240
- tcg_out_bswap32(s, COND_AL, dl, dl);
167
@@ -XXX,XX +XXX,XX @@ static void cris_cpu_set_pc(CPUState *cs, vaddr value)
241
- tcg_out_bswap32(s, COND_AL, dh, dh);
168
cpu->env.pc = value;
242
- }
169
}
243
+ /* Avoid ldrd for user-only emulation, to handle unaligned. */
170
244
+ if (USING_SOFTMMU && use_armv6_instructions
171
+static vaddr cris_cpu_get_pc(CPUState *cs)
245
+ && (datalo & 1) == 0 && datahi == datalo + 1) {
172
+{
246
+ tcg_out_ldrd_8(s, COND_AL, datalo, addrlo, 0);
173
+ CRISCPU *cpu = CRIS_CPU(cs);
247
+ } else if (datalo == addrlo) {
174
+
248
+ tcg_out_ld32_12(s, COND_AL, datahi, addrlo, 4);
175
+ return cpu->env.pc;
249
+ tcg_out_ld32_12(s, COND_AL, datalo, addrlo, 0);
176
+}
250
+ } else {
177
+
251
+ tcg_out_ld32_12(s, COND_AL, datalo, addrlo, 0);
178
static bool cris_cpu_has_work(CPUState *cs)
252
+ tcg_out_ld32_12(s, COND_AL, datahi, addrlo, 4);
179
{
253
}
180
return cs->interrupt_request & (CPU_INTERRUPT_HARD | CPU_INTERRUPT_NMI);
254
break;
181
@@ -XXX,XX +XXX,XX @@ static void cris_cpu_class_init(ObjectClass *oc, void *data)
255
+ default:
182
cc->has_work = cris_cpu_has_work;
256
+ g_assert_not_reached();
183
cc->dump_state = cris_cpu_dump_state;
184
cc->set_pc = cris_cpu_set_pc;
185
+ cc->get_pc = cris_cpu_get_pc;
186
cc->gdb_read_register = cris_cpu_gdb_read_register;
187
cc->gdb_write_register = cris_cpu_gdb_write_register;
188
#ifndef CONFIG_USER_ONLY
189
diff --git a/target/hexagon/cpu.c b/target/hexagon/cpu.c
190
index XXXXXXX..XXXXXXX 100644
191
--- a/target/hexagon/cpu.c
192
+++ b/target/hexagon/cpu.c
193
@@ -XXX,XX +XXX,XX @@ static void hexagon_cpu_set_pc(CPUState *cs, vaddr value)
194
env->gpr[HEX_REG_PC] = value;
195
}
196
197
+static vaddr hexagon_cpu_get_pc(CPUState *cs)
198
+{
199
+ HexagonCPU *cpu = HEXAGON_CPU(cs);
200
+ CPUHexagonState *env = &cpu->env;
201
+ return env->gpr[HEX_REG_PC];
202
+}
203
+
204
static void hexagon_cpu_synchronize_from_tb(CPUState *cs,
205
const TranslationBlock *tb)
206
{
207
@@ -XXX,XX +XXX,XX @@ static void hexagon_cpu_class_init(ObjectClass *c, void *data)
208
cc->has_work = hexagon_cpu_has_work;
209
cc->dump_state = hexagon_dump_state;
210
cc->set_pc = hexagon_cpu_set_pc;
211
+ cc->get_pc = hexagon_cpu_get_pc;
212
cc->gdb_read_register = hexagon_gdb_read_register;
213
cc->gdb_write_register = hexagon_gdb_write_register;
214
cc->gdb_num_core_regs = TOTAL_PER_THREAD_REGS + NUM_VREGS + NUM_QREGS;
215
diff --git a/target/hppa/cpu.c b/target/hppa/cpu.c
216
index XXXXXXX..XXXXXXX 100644
217
--- a/target/hppa/cpu.c
218
+++ b/target/hppa/cpu.c
219
@@ -XXX,XX +XXX,XX @@ static void hppa_cpu_set_pc(CPUState *cs, vaddr value)
220
cpu->env.iaoq_b = value + 4;
221
}
222
223
+static vaddr hppa_cpu_get_pc(CPUState *cs)
224
+{
225
+ HPPACPU *cpu = HPPA_CPU(cs);
226
+
227
+ return cpu->env.iaoq_f;
228
+}
229
+
230
static void hppa_cpu_synchronize_from_tb(CPUState *cs,
231
const TranslationBlock *tb)
232
{
233
@@ -XXX,XX +XXX,XX @@ static void hppa_cpu_class_init(ObjectClass *oc, void *data)
234
cc->has_work = hppa_cpu_has_work;
235
cc->dump_state = hppa_cpu_dump_state;
236
cc->set_pc = hppa_cpu_set_pc;
237
+ cc->get_pc = hppa_cpu_get_pc;
238
cc->gdb_read_register = hppa_cpu_gdb_read_register;
239
cc->gdb_write_register = hppa_cpu_gdb_write_register;
240
#ifndef CONFIG_USER_ONLY
241
diff --git a/target/i386/cpu.c b/target/i386/cpu.c
242
index XXXXXXX..XXXXXXX 100644
243
--- a/target/i386/cpu.c
244
+++ b/target/i386/cpu.c
245
@@ -XXX,XX +XXX,XX @@ static void x86_cpu_set_pc(CPUState *cs, vaddr value)
246
cpu->env.eip = value;
247
}
248
249
+static vaddr x86_cpu_get_pc(CPUState *cs)
250
+{
251
+ X86CPU *cpu = X86_CPU(cs);
252
+
253
+ /* Match cpu_get_tb_cpu_state. */
254
+ return cpu->env.eip + cpu->env.segs[R_CS].base;
255
+}
256
+
257
int x86_cpu_pending_interrupt(CPUState *cs, int interrupt_request)
258
{
259
X86CPU *cpu = X86_CPU(cs);
260
@@ -XXX,XX +XXX,XX @@ static void x86_cpu_common_class_init(ObjectClass *oc, void *data)
261
cc->has_work = x86_cpu_has_work;
262
cc->dump_state = x86_cpu_dump_state;
263
cc->set_pc = x86_cpu_set_pc;
264
+ cc->get_pc = x86_cpu_get_pc;
265
cc->gdb_read_register = x86_cpu_gdb_read_register;
266
cc->gdb_write_register = x86_cpu_gdb_write_register;
267
cc->get_arch_id = x86_cpu_get_arch_id;
268
diff --git a/target/loongarch/cpu.c b/target/loongarch/cpu.c
269
index XXXXXXX..XXXXXXX 100644
270
--- a/target/loongarch/cpu.c
271
+++ b/target/loongarch/cpu.c
272
@@ -XXX,XX +XXX,XX @@ static void loongarch_cpu_set_pc(CPUState *cs, vaddr value)
273
env->pc = value;
274
}
275
276
+static vaddr loongarch_cpu_get_pc(CPUState *cs)
277
+{
278
+ LoongArchCPU *cpu = LOONGARCH_CPU(cs);
279
+ CPULoongArchState *env = &cpu->env;
280
+
281
+ return env->pc;
282
+}
283
+
284
#ifndef CONFIG_USER_ONLY
285
#include "hw/loongarch/virt.h"
286
287
@@ -XXX,XX +XXX,XX @@ static void loongarch_cpu_class_init(ObjectClass *c, void *data)
288
cc->has_work = loongarch_cpu_has_work;
289
cc->dump_state = loongarch_cpu_dump_state;
290
cc->set_pc = loongarch_cpu_set_pc;
291
+ cc->get_pc = loongarch_cpu_get_pc;
292
#ifndef CONFIG_USER_ONLY
293
dc->vmsd = &vmstate_loongarch_cpu;
294
cc->sysemu_ops = &loongarch_sysemu_ops;
295
diff --git a/target/m68k/cpu.c b/target/m68k/cpu.c
296
index XXXXXXX..XXXXXXX 100644
297
--- a/target/m68k/cpu.c
298
+++ b/target/m68k/cpu.c
299
@@ -XXX,XX +XXX,XX @@ static void m68k_cpu_set_pc(CPUState *cs, vaddr value)
300
cpu->env.pc = value;
301
}
302
303
+static vaddr m68k_cpu_get_pc(CPUState *cs)
304
+{
305
+ M68kCPU *cpu = M68K_CPU(cs);
306
+
307
+ return cpu->env.pc;
308
+}
309
+
310
static bool m68k_cpu_has_work(CPUState *cs)
311
{
312
return cs->interrupt_request & CPU_INTERRUPT_HARD;
313
@@ -XXX,XX +XXX,XX @@ static void m68k_cpu_class_init(ObjectClass *c, void *data)
314
cc->has_work = m68k_cpu_has_work;
315
cc->dump_state = m68k_cpu_dump_state;
316
cc->set_pc = m68k_cpu_set_pc;
317
+ cc->get_pc = m68k_cpu_get_pc;
318
cc->gdb_read_register = m68k_cpu_gdb_read_register;
319
cc->gdb_write_register = m68k_cpu_gdb_write_register;
320
#if defined(CONFIG_SOFTMMU)
321
diff --git a/target/microblaze/cpu.c b/target/microblaze/cpu.c
322
index XXXXXXX..XXXXXXX 100644
323
--- a/target/microblaze/cpu.c
324
+++ b/target/microblaze/cpu.c
325
@@ -XXX,XX +XXX,XX @@ static void mb_cpu_set_pc(CPUState *cs, vaddr value)
326
cpu->env.iflags = 0;
327
}
328
329
+static vaddr mb_cpu_get_pc(CPUState *cs)
330
+{
331
+ MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
332
+
333
+ return cpu->env.pc;
334
+}
335
+
336
static void mb_cpu_synchronize_from_tb(CPUState *cs,
337
const TranslationBlock *tb)
338
{
339
@@ -XXX,XX +XXX,XX @@ static void mb_cpu_class_init(ObjectClass *oc, void *data)
340
341
cc->dump_state = mb_cpu_dump_state;
342
cc->set_pc = mb_cpu_set_pc;
343
+ cc->get_pc = mb_cpu_get_pc;
344
cc->gdb_read_register = mb_cpu_gdb_read_register;
345
cc->gdb_write_register = mb_cpu_gdb_write_register;
346
347
diff --git a/target/mips/cpu.c b/target/mips/cpu.c
348
index XXXXXXX..XXXXXXX 100644
349
--- a/target/mips/cpu.c
350
+++ b/target/mips/cpu.c
351
@@ -XXX,XX +XXX,XX @@ static void mips_cpu_set_pc(CPUState *cs, vaddr value)
352
mips_env_set_pc(&cpu->env, value);
353
}
354
355
+static vaddr mips_cpu_get_pc(CPUState *cs)
356
+{
357
+ MIPSCPU *cpu = MIPS_CPU(cs);
358
+
359
+ return cpu->env.active_tc.PC;
360
+}
361
+
362
static bool mips_cpu_has_work(CPUState *cs)
363
{
364
MIPSCPU *cpu = MIPS_CPU(cs);
365
@@ -XXX,XX +XXX,XX @@ static void mips_cpu_class_init(ObjectClass *c, void *data)
366
cc->has_work = mips_cpu_has_work;
367
cc->dump_state = mips_cpu_dump_state;
368
cc->set_pc = mips_cpu_set_pc;
369
+ cc->get_pc = mips_cpu_get_pc;
370
cc->gdb_read_register = mips_cpu_gdb_read_register;
371
cc->gdb_write_register = mips_cpu_gdb_write_register;
372
#ifndef CONFIG_USER_ONLY
373
diff --git a/target/nios2/cpu.c b/target/nios2/cpu.c
374
index XXXXXXX..XXXXXXX 100644
375
--- a/target/nios2/cpu.c
376
+++ b/target/nios2/cpu.c
377
@@ -XXX,XX +XXX,XX @@ static void nios2_cpu_set_pc(CPUState *cs, vaddr value)
378
env->pc = value;
379
}
380
381
+static vaddr nios2_cpu_get_pc(CPUState *cs)
382
+{
383
+ Nios2CPU *cpu = NIOS2_CPU(cs);
384
+ CPUNios2State *env = &cpu->env;
385
+
386
+ return env->pc;
387
+}
388
+
389
static bool nios2_cpu_has_work(CPUState *cs)
390
{
391
return cs->interrupt_request & CPU_INTERRUPT_HARD;
392
@@ -XXX,XX +XXX,XX @@ static void nios2_cpu_class_init(ObjectClass *oc, void *data)
393
cc->has_work = nios2_cpu_has_work;
394
cc->dump_state = nios2_cpu_dump_state;
395
cc->set_pc = nios2_cpu_set_pc;
396
+ cc->get_pc = nios2_cpu_get_pc;
397
cc->disas_set_info = nios2_cpu_disas_set_info;
398
#ifndef CONFIG_USER_ONLY
399
cc->sysemu_ops = &nios2_sysemu_ops;
400
diff --git a/target/openrisc/cpu.c b/target/openrisc/cpu.c
401
index XXXXXXX..XXXXXXX 100644
402
--- a/target/openrisc/cpu.c
403
+++ b/target/openrisc/cpu.c
404
@@ -XXX,XX +XXX,XX @@ static void openrisc_cpu_set_pc(CPUState *cs, vaddr value)
405
cpu->env.dflag = 0;
406
}
407
408
+static vaddr openrisc_cpu_get_pc(CPUState *cs)
409
+{
410
+ OpenRISCCPU *cpu = OPENRISC_CPU(cs);
411
+
412
+ return cpu->env.pc;
413
+}
414
+
415
static void openrisc_cpu_synchronize_from_tb(CPUState *cs,
416
const TranslationBlock *tb)
417
{
418
@@ -XXX,XX +XXX,XX @@ static void openrisc_cpu_class_init(ObjectClass *oc, void *data)
419
cc->has_work = openrisc_cpu_has_work;
420
cc->dump_state = openrisc_cpu_dump_state;
421
cc->set_pc = openrisc_cpu_set_pc;
422
+ cc->get_pc = openrisc_cpu_get_pc;
423
cc->gdb_read_register = openrisc_cpu_gdb_read_register;
424
cc->gdb_write_register = openrisc_cpu_gdb_write_register;
425
#ifndef CONFIG_USER_ONLY
426
diff --git a/target/ppc/cpu_init.c b/target/ppc/cpu_init.c
427
index XXXXXXX..XXXXXXX 100644
428
--- a/target/ppc/cpu_init.c
429
+++ b/target/ppc/cpu_init.c
430
@@ -XXX,XX +XXX,XX @@ static void ppc_cpu_set_pc(CPUState *cs, vaddr value)
431
cpu->env.nip = value;
432
}
433
434
+static vaddr ppc_cpu_get_pc(CPUState *cs)
435
+{
436
+ PowerPCCPU *cpu = POWERPC_CPU(cs);
437
+
438
+ return cpu->env.nip;
439
+}
440
+
441
static bool ppc_cpu_has_work(CPUState *cs)
442
{
443
PowerPCCPU *cpu = POWERPC_CPU(cs);
444
@@ -XXX,XX +XXX,XX @@ static void ppc_cpu_class_init(ObjectClass *oc, void *data)
445
cc->has_work = ppc_cpu_has_work;
446
cc->dump_state = ppc_cpu_dump_state;
447
cc->set_pc = ppc_cpu_set_pc;
448
+ cc->get_pc = ppc_cpu_get_pc;
449
cc->gdb_read_register = ppc_cpu_gdb_read_register;
450
cc->gdb_write_register = ppc_cpu_gdb_write_register;
451
#ifndef CONFIG_USER_ONLY
452
diff --git a/target/riscv/cpu.c b/target/riscv/cpu.c
453
index XXXXXXX..XXXXXXX 100644
454
--- a/target/riscv/cpu.c
455
+++ b/target/riscv/cpu.c
456
@@ -XXX,XX +XXX,XX @@ static void riscv_cpu_set_pc(CPUState *cs, vaddr value)
257
}
457
}
258
}
458
}
259
459
260
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_st_index(TCGContext *s, int cond, MemOp opc,
460
+static vaddr riscv_cpu_get_pc(CPUState *cs)
261
TCGReg datalo, TCGReg datahi,
461
+{
262
TCGReg addrlo, TCGReg addend)
462
+ RISCVCPU *cpu = RISCV_CPU(cs);
263
{
463
+ CPURISCVState *env = &cpu->env;
264
- MemOp bswap = opc & MO_BSWAP;
464
+
265
+ /* Byte swapping is left to middle-end expansion. */
465
+ /* Match cpu_get_tb_cpu_state. */
266
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
466
+ if (env->xl == MXL_RV32) {
267
467
+ return env->pc & UINT32_MAX;
268
switch (opc & MO_SIZE) {
468
+ }
269
case MO_8:
469
+ return env->pc;
270
tcg_out_st8_r(s, cond, datalo, addrlo, addend);
470
+}
271
break;
471
+
272
case MO_16:
472
static void riscv_cpu_synchronize_from_tb(CPUState *cs,
273
- if (bswap) {
473
const TranslationBlock *tb)
274
- tcg_out_bswap16(s, cond, TCG_REG_R0, datalo, 0);
474
{
275
- tcg_out_st16_r(s, cond, TCG_REG_R0, addrlo, addend);
475
@@ -XXX,XX +XXX,XX @@ static void riscv_cpu_class_init(ObjectClass *c, void *data)
276
- } else {
476
cc->has_work = riscv_cpu_has_work;
277
- tcg_out_st16_r(s, cond, datalo, addrlo, addend);
477
cc->dump_state = riscv_cpu_dump_state;
278
- }
478
cc->set_pc = riscv_cpu_set_pc;
279
+ tcg_out_st16_r(s, cond, datalo, addrlo, addend);
479
+ cc->get_pc = riscv_cpu_get_pc;
280
break;
480
cc->gdb_read_register = riscv_cpu_gdb_read_register;
281
case MO_32:
481
cc->gdb_write_register = riscv_cpu_gdb_write_register;
282
- default:
482
cc->gdb_num_core_regs = 33;
283
- if (bswap) {
483
diff --git a/target/rx/cpu.c b/target/rx/cpu.c
284
- tcg_out_bswap32(s, cond, TCG_REG_R0, datalo);
484
index XXXXXXX..XXXXXXX 100644
285
- tcg_out_st32_r(s, cond, TCG_REG_R0, addrlo, addend);
485
--- a/target/rx/cpu.c
286
- } else {
486
+++ b/target/rx/cpu.c
287
- tcg_out_st32_r(s, cond, datalo, addrlo, addend);
487
@@ -XXX,XX +XXX,XX @@ static void rx_cpu_set_pc(CPUState *cs, vaddr value)
288
- }
488
cpu->env.pc = value;
289
+ tcg_out_st32_r(s, cond, datalo, addrlo, addend);
489
}
290
break;
490
291
case MO_64:
491
+static vaddr rx_cpu_get_pc(CPUState *cs)
292
/* Avoid strd for user-only emulation, to handle unaligned. */
492
+{
293
- if (bswap) {
493
+ RXCPU *cpu = RX_CPU(cs);
294
- tcg_out_bswap32(s, cond, TCG_REG_R0, datahi);
494
+
295
- tcg_out_st32_rwb(s, cond, TCG_REG_R0, addend, addrlo);
495
+ return cpu->env.pc;
296
- tcg_out_bswap32(s, cond, TCG_REG_R0, datalo);
496
+}
297
- tcg_out_st32_12(s, cond, TCG_REG_R0, addend, 4);
497
+
298
- } else if (USING_SOFTMMU && use_armv6_instructions
498
static void rx_cpu_synchronize_from_tb(CPUState *cs,
299
- && (datalo & 1) == 0 && datahi == datalo + 1) {
499
const TranslationBlock *tb)
300
+ if (USING_SOFTMMU && use_armv6_instructions
500
{
301
+ && (datalo & 1) == 0 && datahi == datalo + 1) {
501
@@ -XXX,XX +XXX,XX @@ static void rx_cpu_class_init(ObjectClass *klass, void *data)
302
tcg_out_strd_r(s, cond, datalo, addrlo, addend);
502
cc->has_work = rx_cpu_has_work;
303
} else {
503
cc->dump_state = rx_cpu_dump_state;
304
tcg_out_st32_rwb(s, cond, datalo, addend, addrlo);
504
cc->set_pc = rx_cpu_set_pc;
305
tcg_out_st32_12(s, cond, datahi, addend, 4);
505
+ cc->get_pc = rx_cpu_get_pc;
306
}
506
307
break;
507
#ifndef CONFIG_USER_ONLY
308
+ default:
508
cc->sysemu_ops = &rx_sysemu_ops;
309
+ g_assert_not_reached();
509
diff --git a/target/s390x/cpu.c b/target/s390x/cpu.c
310
}
510
index XXXXXXX..XXXXXXX 100644
311
}
511
--- a/target/s390x/cpu.c
312
512
+++ b/target/s390x/cpu.c
313
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc,
513
@@ -XXX,XX +XXX,XX @@ static void s390_cpu_set_pc(CPUState *cs, vaddr value)
314
TCGReg datalo, TCGReg datahi,
514
cpu->env.psw.addr = value;
315
TCGReg addrlo)
515
}
316
{
516
317
- MemOp bswap = opc & MO_BSWAP;
517
+static vaddr s390_cpu_get_pc(CPUState *cs)
318
+ /* Byte swapping is left to middle-end expansion. */
518
+{
319
+ tcg_debug_assert((opc & MO_BSWAP) == 0);
519
+ S390CPU *cpu = S390_CPU(cs);
320
520
+
321
switch (opc & MO_SIZE) {
521
+ return cpu->env.psw.addr;
322
case MO_8:
522
+}
323
tcg_out_st8_12(s, COND_AL, datalo, addrlo, 0);
523
+
324
break;
524
static bool s390_cpu_has_work(CPUState *cs)
325
case MO_16:
525
{
326
- if (bswap) {
526
S390CPU *cpu = S390_CPU(cs);
327
- tcg_out_bswap16(s, COND_AL, TCG_REG_R0, datalo, 0);
527
@@ -XXX,XX +XXX,XX @@ static void s390_cpu_class_init(ObjectClass *oc, void *data)
328
- tcg_out_st16_8(s, COND_AL, TCG_REG_R0, addrlo, 0);
528
cc->has_work = s390_cpu_has_work;
329
- } else {
529
cc->dump_state = s390_cpu_dump_state;
330
- tcg_out_st16_8(s, COND_AL, datalo, addrlo, 0);
530
cc->set_pc = s390_cpu_set_pc;
331
- }
531
+ cc->get_pc = s390_cpu_get_pc;
332
+ tcg_out_st16_8(s, COND_AL, datalo, addrlo, 0);
532
cc->gdb_read_register = s390_cpu_gdb_read_register;
333
break;
533
cc->gdb_write_register = s390_cpu_gdb_write_register;
334
case MO_32:
534
#ifndef CONFIG_USER_ONLY
335
- default:
535
diff --git a/target/sh4/cpu.c b/target/sh4/cpu.c
336
- if (bswap) {
536
index XXXXXXX..XXXXXXX 100644
337
- tcg_out_bswap32(s, COND_AL, TCG_REG_R0, datalo);
537
--- a/target/sh4/cpu.c
338
- tcg_out_st32_12(s, COND_AL, TCG_REG_R0, addrlo, 0);
538
+++ b/target/sh4/cpu.c
339
- } else {
539
@@ -XXX,XX +XXX,XX @@ static void superh_cpu_set_pc(CPUState *cs, vaddr value)
340
- tcg_out_st32_12(s, COND_AL, datalo, addrlo, 0);
540
cpu->env.pc = value;
341
- }
541
}
342
+ tcg_out_st32_12(s, COND_AL, datalo, addrlo, 0);
542
343
break;
543
+static vaddr superh_cpu_get_pc(CPUState *cs)
344
case MO_64:
544
+{
345
/* Avoid strd for user-only emulation, to handle unaligned. */
545
+ SuperHCPU *cpu = SUPERH_CPU(cs);
346
- if (bswap) {
546
+
347
- tcg_out_bswap32(s, COND_AL, TCG_REG_R0, datahi);
547
+ return cpu->env.pc;
348
- tcg_out_st32_12(s, COND_AL, TCG_REG_R0, addrlo, 0);
548
+}
349
- tcg_out_bswap32(s, COND_AL, TCG_REG_R0, datalo);
549
+
350
- tcg_out_st32_12(s, COND_AL, TCG_REG_R0, addrlo, 4);
550
static void superh_cpu_synchronize_from_tb(CPUState *cs,
351
- } else if (USING_SOFTMMU && use_armv6_instructions
551
const TranslationBlock *tb)
352
- && (datalo & 1) == 0 && datahi == datalo + 1) {
552
{
353
+ if (USING_SOFTMMU && use_armv6_instructions
553
@@ -XXX,XX +XXX,XX @@ static void superh_cpu_class_init(ObjectClass *oc, void *data)
354
+ && (datalo & 1) == 0 && datahi == datalo + 1) {
554
cc->has_work = superh_cpu_has_work;
355
tcg_out_strd_8(s, COND_AL, datalo, addrlo, 0);
555
cc->dump_state = superh_cpu_dump_state;
356
} else {
556
cc->set_pc = superh_cpu_set_pc;
357
tcg_out_st32_12(s, COND_AL, datalo, addrlo, 0);
557
+ cc->get_pc = superh_cpu_get_pc;
358
tcg_out_st32_12(s, COND_AL, datahi, addrlo, 4);
558
cc->gdb_read_register = superh_cpu_gdb_read_register;
359
}
559
cc->gdb_write_register = superh_cpu_gdb_write_register;
360
break;
560
#ifndef CONFIG_USER_ONLY
361
+ default:
561
diff --git a/target/sparc/cpu.c b/target/sparc/cpu.c
362
+ g_assert_not_reached();
562
index XXXXXXX..XXXXXXX 100644
363
}
563
--- a/target/sparc/cpu.c
364
}
564
+++ b/target/sparc/cpu.c
365
565
@@ -XXX,XX +XXX,XX @@ static void sparc_cpu_set_pc(CPUState *cs, vaddr value)
566
cpu->env.npc = value + 4;
567
}
568
569
+static vaddr sparc_cpu_get_pc(CPUState *cs)
570
+{
571
+ SPARCCPU *cpu = SPARC_CPU(cs);
572
+
573
+ return cpu->env.pc;
574
+}
575
+
576
static void sparc_cpu_synchronize_from_tb(CPUState *cs,
577
const TranslationBlock *tb)
578
{
579
@@ -XXX,XX +XXX,XX @@ static void sparc_cpu_class_init(ObjectClass *oc, void *data)
580
cc->memory_rw_debug = sparc_cpu_memory_rw_debug;
581
#endif
582
cc->set_pc = sparc_cpu_set_pc;
583
+ cc->get_pc = sparc_cpu_get_pc;
584
cc->gdb_read_register = sparc_cpu_gdb_read_register;
585
cc->gdb_write_register = sparc_cpu_gdb_write_register;
586
#ifndef CONFIG_USER_ONLY
587
diff --git a/target/tricore/cpu.c b/target/tricore/cpu.c
588
index XXXXXXX..XXXXXXX 100644
589
--- a/target/tricore/cpu.c
590
+++ b/target/tricore/cpu.c
591
@@ -XXX,XX +XXX,XX @@ static void tricore_cpu_set_pc(CPUState *cs, vaddr value)
592
env->PC = value & ~(target_ulong)1;
593
}
594
595
+static vaddr tricore_cpu_get_pc(CPUState *cs)
596
+{
597
+ TriCoreCPU *cpu = TRICORE_CPU(cs);
598
+ CPUTriCoreState *env = &cpu->env;
599
+
600
+ return env->PC;
601
+}
602
+
603
static void tricore_cpu_synchronize_from_tb(CPUState *cs,
604
const TranslationBlock *tb)
605
{
606
@@ -XXX,XX +XXX,XX @@ static void tricore_cpu_class_init(ObjectClass *c, void *data)
607
608
cc->dump_state = tricore_cpu_dump_state;
609
cc->set_pc = tricore_cpu_set_pc;
610
+ cc->get_pc = tricore_cpu_get_pc;
611
cc->sysemu_ops = &tricore_sysemu_ops;
612
cc->tcg_ops = &tricore_tcg_ops;
613
}
614
diff --git a/target/xtensa/cpu.c b/target/xtensa/cpu.c
615
index XXXXXXX..XXXXXXX 100644
616
--- a/target/xtensa/cpu.c
617
+++ b/target/xtensa/cpu.c
618
@@ -XXX,XX +XXX,XX @@ static void xtensa_cpu_set_pc(CPUState *cs, vaddr value)
619
cpu->env.pc = value;
620
}
621
622
+static vaddr xtensa_cpu_get_pc(CPUState *cs)
623
+{
624
+ XtensaCPU *cpu = XTENSA_CPU(cs);
625
+
626
+ return cpu->env.pc;
627
+}
628
+
629
static bool xtensa_cpu_has_work(CPUState *cs)
630
{
631
#ifndef CONFIG_USER_ONLY
632
@@ -XXX,XX +XXX,XX @@ static void xtensa_cpu_class_init(ObjectClass *oc, void *data)
633
cc->has_work = xtensa_cpu_has_work;
634
cc->dump_state = xtensa_cpu_dump_state;
635
cc->set_pc = xtensa_cpu_set_pc;
636
+ cc->get_pc = xtensa_cpu_get_pc;
637
cc->gdb_read_register = xtensa_cpu_gdb_read_register;
638
cc->gdb_write_register = xtensa_cpu_gdb_write_register;
639
cc->gdb_stop_before_watchpoint = true;
366
--
640
--
367
2.25.1
641
2.34.1
368
642
369
643
diff view generated by jsdifflib
1
This will eventually simplify front-end usage, and will allow
1
The availability of tb->pc will shortly be conditional.
2
backends to unset TCG_TARGET_HAS_MEMORY_BSWAP without loss of
2
Introduce accessor functions to minimize ifdefs.
3
optimization.
4
3
5
The argument is added during expansion, not currently exposed to the
4
Pass around a known pc to places like tcg_gen_code,
6
front end translators. The backends currently only support a flags
5
where the caller must already have the value.
7
value of either TCG_BSWAP_IZ, or (TCG_BSWAP_IZ | TCG_BSWAP_OZ),
8
since they all require zero top bytes and leave them that way.
9
At the existing call sites we pass in (TCG_BSWAP_IZ | TCG_BSWAP_OZ),
10
except for the flags-ignored cases of a 32-bit swap of a 32-bit
11
value and or a 64-bit swap of a 64-bit value, where we pass 0.
12
6
13
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
14
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
15
---
9
---
16
include/tcg/tcg-opc.h | 10 +++++-----
10
accel/tcg/internal.h | 6 ++++
17
include/tcg/tcg.h | 12 ++++++++++++
11
include/exec/exec-all.h | 6 ++++
18
tcg/tcg-op.c | 13 ++++++++-----
12
include/tcg/tcg.h | 2 +-
19
tcg/tcg.c | 28 ++++++++++++++++++++++++++++
13
accel/tcg/cpu-exec.c | 46 ++++++++++++++-----------
20
tcg/README | 22 ++++++++++++++--------
14
accel/tcg/translate-all.c | 37 +++++++++++---------
21
5 files changed, 67 insertions(+), 18 deletions(-)
15
target/arm/cpu.c | 4 +--
16
target/avr/cpu.c | 2 +-
17
target/hexagon/cpu.c | 2 +-
18
target/hppa/cpu.c | 4 +--
19
target/i386/tcg/tcg-cpu.c | 2 +-
20
target/loongarch/cpu.c | 2 +-
21
target/microblaze/cpu.c | 2 +-
22
target/mips/tcg/exception.c | 2 +-
23
target/mips/tcg/sysemu/special_helper.c | 2 +-
24
target/openrisc/cpu.c | 2 +-
25
target/riscv/cpu.c | 4 +--
26
target/rx/cpu.c | 2 +-
27
target/sh4/cpu.c | 4 +--
28
target/sparc/cpu.c | 2 +-
29
target/tricore/cpu.c | 2 +-
30
tcg/tcg.c | 8 ++---
31
21 files changed, 82 insertions(+), 61 deletions(-)
22
32
23
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
33
diff --git a/accel/tcg/internal.h b/accel/tcg/internal.h
24
index XXXXXXX..XXXXXXX 100644
34
index XXXXXXX..XXXXXXX 100644
25
--- a/include/tcg/tcg-opc.h
35
--- a/accel/tcg/internal.h
26
+++ b/include/tcg/tcg-opc.h
36
+++ b/accel/tcg/internal.h
27
@@ -XXX,XX +XXX,XX @@ DEF(ext8s_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_ext8s_i32))
37
@@ -XXX,XX +XXX,XX @@ G_NORETURN void cpu_io_recompile(CPUState *cpu, uintptr_t retaddr);
28
DEF(ext16s_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_ext16s_i32))
38
void page_init(void);
29
DEF(ext8u_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_ext8u_i32))
39
void tb_htable_init(void);
30
DEF(ext16u_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_ext16u_i32))
40
31
-DEF(bswap16_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_bswap16_i32))
41
+/* Return the current PC from CPU, which may be cached in TB. */
32
-DEF(bswap32_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_bswap32_i32))
42
+static inline target_ulong log_pc(CPUState *cpu, const TranslationBlock *tb)
33
+DEF(bswap16_i32, 1, 1, 1, IMPL(TCG_TARGET_HAS_bswap16_i32))
43
+{
34
+DEF(bswap32_i32, 1, 1, 1, IMPL(TCG_TARGET_HAS_bswap32_i32))
44
+ return tb_pc(tb);
35
DEF(not_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_not_i32))
45
+}
36
DEF(neg_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_neg_i32))
46
+
37
DEF(andc_i32, 1, 2, 0, IMPL(TCG_TARGET_HAS_andc_i32))
47
#endif /* ACCEL_TCG_INTERNAL_H */
38
@@ -XXX,XX +XXX,XX @@ DEF(ext32s_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_ext32s_i64))
48
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
39
DEF(ext8u_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_ext8u_i64))
49
index XXXXXXX..XXXXXXX 100644
40
DEF(ext16u_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_ext16u_i64))
50
--- a/include/exec/exec-all.h
41
DEF(ext32u_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_ext32u_i64))
51
+++ b/include/exec/exec-all.h
42
-DEF(bswap16_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_bswap16_i64))
52
@@ -XXX,XX +XXX,XX @@ struct TranslationBlock {
43
-DEF(bswap32_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_bswap32_i64))
53
uintptr_t jmp_dest[2];
44
-DEF(bswap64_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_bswap64_i64))
54
};
45
+DEF(bswap16_i64, 1, 1, 1, IMPL64 | IMPL(TCG_TARGET_HAS_bswap16_i64))
55
46
+DEF(bswap32_i64, 1, 1, 1, IMPL64 | IMPL(TCG_TARGET_HAS_bswap32_i64))
56
+/* Hide the read to avoid ifdefs for TARGET_TB_PCREL. */
47
+DEF(bswap64_i64, 1, 1, 1, IMPL64 | IMPL(TCG_TARGET_HAS_bswap64_i64))
57
+static inline target_ulong tb_pc(const TranslationBlock *tb)
48
DEF(not_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_not_i64))
58
+{
49
DEF(neg_i64, 1, 1, 0, IMPL64 | IMPL(TCG_TARGET_HAS_neg_i64))
59
+ return tb->pc;
50
DEF(andc_i64, 1, 2, 0, IMPL64 | IMPL(TCG_TARGET_HAS_andc_i64))
60
+}
61
+
62
/* Hide the qatomic_read to make code a little easier on the eyes */
63
static inline uint32_t tb_cflags(const TranslationBlock *tb)
64
{
51
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
65
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
52
index XXXXXXX..XXXXXXX 100644
66
index XXXXXXX..XXXXXXX 100644
53
--- a/include/tcg/tcg.h
67
--- a/include/tcg/tcg.h
54
+++ b/include/tcg/tcg.h
68
+++ b/include/tcg/tcg.h
55
@@ -XXX,XX +XXX,XX @@ typedef TCGv_ptr TCGv_env;
69
@@ -XXX,XX +XXX,XX @@ void tcg_register_thread(void);
56
/* Used to align parameters. See the comment before tcgv_i32_temp. */
70
void tcg_prologue_init(TCGContext *s);
57
#define TCG_CALL_DUMMY_ARG ((TCGArg)0)
71
void tcg_func_start(TCGContext *s);
58
72
59
+/*
73
-int tcg_gen_code(TCGContext *s, TranslationBlock *tb);
60
+ * Flags for the bswap opcodes.
74
+int tcg_gen_code(TCGContext *s, TranslationBlock *tb, target_ulong pc_start);
61
+ * If IZ, the input is zero-extended, otherwise unknown.
75
62
+ * If OZ or OS, the output is zero- or sign-extended respectively,
76
void tcg_set_frame(TCGContext *s, TCGReg reg, intptr_t start, intptr_t size);
63
+ * otherwise the high bits are undefined.
77
64
+ */
78
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
65
+enum {
79
index XXXXXXX..XXXXXXX 100644
66
+ TCG_BSWAP_IZ = 1,
80
--- a/accel/tcg/cpu-exec.c
67
+ TCG_BSWAP_OZ = 2,
81
+++ b/accel/tcg/cpu-exec.c
68
+ TCG_BSWAP_OS = 4,
82
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
69
+};
83
const TranslationBlock *tb = p;
84
const struct tb_desc *desc = d;
85
86
- if (tb->pc == desc->pc &&
87
+ if (tb_pc(tb) == desc->pc &&
88
tb->page_addr[0] == desc->page_addr0 &&
89
tb->cs_base == desc->cs_base &&
90
tb->flags == desc->flags &&
91
@@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
92
return tb;
93
}
94
95
-static inline void log_cpu_exec(target_ulong pc, CPUState *cpu,
96
- const TranslationBlock *tb)
97
+static void log_cpu_exec(target_ulong pc, CPUState *cpu,
98
+ const TranslationBlock *tb)
99
{
100
- if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_CPU | CPU_LOG_EXEC))
101
- && qemu_log_in_addr_range(pc)) {
102
-
103
+ if (qemu_log_in_addr_range(pc)) {
104
qemu_log_mask(CPU_LOG_EXEC,
105
"Trace %d: %p [" TARGET_FMT_lx
106
"/" TARGET_FMT_lx "/%08x/%08x] %s\n",
107
@@ -XXX,XX +XXX,XX @@ const void *HELPER(lookup_tb_ptr)(CPUArchState *env)
108
return tcg_code_gen_epilogue;
109
}
110
111
- log_cpu_exec(pc, cpu, tb);
112
+ if (qemu_loglevel_mask(CPU_LOG_TB_CPU | CPU_LOG_EXEC)) {
113
+ log_cpu_exec(pc, cpu, tb);
114
+ }
115
116
return tb->tc.ptr;
117
}
118
@@ -XXX,XX +XXX,XX @@ cpu_tb_exec(CPUState *cpu, TranslationBlock *itb, int *tb_exit)
119
TranslationBlock *last_tb;
120
const void *tb_ptr = itb->tc.ptr;
121
122
- log_cpu_exec(itb->pc, cpu, itb);
123
+ if (qemu_loglevel_mask(CPU_LOG_TB_CPU | CPU_LOG_EXEC)) {
124
+ log_cpu_exec(log_pc(cpu, itb), cpu, itb);
125
+ }
126
127
qemu_thread_jit_execute();
128
ret = tcg_qemu_tb_exec(env, tb_ptr);
129
@@ -XXX,XX +XXX,XX @@ cpu_tb_exec(CPUState *cpu, TranslationBlock *itb, int *tb_exit)
130
* of the start of the TB.
131
*/
132
CPUClass *cc = CPU_GET_CLASS(cpu);
133
- qemu_log_mask_and_addr(CPU_LOG_EXEC, last_tb->pc,
134
- "Stopped execution of TB chain before %p ["
135
- TARGET_FMT_lx "] %s\n",
136
- last_tb->tc.ptr, last_tb->pc,
137
- lookup_symbol(last_tb->pc));
70
+
138
+
71
typedef enum TCGTempVal {
139
if (cc->tcg_ops->synchronize_from_tb) {
72
TEMP_VAL_DEAD,
140
cc->tcg_ops->synchronize_from_tb(cpu, last_tb);
73
TEMP_VAL_REG,
141
} else {
74
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
142
assert(cc->set_pc);
75
index XXXXXXX..XXXXXXX 100644
143
- cc->set_pc(cpu, last_tb->pc);
76
--- a/tcg/tcg-op.c
144
+ cc->set_pc(cpu, tb_pc(last_tb));
77
+++ b/tcg/tcg-op.c
145
+ }
78
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ext16u_i32(TCGv_i32 ret, TCGv_i32 arg)
146
+ if (qemu_loglevel_mask(CPU_LOG_EXEC)) {
79
void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg)
147
+ target_ulong pc = log_pc(cpu, last_tb);
80
{
148
+ if (qemu_log_in_addr_range(pc)) {
81
if (TCG_TARGET_HAS_bswap16_i32) {
149
+ qemu_log("Stopped execution of TB chain before %p ["
82
- tcg_gen_op2_i32(INDEX_op_bswap16_i32, ret, arg);
150
+ TARGET_FMT_lx "] %s\n",
83
+ tcg_gen_op3i_i32(INDEX_op_bswap16_i32, ret, arg,
151
+ last_tb->tc.ptr, pc, lookup_symbol(pc));
84
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
152
+ }
153
}
154
}
155
156
@@ -XXX,XX +XXX,XX @@ static inline void tb_add_jump(TranslationBlock *tb, int n,
157
158
qemu_spin_unlock(&tb_next->jmp_lock);
159
160
- qemu_log_mask_and_addr(CPU_LOG_EXEC, tb->pc,
161
- "Linking TBs %p [" TARGET_FMT_lx
162
- "] index %d -> %p [" TARGET_FMT_lx "]\n",
163
- tb->tc.ptr, tb->pc, n,
164
- tb_next->tc.ptr, tb_next->pc);
165
+ qemu_log_mask(CPU_LOG_EXEC, "Linking TBs %p index %d -> %p\n",
166
+ tb->tc.ptr, n, tb_next->tc.ptr);
167
return;
168
169
out_unlock_next:
170
@@ -XXX,XX +XXX,XX @@ static inline bool cpu_handle_interrupt(CPUState *cpu,
171
}
172
173
static inline void cpu_loop_exec_tb(CPUState *cpu, TranslationBlock *tb,
174
+ target_ulong pc,
175
TranslationBlock **last_tb, int *tb_exit)
176
{
177
int32_t insns_left;
178
179
- trace_exec_tb(tb, tb->pc);
180
+ trace_exec_tb(tb, pc);
181
tb = cpu_tb_exec(cpu, tb, tb_exit);
182
if (*tb_exit != TB_EXIT_REQUESTED) {
183
*last_tb = tb;
184
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
185
tb_add_jump(last_tb, tb_exit, tb);
186
}
187
188
- cpu_loop_exec_tb(cpu, tb, &last_tb, &tb_exit);
189
+ cpu_loop_exec_tb(cpu, tb, pc, &last_tb, &tb_exit);
190
191
/* Try to align the host and virtual clocks
192
if the guest is in advance */
193
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
194
index XXXXXXX..XXXXXXX 100644
195
--- a/accel/tcg/translate-all.c
196
+++ b/accel/tcg/translate-all.c
197
@@ -XXX,XX +XXX,XX @@ static int encode_search(TranslationBlock *tb, uint8_t *block)
198
199
for (j = 0; j < TARGET_INSN_START_WORDS; ++j) {
200
if (i == 0) {
201
- prev = (j == 0 ? tb->pc : 0);
202
+ prev = (j == 0 ? tb_pc(tb) : 0);
203
} else {
204
prev = tcg_ctx->gen_insn_data[i - 1][j];
205
}
206
@@ -XXX,XX +XXX,XX @@ static int encode_search(TranslationBlock *tb, uint8_t *block)
207
static int cpu_restore_state_from_tb(CPUState *cpu, TranslationBlock *tb,
208
uintptr_t searched_pc, bool reset_icount)
209
{
210
- target_ulong data[TARGET_INSN_START_WORDS] = { tb->pc };
211
+ target_ulong data[TARGET_INSN_START_WORDS] = { tb_pc(tb) };
212
uintptr_t host_pc = (uintptr_t)tb->tc.ptr;
213
CPUArchState *env = cpu->env_ptr;
214
const uint8_t *p = tb->tc.ptr + tb->tc.size;
215
@@ -XXX,XX +XXX,XX @@ static bool tb_cmp(const void *ap, const void *bp)
216
const TranslationBlock *a = ap;
217
const TranslationBlock *b = bp;
218
219
- return a->pc == b->pc &&
220
+ return tb_pc(a) == tb_pc(b) &&
221
a->cs_base == b->cs_base &&
222
a->flags == b->flags &&
223
(tb_cflags(a) & ~CF_INVALID) == (tb_cflags(b) & ~CF_INVALID) &&
224
@@ -XXX,XX +XXX,XX @@ static void do_tb_invalidate_check(void *p, uint32_t hash, void *userp)
225
TranslationBlock *tb = p;
226
target_ulong addr = *(target_ulong *)userp;
227
228
- if (!(addr + TARGET_PAGE_SIZE <= tb->pc || addr >= tb->pc + tb->size)) {
229
+ if (!(addr + TARGET_PAGE_SIZE <= tb_pc(tb) ||
230
+ addr >= tb_pc(tb) + tb->size)) {
231
printf("ERROR invalidate: address=" TARGET_FMT_lx
232
- " PC=%08lx size=%04x\n", addr, (long)tb->pc, tb->size);
233
+ " PC=%08lx size=%04x\n", addr, (long)tb_pc(tb), tb->size);
234
}
235
}
236
237
@@ -XXX,XX +XXX,XX @@ static void do_tb_page_check(void *p, uint32_t hash, void *userp)
238
TranslationBlock *tb = p;
239
int flags1, flags2;
240
241
- flags1 = page_get_flags(tb->pc);
242
- flags2 = page_get_flags(tb->pc + tb->size - 1);
243
+ flags1 = page_get_flags(tb_pc(tb));
244
+ flags2 = page_get_flags(tb_pc(tb) + tb->size - 1);
245
if ((flags1 & PAGE_WRITE) || (flags2 & PAGE_WRITE)) {
246
printf("ERROR page flags: PC=%08lx size=%04x f1=%x f2=%x\n",
247
- (long)tb->pc, tb->size, flags1, flags2);
248
+ (long)tb_pc(tb), tb->size, flags1, flags2);
249
}
250
}
251
252
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
253
254
/* remove the TB from the hash list */
255
phys_pc = tb->page_addr[0];
256
- h = tb_hash_func(phys_pc, tb->pc, tb->flags, orig_cflags,
257
+ h = tb_hash_func(phys_pc, tb_pc(tb), tb->flags, orig_cflags,
258
tb->trace_vcpu_dstate);
259
if (!qht_remove(&tb_ctx.htable, tb, h)) {
260
return;
261
@@ -XXX,XX +XXX,XX @@ tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
262
}
263
264
/* add in the hash table */
265
- h = tb_hash_func(phys_pc, tb->pc, tb->flags, tb->cflags,
266
+ h = tb_hash_func(phys_pc, tb_pc(tb), tb->flags, tb->cflags,
267
tb->trace_vcpu_dstate);
268
qht_insert(&tb_ctx.htable, tb, h, &existing_tb);
269
270
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
271
tcg_ctx->cpu = NULL;
272
max_insns = tb->icount;
273
274
- trace_translate_block(tb, tb->pc, tb->tc.ptr);
275
+ trace_translate_block(tb, pc, tb->tc.ptr);
276
277
/* generate machine code */
278
tb->jmp_reset_offset[0] = TB_JMP_RESET_OFFSET_INVALID;
279
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
280
ti = profile_getclock();
281
#endif
282
283
- gen_code_size = tcg_gen_code(tcg_ctx, tb);
284
+ gen_code_size = tcg_gen_code(tcg_ctx, tb, pc);
285
if (unlikely(gen_code_size < 0)) {
286
error_return:
287
switch (gen_code_size) {
288
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
289
290
#ifdef DEBUG_DISAS
291
if (qemu_loglevel_mask(CPU_LOG_TB_OUT_ASM) &&
292
- qemu_log_in_addr_range(tb->pc)) {
293
+ qemu_log_in_addr_range(pc)) {
294
FILE *logfile = qemu_log_trylock();
295
if (logfile) {
296
int code_size, data_size;
297
@@ -XXX,XX +XXX,XX @@ void cpu_io_recompile(CPUState *cpu, uintptr_t retaddr)
298
*/
299
cpu->cflags_next_tb = curr_cflags(cpu) | CF_MEMI_ONLY | CF_LAST_IO | n;
300
301
- qemu_log_mask_and_addr(CPU_LOG_EXEC, tb->pc,
302
- "cpu_io_recompile: rewound execution of TB to "
303
- TARGET_FMT_lx "\n", tb->pc);
304
+ if (qemu_loglevel_mask(CPU_LOG_EXEC)) {
305
+ target_ulong pc = log_pc(cpu, tb);
306
+ if (qemu_log_in_addr_range(pc)) {
307
+ qemu_log("cpu_io_recompile: rewound execution of TB to "
308
+ TARGET_FMT_lx "\n", pc);
309
+ }
310
+ }
311
312
cpu_loop_exit_noexc(cpu);
313
}
314
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
315
index XXXXXXX..XXXXXXX 100644
316
--- a/target/arm/cpu.c
317
+++ b/target/arm/cpu.c
318
@@ -XXX,XX +XXX,XX @@ void arm_cpu_synchronize_from_tb(CPUState *cs,
319
* never possible for an AArch64 TB to chain to an AArch32 TB.
320
*/
321
if (is_a64(env)) {
322
- env->pc = tb->pc;
323
+ env->pc = tb_pc(tb);
85
} else {
324
} else {
86
TCGv_i32 t0 = tcg_temp_new_i32();
325
- env->regs[15] = tb->pc;
87
326
+ env->regs[15] = tb_pc(tb);
88
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg)
327
}
89
void tcg_gen_bswap32_i32(TCGv_i32 ret, TCGv_i32 arg)
328
}
90
{
329
#endif /* CONFIG_TCG */
91
if (TCG_TARGET_HAS_bswap32_i32) {
330
diff --git a/target/avr/cpu.c b/target/avr/cpu.c
92
- tcg_gen_op2_i32(INDEX_op_bswap32_i32, ret, arg);
331
index XXXXXXX..XXXXXXX 100644
93
+ tcg_gen_op3i_i32(INDEX_op_bswap32_i32, ret, arg, 0);
332
--- a/target/avr/cpu.c
333
+++ b/target/avr/cpu.c
334
@@ -XXX,XX +XXX,XX @@ static void avr_cpu_synchronize_from_tb(CPUState *cs,
335
AVRCPU *cpu = AVR_CPU(cs);
336
CPUAVRState *env = &cpu->env;
337
338
- env->pc_w = tb->pc / 2; /* internally PC points to words */
339
+ env->pc_w = tb_pc(tb) / 2; /* internally PC points to words */
340
}
341
342
static void avr_cpu_reset(DeviceState *ds)
343
diff --git a/target/hexagon/cpu.c b/target/hexagon/cpu.c
344
index XXXXXXX..XXXXXXX 100644
345
--- a/target/hexagon/cpu.c
346
+++ b/target/hexagon/cpu.c
347
@@ -XXX,XX +XXX,XX @@ static void hexagon_cpu_synchronize_from_tb(CPUState *cs,
348
{
349
HexagonCPU *cpu = HEXAGON_CPU(cs);
350
CPUHexagonState *env = &cpu->env;
351
- env->gpr[HEX_REG_PC] = tb->pc;
352
+ env->gpr[HEX_REG_PC] = tb_pc(tb);
353
}
354
355
static bool hexagon_cpu_has_work(CPUState *cs)
356
diff --git a/target/hppa/cpu.c b/target/hppa/cpu.c
357
index XXXXXXX..XXXXXXX 100644
358
--- a/target/hppa/cpu.c
359
+++ b/target/hppa/cpu.c
360
@@ -XXX,XX +XXX,XX @@ static void hppa_cpu_synchronize_from_tb(CPUState *cs,
361
HPPACPU *cpu = HPPA_CPU(cs);
362
363
#ifdef CONFIG_USER_ONLY
364
- cpu->env.iaoq_f = tb->pc;
365
+ cpu->env.iaoq_f = tb_pc(tb);
366
cpu->env.iaoq_b = tb->cs_base;
367
#else
368
/* Recover the IAOQ values from the GVA + PRIV. */
369
@@ -XXX,XX +XXX,XX @@ static void hppa_cpu_synchronize_from_tb(CPUState *cs,
370
int32_t diff = cs_base;
371
372
cpu->env.iasq_f = iasq_f;
373
- cpu->env.iaoq_f = (tb->pc & ~iasq_f) + priv;
374
+ cpu->env.iaoq_f = (tb_pc(tb) & ~iasq_f) + priv;
375
if (diff) {
376
cpu->env.iaoq_b = cpu->env.iaoq_f + diff;
377
}
378
diff --git a/target/i386/tcg/tcg-cpu.c b/target/i386/tcg/tcg-cpu.c
379
index XXXXXXX..XXXXXXX 100644
380
--- a/target/i386/tcg/tcg-cpu.c
381
+++ b/target/i386/tcg/tcg-cpu.c
382
@@ -XXX,XX +XXX,XX @@ static void x86_cpu_synchronize_from_tb(CPUState *cs,
383
{
384
X86CPU *cpu = X86_CPU(cs);
385
386
- cpu->env.eip = tb->pc - tb->cs_base;
387
+ cpu->env.eip = tb_pc(tb) - tb->cs_base;
388
}
389
390
#ifndef CONFIG_USER_ONLY
391
diff --git a/target/loongarch/cpu.c b/target/loongarch/cpu.c
392
index XXXXXXX..XXXXXXX 100644
393
--- a/target/loongarch/cpu.c
394
+++ b/target/loongarch/cpu.c
395
@@ -XXX,XX +XXX,XX @@ static void loongarch_cpu_synchronize_from_tb(CPUState *cs,
396
LoongArchCPU *cpu = LOONGARCH_CPU(cs);
397
CPULoongArchState *env = &cpu->env;
398
399
- env->pc = tb->pc;
400
+ env->pc = tb_pc(tb);
401
}
402
#endif /* CONFIG_TCG */
403
404
diff --git a/target/microblaze/cpu.c b/target/microblaze/cpu.c
405
index XXXXXXX..XXXXXXX 100644
406
--- a/target/microblaze/cpu.c
407
+++ b/target/microblaze/cpu.c
408
@@ -XXX,XX +XXX,XX @@ static void mb_cpu_synchronize_from_tb(CPUState *cs,
409
{
410
MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
411
412
- cpu->env.pc = tb->pc;
413
+ cpu->env.pc = tb_pc(tb);
414
cpu->env.iflags = tb->flags & IFLAGS_TB_MASK;
415
}
416
417
diff --git a/target/mips/tcg/exception.c b/target/mips/tcg/exception.c
418
index XXXXXXX..XXXXXXX 100644
419
--- a/target/mips/tcg/exception.c
420
+++ b/target/mips/tcg/exception.c
421
@@ -XXX,XX +XXX,XX @@ void mips_cpu_synchronize_from_tb(CPUState *cs, const TranslationBlock *tb)
422
MIPSCPU *cpu = MIPS_CPU(cs);
423
CPUMIPSState *env = &cpu->env;
424
425
- env->active_tc.PC = tb->pc;
426
+ env->active_tc.PC = tb_pc(tb);
427
env->hflags &= ~MIPS_HFLAG_BMASK;
428
env->hflags |= tb->flags & MIPS_HFLAG_BMASK;
429
}
430
diff --git a/target/mips/tcg/sysemu/special_helper.c b/target/mips/tcg/sysemu/special_helper.c
431
index XXXXXXX..XXXXXXX 100644
432
--- a/target/mips/tcg/sysemu/special_helper.c
433
+++ b/target/mips/tcg/sysemu/special_helper.c
434
@@ -XXX,XX +XXX,XX @@ bool mips_io_recompile_replay_branch(CPUState *cs, const TranslationBlock *tb)
435
CPUMIPSState *env = &cpu->env;
436
437
if ((env->hflags & MIPS_HFLAG_BMASK) != 0
438
- && env->active_tc.PC != tb->pc) {
439
+ && env->active_tc.PC != tb_pc(tb)) {
440
env->active_tc.PC -= (env->hflags & MIPS_HFLAG_B16 ? 2 : 4);
441
env->hflags &= ~MIPS_HFLAG_BMASK;
442
return true;
443
diff --git a/target/openrisc/cpu.c b/target/openrisc/cpu.c
444
index XXXXXXX..XXXXXXX 100644
445
--- a/target/openrisc/cpu.c
446
+++ b/target/openrisc/cpu.c
447
@@ -XXX,XX +XXX,XX @@ static void openrisc_cpu_synchronize_from_tb(CPUState *cs,
448
{
449
OpenRISCCPU *cpu = OPENRISC_CPU(cs);
450
451
- cpu->env.pc = tb->pc;
452
+ cpu->env.pc = tb_pc(tb);
453
}
454
455
456
diff --git a/target/riscv/cpu.c b/target/riscv/cpu.c
457
index XXXXXXX..XXXXXXX 100644
458
--- a/target/riscv/cpu.c
459
+++ b/target/riscv/cpu.c
460
@@ -XXX,XX +XXX,XX @@ static void riscv_cpu_synchronize_from_tb(CPUState *cs,
461
RISCVMXL xl = FIELD_EX32(tb->flags, TB_FLAGS, XL);
462
463
if (xl == MXL_RV32) {
464
- env->pc = (int32_t)tb->pc;
465
+ env->pc = (int32_t)tb_pc(tb);
94
} else {
466
} else {
95
TCGv_i32 t0 = tcg_temp_new_i32();
467
- env->pc = tb->pc;
96
TCGv_i32 t1 = tcg_temp_new_i32();
468
+ env->pc = tb_pc(tb);
97
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap16_i64(TCGv_i64 ret, TCGv_i64 arg)
469
}
98
tcg_gen_bswap16_i32(TCGV_LOW(ret), TCGV_LOW(arg));
470
}
99
tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
471
100
} else if (TCG_TARGET_HAS_bswap16_i64) {
472
diff --git a/target/rx/cpu.c b/target/rx/cpu.c
101
- tcg_gen_op2_i64(INDEX_op_bswap16_i64, ret, arg);
473
index XXXXXXX..XXXXXXX 100644
102
+ tcg_gen_op3i_i64(INDEX_op_bswap16_i64, ret, arg,
474
--- a/target/rx/cpu.c
103
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
475
+++ b/target/rx/cpu.c
104
} else {
476
@@ -XXX,XX +XXX,XX @@ static void rx_cpu_synchronize_from_tb(CPUState *cs,
105
TCGv_i64 t0 = tcg_temp_new_i64();
477
{
106
478
RXCPU *cpu = RX_CPU(cs);
107
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg)
479
108
tcg_gen_bswap32_i32(TCGV_LOW(ret), TCGV_LOW(arg));
480
- cpu->env.pc = tb->pc;
109
tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
481
+ cpu->env.pc = tb_pc(tb);
110
} else if (TCG_TARGET_HAS_bswap32_i64) {
482
}
111
- tcg_gen_op2_i64(INDEX_op_bswap32_i64, ret, arg);
483
112
+ tcg_gen_op3i_i64(INDEX_op_bswap32_i64, ret, arg,
484
static bool rx_cpu_has_work(CPUState *cs)
113
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
485
diff --git a/target/sh4/cpu.c b/target/sh4/cpu.c
114
} else {
486
index XXXXXXX..XXXXXXX 100644
115
TCGv_i64 t0 = tcg_temp_new_i64();
487
--- a/target/sh4/cpu.c
116
TCGv_i64 t1 = tcg_temp_new_i64();
488
+++ b/target/sh4/cpu.c
117
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap64_i64(TCGv_i64 ret, TCGv_i64 arg)
489
@@ -XXX,XX +XXX,XX @@ static void superh_cpu_synchronize_from_tb(CPUState *cs,
118
tcg_temp_free_i32(t0);
490
{
119
tcg_temp_free_i32(t1);
491
SuperHCPU *cpu = SUPERH_CPU(cs);
120
} else if (TCG_TARGET_HAS_bswap64_i64) {
492
121
- tcg_gen_op2_i64(INDEX_op_bswap64_i64, ret, arg);
493
- cpu->env.pc = tb->pc;
122
+ tcg_gen_op3i_i64(INDEX_op_bswap64_i64, ret, arg, 0);
494
+ cpu->env.pc = tb_pc(tb);
123
} else {
495
cpu->env.flags = tb->flags & TB_FLAG_ENVFLAGS_MASK;
124
TCGv_i64 t0 = tcg_temp_new_i64();
496
}
125
TCGv_i64 t1 = tcg_temp_new_i64();
497
498
@@ -XXX,XX +XXX,XX @@ static bool superh_io_recompile_replay_branch(CPUState *cs,
499
CPUSH4State *env = &cpu->env;
500
501
if ((env->flags & ((DELAY_SLOT | DELAY_SLOT_CONDITIONAL))) != 0
502
- && env->pc != tb->pc) {
503
+ && env->pc != tb_pc(tb)) {
504
env->pc -= 2;
505
env->flags &= ~(DELAY_SLOT | DELAY_SLOT_CONDITIONAL);
506
return true;
507
diff --git a/target/sparc/cpu.c b/target/sparc/cpu.c
508
index XXXXXXX..XXXXXXX 100644
509
--- a/target/sparc/cpu.c
510
+++ b/target/sparc/cpu.c
511
@@ -XXX,XX +XXX,XX @@ static void sparc_cpu_synchronize_from_tb(CPUState *cs,
512
{
513
SPARCCPU *cpu = SPARC_CPU(cs);
514
515
- cpu->env.pc = tb->pc;
516
+ cpu->env.pc = tb_pc(tb);
517
cpu->env.npc = tb->cs_base;
518
}
519
520
diff --git a/target/tricore/cpu.c b/target/tricore/cpu.c
521
index XXXXXXX..XXXXXXX 100644
522
--- a/target/tricore/cpu.c
523
+++ b/target/tricore/cpu.c
524
@@ -XXX,XX +XXX,XX @@ static void tricore_cpu_synchronize_from_tb(CPUState *cs,
525
TriCoreCPU *cpu = TRICORE_CPU(cs);
526
CPUTriCoreState *env = &cpu->env;
527
528
- env->PC = tb->pc;
529
+ env->PC = tb_pc(tb);
530
}
531
532
static void tricore_cpu_reset(DeviceState *dev)
126
diff --git a/tcg/tcg.c b/tcg/tcg.c
533
diff --git a/tcg/tcg.c b/tcg/tcg.c
127
index XXXXXXX..XXXXXXX 100644
534
index XXXXXXX..XXXXXXX 100644
128
--- a/tcg/tcg.c
535
--- a/tcg/tcg.c
129
+++ b/tcg/tcg.c
536
+++ b/tcg/tcg.c
130
@@ -XXX,XX +XXX,XX @@ static const char * const alignment_name[(MO_AMASK >> MO_ASHIFT) + 1] = {
537
@@ -XXX,XX +XXX,XX @@ int64_t tcg_cpu_exec_time(void)
131
[MO_ALIGN_64 >> MO_ASHIFT] = "al64+",
538
#endif
132
};
539
133
540
134
+static const char bswap_flag_name[][6] = {
541
-int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
135
+ [TCG_BSWAP_IZ] = "iz",
542
+int tcg_gen_code(TCGContext *s, TranslationBlock *tb, target_ulong pc_start)
136
+ [TCG_BSWAP_OZ] = "oz",
543
{
137
+ [TCG_BSWAP_OS] = "os",
544
#ifdef CONFIG_PROFILER
138
+ [TCG_BSWAP_IZ | TCG_BSWAP_OZ] = "iz,oz",
545
TCGProfile *prof = &s->prof;
139
+ [TCG_BSWAP_IZ | TCG_BSWAP_OS] = "iz,os",
546
@@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
140
+};
547
141
+
548
#ifdef DEBUG_DISAS
142
static inline bool tcg_regset_single(TCGRegSet d)
549
if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP)
143
{
550
- && qemu_log_in_addr_range(tb->pc))) {
144
return (d & (d - 1)) == 0;
551
+ && qemu_log_in_addr_range(pc_start))) {
145
@@ -XXX,XX +XXX,XX @@ static void tcg_dump_ops(TCGContext *s, bool have_prefs)
552
FILE *logfile = qemu_log_trylock();
146
i = 1;
553
if (logfile) {
147
}
554
fprintf(logfile, "OP:\n");
148
break;
555
@@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
149
+ case INDEX_op_bswap16_i32:
556
if (s->nb_indirects > 0) {
150
+ case INDEX_op_bswap16_i64:
557
#ifdef DEBUG_DISAS
151
+ case INDEX_op_bswap32_i32:
558
if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_IND)
152
+ case INDEX_op_bswap32_i64:
559
- && qemu_log_in_addr_range(tb->pc))) {
153
+ case INDEX_op_bswap64_i64:
560
+ && qemu_log_in_addr_range(pc_start))) {
154
+ {
561
FILE *logfile = qemu_log_trylock();
155
+ TCGArg flags = op->args[k];
562
if (logfile) {
156
+ const char *name = NULL;
563
fprintf(logfile, "OP before indirect lowering:\n");
157
+
564
@@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
158
+ if (flags < ARRAY_SIZE(bswap_flag_name)) {
565
159
+ name = bswap_flag_name[flags];
566
#ifdef DEBUG_DISAS
160
+ }
567
if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_OPT)
161
+ if (name) {
568
- && qemu_log_in_addr_range(tb->pc))) {
162
+ col += qemu_log(",%s", name);
569
+ && qemu_log_in_addr_range(pc_start))) {
163
+ } else {
570
FILE *logfile = qemu_log_trylock();
164
+ col += qemu_log(",$0x%" TCG_PRIlx, flags);
571
if (logfile) {
165
+ }
572
fprintf(logfile, "OP after optimization and liveness analysis:\n");
166
+ i = k = 1;
167
+ }
168
+ break;
169
default:
170
i = 0;
171
break;
172
diff --git a/tcg/README b/tcg/README
173
index XXXXXXX..XXXXXXX 100644
174
--- a/tcg/README
175
+++ b/tcg/README
176
@@ -XXX,XX +XXX,XX @@ ext32u_i64 t0, t1
177
178
8, 16 or 32 bit sign/zero extension (both operands must have the same type)
179
180
-* bswap16_i32/i64 t0, t1
181
+* bswap16_i32/i64 t0, t1, flags
182
183
-16 bit byte swap on a 32/64 bit value. It assumes that the two/six high order
184
-bytes are set to zero.
185
+16 bit byte swap on the low bits of a 32/64 bit input.
186
+If flags & TCG_BSWAP_IZ, then t1 is known to be zero-extended from bit 15.
187
+If flags & TCG_BSWAP_OZ, then t0 will be zero-extended from bit 15.
188
+If flags & TCG_BSWAP_OS, then t0 will be sign-extended from bit 15.
189
+If neither TCG_BSWAP_OZ nor TCG_BSWAP_OS are set, then the bits of
190
+t0 above bit 15 may contain any value.
191
192
-* bswap32_i32/i64 t0, t1
193
+* bswap32_i64 t0, t1, flags
194
195
-32 bit byte swap on a 32/64 bit value. With a 64 bit value, it assumes that
196
-the four high order bytes are set to zero.
197
+32 bit byte swap on a 64-bit value. The flags are the same as for bswap16,
198
+except they apply from bit 31 instead of bit 15.
199
200
-* bswap64_i64 t0, t1
201
+* bswap32_i32 t0, t1, flags
202
+* bswap64_i64 t0, t1, flags
203
204
-64 bit byte swap
205
+32/64 bit byte swap. The flags are ignored, but still present
206
+for consistency with the other bswap opcodes.
207
208
* discard_i32/i64 t0
209
210
--
573
--
211
2.25.1
574
2.34.1
212
575
213
576
diff view generated by jsdifflib
1
Tested-by: Michael Rolnik <mrolnik@gmail.com>
1
Prepare for targets to be able to produce TBs that can
2
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
2
run in more than one virtual context.
3
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
6
---
5
target/avr/translate.c | 234 ++++++++++++++++++++++-------------------
7
accel/tcg/internal.h | 4 +++
6
1 file changed, 128 insertions(+), 106 deletions(-)
8
accel/tcg/tb-jmp-cache.h | 41 +++++++++++++++++++++++++
9
include/exec/cpu-defs.h | 3 ++
10
include/exec/exec-all.h | 32 ++++++++++++++++++--
11
accel/tcg/cpu-exec.c | 16 ++++++----
12
accel/tcg/translate-all.c | 64 ++++++++++++++++++++++++++-------------
13
6 files changed, 131 insertions(+), 29 deletions(-)
7
14
8
diff --git a/target/avr/translate.c b/target/avr/translate.c
15
diff --git a/accel/tcg/internal.h b/accel/tcg/internal.h
9
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
10
--- a/target/avr/translate.c
17
--- a/accel/tcg/internal.h
11
+++ b/target/avr/translate.c
18
+++ b/accel/tcg/internal.h
12
@@ -XXX,XX +XXX,XX @@ static bool canonicalize_skip(DisasContext *ctx)
19
@@ -XXX,XX +XXX,XX @@ void tb_htable_init(void);
13
return true;
20
/* Return the current PC from CPU, which may be cached in TB. */
14
}
21
static inline target_ulong log_pc(CPUState *cpu, const TranslationBlock *tb)
15
22
{
16
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
23
+#if TARGET_TB_PCREL
17
+static void gen_breakpoint(DisasContext *ctx)
24
+ return cpu->cc->get_pc(cpu);
18
{
25
+#else
19
+ canonicalize_skip(ctx);
26
return tb_pc(tb);
20
+ tcg_gen_movi_tl(cpu_pc, ctx->npc);
27
+#endif
21
+ gen_helper_debug(cpu_env);
28
}
22
+ ctx->base.is_jmp = DISAS_NORETURN;
29
30
#endif /* ACCEL_TCG_INTERNAL_H */
31
diff --git a/accel/tcg/tb-jmp-cache.h b/accel/tcg/tb-jmp-cache.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/accel/tcg/tb-jmp-cache.h
34
+++ b/accel/tcg/tb-jmp-cache.h
35
@@ -XXX,XX +XXX,XX @@
36
37
/*
38
* Accessed in parallel; all accesses to 'tb' must be atomic.
39
+ * For TARGET_TB_PCREL, accesses to 'pc' must be protected by
40
+ * a load_acquire/store_release to 'tb'.
41
*/
42
struct CPUJumpCache {
43
struct {
44
TranslationBlock *tb;
45
+#if TARGET_TB_PCREL
46
+ target_ulong pc;
47
+#endif
48
} array[TB_JMP_CACHE_SIZE];
49
};
50
51
+static inline TranslationBlock *
52
+tb_jmp_cache_get_tb(CPUJumpCache *jc, uint32_t hash)
53
+{
54
+#if TARGET_TB_PCREL
55
+ /* Use acquire to ensure current load of pc from jc. */
56
+ return qatomic_load_acquire(&jc->array[hash].tb);
57
+#else
58
+ /* Use rcu_read to ensure current load of pc from *tb. */
59
+ return qatomic_rcu_read(&jc->array[hash].tb);
60
+#endif
23
+}
61
+}
24
+
62
+
25
+static void avr_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
63
+static inline target_ulong
64
+tb_jmp_cache_get_pc(CPUJumpCache *jc, uint32_t hash, TranslationBlock *tb)
26
+{
65
+{
27
+ DisasContext *ctx = container_of(dcbase, DisasContext, base);
66
+#if TARGET_TB_PCREL
28
CPUAVRState *env = cs->env_ptr;
67
+ return jc->array[hash].pc;
29
- DisasContext ctx1 = {
68
+#else
30
- .base.tb = tb,
69
+ return tb_pc(tb);
31
- .base.is_jmp = DISAS_NEXT,
70
+#endif
32
- .base.pc_first = tb->pc,
71
+}
33
- .base.pc_next = tb->pc,
72
+
34
- .base.singlestep_enabled = cs->singlestep_enabled,
73
+static inline void
35
- .cs = cs,
74
+tb_jmp_cache_set(CPUJumpCache *jc, uint32_t hash,
36
- .env = env,
75
+ TranslationBlock *tb, target_ulong pc)
37
- .memidx = 0,
76
+{
38
- .skip_cond = TCG_COND_NEVER,
77
+#if TARGET_TB_PCREL
39
- };
78
+ jc->array[hash].pc = pc;
40
- DisasContext *ctx = &ctx1;
79
+ /* Use store_release on tb to ensure pc is written first. */
41
- target_ulong pc_start = tb->pc / 2;
80
+ qatomic_store_release(&jc->array[hash].tb, tb);
42
- int num_insns = 0;
81
+#else
43
+ uint32_t tb_flags = ctx->base.tb->flags;
82
+ /* Use the pc value already stored in tb->pc. */
44
83
+ qatomic_set(&jc->array[hash].tb, tb);
45
- if (tb->flags & TB_FLAGS_FULL_ACCESS) {
84
+#endif
46
- /*
85
+}
47
- * This flag is set by ST/LD instruction we will regenerate it ONLY
86
+
48
- * with mem/cpu memory access instead of mem access
87
#endif /* ACCEL_TCG_TB_JMP_CACHE_H */
49
- */
88
diff --git a/include/exec/cpu-defs.h b/include/exec/cpu-defs.h
50
- max_insns = 1;
89
index XXXXXXX..XXXXXXX 100644
51
- }
90
--- a/include/exec/cpu-defs.h
52
- if (ctx->base.singlestep_enabled) {
91
+++ b/include/exec/cpu-defs.h
53
- max_insns = 1;
92
@@ -XXX,XX +XXX,XX @@
54
- }
93
# error TARGET_PAGE_BITS must be defined in cpu-param.h
55
+ ctx->cs = cs;
94
# endif
56
+ ctx->env = env;
95
#endif
57
+ ctx->npc = ctx->base.pc_first / 2;
96
+#ifndef TARGET_TB_PCREL
58
97
+# define TARGET_TB_PCREL 0
59
- gen_tb_start(tb);
98
+#endif
60
-
99
61
- ctx->npc = pc_start;
100
#define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
62
- if (tb->flags & TB_FLAGS_SKIP) {
101
63
+ ctx->skip_cond = TCG_COND_NEVER;
102
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
64
+ if (tb_flags & TB_FLAGS_SKIP) {
103
index XXXXXXX..XXXXXXX 100644
65
ctx->skip_cond = TCG_COND_ALWAYS;
104
--- a/include/exec/exec-all.h
66
ctx->skip_var0 = cpu_skip;
105
+++ b/include/exec/exec-all.h
67
}
106
@@ -XXX,XX +XXX,XX @@ struct tb_tc {
68
107
};
69
- do {
108
70
- TCGLabel *skip_label = NULL;
109
struct TranslationBlock {
71
-
110
- target_ulong pc; /* simulated PC corresponding to this block (EIP + CS base) */
72
- /* translate current instruction */
111
- target_ulong cs_base; /* CS base for this block */
73
- tcg_gen_insn_start(ctx->npc);
112
+#if !TARGET_TB_PCREL
74
- num_insns++;
113
+ /*
75
-
114
+ * Guest PC corresponding to this block. This must be the true
76
+ if (tb_flags & TB_FLAGS_FULL_ACCESS) {
115
+ * virtual address. Therefore e.g. x86 stores EIP + CS_BASE, and
77
/*
116
+ * targets like Arm, MIPS, HP-PA, which reuse low bits for ISA or
78
- * this is due to some strange GDB behavior
117
+ * privilege, must store those bits elsewhere.
79
- * let's assume main has address 0x100
118
+ *
80
- * b main - sets breakpoint at address 0x00000100 (code)
119
+ * If TARGET_TB_PCREL, the opcodes for the TranslationBlock are
81
- * b *0x100 - sets breakpoint at address 0x00800100 (data)
120
+ * written such that the TB is associated only with the physical
82
+ * This flag is set by ST/LD instruction we will regenerate it ONLY
121
+ * page and may be run in any virtual address context. In this case,
83
+ * with mem/cpu memory access instead of mem access
122
+ * PC must always be taken from ENV in a target-specific manner.
84
*/
123
+ * Unwind information is taken as offsets from the page, to be
85
- if (unlikely(!ctx->base.singlestep_enabled &&
124
+ * deposited into the "current" PC.
86
- (cpu_breakpoint_test(cs, OFFSET_CODE + ctx->npc * 2, BP_ANY) ||
125
+ */
87
- cpu_breakpoint_test(cs, OFFSET_DATA + ctx->npc * 2, BP_ANY)))) {
126
+ target_ulong pc;
88
- canonicalize_skip(ctx);
127
+#endif
89
- tcg_gen_movi_tl(cpu_pc, ctx->npc);
128
+
90
- gen_helper_debug(cpu_env);
129
+ /*
91
- goto done_generating;
130
+ * Target-specific data associated with the TranslationBlock, e.g.:
92
- }
131
+ * x86: the original user, the Code Segment virtual base,
93
+ ctx->base.max_insns = 1;
132
+ * arm: an extension of tb->flags,
133
+ * s390x: instruction data for EXECUTE,
134
+ * sparc: the next pc of the instruction queue (for delay slots).
135
+ */
136
+ target_ulong cs_base;
137
+
138
uint32_t flags; /* flags defining in which context the code was generated */
139
uint32_t cflags; /* compile flags */
140
141
@@ -XXX,XX +XXX,XX @@ struct TranslationBlock {
142
/* Hide the read to avoid ifdefs for TARGET_TB_PCREL. */
143
static inline target_ulong tb_pc(const TranslationBlock *tb)
144
{
145
+#if TARGET_TB_PCREL
146
+ qemu_build_not_reached();
147
+#else
148
return tb->pc;
149
+#endif
150
}
151
152
/* Hide the qatomic_read to make code a little easier on the eyes */
153
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
154
index XXXXXXX..XXXXXXX 100644
155
--- a/accel/tcg/cpu-exec.c
156
+++ b/accel/tcg/cpu-exec.c
157
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
158
const TranslationBlock *tb = p;
159
const struct tb_desc *desc = d;
160
161
- if (tb_pc(tb) == desc->pc &&
162
+ if ((TARGET_TB_PCREL || tb_pc(tb) == desc->pc) &&
163
tb->page_addr[0] == desc->page_addr0 &&
164
tb->cs_base == desc->cs_base &&
165
tb->flags == desc->flags &&
166
@@ -XXX,XX +XXX,XX @@ static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
167
return NULL;
168
}
169
desc.page_addr0 = phys_pc;
170
- h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
171
+ h = tb_hash_func(phys_pc, (TARGET_TB_PCREL ? 0 : pc),
172
+ flags, cflags, *cpu->trace_dstate);
173
return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
174
}
175
176
@@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
177
uint32_t flags, uint32_t cflags)
178
{
179
TranslationBlock *tb;
180
+ CPUJumpCache *jc;
181
uint32_t hash;
182
183
/* we should never be trying to look up an INVALID tb */
184
tcg_debug_assert(!(cflags & CF_INVALID));
185
186
hash = tb_jmp_cache_hash_func(pc);
187
- tb = qatomic_rcu_read(&cpu->tb_jmp_cache->array[hash].tb);
188
+ jc = cpu->tb_jmp_cache;
189
+ tb = tb_jmp_cache_get_tb(jc, hash);
190
191
if (likely(tb &&
192
- tb->pc == pc &&
193
+ tb_jmp_cache_get_pc(jc, hash, tb) == pc &&
194
tb->cs_base == cs_base &&
195
tb->flags == flags &&
196
tb->trace_vcpu_dstate == *cpu->trace_dstate &&
197
@@ -XXX,XX +XXX,XX @@ static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
198
if (tb == NULL) {
199
return NULL;
200
}
201
- qatomic_set(&cpu->tb_jmp_cache->array[hash].tb, tb);
202
+ tb_jmp_cache_set(jc, hash, tb, pc);
203
return tb;
204
}
205
206
@@ -XXX,XX +XXX,XX @@ cpu_tb_exec(CPUState *cpu, TranslationBlock *itb, int *tb_exit)
207
if (cc->tcg_ops->synchronize_from_tb) {
208
cc->tcg_ops->synchronize_from_tb(cpu, last_tb);
209
} else {
210
+ assert(!TARGET_TB_PCREL);
211
assert(cc->set_pc);
212
cc->set_pc(cpu, tb_pc(last_tb));
213
}
214
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
215
* for the fast lookup
216
*/
217
h = tb_jmp_cache_hash_func(pc);
218
- qatomic_set(&cpu->tb_jmp_cache->array[h].tb, tb);
219
+ tb_jmp_cache_set(cpu->tb_jmp_cache, h, tb, pc);
220
}
221
222
#ifndef CONFIG_USER_ONLY
223
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
224
index XXXXXXX..XXXXXXX 100644
225
--- a/accel/tcg/translate-all.c
226
+++ b/accel/tcg/translate-all.c
227
@@ -XXX,XX +XXX,XX @@ static int encode_search(TranslationBlock *tb, uint8_t *block)
228
229
for (j = 0; j < TARGET_INSN_START_WORDS; ++j) {
230
if (i == 0) {
231
- prev = (j == 0 ? tb_pc(tb) : 0);
232
+ prev = (!TARGET_TB_PCREL && j == 0 ? tb_pc(tb) : 0);
233
} else {
234
prev = tcg_ctx->gen_insn_data[i - 1][j];
235
}
236
@@ -XXX,XX +XXX,XX @@ static int encode_search(TranslationBlock *tb, uint8_t *block)
237
static int cpu_restore_state_from_tb(CPUState *cpu, TranslationBlock *tb,
238
uintptr_t searched_pc, bool reset_icount)
239
{
240
- target_ulong data[TARGET_INSN_START_WORDS] = { tb_pc(tb) };
241
+ target_ulong data[TARGET_INSN_START_WORDS];
242
uintptr_t host_pc = (uintptr_t)tb->tc.ptr;
243
CPUArchState *env = cpu->env_ptr;
244
const uint8_t *p = tb->tc.ptr + tb->tc.size;
245
@@ -XXX,XX +XXX,XX @@ static int cpu_restore_state_from_tb(CPUState *cpu, TranslationBlock *tb,
246
return -1;
247
}
248
249
+ memset(data, 0, sizeof(data));
250
+ if (!TARGET_TB_PCREL) {
251
+ data[0] = tb_pc(tb);
94
+ }
252
+ }
95
+}
253
+
96
254
/* Reconstruct the stored insn data while looking for the point at
97
- /* Conditionally skip the next instruction, if indicated. */
255
which the end of the insn exceeds the searched_pc. */
98
- if (ctx->skip_cond != TCG_COND_NEVER) {
256
for (i = 0; i < num_insns; ++i) {
99
- skip_label = gen_new_label();
257
@@ -XXX,XX +XXX,XX @@ static bool tb_cmp(const void *ap, const void *bp)
100
- if (ctx->skip_var0 == cpu_skip) {
258
const TranslationBlock *a = ap;
101
- /*
259
const TranslationBlock *b = bp;
102
- * Copy cpu_skip so that we may zero it before the branch.
260
103
- * This ensures that cpu_skip is non-zero after the label
261
- return tb_pc(a) == tb_pc(b) &&
104
- * if and only if the skipped insn itself sets a skip.
262
- a->cs_base == b->cs_base &&
105
- */
263
- a->flags == b->flags &&
106
- ctx->free_skip_var0 = true;
264
- (tb_cflags(a) & ~CF_INVALID) == (tb_cflags(b) & ~CF_INVALID) &&
107
- ctx->skip_var0 = tcg_temp_new();
265
- a->trace_vcpu_dstate == b->trace_vcpu_dstate &&
108
- tcg_gen_mov_tl(ctx->skip_var0, cpu_skip);
266
- a->page_addr[0] == b->page_addr[0] &&
109
- tcg_gen_movi_tl(cpu_skip, 0);
267
- a->page_addr[1] == b->page_addr[1];
110
- }
268
+ return ((TARGET_TB_PCREL || tb_pc(a) == tb_pc(b)) &&
111
- if (ctx->skip_var1 == NULL) {
269
+ a->cs_base == b->cs_base &&
112
- tcg_gen_brcondi_tl(ctx->skip_cond, ctx->skip_var0,
270
+ a->flags == b->flags &&
113
- 0, skip_label);
271
+ (tb_cflags(a) & ~CF_INVALID) == (tb_cflags(b) & ~CF_INVALID) &&
114
- } else {
272
+ a->trace_vcpu_dstate == b->trace_vcpu_dstate &&
115
- tcg_gen_brcond_tl(ctx->skip_cond, ctx->skip_var0,
273
+ a->page_addr[0] == b->page_addr[0] &&
116
- ctx->skip_var1, skip_label);
274
+ a->page_addr[1] == b->page_addr[1]);
117
- ctx->skip_var1 = NULL;
275
}
118
- }
276
119
- if (ctx->free_skip_var0) {
277
void tb_htable_init(void)
120
- tcg_temp_free(ctx->skip_var0);
278
@@ -XXX,XX +XXX,XX @@ static inline void tb_jmp_unlink(TranslationBlock *dest)
121
- ctx->free_skip_var0 = false;
279
qemu_spin_unlock(&dest->jmp_lock);
122
- }
280
}
123
- ctx->skip_cond = TCG_COND_NEVER;
281
124
- ctx->skip_var0 = NULL;
282
+static void tb_jmp_cache_inval_tb(TranslationBlock *tb)
125
- }
126
+static void avr_tr_tb_start(DisasContextBase *db, CPUState *cs)
127
+{
283
+{
128
+}
284
+ CPUState *cpu;
129
285
+
130
- translate(ctx);
286
+ if (TARGET_TB_PCREL) {
131
+static void avr_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
287
+ /* A TB may be at any virtual address */
132
+{
288
+ CPU_FOREACH(cpu) {
133
+ DisasContext *ctx = container_of(dcbase, DisasContext, base);
289
+ tcg_flush_jmp_cache(cpu);
134
135
- if (skip_label) {
136
- canonicalize_skip(ctx);
137
- gen_set_label(skip_label);
138
- if (ctx->base.is_jmp == DISAS_NORETURN) {
139
- ctx->base.is_jmp = DISAS_CHAIN;
140
- }
141
- }
142
- } while (ctx->base.is_jmp == DISAS_NEXT
143
- && num_insns < max_insns
144
- && (ctx->npc - pc_start) * 2 < TARGET_PAGE_SIZE - 4
145
- && !tcg_op_buf_full());
146
+ tcg_gen_insn_start(ctx->npc);
147
+}
148
149
- if (tb->cflags & CF_LAST_IO) {
150
- gen_io_end();
151
+static bool avr_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cs,
152
+ const CPUBreakpoint *bp)
153
+{
154
+ DisasContext *ctx = container_of(dcbase, DisasContext, base);
155
+
156
+ gen_breakpoint(ctx);
157
+ return true;
158
+}
159
+
160
+static void avr_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
161
+{
162
+ DisasContext *ctx = container_of(dcbase, DisasContext, base);
163
+ TCGLabel *skip_label = NULL;
164
+
165
+ /*
166
+ * This is due to some strange GDB behavior
167
+ * Let's assume main has address 0x100:
168
+ * b main - sets breakpoint at address 0x00000100 (code)
169
+ * b *0x100 - sets breakpoint at address 0x00800100 (data)
170
+ *
171
+ * The translator driver has already taken care of the code pointer.
172
+ */
173
+ if (!ctx->base.singlestep_enabled &&
174
+ cpu_breakpoint_test(cs, OFFSET_DATA + ctx->base.pc_next, BP_ANY)) {
175
+ gen_breakpoint(ctx);
176
+ return;
177
}
178
179
+ /* Conditionally skip the next instruction, if indicated. */
180
+ if (ctx->skip_cond != TCG_COND_NEVER) {
181
+ skip_label = gen_new_label();
182
+ if (ctx->skip_var0 == cpu_skip) {
183
+ /*
184
+ * Copy cpu_skip so that we may zero it before the branch.
185
+ * This ensures that cpu_skip is non-zero after the label
186
+ * if and only if the skipped insn itself sets a skip.
187
+ */
188
+ ctx->free_skip_var0 = true;
189
+ ctx->skip_var0 = tcg_temp_new();
190
+ tcg_gen_mov_tl(ctx->skip_var0, cpu_skip);
191
+ tcg_gen_movi_tl(cpu_skip, 0);
192
+ }
290
+ }
193
+ if (ctx->skip_var1 == NULL) {
291
+ } else {
194
+ tcg_gen_brcondi_tl(ctx->skip_cond, ctx->skip_var0, 0, skip_label);
292
+ uint32_t h = tb_jmp_cache_hash_func(tb_pc(tb));
195
+ } else {
293
+
196
+ tcg_gen_brcond_tl(ctx->skip_cond, ctx->skip_var0,
294
+ CPU_FOREACH(cpu) {
197
+ ctx->skip_var1, skip_label);
295
+ CPUJumpCache *jc = cpu->tb_jmp_cache;
198
+ ctx->skip_var1 = NULL;
296
+
199
+ }
297
+ if (qatomic_read(&jc->array[h].tb) == tb) {
200
+ if (ctx->free_skip_var0) {
298
+ qatomic_set(&jc->array[h].tb, NULL);
201
+ tcg_temp_free(ctx->skip_var0);
299
+ }
202
+ ctx->free_skip_var0 = false;
203
+ }
204
+ ctx->skip_cond = TCG_COND_NEVER;
205
+ ctx->skip_var0 = NULL;
206
+ }
207
+
208
+ translate(ctx);
209
+
210
+ ctx->base.pc_next = ctx->npc * 2;
211
+
212
+ if (skip_label) {
213
+ canonicalize_skip(ctx);
214
+ gen_set_label(skip_label);
215
+ if (ctx->base.is_jmp == DISAS_NORETURN) {
216
+ ctx->base.is_jmp = DISAS_CHAIN;
217
+ }
218
+ }
219
+
220
+ if (ctx->base.is_jmp == DISAS_NEXT) {
221
+ target_ulong page_first = ctx->base.pc_first & TARGET_PAGE_MASK;
222
+
223
+ if ((ctx->base.pc_next - page_first) >= TARGET_PAGE_SIZE - 4) {
224
+ ctx->base.is_jmp = DISAS_TOO_MANY;
225
+ }
300
+ }
226
+ }
301
+ }
227
+}
302
+}
228
+
303
+
229
+static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
304
/*
230
+{
305
* In user-mode, call with mmap_lock held.
231
+ DisasContext *ctx = container_of(dcbase, DisasContext, base);
306
* In !user-mode, if @rm_from_page_list is set, call with the TB's pages'
232
bool nonconst_skip = canonicalize_skip(ctx);
307
@@ -XXX,XX +XXX,XX @@ static inline void tb_jmp_unlink(TranslationBlock *dest)
233
308
*/
234
switch (ctx->base.is_jmp) {
309
static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
235
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
310
{
236
default:
311
- CPUState *cpu;
237
g_assert_not_reached();
312
PageDesc *p;
238
}
313
uint32_t h;
239
+}
314
tb_page_addr_t phys_pc;
240
315
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
241
-done_generating:
316
242
- gen_tb_end(tb, num_insns);
317
/* remove the TB from the hash list */
243
+static void avr_tr_disas_log(const DisasContextBase *dcbase, CPUState *cs)
318
phys_pc = tb->page_addr[0];
244
+{
319
- h = tb_hash_func(phys_pc, tb_pc(tb), tb->flags, orig_cflags,
245
+ qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
320
- tb->trace_vcpu_dstate);
246
+ log_target_disas(cs, dcbase->pc_first, dcbase->tb->size);
321
+ h = tb_hash_func(phys_pc, (TARGET_TB_PCREL ? 0 : tb_pc(tb)),
247
+}
322
+ tb->flags, orig_cflags, tb->trace_vcpu_dstate);
248
323
if (!qht_remove(&tb_ctx.htable, tb, h)) {
249
- tb->size = (ctx->npc - pc_start) * 2;
324
return;
250
- tb->icount = num_insns;
325
}
251
+static const TranslatorOps avr_tr_ops = {
326
@@ -XXX,XX +XXX,XX @@ static void do_tb_phys_invalidate(TranslationBlock *tb, bool rm_from_page_list)
252
+ .init_disas_context = avr_tr_init_disas_context,
327
}
253
+ .tb_start = avr_tr_tb_start,
328
254
+ .insn_start = avr_tr_insn_start,
329
/* remove the TB from the hash list */
255
+ .breakpoint_check = avr_tr_breakpoint_check,
330
- h = tb_jmp_cache_hash_func(tb->pc);
256
+ .translate_insn = avr_tr_translate_insn,
331
- CPU_FOREACH(cpu) {
257
+ .tb_stop = avr_tr_tb_stop,
332
- CPUJumpCache *jc = cpu->tb_jmp_cache;
258
+ .disas_log = avr_tr_disas_log,
333
- if (qatomic_read(&jc->array[h].tb) == tb) {
259
+};
334
- qatomic_set(&jc->array[h].tb, NULL);
260
335
- }
261
-#ifdef DEBUG_DISAS
262
- if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
263
- && qemu_log_in_addr_range(tb->pc)) {
264
- FILE *fd;
265
- fd = qemu_log_lock();
266
- qemu_log("IN: %s\n", lookup_symbol(tb->pc));
267
- log_target_disas(cs, tb->pc, tb->size);
268
- qemu_log("\n");
269
- qemu_log_unlock(fd);
270
- }
336
- }
271
-#endif
337
+ tb_jmp_cache_inval_tb(tb);
272
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
338
273
+{
339
/* suppress this TB from the two jump lists */
274
+ DisasContext dc = { };
340
tb_remove_from_jmp_list(tb, 0);
275
+ translator_loop(&avr_tr_ops, &dc.base, cs, tb, max_insns);
341
@@ -XXX,XX +XXX,XX @@ tb_link_page(TranslationBlock *tb, tb_page_addr_t phys_pc,
276
}
342
}
277
343
278
void restore_state_to_opc(CPUAVRState *env, TranslationBlock *tb,
344
/* add in the hash table */
345
- h = tb_hash_func(phys_pc, tb_pc(tb), tb->flags, tb->cflags,
346
- tb->trace_vcpu_dstate);
347
+ h = tb_hash_func(phys_pc, (TARGET_TB_PCREL ? 0 : tb_pc(tb)),
348
+ tb->flags, tb->cflags, tb->trace_vcpu_dstate);
349
qht_insert(&tb_ctx.htable, tb, h, &existing_tb);
350
351
/* remove TB from the page(s) if we couldn't insert it */
352
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
353
354
gen_code_buf = tcg_ctx->code_gen_ptr;
355
tb->tc.ptr = tcg_splitwx_to_rx(gen_code_buf);
356
+#if !TARGET_TB_PCREL
357
tb->pc = pc;
358
+#endif
359
tb->cs_base = cs_base;
360
tb->flags = flags;
361
tb->cflags = cflags;
279
--
362
--
280
2.25.1
363
2.34.1
281
364
282
365
diff view generated by jsdifflib
Deleted patch
1
Migrate the is_jmp, tb and singlestep_enabled fields
2
from DisasContext into the base.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 49 +++++++++++++++++----------------
9
target/cris/translate_v10.c.inc | 4 +--
10
2 files changed, 27 insertions(+), 26 deletions(-)
11
12
diff --git a/target/cris/translate.c b/target/cris/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/cris/translate.c
15
+++ b/target/cris/translate.c
16
@@ -XXX,XX +XXX,XX @@ static TCGv env_pc;
17
18
/* This is the state at translation time. */
19
typedef struct DisasContext {
20
+ DisasContextBase base;
21
+
22
CRISCPU *cpu;
23
target_ulong pc, ppc;
24
25
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
26
int clear_locked_irq; /* Clear the irq lockout. */
27
int cpustate_changed;
28
unsigned int tb_flags; /* tb dependent flags. */
29
- int is_jmp;
30
31
#define JMP_NOJMP 0
32
#define JMP_DIRECT 1
33
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
34
uint32_t jmp_pc;
35
36
int delayed_branch;
37
-
38
- TranslationBlock *tb;
39
- int singlestep_enabled;
40
} DisasContext;
41
42
static void gen_BUG(DisasContext *dc, const char *file, int line)
43
@@ -XXX,XX +XXX,XX @@ static void t_gen_cc_jmp(TCGv pc_true, TCGv pc_false)
44
static inline bool use_goto_tb(DisasContext *dc, target_ulong dest)
45
{
46
#ifndef CONFIG_USER_ONLY
47
- return (dc->tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) ||
48
+ return (dc->base.pc_first & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) ||
49
(dc->ppc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
50
#else
51
return true;
52
@@ -XXX,XX +XXX,XX @@ static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
53
if (use_goto_tb(dc, dest)) {
54
tcg_gen_goto_tb(n);
55
tcg_gen_movi_tl(env_pc, dest);
56
- tcg_gen_exit_tb(dc->tb, n);
57
+ tcg_gen_exit_tb(dc->base.tb, n);
58
} else {
59
tcg_gen_movi_tl(env_pc, dest);
60
tcg_gen_exit_tb(NULL, 0);
61
@@ -XXX,XX +XXX,XX @@ static int dec_setclrf(CPUCRISState *env, DisasContext *dc)
62
/* Break the TB if any of the SPI flag changes. */
63
if (flags & (P_FLAG | S_FLAG)) {
64
tcg_gen_movi_tl(env_pc, dc->pc + 2);
65
- dc->is_jmp = DISAS_UPDATE;
66
+ dc->base.is_jmp = DISAS_UPDATE;
67
dc->cpustate_changed = 1;
68
}
69
70
/* For the I flag, only act on posedge. */
71
if ((flags & I_FLAG)) {
72
tcg_gen_movi_tl(env_pc, dc->pc + 2);
73
- dc->is_jmp = DISAS_UPDATE;
74
+ dc->base.is_jmp = DISAS_UPDATE;
75
dc->cpustate_changed = 1;
76
}
77
78
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
79
LOG_DIS("rfe\n");
80
cris_evaluate_flags(dc);
81
gen_helper_rfe(cpu_env);
82
- dc->is_jmp = DISAS_UPDATE;
83
+ dc->base.is_jmp = DISAS_UPDATE;
84
break;
85
case 5:
86
/* rfn. */
87
LOG_DIS("rfn\n");
88
cris_evaluate_flags(dc);
89
gen_helper_rfn(cpu_env);
90
- dc->is_jmp = DISAS_UPDATE;
91
+ dc->base.is_jmp = DISAS_UPDATE;
92
break;
93
case 6:
94
LOG_DIS("break %d\n", dc->op1);
95
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
96
/* Breaks start at 16 in the exception vector. */
97
t_gen_movi_env_TN(trap_vector, dc->op1 + 16);
98
t_gen_raise_exception(EXCP_BREAK);
99
- dc->is_jmp = DISAS_UPDATE;
100
+ dc->base.is_jmp = DISAS_UPDATE;
101
break;
102
default:
103
printf("op2=%x\n", dc->op2);
104
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
105
* delayslot, like in real hw.
106
*/
107
pc_start = tb->pc & ~1;
108
- dc->cpu = env_archcpu(env);
109
- dc->tb = tb;
110
111
- dc->is_jmp = DISAS_NEXT;
112
+ dc->base.tb = tb;
113
+ dc->base.pc_first = pc_start;
114
+ dc->base.pc_next = pc_start;
115
+ dc->base.is_jmp = DISAS_NEXT;
116
+ dc->base.singlestep_enabled = cs->singlestep_enabled;
117
+
118
+ dc->cpu = env_archcpu(env);
119
dc->ppc = pc_start;
120
dc->pc = pc_start;
121
- dc->singlestep_enabled = cs->singlestep_enabled;
122
dc->flags_uptodate = 1;
123
dc->flagx_known = 1;
124
dc->flags_x = tb->flags & X_FLAG;
125
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
126
cris_evaluate_flags(dc);
127
tcg_gen_movi_tl(env_pc, dc->pc);
128
t_gen_raise_exception(EXCP_DEBUG);
129
- dc->is_jmp = DISAS_UPDATE;
130
+ dc->base.is_jmp = DISAS_UPDATE;
131
/* The address covered by the breakpoint must be included in
132
[tb->pc, tb->pc + tb->size) in order to for it to be
133
properly cleared -- thus we increment the PC here so that
134
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
135
gen_goto_tb(dc, 1, dc->jmp_pc);
136
gen_set_label(l1);
137
gen_goto_tb(dc, 0, dc->pc);
138
- dc->is_jmp = DISAS_TB_JUMP;
139
+ dc->base.is_jmp = DISAS_TB_JUMP;
140
dc->jmp = JMP_NOJMP;
141
} else if (dc->jmp == JMP_DIRECT) {
142
cris_evaluate_flags(dc);
143
gen_goto_tb(dc, 0, dc->jmp_pc);
144
- dc->is_jmp = DISAS_TB_JUMP;
145
+ dc->base.is_jmp = DISAS_TB_JUMP;
146
dc->jmp = JMP_NOJMP;
147
} else {
148
TCGv c = tcg_const_tl(dc->pc);
149
t_gen_cc_jmp(env_btarget, c);
150
tcg_temp_free(c);
151
- dc->is_jmp = DISAS_JUMP;
152
+ dc->base.is_jmp = DISAS_JUMP;
153
}
154
break;
155
}
156
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
157
if (!(tb->pc & 1) && cs->singlestep_enabled) {
158
break;
159
}
160
- } while (!dc->is_jmp && !dc->cpustate_changed
161
+ } while (!dc->base.is_jmp && !dc->cpustate_changed
162
&& !tcg_op_buf_full()
163
&& !singlestep
164
&& (dc->pc - page_start < TARGET_PAGE_SIZE)
165
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
166
npc = dc->pc;
167
168
/* Force an update if the per-tb cpu state has changed. */
169
- if (dc->is_jmp == DISAS_NEXT
170
+ if (dc->base.is_jmp == DISAS_NEXT
171
&& (dc->cpustate_changed || !dc->flagx_known
172
|| (dc->flags_x != (tb->flags & X_FLAG)))) {
173
- dc->is_jmp = DISAS_UPDATE;
174
+ dc->base.is_jmp = DISAS_UPDATE;
175
tcg_gen_movi_tl(env_pc, npc);
176
}
177
/* Broken branch+delayslot sequence. */
178
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
179
cris_evaluate_flags(dc);
180
181
if (unlikely(cs->singlestep_enabled)) {
182
- if (dc->is_jmp == DISAS_NEXT) {
183
+ if (dc->base.is_jmp == DISAS_NEXT) {
184
tcg_gen_movi_tl(env_pc, npc);
185
}
186
t_gen_raise_exception(EXCP_DEBUG);
187
} else {
188
- switch (dc->is_jmp) {
189
+ switch (dc->base.is_jmp) {
190
case DISAS_NEXT:
191
gen_goto_tb(dc, 1, npc);
192
break;
193
diff --git a/target/cris/translate_v10.c.inc b/target/cris/translate_v10.c.inc
194
index XXXXXXX..XXXXXXX 100644
195
--- a/target/cris/translate_v10.c.inc
196
+++ b/target/cris/translate_v10.c.inc
197
@@ -XXX,XX +XXX,XX @@ static unsigned int dec10_ind(CPUCRISState *env, DisasContext *dc)
198
t_gen_mov_env_TN(trap_vector, c);
199
tcg_temp_free(c);
200
t_gen_raise_exception(EXCP_BREAK);
201
- dc->is_jmp = DISAS_UPDATE;
202
+ dc->base.is_jmp = DISAS_UPDATE;
203
return insn_len;
204
}
205
LOG_DIS("%d: jump.%d %d r%d r%d\n", __LINE__, size,
206
@@ -XXX,XX +XXX,XX @@ static unsigned int crisv10_decoder(CPUCRISState *env, DisasContext *dc)
207
if (dc->clear_prefix && dc->tb_flags & PFIX_FLAG) {
208
dc->tb_flags &= ~PFIX_FLAG;
209
tcg_gen_andi_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], ~PFIX_FLAG);
210
- if (dc->tb_flags != dc->tb->flags) {
211
+ if (dc->tb_flags != dc->base.tb->flags) {
212
dc->cpustate_changed = 1;
213
}
214
}
215
--
216
2.25.1
217
218
diff view generated by jsdifflib
Deleted patch
1
This value is unused.
2
1
3
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
4
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
target/cris/translate.c | 2 --
8
1 file changed, 2 deletions(-)
9
10
diff --git a/target/cris/translate.c b/target/cris/translate.c
11
index XXXXXXX..XXXXXXX 100644
12
--- a/target/cris/translate.c
13
+++ b/target/cris/translate.c
14
@@ -XXX,XX +XXX,XX @@
15
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
16
#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
17
#define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */
18
-#define DISAS_SWI DISAS_TARGET_3
19
20
/* Used by the decoder. */
21
#define EXTRACT_FIELD(src, start, end) \
22
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
23
to find the next TB */
24
tcg_gen_exit_tb(NULL, 0);
25
break;
26
- case DISAS_SWI:
27
case DISAS_TB_JUMP:
28
/* nothing more to generate */
29
break;
30
--
31
2.25.1
32
33
diff view generated by jsdifflib
Deleted patch
1
The only semantic of DISAS_TB_JUMP is that we've done goto_tb,
2
which is the same as DISAS_NORETURN -- we've exited the tb.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 7 +++----
9
1 file changed, 3 insertions(+), 4 deletions(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@
16
/* is_jmp field values */
17
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
18
#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
19
-#define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */
20
21
/* Used by the decoder. */
22
#define EXTRACT_FIELD(src, start, end) \
23
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
24
gen_goto_tb(dc, 1, dc->jmp_pc);
25
gen_set_label(l1);
26
gen_goto_tb(dc, 0, dc->pc);
27
- dc->base.is_jmp = DISAS_TB_JUMP;
28
+ dc->base.is_jmp = DISAS_NORETURN;
29
dc->jmp = JMP_NOJMP;
30
} else if (dc->jmp == JMP_DIRECT) {
31
cris_evaluate_flags(dc);
32
gen_goto_tb(dc, 0, dc->jmp_pc);
33
- dc->base.is_jmp = DISAS_TB_JUMP;
34
+ dc->base.is_jmp = DISAS_NORETURN;
35
dc->jmp = JMP_NOJMP;
36
} else {
37
TCGv c = tcg_const_tl(dc->pc);
38
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
39
to find the next TB */
40
tcg_gen_exit_tb(NULL, 0);
41
break;
42
- case DISAS_TB_JUMP:
43
+ case DISAS_NORETURN:
44
/* nothing more to generate */
45
break;
46
}
47
--
48
2.25.1
49
50
diff view generated by jsdifflib
Deleted patch
1
After we've raised the exception, we have left the TB.
2
1
3
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
4
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
target/cris/translate.c | 5 +++--
8
target/cris/translate_v10.c.inc | 3 ++-
9
2 files changed, 5 insertions(+), 3 deletions(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
16
-offsetof(CRISCPU, env) + offsetof(CPUState, halted));
17
tcg_gen_movi_tl(env_pc, dc->pc + 2);
18
t_gen_raise_exception(EXCP_HLT);
19
+ dc->base.is_jmp = DISAS_NORETURN;
20
return 2;
21
}
22
23
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
24
/* Breaks start at 16 in the exception vector. */
25
t_gen_movi_env_TN(trap_vector, dc->op1 + 16);
26
t_gen_raise_exception(EXCP_BREAK);
27
- dc->base.is_jmp = DISAS_UPDATE;
28
+ dc->base.is_jmp = DISAS_NORETURN;
29
break;
30
default:
31
printf("op2=%x\n", dc->op2);
32
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
33
cris_evaluate_flags(dc);
34
tcg_gen_movi_tl(env_pc, dc->pc);
35
t_gen_raise_exception(EXCP_DEBUG);
36
- dc->base.is_jmp = DISAS_UPDATE;
37
+ dc->base.is_jmp = DISAS_NORETURN;
38
/* The address covered by the breakpoint must be included in
39
[tb->pc, tb->pc + tb->size) in order to for it to be
40
properly cleared -- thus we increment the PC here so that
41
diff --git a/target/cris/translate_v10.c.inc b/target/cris/translate_v10.c.inc
42
index XXXXXXX..XXXXXXX 100644
43
--- a/target/cris/translate_v10.c.inc
44
+++ b/target/cris/translate_v10.c.inc
45
@@ -XXX,XX +XXX,XX @@ static inline void cris_illegal_insn(DisasContext *dc)
46
{
47
qemu_log_mask(LOG_GUEST_ERROR, "illegal insn at pc=%x\n", dc->pc);
48
t_gen_raise_exception(EXCP_BREAK);
49
+ dc->base.is_jmp = DISAS_NORETURN;
50
}
51
52
static void gen_store_v10_conditional(DisasContext *dc, TCGv addr, TCGv val,
53
@@ -XXX,XX +XXX,XX @@ static unsigned int dec10_ind(CPUCRISState *env, DisasContext *dc)
54
t_gen_mov_env_TN(trap_vector, c);
55
tcg_temp_free(c);
56
t_gen_raise_exception(EXCP_BREAK);
57
- dc->base.is_jmp = DISAS_UPDATE;
58
+ dc->base.is_jmp = DISAS_NORETURN;
59
return insn_len;
60
}
61
LOG_DIS("%d: jump.%d %d r%d r%d\n", __LINE__, size,
62
--
63
2.25.1
64
65
diff view generated by jsdifflib
Deleted patch
1
Do not skip the page check for user-only -- mmap/mprotect can
2
still change page mappings. Only check dc->base.pc_first, not
3
dc->ppc -- the start page is the only one that's relevant.
4
1
5
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
target/cris/translate.c | 9 ++-------
10
1 file changed, 2 insertions(+), 7 deletions(-)
11
12
diff --git a/target/cris/translate.c b/target/cris/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/cris/translate.c
15
+++ b/target/cris/translate.c
16
@@ -XXX,XX +XXX,XX @@ static void t_gen_cc_jmp(TCGv pc_true, TCGv pc_false)
17
gen_set_label(l1);
18
}
19
20
-static inline bool use_goto_tb(DisasContext *dc, target_ulong dest)
21
+static bool use_goto_tb(DisasContext *dc, target_ulong dest)
22
{
23
-#ifndef CONFIG_USER_ONLY
24
- return (dc->base.pc_first & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) ||
25
- (dc->ppc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
26
-#else
27
- return true;
28
-#endif
29
+ return ((dest ^ dc->base.pc_first) & TARGET_PAGE_MASK) == 0;
30
}
31
32
static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
33
--
34
2.25.1
35
36
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
2
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
5
target/cris/translate.c | 317 ++++++++++++++++++++++------------------
6
1 file changed, 174 insertions(+), 143 deletions(-)
7
1
8
diff --git a/target/cris/translate.c b/target/cris/translate.c
9
index XXXXXXX..XXXXXXX 100644
10
--- a/target/cris/translate.c
11
+++ b/target/cris/translate.c
12
@@ -XXX,XX +XXX,XX @@ static unsigned int crisv32_decoder(CPUCRISState *env, DisasContext *dc)
13
*
14
*/
15
16
-/* generate intermediate code for basic block 'tb'. */
17
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
18
+static void cris_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
19
{
20
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
21
CPUCRISState *env = cs->env_ptr;
22
+ uint32_t tb_flags = dc->base.tb->flags;
23
uint32_t pc_start;
24
- unsigned int insn_len;
25
- struct DisasContext ctx;
26
- struct DisasContext *dc = &ctx;
27
- uint32_t page_start;
28
- target_ulong npc;
29
- int num_insns;
30
31
if (env->pregs[PR_VR] == 32) {
32
dc->decoder = crisv32_decoder;
33
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
34
dc->clear_locked_irq = 1;
35
}
36
37
- /* Odd PC indicates that branch is rexecuting due to exception in the
38
+ /*
39
+ * Odd PC indicates that branch is rexecuting due to exception in the
40
* delayslot, like in real hw.
41
*/
42
- pc_start = tb->pc & ~1;
43
-
44
- dc->base.tb = tb;
45
+ pc_start = dc->base.pc_first & ~1;
46
dc->base.pc_first = pc_start;
47
dc->base.pc_next = pc_start;
48
- dc->base.is_jmp = DISAS_NEXT;
49
- dc->base.singlestep_enabled = cs->singlestep_enabled;
50
51
dc->cpu = env_archcpu(env);
52
dc->ppc = pc_start;
53
dc->pc = pc_start;
54
dc->flags_uptodate = 1;
55
dc->flagx_known = 1;
56
- dc->flags_x = tb->flags & X_FLAG;
57
+ dc->flags_x = tb_flags & X_FLAG;
58
dc->cc_x_uptodate = 0;
59
dc->cc_mask = 0;
60
dc->update_cc = 0;
61
dc->clear_prefix = 0;
62
+ dc->cpustate_changed = 0;
63
64
cris_update_cc_op(dc, CC_OP_FLAGS, 4);
65
dc->cc_size_uptodate = -1;
66
67
/* Decode TB flags. */
68
- dc->tb_flags = tb->flags & (S_FLAG | P_FLAG | U_FLAG \
69
- | X_FLAG | PFIX_FLAG);
70
- dc->delayed_branch = !!(tb->flags & 7);
71
+ dc->tb_flags = tb_flags & (S_FLAG | P_FLAG | U_FLAG | X_FLAG | PFIX_FLAG);
72
+ dc->delayed_branch = !!(tb_flags & 7);
73
if (dc->delayed_branch) {
74
dc->jmp = JMP_INDIRECT;
75
} else {
76
dc->jmp = JMP_NOJMP;
77
}
78
+}
79
80
- dc->cpustate_changed = 0;
81
+static void cris_tr_tb_start(DisasContextBase *db, CPUState *cpu)
82
+{
83
+}
84
85
- page_start = pc_start & TARGET_PAGE_MASK;
86
- num_insns = 0;
87
+static void cris_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
88
+{
89
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
90
91
- gen_tb_start(tb);
92
- do {
93
- tcg_gen_insn_start(dc->delayed_branch == 1
94
- ? dc->ppc | 1 : dc->pc);
95
- num_insns++;
96
+ tcg_gen_insn_start(dc->delayed_branch == 1 ? dc->ppc | 1 : dc->pc);
97
+}
98
99
- if (unlikely(cpu_breakpoint_test(cs, dc->pc, BP_ANY))) {
100
+static bool cris_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cpu,
101
+ const CPUBreakpoint *bp)
102
+{
103
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
104
+
105
+ cris_evaluate_flags(dc);
106
+ tcg_gen_movi_tl(env_pc, dc->pc);
107
+ t_gen_raise_exception(EXCP_DEBUG);
108
+ dc->base.is_jmp = DISAS_NORETURN;
109
+ /*
110
+ * The address covered by the breakpoint must be included in
111
+ * [tb->pc, tb->pc + tb->size) in order to for it to be
112
+ * properly cleared -- thus we increment the PC here so that
113
+ * the logic setting tb->size below does the right thing.
114
+ */
115
+ dc->pc += 2;
116
+ return true;
117
+}
118
+
119
+static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
120
+{
121
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
122
+ CPUCRISState *env = cs->env_ptr;
123
+ unsigned int insn_len;
124
+
125
+ /* Pretty disas. */
126
+ LOG_DIS("%8.8x:\t", dc->pc);
127
+
128
+ dc->clear_x = 1;
129
+
130
+ insn_len = dc->decoder(env, dc);
131
+ dc->ppc = dc->pc;
132
+ dc->pc += insn_len;
133
+ dc->base.pc_next += insn_len;
134
+
135
+ if (dc->base.is_jmp == DISAS_NORETURN) {
136
+ return;
137
+ }
138
+
139
+ if (dc->clear_x) {
140
+ cris_clear_x_flag(dc);
141
+ }
142
+
143
+ /*
144
+ * Check for delayed branches here. If we do it before
145
+ * actually generating any host code, the simulator will just
146
+ * loop doing nothing for on this program location.
147
+ */
148
+ if (dc->delayed_branch && --dc->delayed_branch == 0) {
149
+ if (dc->base.tb->flags & 7) {
150
+ t_gen_movi_env_TN(dslot, 0);
151
+ }
152
+
153
+ if (dc->cpustate_changed
154
+ || !dc->flagx_known
155
+ || (dc->flags_x != (dc->base.tb->flags & X_FLAG))) {
156
+ cris_store_direct_jmp(dc);
157
+ }
158
+
159
+ if (dc->clear_locked_irq) {
160
+ dc->clear_locked_irq = 0;
161
+ t_gen_movi_env_TN(locked_irq, 0);
162
+ }
163
+
164
+ if (dc->jmp == JMP_DIRECT_CC) {
165
+ TCGLabel *l1 = gen_new_label();
166
cris_evaluate_flags(dc);
167
- tcg_gen_movi_tl(env_pc, dc->pc);
168
- t_gen_raise_exception(EXCP_DEBUG);
169
+
170
+ /* Conditional jmp. */
171
+ tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, l1);
172
+ gen_goto_tb(dc, 1, dc->jmp_pc);
173
+ gen_set_label(l1);
174
+ gen_goto_tb(dc, 0, dc->pc);
175
dc->base.is_jmp = DISAS_NORETURN;
176
- /* The address covered by the breakpoint must be included in
177
- [tb->pc, tb->pc + tb->size) in order to for it to be
178
- properly cleared -- thus we increment the PC here so that
179
- the logic setting tb->size below does the right thing. */
180
- dc->pc += 2;
181
- break;
182
+ dc->jmp = JMP_NOJMP;
183
+ } else if (dc->jmp == JMP_DIRECT) {
184
+ cris_evaluate_flags(dc);
185
+ gen_goto_tb(dc, 0, dc->jmp_pc);
186
+ dc->base.is_jmp = DISAS_NORETURN;
187
+ dc->jmp = JMP_NOJMP;
188
+ } else {
189
+ TCGv c = tcg_const_tl(dc->pc);
190
+ t_gen_cc_jmp(env_btarget, c);
191
+ tcg_temp_free(c);
192
+ dc->base.is_jmp = DISAS_JUMP;
193
}
194
+ }
195
196
- /* Pretty disas. */
197
- LOG_DIS("%8.8x:\t", dc->pc);
198
+ /* Force an update if the per-tb cpu state has changed. */
199
+ if (dc->base.is_jmp == DISAS_NEXT
200
+ && (dc->cpustate_changed
201
+ || !dc->flagx_known
202
+ || (dc->flags_x != (dc->base.tb->flags & X_FLAG)))) {
203
+ dc->base.is_jmp = DISAS_UPDATE;
204
+ tcg_gen_movi_tl(env_pc, dc->pc);
205
+ }
206
207
- if (num_insns == max_insns && (tb_cflags(tb) & CF_LAST_IO)) {
208
- gen_io_start();
209
- }
210
- dc->clear_x = 1;
211
+ /*
212
+ * FIXME: Only the first insn in the TB should cross a page boundary.
213
+ * If we can detect the length of the next insn easily, we should.
214
+ * In the meantime, simply stop when we do cross.
215
+ */
216
+ if (dc->base.is_jmp == DISAS_NEXT
217
+ && ((dc->pc ^ dc->base.pc_first) & TARGET_PAGE_MASK) != 0) {
218
+ dc->base.is_jmp = DISAS_TOO_MANY;
219
+ }
220
+}
221
222
- insn_len = dc->decoder(env, dc);
223
- dc->ppc = dc->pc;
224
- dc->pc += insn_len;
225
- if (dc->clear_x) {
226
- cris_clear_x_flag(dc);
227
- }
228
+static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
229
+{
230
+ DisasContext *dc = container_of(dcbase, DisasContext, base);
231
+ DisasJumpType is_jmp = dc->base.is_jmp;
232
+ target_ulong npc = dc->pc;
233
234
- /* Check for delayed branches here. If we do it before
235
- actually generating any host code, the simulator will just
236
- loop doing nothing for on this program location. */
237
- if (dc->delayed_branch) {
238
- dc->delayed_branch--;
239
- if (dc->delayed_branch == 0) {
240
- if (tb->flags & 7) {
241
- t_gen_movi_env_TN(dslot, 0);
242
- }
243
- if (dc->cpustate_changed || !dc->flagx_known
244
- || (dc->flags_x != (tb->flags & X_FLAG))) {
245
- cris_store_direct_jmp(dc);
246
- }
247
-
248
- if (dc->clear_locked_irq) {
249
- dc->clear_locked_irq = 0;
250
- t_gen_movi_env_TN(locked_irq, 0);
251
- }
252
-
253
- if (dc->jmp == JMP_DIRECT_CC) {
254
- TCGLabel *l1 = gen_new_label();
255
- cris_evaluate_flags(dc);
256
-
257
- /* Conditional jmp. */
258
- tcg_gen_brcondi_tl(TCG_COND_EQ,
259
- env_btaken, 0, l1);
260
- gen_goto_tb(dc, 1, dc->jmp_pc);
261
- gen_set_label(l1);
262
- gen_goto_tb(dc, 0, dc->pc);
263
- dc->base.is_jmp = DISAS_NORETURN;
264
- dc->jmp = JMP_NOJMP;
265
- } else if (dc->jmp == JMP_DIRECT) {
266
- cris_evaluate_flags(dc);
267
- gen_goto_tb(dc, 0, dc->jmp_pc);
268
- dc->base.is_jmp = DISAS_NORETURN;
269
- dc->jmp = JMP_NOJMP;
270
- } else {
271
- TCGv c = tcg_const_tl(dc->pc);
272
- t_gen_cc_jmp(env_btarget, c);
273
- tcg_temp_free(c);
274
- dc->base.is_jmp = DISAS_JUMP;
275
- }
276
- break;
277
- }
278
- }
279
-
280
- /* If we are rexecuting a branch due to exceptions on
281
- delay slots don't break. */
282
- if (!(tb->pc & 1) && cs->singlestep_enabled) {
283
- break;
284
- }
285
- } while (!dc->base.is_jmp && !dc->cpustate_changed
286
- && !tcg_op_buf_full()
287
- && !singlestep
288
- && (dc->pc - page_start < TARGET_PAGE_SIZE)
289
- && num_insns < max_insns);
290
+ if (is_jmp == DISAS_NORETURN) {
291
+ /* If we have a broken branch+delayslot sequence, it's too late. */
292
+ assert(dc->delayed_branch != 1);
293
+ return;
294
+ }
295
296
if (dc->clear_locked_irq) {
297
t_gen_movi_env_TN(locked_irq, 0);
298
}
299
300
- npc = dc->pc;
301
-
302
- /* Force an update if the per-tb cpu state has changed. */
303
- if (dc->base.is_jmp == DISAS_NEXT
304
- && (dc->cpustate_changed || !dc->flagx_known
305
- || (dc->flags_x != (tb->flags & X_FLAG)))) {
306
- dc->base.is_jmp = DISAS_UPDATE;
307
- tcg_gen_movi_tl(env_pc, npc);
308
- }
309
/* Broken branch+delayslot sequence. */
310
if (dc->delayed_branch == 1) {
311
/* Set env->dslot to the size of the branch insn. */
312
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
313
314
cris_evaluate_flags(dc);
315
316
- if (unlikely(cs->singlestep_enabled)) {
317
- if (dc->base.is_jmp == DISAS_NEXT) {
318
+ if (unlikely(dc->base.singlestep_enabled)) {
319
+ switch (is_jmp) {
320
+ case DISAS_TOO_MANY:
321
tcg_gen_movi_tl(env_pc, npc);
322
- }
323
- t_gen_raise_exception(EXCP_DEBUG);
324
- } else {
325
- switch (dc->base.is_jmp) {
326
- case DISAS_NEXT:
327
- gen_goto_tb(dc, 1, npc);
328
- break;
329
- default:
330
+ /* fall through */
331
case DISAS_JUMP:
332
case DISAS_UPDATE:
333
- /* indicate that the hash table must be used
334
- to find the next TB */
335
- tcg_gen_exit_tb(NULL, 0);
336
- break;
337
- case DISAS_NORETURN:
338
- /* nothing more to generate */
339
+ t_gen_raise_exception(EXCP_DEBUG);
340
+ return;
341
+ default:
342
break;
343
}
344
+ g_assert_not_reached();
345
}
346
- gen_tb_end(tb, num_insns);
347
348
- tb->size = dc->pc - pc_start;
349
- tb->icount = num_insns;
350
-
351
-#ifdef DEBUG_DISAS
352
-#if !DISAS_CRIS
353
- if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
354
- && qemu_log_in_addr_range(pc_start)) {
355
- FILE *logfile = qemu_log_lock();
356
- qemu_log("--------------\n");
357
- qemu_log("IN: %s\n", lookup_symbol(pc_start));
358
- log_target_disas(cs, pc_start, dc->pc - pc_start);
359
- qemu_log_unlock(logfile);
360
+ switch (is_jmp) {
361
+ case DISAS_TOO_MANY:
362
+ gen_goto_tb(dc, 0, npc);
363
+ break;
364
+ case DISAS_JUMP:
365
+ case DISAS_UPDATE:
366
+ /* Indicate that interupts must be re-evaluated before the next TB. */
367
+ tcg_gen_exit_tb(NULL, 0);
368
+ break;
369
+ default:
370
+ g_assert_not_reached();
371
}
372
-#endif
373
-#endif
374
+}
375
+
376
+static void cris_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
377
+{
378
+ if (!DISAS_CRIS) {
379
+ qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
380
+ log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size);
381
+ }
382
+}
383
+
384
+static const TranslatorOps cris_tr_ops = {
385
+ .init_disas_context = cris_tr_init_disas_context,
386
+ .tb_start = cris_tr_tb_start,
387
+ .insn_start = cris_tr_insn_start,
388
+ .breakpoint_check = cris_tr_breakpoint_check,
389
+ .translate_insn = cris_tr_translate_insn,
390
+ .tb_stop = cris_tr_tb_stop,
391
+ .disas_log = cris_tr_disas_log,
392
+};
393
+
394
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
395
+{
396
+ DisasContext dc;
397
+ translator_loop(&cris_tr_ops, &dc.base, cs, tb, max_insns);
398
}
399
400
void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
401
--
402
2.25.1
403
404
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
2
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
5
target/cris/helper.h | 2 +-
6
1 file changed, 1 insertion(+), 1 deletion(-)
7
1
8
diff --git a/target/cris/helper.h b/target/cris/helper.h
9
index XXXXXXX..XXXXXXX 100644
10
--- a/target/cris/helper.h
11
+++ b/target/cris/helper.h
12
@@ -XXX,XX +XXX,XX @@
13
-DEF_HELPER_2(raise_exception, void, env, i32)
14
+DEF_HELPER_2(raise_exception, noreturn, env, i32)
15
DEF_HELPER_2(tlb_flush_pid, void, env, i32)
16
DEF_HELPER_2(spc_write, void, env, i32)
17
DEF_HELPER_1(rfe, void, env)
18
--
19
2.25.1
20
21
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
2
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
5
target/cris/translate.c | 19 ++++++++++---------
6
target/cris/translate_v10.c.inc | 6 +++---
7
2 files changed, 13 insertions(+), 12 deletions(-)
8
1
9
diff --git a/target/cris/translate.c b/target/cris/translate.c
10
index XXXXXXX..XXXXXXX 100644
11
--- a/target/cris/translate.c
12
+++ b/target/cris/translate.c
13
@@ -XXX,XX +XXX,XX @@ static void gen_BUG(DisasContext *dc, const char *file, int line)
14
cpu_abort(CPU(dc->cpu), "%s:%d pc=%x\n", file, line, dc->pc);
15
}
16
17
-static const char *regnames_v32[] =
18
+static const char * const regnames_v32[] =
19
{
20
"$r0", "$r1", "$r2", "$r3",
21
"$r4", "$r5", "$r6", "$r7",
22
"$r8", "$r9", "$r10", "$r11",
23
"$r12", "$r13", "$sp", "$acr",
24
};
25
-static const char *pregnames_v32[] =
26
+
27
+static const char * const pregnames_v32[] =
28
{
29
"$bz", "$vr", "$pid", "$srs",
30
"$wz", "$exs", "$eda", "$mof",
31
@@ -XXX,XX +XXX,XX @@ static const char *pregnames_v32[] =
32
};
33
34
/* We need this table to handle preg-moves with implicit width. */
35
-static int preg_sizes[] = {
36
+static const int preg_sizes[] = {
37
1, /* bz. */
38
1, /* vr. */
39
4, /* pid. */
40
@@ -XXX,XX +XXX,XX @@ static inline void t_gen_swapw(TCGv d, TCGv s)
41
((T0 >> 5) & 0x02020202) |
42
((T0 >> 7) & 0x01010101));
43
*/
44
-static inline void t_gen_swapr(TCGv d, TCGv s)
45
+static void t_gen_swapr(TCGv d, TCGv s)
46
{
47
- struct {
48
+ static const struct {
49
int shift; /* LSL when positive, LSR when negative. */
50
uint32_t mask;
51
} bitrev[] = {
52
@@ -XXX,XX +XXX,XX @@ static int dec_prep_alu_m(CPUCRISState *env, DisasContext *dc,
53
#if DISAS_CRIS
54
static const char *cc_name(int cc)
55
{
56
- static const char *cc_names[16] = {
57
+ static const char * const cc_names[16] = {
58
"cc", "cs", "ne", "eq", "vc", "vs", "pl", "mi",
59
"ls", "hi", "ge", "lt", "gt", "le", "a", "p"
60
};
61
@@ -XXX,XX +XXX,XX @@ static int dec_null(CPUCRISState *env, DisasContext *dc)
62
return 2;
63
}
64
65
-static struct decoder_info {
66
+static const struct decoder_info {
67
struct {
68
uint32_t bits;
69
uint32_t mask;
70
@@ -XXX,XX +XXX,XX @@ void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
71
{
72
CRISCPU *cpu = CRIS_CPU(cs);
73
CPUCRISState *env = &cpu->env;
74
- const char **regnames;
75
- const char **pregnames;
76
+ const char * const *regnames;
77
+ const char * const *pregnames;
78
int i;
79
80
if (!env) {
81
diff --git a/target/cris/translate_v10.c.inc b/target/cris/translate_v10.c.inc
82
index XXXXXXX..XXXXXXX 100644
83
--- a/target/cris/translate_v10.c.inc
84
+++ b/target/cris/translate_v10.c.inc
85
@@ -XXX,XX +XXX,XX @@
86
#include "qemu/osdep.h"
87
#include "crisv10-decode.h"
88
89
-static const char *regnames_v10[] =
90
+static const char * const regnames_v10[] =
91
{
92
"$r0", "$r1", "$r2", "$r3",
93
"$r4", "$r5", "$r6", "$r7",
94
@@ -XXX,XX +XXX,XX @@ static const char *regnames_v10[] =
95
"$r12", "$r13", "$sp", "$pc",
96
};
97
98
-static const char *pregnames_v10[] =
99
+static const char * const pregnames_v10[] =
100
{
101
"$bz", "$vr", "$p2", "$p3",
102
"$wz", "$ccr", "$p6-prefix", "$mof",
103
@@ -XXX,XX +XXX,XX @@ static const char *pregnames_v10[] =
104
};
105
106
/* We need this table to handle preg-moves with implicit width. */
107
-static int preg_sizes_v10[] = {
108
+static const int preg_sizes_v10[] = {
109
1, /* bz. */
110
1, /* vr. */
111
1, /* pid. */
112
--
113
2.25.1
114
115
diff view generated by jsdifflib
Deleted patch
1
We really do this already, by including them into the same test.
2
This just hoists the expression up a bit.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 13 ++++++-------
9
1 file changed, 6 insertions(+), 7 deletions(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
16
cris_clear_x_flag(dc);
17
}
18
19
+ /* Fold unhandled changes to X_FLAG into cpustate_changed. */
20
+ dc->cpustate_changed |= !dc->flagx_known;
21
+ dc->cpustate_changed |= dc->flags_x != (dc->base.tb->flags & X_FLAG);
22
+
23
/*
24
* Check for delayed branches here. If we do it before
25
* actually generating any host code, the simulator will just
26
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
27
t_gen_movi_env_TN(dslot, 0);
28
}
29
30
- if (dc->cpustate_changed
31
- || !dc->flagx_known
32
- || (dc->flags_x != (dc->base.tb->flags & X_FLAG))) {
33
+ if (dc->cpustate_changed) {
34
cris_store_direct_jmp(dc);
35
}
36
37
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
38
}
39
40
/* Force an update if the per-tb cpu state has changed. */
41
- if (dc->base.is_jmp == DISAS_NEXT
42
- && (dc->cpustate_changed
43
- || !dc->flagx_known
44
- || (dc->flags_x != (dc->base.tb->flags & X_FLAG)))) {
45
+ if (dc->base.is_jmp == DISAS_NEXT && dc->cpustate_changed) {
46
dc->base.is_jmp = DISAS_UPDATE;
47
tcg_gen_movi_tl(env_pc, dc->pc);
48
}
49
--
50
2.25.1
51
52
diff view generated by jsdifflib
Deleted patch
1
These insns set DISAS_UPDATE without cpustate_changed,
2
which isn't quite right.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 2 ++
9
1 file changed, 2 insertions(+)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
16
cris_evaluate_flags(dc);
17
gen_helper_rfe(cpu_env);
18
dc->base.is_jmp = DISAS_UPDATE;
19
+ dc->cpustate_changed = true;
20
break;
21
case 5:
22
/* rfn. */
23
@@ -XXX,XX +XXX,XX @@ static int dec_rfe_etc(CPUCRISState *env, DisasContext *dc)
24
cris_evaluate_flags(dc);
25
gen_helper_rfn(cpu_env);
26
dc->base.is_jmp = DISAS_UPDATE;
27
+ dc->cpustate_changed = true;
28
break;
29
case 6:
30
LOG_DIS("break %d\n", dc->op1);
31
--
32
2.25.1
33
34
diff view generated by jsdifflib
Deleted patch
1
Move this pc update into tb_stop.
2
We will be able to re-use this code shortly.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 20 +++++++++++++++-----
9
1 file changed, 15 insertions(+), 5 deletions(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@
16
#define BUG() (gen_BUG(dc, __FILE__, __LINE__))
17
#define BUG_ON(x) ({if (x) BUG();})
18
19
-/* is_jmp field values */
20
-#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
21
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
22
+/*
23
+ * Target-specific is_jmp field values
24
+ */
25
+/* Only pc was modified dynamically */
26
+#define DISAS_JUMP DISAS_TARGET_0
27
+/* Cpu state was modified dynamically, including pc */
28
+#define DISAS_UPDATE DISAS_TARGET_1
29
+/* Cpu state was modified dynamically, excluding pc -- use npc */
30
+#define DISAS_UPDATE_NEXT DISAS_TARGET_2
31
32
/* Used by the decoder. */
33
#define EXTRACT_FIELD(src, start, end) \
34
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
35
36
/* Force an update if the per-tb cpu state has changed. */
37
if (dc->base.is_jmp == DISAS_NEXT && dc->cpustate_changed) {
38
- dc->base.is_jmp = DISAS_UPDATE;
39
- tcg_gen_movi_tl(env_pc, dc->pc);
40
+ dc->base.is_jmp = DISAS_UPDATE_NEXT;
41
+ return;
42
}
43
44
/*
45
@@ -XXX,XX +XXX,XX @@ static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
46
if (unlikely(dc->base.singlestep_enabled)) {
47
switch (is_jmp) {
48
case DISAS_TOO_MANY:
49
+ case DISAS_UPDATE_NEXT:
50
tcg_gen_movi_tl(env_pc, npc);
51
/* fall through */
52
case DISAS_JUMP:
53
@@ -XXX,XX +XXX,XX @@ static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
54
case DISAS_TOO_MANY:
55
gen_goto_tb(dc, 0, npc);
56
break;
57
+ case DISAS_UPDATE_NEXT:
58
+ tcg_gen_movi_tl(env_pc, npc);
59
+ /* fall through */
60
case DISAS_JUMP:
61
case DISAS_UPDATE:
62
/* Indicate that interupts must be re-evaluated before the next TB. */
63
--
64
2.25.1
65
66
diff view generated by jsdifflib
Deleted patch
1
Move delayed branch handling to tb_stop, where we can re-use other
2
end-of-tb code, e.g. the evaluation of flags. Honor single stepping.
3
Validate that we aren't losing state by overwriting is_jmp.
4
1
5
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
target/cris/translate.c | 96 ++++++++++++++++++++++++-----------------
10
1 file changed, 56 insertions(+), 40 deletions(-)
11
12
diff --git a/target/cris/translate.c b/target/cris/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/cris/translate.c
15
+++ b/target/cris/translate.c
16
@@ -XXX,XX +XXX,XX @@
17
#define DISAS_UPDATE DISAS_TARGET_1
18
/* Cpu state was modified dynamically, excluding pc -- use npc */
19
#define DISAS_UPDATE_NEXT DISAS_TARGET_2
20
+/* PC update for delayed branch, see cpustate_changed otherwise */
21
+#define DISAS_DBRANCH DISAS_TARGET_3
22
23
/* Used by the decoder. */
24
#define EXTRACT_FIELD(src, start, end) \
25
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
26
dc->cpustate_changed |= dc->flags_x != (dc->base.tb->flags & X_FLAG);
27
28
/*
29
- * Check for delayed branches here. If we do it before
30
- * actually generating any host code, the simulator will just
31
- * loop doing nothing for on this program location.
32
+ * All branches are delayed branches, handled immediately below.
33
+ * We don't expect to see odd combinations of exit conditions.
34
*/
35
+ assert(dc->base.is_jmp == DISAS_NEXT || dc->cpustate_changed);
36
+
37
if (dc->delayed_branch && --dc->delayed_branch == 0) {
38
- if (dc->base.tb->flags & 7) {
39
- t_gen_movi_env_TN(dslot, 0);
40
- }
41
+ dc->base.is_jmp = DISAS_DBRANCH;
42
+ return;
43
+ }
44
45
- if (dc->cpustate_changed) {
46
- cris_store_direct_jmp(dc);
47
- }
48
-
49
- if (dc->clear_locked_irq) {
50
- dc->clear_locked_irq = 0;
51
- t_gen_movi_env_TN(locked_irq, 0);
52
- }
53
-
54
- if (dc->jmp == JMP_DIRECT_CC) {
55
- TCGLabel *l1 = gen_new_label();
56
- cris_evaluate_flags(dc);
57
-
58
- /* Conditional jmp. */
59
- tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, l1);
60
- gen_goto_tb(dc, 1, dc->jmp_pc);
61
- gen_set_label(l1);
62
- gen_goto_tb(dc, 0, dc->pc);
63
- dc->base.is_jmp = DISAS_NORETURN;
64
- dc->jmp = JMP_NOJMP;
65
- } else if (dc->jmp == JMP_DIRECT) {
66
- cris_evaluate_flags(dc);
67
- gen_goto_tb(dc, 0, dc->jmp_pc);
68
- dc->base.is_jmp = DISAS_NORETURN;
69
- dc->jmp = JMP_NOJMP;
70
- } else {
71
- TCGv c = tcg_const_tl(dc->pc);
72
- t_gen_cc_jmp(env_btarget, c);
73
- tcg_temp_free(c);
74
- dc->base.is_jmp = DISAS_JUMP;
75
- }
76
+ if (dc->base.is_jmp != DISAS_NEXT) {
77
+ return;
78
}
79
80
/* Force an update if the per-tb cpu state has changed. */
81
- if (dc->base.is_jmp == DISAS_NEXT && dc->cpustate_changed) {
82
+ if (dc->cpustate_changed) {
83
dc->base.is_jmp = DISAS_UPDATE_NEXT;
84
return;
85
}
86
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
87
* If we can detect the length of the next insn easily, we should.
88
* In the meantime, simply stop when we do cross.
89
*/
90
- if (dc->base.is_jmp == DISAS_NEXT
91
- && ((dc->pc ^ dc->base.pc_first) & TARGET_PAGE_MASK) != 0) {
92
+ if ((dc->pc ^ dc->base.pc_first) & TARGET_PAGE_MASK) {
93
dc->base.is_jmp = DISAS_TOO_MANY;
94
}
95
}
96
@@ -XXX,XX +XXX,XX @@ static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
97
98
cris_evaluate_flags(dc);
99
100
+ /* Evaluate delayed branch destination and fold to another is_jmp case. */
101
+ if (is_jmp == DISAS_DBRANCH) {
102
+ if (dc->base.tb->flags & 7) {
103
+ t_gen_movi_env_TN(dslot, 0);
104
+ }
105
+
106
+ switch (dc->jmp) {
107
+ case JMP_DIRECT:
108
+ npc = dc->jmp_pc;
109
+ is_jmp = dc->cpustate_changed ? DISAS_UPDATE_NEXT : DISAS_TOO_MANY;
110
+ break;
111
+
112
+ case JMP_DIRECT_CC:
113
+ /*
114
+ * Use a conditional branch if either taken or not-taken path
115
+ * can use goto_tb. If neither can, then treat it as indirect.
116
+ */
117
+ if (likely(!dc->base.singlestep_enabled)
118
+ && likely(!dc->cpustate_changed)
119
+ && (use_goto_tb(dc, dc->jmp_pc) || use_goto_tb(dc, npc))) {
120
+ TCGLabel *not_taken = gen_new_label();
121
+
122
+ tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, not_taken);
123
+ gen_goto_tb(dc, 1, dc->jmp_pc);
124
+ gen_set_label(not_taken);
125
+
126
+ /* not-taken case handled below. */
127
+ is_jmp = DISAS_TOO_MANY;
128
+ break;
129
+ }
130
+ tcg_gen_movi_tl(env_btarget, dc->jmp_pc);
131
+ /* fall through */
132
+
133
+ case JMP_INDIRECT:
134
+ t_gen_cc_jmp(env_btarget, tcg_constant_tl(npc));
135
+ is_jmp = dc->cpustate_changed ? DISAS_UPDATE : DISAS_JUMP;
136
+ break;
137
+
138
+ default:
139
+ g_assert_not_reached();
140
+ }
141
+ }
142
+
143
if (unlikely(dc->base.singlestep_enabled)) {
144
switch (is_jmp) {
145
case DISAS_TOO_MANY:
146
--
147
2.25.1
148
149
diff view generated by jsdifflib
Deleted patch
1
We can use this in gen_goto_tb and for DISAS_JUMP
2
to indirectly chain to the next TB.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 4 +++-
9
1 file changed, 3 insertions(+), 1 deletion(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@ static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
16
tcg_gen_exit_tb(dc->base.tb, n);
17
} else {
18
tcg_gen_movi_tl(env_pc, dest);
19
- tcg_gen_exit_tb(NULL, 0);
20
+ tcg_gen_lookup_and_goto_ptr();
21
}
22
}
23
24
@@ -XXX,XX +XXX,XX @@ static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
25
tcg_gen_movi_tl(env_pc, npc);
26
/* fall through */
27
case DISAS_JUMP:
28
+ tcg_gen_lookup_and_goto_ptr();
29
+ break;
30
case DISAS_UPDATE:
31
/* Indicate that interupts must be re-evaluated before the next TB. */
32
tcg_gen_exit_tb(NULL, 0);
33
--
34
2.25.1
35
36
diff view generated by jsdifflib
Deleted patch
1
Use movcond instead of brcond to set env_pc.
2
Discard the btarget and btaken variables to improve
3
register allocation and avoid unnecessary writeback.
4
1
5
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
target/cris/translate.c | 22 ++++++++++------------
10
1 file changed, 10 insertions(+), 12 deletions(-)
11
12
diff --git a/target/cris/translate.c b/target/cris/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/cris/translate.c
15
+++ b/target/cris/translate.c
16
@@ -XXX,XX +XXX,XX @@ static void t_gen_swapr(TCGv d, TCGv s)
17
tcg_temp_free(org_s);
18
}
19
20
-static void t_gen_cc_jmp(TCGv pc_true, TCGv pc_false)
21
-{
22
- TCGLabel *l1 = gen_new_label();
23
-
24
- /* Conditional jmp. */
25
- tcg_gen_mov_tl(env_pc, pc_false);
26
- tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, l1);
27
- tcg_gen_mov_tl(env_pc, pc_true);
28
- gen_set_label(l1);
29
-}
30
-
31
static bool use_goto_tb(DisasContext *dc, target_ulong dest)
32
{
33
return ((dest ^ dc->base.pc_first) & TARGET_PAGE_MASK) == 0;
34
@@ -XXX,XX +XXX,XX @@ static void cris_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
35
/* fall through */
36
37
case JMP_INDIRECT:
38
- t_gen_cc_jmp(env_btarget, tcg_constant_tl(npc));
39
+ tcg_gen_movcond_tl(TCG_COND_NE, env_pc,
40
+ env_btaken, tcg_constant_tl(0),
41
+ env_btarget, tcg_constant_tl(npc));
42
is_jmp = dc->cpustate_changed ? DISAS_UPDATE : DISAS_JUMP;
43
+
44
+ /*
45
+ * We have now consumed btaken and btarget. Hint to the
46
+ * tcg compiler that the writeback to env may be dropped.
47
+ */
48
+ tcg_gen_discard_tl(env_btaken);
49
+ tcg_gen_discard_tl(env_btarget);
50
break;
51
52
default:
53
--
54
2.25.1
55
56
diff view generated by jsdifflib
Deleted patch
1
Ever since 2a44f7f17364, flagx_known is always true.
2
Fold away all of the tests against the flag.
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 99 ++++++++-------------------------
9
target/cris/translate_v10.c.inc | 6 +-
10
2 files changed, 24 insertions(+), 81 deletions(-)
11
12
diff --git a/target/cris/translate.c b/target/cris/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/cris/translate.c
15
+++ b/target/cris/translate.c
16
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
17
18
int cc_x_uptodate; /* 1 - ccs, 2 - known | X_FLAG. 0 not up-to-date. */
19
int flags_uptodate; /* Whether or not $ccs is up-to-date. */
20
- int flagx_known; /* Whether or not flags_x has the x flag known at
21
- translation time. */
22
int flags_x;
23
24
int clear_x; /* Clear x after this insn? */
25
@@ -XXX,XX +XXX,XX @@ static inline void t_gen_add_flag(TCGv d, int flag)
26
27
static inline void t_gen_addx_carry(DisasContext *dc, TCGv d)
28
{
29
- if (dc->flagx_known) {
30
- if (dc->flags_x) {
31
- TCGv c;
32
-
33
- c = tcg_temp_new();
34
- t_gen_mov_TN_preg(c, PR_CCS);
35
- /* C flag is already at bit 0. */
36
- tcg_gen_andi_tl(c, c, C_FLAG);
37
- tcg_gen_add_tl(d, d, c);
38
- tcg_temp_free(c);
39
- }
40
- } else {
41
- TCGv x, c;
42
+ if (dc->flags_x) {
43
+ TCGv c = tcg_temp_new();
44
45
- x = tcg_temp_new();
46
- c = tcg_temp_new();
47
- t_gen_mov_TN_preg(x, PR_CCS);
48
- tcg_gen_mov_tl(c, x);
49
-
50
- /* Propagate carry into d if X is set. Branch free. */
51
+ t_gen_mov_TN_preg(c, PR_CCS);
52
+ /* C flag is already at bit 0. */
53
tcg_gen_andi_tl(c, c, C_FLAG);
54
- tcg_gen_andi_tl(x, x, X_FLAG);
55
- tcg_gen_shri_tl(x, x, 4);
56
-
57
- tcg_gen_and_tl(x, x, c);
58
- tcg_gen_add_tl(d, d, x);
59
- tcg_temp_free(x);
60
+ tcg_gen_add_tl(d, d, c);
61
tcg_temp_free(c);
62
}
63
}
64
65
static inline void t_gen_subx_carry(DisasContext *dc, TCGv d)
66
{
67
- if (dc->flagx_known) {
68
- if (dc->flags_x) {
69
- TCGv c;
70
-
71
- c = tcg_temp_new();
72
- t_gen_mov_TN_preg(c, PR_CCS);
73
- /* C flag is already at bit 0. */
74
- tcg_gen_andi_tl(c, c, C_FLAG);
75
- tcg_gen_sub_tl(d, d, c);
76
- tcg_temp_free(c);
77
- }
78
- } else {
79
- TCGv x, c;
80
+ if (dc->flags_x) {
81
+ TCGv c = tcg_temp_new();
82
83
- x = tcg_temp_new();
84
- c = tcg_temp_new();
85
- t_gen_mov_TN_preg(x, PR_CCS);
86
- tcg_gen_mov_tl(c, x);
87
-
88
- /* Propagate carry into d if X is set. Branch free. */
89
+ t_gen_mov_TN_preg(c, PR_CCS);
90
+ /* C flag is already at bit 0. */
91
tcg_gen_andi_tl(c, c, C_FLAG);
92
- tcg_gen_andi_tl(x, x, X_FLAG);
93
- tcg_gen_shri_tl(x, x, 4);
94
-
95
- tcg_gen_and_tl(x, x, c);
96
- tcg_gen_sub_tl(d, d, x);
97
- tcg_temp_free(x);
98
+ tcg_gen_sub_tl(d, d, c);
99
tcg_temp_free(c);
100
}
101
}
102
@@ -XXX,XX +XXX,XX @@ static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
103
104
static inline void cris_clear_x_flag(DisasContext *dc)
105
{
106
- if (dc->flagx_known && dc->flags_x) {
107
+ if (dc->flags_x) {
108
dc->flags_uptodate = 0;
109
}
110
-
111
- dc->flagx_known = 1;
112
dc->flags_x = 0;
113
}
114
115
@@ -XXX,XX +XXX,XX @@ static void cris_evaluate_flags(DisasContext *dc)
116
break;
117
}
118
119
- if (dc->flagx_known) {
120
- if (dc->flags_x) {
121
- tcg_gen_ori_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], X_FLAG);
122
- } else if (dc->cc_op == CC_OP_FLAGS) {
123
- tcg_gen_andi_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], ~X_FLAG);
124
- }
125
+ if (dc->flags_x) {
126
+ tcg_gen_ori_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], X_FLAG);
127
+ } else if (dc->cc_op == CC_OP_FLAGS) {
128
+ tcg_gen_andi_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], ~X_FLAG);
129
}
130
dc->flags_uptodate = 1;
131
}
132
@@ -XXX,XX +XXX,XX @@ static void cris_update_cc_op(DisasContext *dc, int op, int size)
133
static inline void cris_update_cc_x(DisasContext *dc)
134
{
135
/* Save the x flag state at the time of the cc snapshot. */
136
- if (dc->flagx_known) {
137
- if (dc->cc_x_uptodate == (2 | dc->flags_x)) {
138
- return;
139
- }
140
- tcg_gen_movi_tl(cc_x, dc->flags_x);
141
- dc->cc_x_uptodate = 2 | dc->flags_x;
142
- } else {
143
- tcg_gen_andi_tl(cc_x, cpu_PR[PR_CCS], X_FLAG);
144
- dc->cc_x_uptodate = 1;
145
+ if (dc->cc_x_uptodate == (2 | dc->flags_x)) {
146
+ return;
147
}
148
+ tcg_gen_movi_tl(cc_x, dc->flags_x);
149
+ dc->cc_x_uptodate = 2 | dc->flags_x;
150
}
151
152
/* Update cc prior to executing ALU op. Needs source operands untouched. */
153
@@ -XXX,XX +XXX,XX @@ static void gen_store (DisasContext *dc, TCGv addr, TCGv val,
154
155
/* Conditional writes. We only support the kind were X and P are known
156
at translation time. */
157
- if (dc->flagx_known && dc->flags_x && (dc->tb_flags & P_FLAG)) {
158
+ if (dc->flags_x && (dc->tb_flags & P_FLAG)) {
159
dc->postinc = 0;
160
cris_evaluate_flags(dc);
161
tcg_gen_ori_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], C_FLAG);
162
@@ -XXX,XX +XXX,XX @@ static void gen_store (DisasContext *dc, TCGv addr, TCGv val,
163
164
tcg_gen_qemu_st_tl(val, addr, mem_index, MO_TE + ctz32(size));
165
166
- if (dc->flagx_known && dc->flags_x) {
167
+ if (dc->flags_x) {
168
cris_evaluate_flags(dc);
169
tcg_gen_andi_tl(cpu_PR[PR_CCS], cpu_PR[PR_CCS], ~C_FLAG);
170
}
171
@@ -XXX,XX +XXX,XX @@ static int dec_addc_r(CPUCRISState *env, DisasContext *dc)
172
LOG_DIS("addc $r%u, $r%u\n",
173
dc->op1, dc->op2);
174
cris_evaluate_flags(dc);
175
+
176
/* Set for this insn. */
177
- dc->flagx_known = 1;
178
dc->flags_x = X_FLAG;
179
180
cris_cc_mask(dc, CC_MASK_NZVC);
181
@@ -XXX,XX +XXX,XX @@ static int dec_setclrf(CPUCRISState *env, DisasContext *dc)
182
}
183
184
if (flags & X_FLAG) {
185
- dc->flagx_known = 1;
186
if (set) {
187
dc->flags_x = X_FLAG;
188
} else {
189
@@ -XXX,XX +XXX,XX @@ static int dec_addc_mr(CPUCRISState *env, DisasContext *dc)
190
cris_evaluate_flags(dc);
191
192
/* Set for this insn. */
193
- dc->flagx_known = 1;
194
dc->flags_x = X_FLAG;
195
196
cris_alu_m_alloc_temps(t);
197
@@ -XXX,XX +XXX,XX @@ static void cris_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
198
dc->ppc = pc_start;
199
dc->pc = pc_start;
200
dc->flags_uptodate = 1;
201
- dc->flagx_known = 1;
202
dc->flags_x = tb_flags & X_FLAG;
203
dc->cc_x_uptodate = 0;
204
dc->cc_mask = 0;
205
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
206
}
207
208
/* Fold unhandled changes to X_FLAG into cpustate_changed. */
209
- dc->cpustate_changed |= !dc->flagx_known;
210
dc->cpustate_changed |= dc->flags_x != (dc->base.tb->flags & X_FLAG);
211
212
/*
213
diff --git a/target/cris/translate_v10.c.inc b/target/cris/translate_v10.c.inc
214
index XXXXXXX..XXXXXXX 100644
215
--- a/target/cris/translate_v10.c.inc
216
+++ b/target/cris/translate_v10.c.inc
217
@@ -XXX,XX +XXX,XX @@ static void gen_store_v10(DisasContext *dc, TCGv addr, TCGv val,
218
cris_store_direct_jmp(dc);
219
}
220
221
- /* Conditional writes. We only support the kind were X is known
222
- at translation time. */
223
- if (dc->flagx_known && dc->flags_x) {
224
+ /* Conditional writes. */
225
+ if (dc->flags_x) {
226
gen_store_v10_conditional(dc, addr, val, size, mem_index);
227
return;
228
}
229
@@ -XXX,XX +XXX,XX @@ static unsigned int dec10_setclrf(DisasContext *dc)
230
231
232
if (flags & X_FLAG) {
233
- dc->flagx_known = 1;
234
if (set)
235
dc->flags_x = X_FLAG;
236
else
237
--
238
2.25.1
239
240
diff view generated by jsdifflib
Deleted patch
1
We always know the exact value of X, that's all that matters.
2
This avoids splitting the TB e.g. between "ax" and "addq".
3
1
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Tested-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/cris/translate.c | 3 ---
9
1 file changed, 3 deletions(-)
10
11
diff --git a/target/cris/translate.c b/target/cris/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/cris/translate.c
14
+++ b/target/cris/translate.c
15
@@ -XXX,XX +XXX,XX @@ static void cris_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
16
cris_clear_x_flag(dc);
17
}
18
19
- /* Fold unhandled changes to X_FLAG into cpustate_changed. */
20
- dc->cpustate_changed |= dc->flags_x != (dc->base.tb->flags & X_FLAG);
21
-
22
/*
23
* All branches are delayed branches, handled immediately below.
24
* We don't expect to see odd combinations of exit conditions.
25
--
26
2.25.1
27
28
diff view generated by jsdifflib
Deleted patch
1
From: LIU Zhiwei <zhiwei_liu@c-sky.com>
2
1
3
Implement tcg_gen_vec_add{sub}16_tl by adding corresponding i32 OP.
4
5
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
6
Message-Id: <20210624105023.3852-2-zhiwei_liu@c-sky.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
include/tcg/tcg-op-gvec.h | 13 +++++++++++++
10
tcg/tcg-op-gvec.c | 28 ++++++++++++++++++++++++++++
11
2 files changed, 41 insertions(+)
12
13
diff --git a/include/tcg/tcg-op-gvec.h b/include/tcg/tcg-op-gvec.h
14
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg-op-gvec.h
16
+++ b/include/tcg/tcg-op-gvec.h
17
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t);
18
void tcg_gen_vec_rotl8i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c);
19
void tcg_gen_vec_rotl16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c);
20
21
+/* 32-bit vector operations. */
22
+void tcg_gen_vec_add16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
23
+
24
+void tcg_gen_vec_sub16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
25
+
26
+#if TARGET_LONG_BITS == 64
27
+#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i64
28
+#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i64
29
+#else
30
+#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i32
31
+#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i32
32
+#endif
33
+
34
#endif
35
diff --git a/tcg/tcg-op-gvec.c b/tcg/tcg-op-gvec.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/tcg/tcg-op-gvec.c
38
+++ b/tcg/tcg-op-gvec.c
39
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_add16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
40
gen_addv_mask(d, a, b, m);
41
}
42
43
+void tcg_gen_vec_add16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
44
+{
45
+ TCGv_i32 t1 = tcg_temp_new_i32();
46
+ TCGv_i32 t2 = tcg_temp_new_i32();
47
+
48
+ tcg_gen_andi_i32(t1, a, ~0xffff);
49
+ tcg_gen_add_i32(t2, a, b);
50
+ tcg_gen_add_i32(t1, t1, b);
51
+ tcg_gen_deposit_i32(d, t1, t2, 0, 16);
52
+
53
+ tcg_temp_free_i32(t1);
54
+ tcg_temp_free_i32(t2);
55
+}
56
+
57
void tcg_gen_vec_add32_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
58
{
59
TCGv_i64 t1 = tcg_temp_new_i64();
60
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sub16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
61
gen_subv_mask(d, a, b, m);
62
}
63
64
+void tcg_gen_vec_sub16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
65
+{
66
+ TCGv_i32 t1 = tcg_temp_new_i32();
67
+ TCGv_i32 t2 = tcg_temp_new_i32();
68
+
69
+ tcg_gen_andi_i32(t1, b, ~0xffff);
70
+ tcg_gen_sub_i32(t2, a, b);
71
+ tcg_gen_sub_i32(t1, a, t1);
72
+ tcg_gen_deposit_i32(d, t1, t2, 0, 16);
73
+
74
+ tcg_temp_free_i32(t1);
75
+ tcg_temp_free_i32(t2);
76
+}
77
+
78
void tcg_gen_vec_sub32_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
79
{
80
TCGv_i64 t1 = tcg_temp_new_i64();
81
--
82
2.25.1
83
84
diff view generated by jsdifflib
Deleted patch
1
From: LIU Zhiwei <zhiwei_liu@c-sky.com>
2
1
3
Implement tcg_gen_vec_add{sub}8_tl by adding corresponging i32 OP.
4
5
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
6
Message-Id: <20210624105023.3852-3-zhiwei_liu@c-sky.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
include/tcg/tcg-op-gvec.h | 6 ++++++
10
tcg/tcg-op-gvec.c | 38 ++++++++++++++++++++++++++++++++++++++
11
2 files changed, 44 insertions(+)
12
13
diff --git a/include/tcg/tcg-op-gvec.h b/include/tcg/tcg-op-gvec.h
14
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg-op-gvec.h
16
+++ b/include/tcg/tcg-op-gvec.h
17
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_rotl8i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c);
18
void tcg_gen_vec_rotl16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c);
19
20
/* 32-bit vector operations. */
21
+void tcg_gen_vec_add8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
22
void tcg_gen_vec_add16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
23
24
+void tcg_gen_vec_sub8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
25
void tcg_gen_vec_sub16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
26
27
#if TARGET_LONG_BITS == 64
28
+#define tcg_gen_vec_add8_tl tcg_gen_vec_add8_i64
29
+#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i64
30
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i64
31
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i64
32
#else
33
+#define tcg_gen_vec_add8_tl tcg_gen_vec_add8_i32
34
+#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i32
35
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i32
36
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i32
37
#endif
38
diff --git a/tcg/tcg-op-gvec.c b/tcg/tcg-op-gvec.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/tcg/tcg-op-gvec.c
41
+++ b/tcg/tcg-op-gvec.c
42
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_add8_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
43
gen_addv_mask(d, a, b, m);
44
}
45
46
+void tcg_gen_vec_add8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
47
+{
48
+ TCGv_i32 m = tcg_constant_i32((int32_t)dup_const(MO_8, 0x80));
49
+ TCGv_i32 t1 = tcg_temp_new_i32();
50
+ TCGv_i32 t2 = tcg_temp_new_i32();
51
+ TCGv_i32 t3 = tcg_temp_new_i32();
52
+
53
+ tcg_gen_andc_i32(t1, a, m);
54
+ tcg_gen_andc_i32(t2, b, m);
55
+ tcg_gen_xor_i32(t3, a, b);
56
+ tcg_gen_add_i32(d, t1, t2);
57
+ tcg_gen_and_i32(t3, t3, m);
58
+ tcg_gen_xor_i32(d, d, t3);
59
+
60
+ tcg_temp_free_i32(t1);
61
+ tcg_temp_free_i32(t2);
62
+ tcg_temp_free_i32(t3);
63
+}
64
+
65
void tcg_gen_vec_add16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
66
{
67
TCGv_i64 m = tcg_constant_i64(dup_const(MO_16, 0x8000));
68
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sub8_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
69
gen_subv_mask(d, a, b, m);
70
}
71
72
+void tcg_gen_vec_sub8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
73
+{
74
+ TCGv_i32 m = tcg_constant_i32((int32_t)dup_const(MO_8, 0x80));
75
+ TCGv_i32 t1 = tcg_temp_new_i32();
76
+ TCGv_i32 t2 = tcg_temp_new_i32();
77
+ TCGv_i32 t3 = tcg_temp_new_i32();
78
+
79
+ tcg_gen_or_i32(t1, a, m);
80
+ tcg_gen_andc_i32(t2, b, m);
81
+ tcg_gen_eqv_i32(t3, a, b);
82
+ tcg_gen_sub_i32(d, t1, t2);
83
+ tcg_gen_and_i32(t3, t3, m);
84
+ tcg_gen_xor_i32(d, d, t3);
85
+
86
+ tcg_temp_free_i32(t1);
87
+ tcg_temp_free_i32(t2);
88
+ tcg_temp_free_i32(t3);
89
+}
90
+
91
void tcg_gen_vec_sub16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
92
{
93
TCGv_i64 m = tcg_constant_i64(dup_const(MO_16, 0x8000));
94
--
95
2.25.1
96
97
diff view generated by jsdifflib
Deleted patch
1
From: LIU Zhiwei <zhiwei_liu@c-sky.com>
2
1
3
Implement tcg_gen_vec_shl{shr}{sar}16i_tl by adding corresponging i32 OP.
4
5
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
6
Message-Id: <20210624105023.3852-4-zhiwei_liu@c-sky.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
include/tcg/tcg-op-gvec.h | 10 ++++++++++
10
tcg/tcg-op-gvec.c | 28 ++++++++++++++++++++++++++++
11
2 files changed, 38 insertions(+)
12
13
diff --git a/include/tcg/tcg-op-gvec.h b/include/tcg/tcg-op-gvec.h
14
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg-op-gvec.h
16
+++ b/include/tcg/tcg-op-gvec.h
17
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_add16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
18
void tcg_gen_vec_sub8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
19
void tcg_gen_vec_sub16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
20
21
+void tcg_gen_vec_shl16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
22
+void tcg_gen_vec_shr16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
23
+void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
24
+
25
#if TARGET_LONG_BITS == 64
26
#define tcg_gen_vec_add8_tl tcg_gen_vec_add8_i64
27
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i64
28
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i64
29
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i64
30
+#define tcg_gen_vec_shl16i_tl tcg_gen_vec_shl16i_i64
31
+#define tcg_gen_vec_shr16i_tl tcg_gen_vec_shr16i_i64
32
+#define tcg_gen_vec_sar16i_tl tcg_gen_vec_sar16i_i64
33
#else
34
#define tcg_gen_vec_add8_tl tcg_gen_vec_add8_i32
35
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i32
36
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i32
37
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i32
38
+#define tcg_gen_vec_shl16i_tl tcg_gen_vec_shl16i_i32
39
+#define tcg_gen_vec_shr16i_tl tcg_gen_vec_shr16i_i32
40
+#define tcg_gen_vec_sar16i_tl tcg_gen_vec_sar16i_i32
41
#endif
42
43
#endif
44
diff --git a/tcg/tcg-op-gvec.c b/tcg/tcg-op-gvec.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/tcg/tcg-op-gvec.c
47
+++ b/tcg/tcg-op-gvec.c
48
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_shl16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
49
tcg_gen_andi_i64(d, d, mask);
50
}
51
52
+void tcg_gen_vec_shl16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
53
+{
54
+ uint32_t mask = dup_const(MO_16, 0xffff << c);
55
+ tcg_gen_shli_i32(d, a, c);
56
+ tcg_gen_andi_i32(d, d, mask);
57
+}
58
+
59
void tcg_gen_gvec_shli(unsigned vece, uint32_t dofs, uint32_t aofs,
60
int64_t shift, uint32_t oprsz, uint32_t maxsz)
61
{
62
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_shr16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
63
tcg_gen_andi_i64(d, d, mask);
64
}
65
66
+void tcg_gen_vec_shr16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
67
+{
68
+ uint32_t mask = dup_const(MO_16, 0xffff >> c);
69
+ tcg_gen_shri_i32(d, a, c);
70
+ tcg_gen_andi_i32(d, d, mask);
71
+}
72
+
73
void tcg_gen_gvec_shri(unsigned vece, uint32_t dofs, uint32_t aofs,
74
int64_t shift, uint32_t oprsz, uint32_t maxsz)
75
{
76
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
77
tcg_temp_free_i64(s);
78
}
79
80
+void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
81
+{
82
+ uint32_t s_mask = dup_const(MO_16, 0x8000 >> c);
83
+ uint32_t c_mask = dup_const(MO_16, 0xffff >> c);
84
+ TCGv_i32 s = tcg_temp_new_i32();
85
+
86
+ tcg_gen_shri_i32(d, a, c);
87
+ tcg_gen_andi_i32(s, d, s_mask); /* isolate (shifted) sign bit */
88
+ tcg_gen_andi_i32(d, d, c_mask); /* clear out bits above sign */
89
+ tcg_gen_muli_i32(s, s, (2 << c) - 2); /* replicate isolated signs */
90
+ tcg_gen_or_i32(d, d, s); /* include sign extension */
91
+ tcg_temp_free_i32(s);
92
+}
93
+
94
void tcg_gen_gvec_sari(unsigned vece, uint32_t dofs, uint32_t aofs,
95
int64_t shift, uint32_t oprsz, uint32_t maxsz)
96
{
97
--
98
2.25.1
99
100
diff view generated by jsdifflib
Deleted patch
1
From: LIU Zhiwei <zhiwei_liu@c-sky.com>
2
1
3
Implement tcg_gen_vec_shl{shr}{sar}8i_tl by adding corresponging i32 OP.
4
5
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
6
Message-Id: <20210624105023.3852-5-zhiwei_liu@c-sky.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
include/tcg/tcg-op-gvec.h | 10 ++++++++++
10
tcg/tcg-op-gvec.c | 28 ++++++++++++++++++++++++++++
11
2 files changed, 38 insertions(+)
12
13
diff --git a/include/tcg/tcg-op-gvec.h b/include/tcg/tcg-op-gvec.h
14
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg-op-gvec.h
16
+++ b/include/tcg/tcg-op-gvec.h
17
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_add16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
18
void tcg_gen_vec_sub8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
19
void tcg_gen_vec_sub16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b);
20
21
+void tcg_gen_vec_shl8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
22
void tcg_gen_vec_shl16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
23
+void tcg_gen_vec_shr8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
24
void tcg_gen_vec_shr16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
25
+void tcg_gen_vec_sar8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
26
void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
27
28
#if TARGET_LONG_BITS == 64
29
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
30
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i64
31
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i64
32
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i64
33
+#define tcg_gen_vec_shl8i_tl tcg_gen_vec_shl8i_i64
34
+#define tcg_gen_vec_shr8i_tl tcg_gen_vec_shr8i_i64
35
+#define tcg_gen_vec_sar8i_tl tcg_gen_vec_sar8i_i64
36
#define tcg_gen_vec_shl16i_tl tcg_gen_vec_shl16i_i64
37
#define tcg_gen_vec_shr16i_tl tcg_gen_vec_shr16i_i64
38
#define tcg_gen_vec_sar16i_tl tcg_gen_vec_sar16i_i64
39
+
40
#else
41
#define tcg_gen_vec_add8_tl tcg_gen_vec_add8_i32
42
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i32
43
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i32
44
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i32
45
+#define tcg_gen_vec_shl8i_tl tcg_gen_vec_shl8i_i32
46
+#define tcg_gen_vec_shr8i_tl tcg_gen_vec_shr8i_i32
47
+#define tcg_gen_vec_sar8i_tl tcg_gen_vec_sar8i_i32
48
#define tcg_gen_vec_shl16i_tl tcg_gen_vec_shl16i_i32
49
#define tcg_gen_vec_shr16i_tl tcg_gen_vec_shr16i_i32
50
#define tcg_gen_vec_sar16i_tl tcg_gen_vec_sar16i_i32
51
diff --git a/tcg/tcg-op-gvec.c b/tcg/tcg-op-gvec.c
52
index XXXXXXX..XXXXXXX 100644
53
--- a/tcg/tcg-op-gvec.c
54
+++ b/tcg/tcg-op-gvec.c
55
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_shl16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
56
tcg_gen_andi_i64(d, d, mask);
57
}
58
59
+void tcg_gen_vec_shl8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
60
+{
61
+ uint32_t mask = dup_const(MO_8, 0xff << c);
62
+ tcg_gen_shli_i32(d, a, c);
63
+ tcg_gen_andi_i32(d, d, mask);
64
+}
65
+
66
void tcg_gen_vec_shl16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
67
{
68
uint32_t mask = dup_const(MO_16, 0xffff << c);
69
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_shr16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
70
tcg_gen_andi_i64(d, d, mask);
71
}
72
73
+void tcg_gen_vec_shr8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
74
+{
75
+ uint32_t mask = dup_const(MO_8, 0xff >> c);
76
+ tcg_gen_shri_i32(d, a, c);
77
+ tcg_gen_andi_i32(d, d, mask);
78
+}
79
+
80
void tcg_gen_vec_shr16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
81
{
82
uint32_t mask = dup_const(MO_16, 0xffff >> c);
83
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
84
tcg_temp_free_i64(s);
85
}
86
87
+void tcg_gen_vec_sar8i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
88
+{
89
+ uint32_t s_mask = dup_const(MO_8, 0x80 >> c);
90
+ uint32_t c_mask = dup_const(MO_8, 0xff >> c);
91
+ TCGv_i32 s = tcg_temp_new_i32();
92
+
93
+ tcg_gen_shri_i32(d, a, c);
94
+ tcg_gen_andi_i32(s, d, s_mask); /* isolate (shifted) sign bit */
95
+ tcg_gen_muli_i32(s, s, (2 << c) - 2); /* replicate isolated signs */
96
+ tcg_gen_andi_i32(d, d, c_mask); /* clear out bits above sign */
97
+ tcg_gen_or_i32(d, d, s); /* include sign extension */
98
+ tcg_temp_free_i32(s);
99
+}
100
+
101
void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t c)
102
{
103
uint32_t s_mask = dup_const(MO_16, 0x8000 >> c);
104
--
105
2.25.1
106
107
diff view generated by jsdifflib
Deleted patch
1
From: LIU Zhiwei <zhiwei_liu@c-sky.com>
2
1
3
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
4
Message-Id: <20210624105023.3852-6-zhiwei_liu@c-sky.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
include/tcg/tcg-op-gvec.h | 4 ++++
8
1 file changed, 4 insertions(+)
9
10
diff --git a/include/tcg/tcg-op-gvec.h b/include/tcg/tcg-op-gvec.h
11
index XXXXXXX..XXXXXXX 100644
12
--- a/include/tcg/tcg-op-gvec.h
13
+++ b/include/tcg/tcg-op-gvec.h
14
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
15
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i64
16
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i64
17
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i64
18
+#define tcg_gen_vec_add32_tl tcg_gen_vec_add32_i64
19
+#define tcg_gen_vec_sub32_tl tcg_gen_vec_sub32_i64
20
#define tcg_gen_vec_shl8i_tl tcg_gen_vec_shl8i_i64
21
#define tcg_gen_vec_shr8i_tl tcg_gen_vec_shr8i_i64
22
#define tcg_gen_vec_sar8i_tl tcg_gen_vec_sar8i_i64
23
@@ -XXX,XX +XXX,XX @@ void tcg_gen_vec_sar16i_i32(TCGv_i32 d, TCGv_i32 a, int32_t);
24
#define tcg_gen_vec_sub8_tl tcg_gen_vec_sub8_i32
25
#define tcg_gen_vec_add16_tl tcg_gen_vec_add16_i32
26
#define tcg_gen_vec_sub16_tl tcg_gen_vec_sub16_i32
27
+#define tcg_gen_vec_add32_tl tcg_gen_add_i32
28
+#define tcg_gen_vec_sub32_tl tcg_gen_sub_i32
29
#define tcg_gen_vec_shl8i_tl tcg_gen_vec_shl8i_i32
30
#define tcg_gen_vec_shr8i_tl tcg_gen_vec_shr8i_i32
31
#define tcg_gen_vec_sar8i_tl tcg_gen_vec_sar8i_i32
32
--
33
2.25.1
34
35
diff view generated by jsdifflib
Deleted patch
1
From: Warner Losh <imp@bsdimp.com>
2
1
3
The trap number for a page fault on BSD systems is T_PAGEFLT
4
not 0xe -- 0xe is used by Linux and represents the intel hardware
5
trap vector. The BSD kernels, however, translate this to T_PAGEFLT
6
in their Xpage, Xtrap0e, Xtrap14, etc fault handlers. This is true
7
for i386 and x86_64, though the name of the trap hanlder can very
8
on the flavor of BSD. As far as I can tell, Linux doesn't provide
9
a define for this value. Invent a new one (PAGE_FAULT_TRAP) and
10
use it instead to avoid uglier ifdefs.
11
12
Signed-off-by: Mark Johnston <markj@FreeBSD.org>
13
Signed-off-by: Juergen Lock <nox@FreeBSD.org>
14
[ Rework to avoid ifdefs and expand it to i386 ]
15
Signed-off-by: Warner Losh <imp@bsdimp.com>
16
Message-Id: <20210625045707.84534-3-imp@bsdimp.com>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
18
---
19
accel/tcg/user-exec.c | 20 ++++++++++++++++++--
20
1 file changed, 18 insertions(+), 2 deletions(-)
21
22
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
23
index XXXXXXX..XXXXXXX 100644
24
--- a/accel/tcg/user-exec.c
25
+++ b/accel/tcg/user-exec.c
26
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
27
28
#if defined(__NetBSD__)
29
#include <ucontext.h>
30
+#include <machine/trap.h>
31
32
#define EIP_sig(context) ((context)->uc_mcontext.__gregs[_REG_EIP])
33
#define TRAP_sig(context) ((context)->uc_mcontext.__gregs[_REG_TRAPNO])
34
#define ERROR_sig(context) ((context)->uc_mcontext.__gregs[_REG_ERR])
35
#define MASK_sig(context) ((context)->uc_sigmask)
36
+#define PAGE_FAULT_TRAP T_PAGEFLT
37
#elif defined(__FreeBSD__) || defined(__DragonFly__)
38
#include <ucontext.h>
39
+#include <machine/trap.h>
40
41
#define EIP_sig(context) (*((unsigned long *)&(context)->uc_mcontext.mc_eip))
42
#define TRAP_sig(context) ((context)->uc_mcontext.mc_trapno)
43
#define ERROR_sig(context) ((context)->uc_mcontext.mc_err)
44
#define MASK_sig(context) ((context)->uc_sigmask)
45
+#define PAGE_FAULT_TRAP T_PAGEFLT
46
#elif defined(__OpenBSD__)
47
+#include <machine/trap.h>
48
#define EIP_sig(context) ((context)->sc_eip)
49
#define TRAP_sig(context) ((context)->sc_trapno)
50
#define ERROR_sig(context) ((context)->sc_err)
51
#define MASK_sig(context) ((context)->sc_mask)
52
+#define PAGE_FAULT_TRAP T_PAGEFLT
53
#else
54
#define EIP_sig(context) ((context)->uc_mcontext.gregs[REG_EIP])
55
#define TRAP_sig(context) ((context)->uc_mcontext.gregs[REG_TRAPNO])
56
#define ERROR_sig(context) ((context)->uc_mcontext.gregs[REG_ERR])
57
#define MASK_sig(context) ((context)->uc_sigmask)
58
+#define PAGE_FAULT_TRAP 0xe
59
#endif
60
61
int cpu_signal_handler(int host_signum, void *pinfo,
62
@@ -XXX,XX +XXX,XX @@ int cpu_signal_handler(int host_signum, void *pinfo,
63
pc = EIP_sig(uc);
64
trapno = TRAP_sig(uc);
65
return handle_cpu_signal(pc, info,
66
- trapno == 0xe ? (ERROR_sig(uc) >> 1) & 1 : 0,
67
+ trapno == PAGE_FAULT_TRAP ?
68
+ (ERROR_sig(uc) >> 1) & 1 : 0,
69
&MASK_sig(uc));
70
}
71
72
#elif defined(__x86_64__)
73
74
#ifdef __NetBSD__
75
+#include <machine/trap.h>
76
#define PC_sig(context) _UC_MACHINE_PC(context)
77
#define TRAP_sig(context) ((context)->uc_mcontext.__gregs[_REG_TRAPNO])
78
#define ERROR_sig(context) ((context)->uc_mcontext.__gregs[_REG_ERR])
79
#define MASK_sig(context) ((context)->uc_sigmask)
80
+#define PAGE_FAULT_TRAP T_PAGEFLT
81
#elif defined(__OpenBSD__)
82
+#include <machine/trap.h>
83
#define PC_sig(context) ((context)->sc_rip)
84
#define TRAP_sig(context) ((context)->sc_trapno)
85
#define ERROR_sig(context) ((context)->sc_err)
86
#define MASK_sig(context) ((context)->sc_mask)
87
+#define PAGE_FAULT_TRAP T_PAGEFLT
88
#elif defined(__FreeBSD__) || defined(__DragonFly__)
89
#include <ucontext.h>
90
+#include <machine/trap.h>
91
92
#define PC_sig(context) (*((unsigned long *)&(context)->uc_mcontext.mc_rip))
93
#define TRAP_sig(context) ((context)->uc_mcontext.mc_trapno)
94
#define ERROR_sig(context) ((context)->uc_mcontext.mc_err)
95
#define MASK_sig(context) ((context)->uc_sigmask)
96
+#define PAGE_FAULT_TRAP T_PAGEFLT
97
#else
98
#define PC_sig(context) ((context)->uc_mcontext.gregs[REG_RIP])
99
#define TRAP_sig(context) ((context)->uc_mcontext.gregs[REG_TRAPNO])
100
#define ERROR_sig(context) ((context)->uc_mcontext.gregs[REG_ERR])
101
#define MASK_sig(context) ((context)->uc_sigmask)
102
+#define PAGE_FAULT_TRAP 0xe
103
#endif
104
105
int cpu_signal_handler(int host_signum, void *pinfo,
106
@@ -XXX,XX +XXX,XX @@ int cpu_signal_handler(int host_signum, void *pinfo,
107
108
pc = PC_sig(uc);
109
return handle_cpu_signal(pc, info,
110
- TRAP_sig(uc) == 0xe ? (ERROR_sig(uc) >> 1) & 1 : 0,
111
+ TRAP_sig(uc) == PAGE_FAULT_TRAP ?
112
+ (ERROR_sig(uc) >> 1) & 1 : 0,
113
&MASK_sig(uc));
114
}
115
116
--
117
2.25.1
118
119
diff view generated by jsdifflib
Deleted patch
1
Retain the current rorw bswap16 expansion for the zero-in/zero-out case.
2
Otherwise, perform a wider bswap plus a right-shift or extend.
3
1
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/i386/tcg-target.c.inc | 20 +++++++++++++++++++-
8
1 file changed, 19 insertions(+), 1 deletion(-)
9
10
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.c.inc
13
+++ b/tcg/i386/tcg-target.c.inc
14
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
15
break;
16
17
OP_32_64(bswap16):
18
- tcg_out_rolw_8(s, a0);
19
+ if (a2 & TCG_BSWAP_OS) {
20
+ /* Output must be sign-extended. */
21
+ if (rexw) {
22
+ tcg_out_bswap64(s, a0);
23
+ tcg_out_shifti(s, SHIFT_SAR + rexw, a0, 48);
24
+ } else {
25
+ tcg_out_bswap32(s, a0);
26
+ tcg_out_shifti(s, SHIFT_SAR, a0, 16);
27
+ }
28
+ } else if ((a2 & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
29
+ /* Output must be zero-extended, but input isn't. */
30
+ tcg_out_bswap32(s, a0);
31
+ tcg_out_shifti(s, SHIFT_SHR, a0, 16);
32
+ } else {
33
+ tcg_out_rolw_8(s, a0);
34
+ }
35
break;
36
OP_32_64(bswap32):
37
tcg_out_bswap32(s, a0);
38
+ if (rexw && (a2 & TCG_BSWAP_OS)) {
39
+ tcg_out_ext32s(s, a0, a0);
40
+ }
41
break;
42
43
OP_32_64(neg):
44
--
45
2.25.1
46
47
diff view generated by jsdifflib
Deleted patch
1
Pass in the input and output size. We currently use 3 of the 5
2
possible combinations; the others may be used by new tcg opcodes.
3
1
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/aarch64/tcg-target.c.inc | 42 ++++++++++++++----------------------
8
1 file changed, 16 insertions(+), 26 deletions(-)
9
10
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/aarch64/tcg-target.c.inc
13
+++ b/tcg/aarch64/tcg-target.c.inc
14
@@ -XXX,XX +XXX,XX @@ typedef enum {
15
/* Data-processing (1 source) instructions. */
16
I3507_CLZ = 0x5ac01000,
17
I3507_RBIT = 0x5ac00000,
18
- I3507_REV16 = 0x5ac00400,
19
- I3507_REV32 = 0x5ac00800,
20
- I3507_REV64 = 0x5ac00c00,
21
+ I3507_REV = 0x5ac00000, /* + size << 10 */
22
23
/* Data-processing (2 source) instructions. */
24
I3508_LSLV = 0x1ac02000,
25
@@ -XXX,XX +XXX,XX @@ static void tcg_out_brcond(TCGContext *s, TCGType ext, TCGCond c, TCGArg a,
26
}
27
}
28
29
-static inline void tcg_out_rev64(TCGContext *s, TCGReg rd, TCGReg rn)
30
+static inline void tcg_out_rev(TCGContext *s, int ext, MemOp s_bits,
31
+ TCGReg rd, TCGReg rn)
32
{
33
- tcg_out_insn(s, 3507, REV64, TCG_TYPE_I64, rd, rn);
34
-}
35
-
36
-static inline void tcg_out_rev32(TCGContext *s, TCGReg rd, TCGReg rn)
37
-{
38
- tcg_out_insn(s, 3507, REV32, TCG_TYPE_I32, rd, rn);
39
-}
40
-
41
-static inline void tcg_out_rev16(TCGContext *s, TCGReg rd, TCGReg rn)
42
-{
43
- tcg_out_insn(s, 3507, REV16, TCG_TYPE_I32, rd, rn);
44
+ /* REV, REV16, REV32 */
45
+ tcg_out_insn_3507(s, I3507_REV | (s_bits << 10), ext, rd, rn);
46
}
47
48
static inline void tcg_out_sxt(TCGContext *s, TCGType ext, MemOp s_bits,
49
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
50
case MO_UW:
51
tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
52
if (bswap) {
53
- tcg_out_rev16(s, data_r, data_r);
54
+ tcg_out_rev(s, TCG_TYPE_I32, MO_16, data_r, data_r);
55
}
56
break;
57
case MO_SW:
58
if (bswap) {
59
tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
60
- tcg_out_rev16(s, data_r, data_r);
61
+ tcg_out_rev(s, TCG_TYPE_I32, MO_16, data_r, data_r);
62
tcg_out_sxt(s, ext, MO_16, data_r, data_r);
63
} else {
64
tcg_out_ldst_r(s, (ext ? I3312_LDRSHX : I3312_LDRSHW),
65
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
66
case MO_UL:
67
tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
68
if (bswap) {
69
- tcg_out_rev32(s, data_r, data_r);
70
+ tcg_out_rev(s, TCG_TYPE_I32, MO_32, data_r, data_r);
71
}
72
break;
73
case MO_SL:
74
if (bswap) {
75
tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
76
- tcg_out_rev32(s, data_r, data_r);
77
+ tcg_out_rev(s, TCG_TYPE_I32, MO_32, data_r, data_r);
78
tcg_out_sxt(s, TCG_TYPE_I64, MO_32, data_r, data_r);
79
} else {
80
tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, otype, off_r);
81
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
82
case MO_Q:
83
tcg_out_ldst_r(s, I3312_LDRX, data_r, addr_r, otype, off_r);
84
if (bswap) {
85
- tcg_out_rev64(s, data_r, data_r);
86
+ tcg_out_rev(s, TCG_TYPE_I64, MO_64, data_r, data_r);
87
}
88
break;
89
default:
90
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp memop,
91
break;
92
case MO_16:
93
if (bswap && data_r != TCG_REG_XZR) {
94
- tcg_out_rev16(s, TCG_REG_TMP, data_r);
95
+ tcg_out_rev(s, TCG_TYPE_I32, MO_16, TCG_REG_TMP, data_r);
96
data_r = TCG_REG_TMP;
97
}
98
tcg_out_ldst_r(s, I3312_STRH, data_r, addr_r, otype, off_r);
99
break;
100
case MO_32:
101
if (bswap && data_r != TCG_REG_XZR) {
102
- tcg_out_rev32(s, TCG_REG_TMP, data_r);
103
+ tcg_out_rev(s, TCG_TYPE_I32, MO_32, TCG_REG_TMP, data_r);
104
data_r = TCG_REG_TMP;
105
}
106
tcg_out_ldst_r(s, I3312_STRW, data_r, addr_r, otype, off_r);
107
break;
108
case MO_64:
109
if (bswap && data_r != TCG_REG_XZR) {
110
- tcg_out_rev64(s, TCG_REG_TMP, data_r);
111
+ tcg_out_rev(s, TCG_TYPE_I64, MO_64, TCG_REG_TMP, data_r);
112
data_r = TCG_REG_TMP;
113
}
114
tcg_out_ldst_r(s, I3312_STRX, data_r, addr_r, otype, off_r);
115
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
116
break;
117
118
case INDEX_op_bswap64_i64:
119
- tcg_out_rev64(s, a0, a1);
120
+ tcg_out_rev(s, TCG_TYPE_I64, MO_64, a0, a1);
121
break;
122
case INDEX_op_bswap32_i64:
123
case INDEX_op_bswap32_i32:
124
- tcg_out_rev32(s, a0, a1);
125
+ tcg_out_rev(s, TCG_TYPE_I32, MO_32, a0, a1);
126
break;
127
case INDEX_op_bswap16_i64:
128
case INDEX_op_bswap16_i32:
129
- tcg_out_rev16(s, a0, a1);
130
+ tcg_out_rev(s, TCG_TYPE_I32, MO_16, a0, a1);
131
break;
132
133
case INDEX_op_ext8s_i64:
134
--
135
2.25.1
136
137
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
5
tcg/aarch64/tcg-target.c.inc | 12 ++++++++++++
6
1 file changed, 12 insertions(+)
7
1
8
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
9
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/aarch64/tcg-target.c.inc
11
+++ b/tcg/aarch64/tcg-target.c.inc
12
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
13
tcg_out_rev(s, TCG_TYPE_I64, MO_64, a0, a1);
14
break;
15
case INDEX_op_bswap32_i64:
16
+ tcg_out_rev(s, TCG_TYPE_I32, MO_32, a0, a1);
17
+ if (a2 & TCG_BSWAP_OS) {
18
+ tcg_out_sxt(s, TCG_TYPE_I64, MO_32, a0, a0);
19
+ }
20
+ break;
21
case INDEX_op_bswap32_i32:
22
tcg_out_rev(s, TCG_TYPE_I32, MO_32, a0, a1);
23
break;
24
case INDEX_op_bswap16_i64:
25
case INDEX_op_bswap16_i32:
26
tcg_out_rev(s, TCG_TYPE_I32, MO_16, a0, a1);
27
+ if (a2 & TCG_BSWAP_OS) {
28
+ /* Output must be sign-extended. */
29
+ tcg_out_sxt(s, ext, MO_16, a0, a0);
30
+ } else if ((a2 & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
31
+ /* Output must be zero-extended, but input isn't. */
32
+ tcg_out_uxt(s, MO_16, a0, a0);
33
+ }
34
break;
35
36
case INDEX_op_ext8s_i64:
37
--
38
2.25.1
39
40
diff view generated by jsdifflib
Deleted patch
1
We will shortly require these in other context;
2
make the expansion as clear as possible.
3
1
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/ppc/tcg-target.c.inc | 31 +++++++++++++++++++++----------
9
1 file changed, 21 insertions(+), 10 deletions(-)
10
11
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/ppc/tcg-target.c.inc
14
+++ b/tcg/ppc/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_rlw(TCGContext *s, int op, TCGReg ra, TCGReg rs,
16
tcg_out32(s, op | RA(ra) | RS(rs) | SH(sh) | MB(mb) | ME(me));
17
}
18
19
+static inline void tcg_out_ext8s(TCGContext *s, TCGReg dst, TCGReg src)
20
+{
21
+ tcg_out32(s, EXTSB | RA(dst) | RS(src));
22
+}
23
+
24
+static inline void tcg_out_ext16s(TCGContext *s, TCGReg dst, TCGReg src)
25
+{
26
+ tcg_out32(s, EXTSH | RA(dst) | RS(src));
27
+}
28
+
29
+static inline void tcg_out_ext32s(TCGContext *s, TCGReg dst, TCGReg src)
30
+{
31
+ tcg_out32(s, EXTSW | RA(dst) | RS(src));
32
+}
33
+
34
static inline void tcg_out_ext32u(TCGContext *s, TCGReg dst, TCGReg src)
35
{
36
tcg_out_rld(s, RLDICL, dst, src, 0, 32);
37
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
38
const int const_args[TCG_MAX_OP_ARGS])
39
{
40
TCGArg a0, a1, a2;
41
- int c;
42
43
switch (opc) {
44
case INDEX_op_exit_tb:
45
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
46
case INDEX_op_ld8s_i32:
47
case INDEX_op_ld8s_i64:
48
tcg_out_mem_long(s, LBZ, LBZX, args[0], args[1], args[2]);
49
- tcg_out32(s, EXTSB | RS(args[0]) | RA(args[0]));
50
+ tcg_out_ext8s(s, args[0], args[0]);
51
break;
52
case INDEX_op_ld16u_i32:
53
case INDEX_op_ld16u_i64:
54
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
55
56
case INDEX_op_ext8s_i32:
57
case INDEX_op_ext8s_i64:
58
- c = EXTSB;
59
- goto gen_ext;
60
+ tcg_out_ext8s(s, args[0], args[1]);
61
+ break;
62
case INDEX_op_ext16s_i32:
63
case INDEX_op_ext16s_i64:
64
- c = EXTSH;
65
- goto gen_ext;
66
+ tcg_out_ext16s(s, args[0], args[1]);
67
+ break;
68
case INDEX_op_ext_i32_i64:
69
case INDEX_op_ext32s_i64:
70
- c = EXTSW;
71
- goto gen_ext;
72
- gen_ext:
73
- tcg_out32(s, c | RS(args[1]) | RA(args[0]));
74
+ tcg_out_ext32s(s, args[0], args[1]);
75
break;
76
case INDEX_op_extu_i32_i64:
77
tcg_out_ext32u(s, args[0], args[1]);
78
--
79
2.25.1
80
81
diff view generated by jsdifflib
Deleted patch
1
We will shortly require sari in other context;
2
split out both for cleanliness sake.
3
1
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/ppc/tcg-target.c.inc | 17 +++++++++++++----
8
1 file changed, 13 insertions(+), 4 deletions(-)
9
10
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/ppc/tcg-target.c.inc
13
+++ b/tcg/ppc/tcg-target.c.inc
14
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_shli64(TCGContext *s, TCGReg dst, TCGReg src, int c)
15
tcg_out_rld(s, RLDICR, dst, src, c, 63 - c);
16
}
17
18
+static inline void tcg_out_sari32(TCGContext *s, TCGReg dst, TCGReg src, int c)
19
+{
20
+ /* Limit immediate shift count lest we create an illegal insn. */
21
+ tcg_out32(s, SRAWI | RA(dst) | RS(src) | SH(c & 31));
22
+}
23
+
24
static inline void tcg_out_shri32(TCGContext *s, TCGReg dst, TCGReg src, int c)
25
{
26
tcg_out_rlw(s, RLWINM, dst, src, 32 - c, c, 31);
27
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_shri64(TCGContext *s, TCGReg dst, TCGReg src, int c)
28
tcg_out_rld(s, RLDICL, dst, src, 64 - c, c);
29
}
30
31
+static inline void tcg_out_sari64(TCGContext *s, TCGReg dst, TCGReg src, int c)
32
+{
33
+ tcg_out32(s, SRADI | RA(dst) | RS(src) | SH(c & 0x1f) | ((c >> 4) & 2));
34
+}
35
+
36
/* Emit a move into ret of arg, if it can be done in one insn. */
37
static bool tcg_out_movi_one(TCGContext *s, TCGReg ret, tcg_target_long arg)
38
{
39
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
40
break;
41
case INDEX_op_sar_i32:
42
if (const_args[2]) {
43
- /* Limit immediate shift count lest we create an illegal insn. */
44
- tcg_out32(s, SRAWI | RS(args[1]) | RA(args[0]) | SH(args[2] & 31));
45
+ tcg_out_sari32(s, args[0], args[1], args[2]);
46
} else {
47
tcg_out32(s, SRAW | SAB(args[1], args[0], args[2]));
48
}
49
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
50
break;
51
case INDEX_op_sar_i64:
52
if (const_args[2]) {
53
- int sh = SH(args[2] & 0x1f) | (((args[2] >> 5) & 1) << 1);
54
- tcg_out32(s, SRADI | RA(args[0]) | RS(args[1]) | sh);
55
+ tcg_out_sari64(s, args[0], args[1], args[2]);
56
} else {
57
tcg_out32(s, SRAD | SAB(args[1], args[0], args[2]));
58
}
59
--
60
2.25.1
61
62
diff view generated by jsdifflib
Deleted patch
1
With the use of a suitable temporary, we can use the same
2
algorithm when src overlaps dst. The result is the same
3
number of instructions either way.
4
1
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/ppc/tcg-target.c.inc | 34 +++++++++++++++++++---------------
9
1 file changed, 19 insertions(+), 15 deletions(-)
10
11
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/ppc/tcg-target.c.inc
14
+++ b/tcg/ppc/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_sari64(TCGContext *s, TCGReg dst, TCGReg src, int c)
16
tcg_out32(s, SRADI | RA(dst) | RS(src) | SH(c & 0x1f) | ((c >> 4) & 2));
17
}
18
19
+static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src)
20
+{
21
+ TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
22
+
23
+ /*
24
+ * In the following,
25
+ * dep(a, b, m) -> (a & ~m) | (b & m)
26
+ *
27
+ * Begin with: src = xxxxabcd
28
+ */
29
+ /* tmp = rol32(src, 24) & 0x000000ff = 0000000c */
30
+ tcg_out_rlw(s, RLWINM, tmp, src, 24, 24, 31);
31
+ /* tmp = dep(tmp, rol32(src, 8), 0x0000ff00) = 000000dc */
32
+ tcg_out_rlw(s, RLWIMI, tmp, src, 8, 16, 23);
33
+
34
+ tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
35
+}
36
+
37
/* Emit a move into ret of arg, if it can be done in one insn. */
38
static bool tcg_out_movi_one(TCGContext *s, TCGReg ret, tcg_target_long arg)
39
{
40
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
41
42
case INDEX_op_bswap16_i32:
43
case INDEX_op_bswap16_i64:
44
- a0 = args[0], a1 = args[1];
45
- /* a1 = abcd */
46
- if (a0 != a1) {
47
- /* a0 = (a1 r<< 24) & 0xff # 000c */
48
- tcg_out_rlw(s, RLWINM, a0, a1, 24, 24, 31);
49
- /* a0 = (a0 & ~0xff00) | (a1 r<< 8) & 0xff00 # 00dc */
50
- tcg_out_rlw(s, RLWIMI, a0, a1, 8, 16, 23);
51
- } else {
52
- /* r0 = (a1 r<< 8) & 0xff00 # 00d0 */
53
- tcg_out_rlw(s, RLWINM, TCG_REG_R0, a1, 8, 16, 23);
54
- /* a0 = (a1 r<< 24) & 0xff # 000c */
55
- tcg_out_rlw(s, RLWINM, a0, a1, 24, 24, 31);
56
- /* a0 = a0 | r0 # 00dc */
57
- tcg_out32(s, OR | SAB(TCG_REG_R0, a0, a0));
58
- }
59
+ tcg_out_bswap16(s, args[0], args[1]);
60
break;
61
62
case INDEX_op_bswap32_i32:
63
--
64
2.25.1
65
66
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
4
tcg/ppc/tcg-target.c.inc | 38 ++++++++++++++++++++++----------------
5
1 file changed, 22 insertions(+), 16 deletions(-)
6
1
7
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
8
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/ppc/tcg-target.c.inc
10
+++ b/tcg/ppc/tcg-target.c.inc
11
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src)
12
tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
13
}
14
15
+static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src)
16
+{
17
+ TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
18
+
19
+ /*
20
+ * Stolen from gcc's builtin_bswap32.
21
+ * In the following,
22
+ * dep(a, b, m) -> (a & ~m) | (b & m)
23
+ *
24
+ * Begin with: src = xxxxabcd
25
+ */
26
+ /* tmp = rol32(src, 8) & 0xffffffff = 0000bcda */
27
+ tcg_out_rlw(s, RLWINM, tmp, src, 8, 0, 31);
28
+ /* tmp = dep(tmp, rol32(src, 24), 0xff000000) = 0000dcda */
29
+ tcg_out_rlw(s, RLWIMI, tmp, src, 24, 0, 7);
30
+ /* tmp = dep(tmp, rol32(src, 24), 0x0000ff00) = 0000dcba */
31
+ tcg_out_rlw(s, RLWIMI, tmp, src, 24, 16, 23);
32
+
33
+ tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
34
+}
35
+
36
/* Emit a move into ret of arg, if it can be done in one insn. */
37
static bool tcg_out_movi_one(TCGContext *s, TCGReg ret, tcg_target_long arg)
38
{
39
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
40
case INDEX_op_bswap16_i64:
41
tcg_out_bswap16(s, args[0], args[1]);
42
break;
43
-
44
case INDEX_op_bswap32_i32:
45
case INDEX_op_bswap32_i64:
46
- /* Stolen from gcc's builtin_bswap32 */
47
- a1 = args[1];
48
- a0 = args[0] == a1 ? TCG_REG_R0 : args[0];
49
-
50
- /* a1 = args[1] # abcd */
51
- /* a0 = rotate_left (a1, 8) # bcda */
52
- tcg_out_rlw(s, RLWINM, a0, a1, 8, 0, 31);
53
- /* a0 = (a0 & ~0xff000000) | ((a1 r<< 24) & 0xff000000) # dcda */
54
- tcg_out_rlw(s, RLWIMI, a0, a1, 24, 0, 7);
55
- /* a0 = (a0 & ~0x0000ff00) | ((a1 r<< 24) & 0x0000ff00) # dcba */
56
- tcg_out_rlw(s, RLWIMI, a0, a1, 24, 16, 23);
57
-
58
- if (a0 == TCG_REG_R0) {
59
- tcg_out_mov(s, TCG_TYPE_REG, args[0], a0);
60
- }
61
+ tcg_out_bswap32(s, args[0], args[1]);
62
break;
63
64
case INDEX_op_bswap64_i64:
65
--
66
2.25.1
67
68
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
4
tcg/ppc/tcg-target.c.inc | 64 +++++++++++++++++++++-------------------
5
1 file changed, 34 insertions(+), 30 deletions(-)
6
1
7
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
8
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/ppc/tcg-target.c.inc
10
+++ b/tcg/ppc/tcg-target.c.inc
11
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src)
12
tcg_out_mov(s, TCG_TYPE_REG, dst, tmp);
13
}
14
15
+static void tcg_out_bswap64(TCGContext *s, TCGReg dst, TCGReg src)
16
+{
17
+ TCGReg t0 = dst == src ? TCG_REG_R0 : dst;
18
+ TCGReg t1 = dst == src ? dst : TCG_REG_R0;
19
+
20
+ /*
21
+ * In the following,
22
+ * dep(a, b, m) -> (a & ~m) | (b & m)
23
+ *
24
+ * Begin with: src = abcdefgh
25
+ */
26
+ /* t0 = rol32(src, 8) & 0xffffffff = 0000fghe */
27
+ tcg_out_rlw(s, RLWINM, t0, src, 8, 0, 31);
28
+ /* t0 = dep(t0, rol32(src, 24), 0xff000000) = 0000hghe */
29
+ tcg_out_rlw(s, RLWIMI, t0, src, 24, 0, 7);
30
+ /* t0 = dep(t0, rol32(src, 24), 0x0000ff00) = 0000hgfe */
31
+ tcg_out_rlw(s, RLWIMI, t0, src, 24, 16, 23);
32
+
33
+ /* t0 = rol64(t0, 32) = hgfe0000 */
34
+ tcg_out_rld(s, RLDICL, t0, t0, 32, 0);
35
+ /* t1 = rol64(src, 32) = efghabcd */
36
+ tcg_out_rld(s, RLDICL, t1, src, 32, 0);
37
+
38
+ /* t0 = dep(t0, rol32(t1, 24), 0xffffffff) = hgfebcda */
39
+ tcg_out_rlw(s, RLWIMI, t0, t1, 8, 0, 31);
40
+ /* t0 = dep(t0, rol32(t1, 24), 0xff000000) = hgfedcda */
41
+ tcg_out_rlw(s, RLWIMI, t0, t1, 24, 0, 7);
42
+ /* t0 = dep(t0, rol32(t1, 24), 0x0000ff00) = hgfedcba */
43
+ tcg_out_rlw(s, RLWIMI, t0, t1, 24, 16, 23);
44
+
45
+ tcg_out_mov(s, TCG_TYPE_REG, dst, t0);
46
+}
47
+
48
/* Emit a move into ret of arg, if it can be done in one insn. */
49
static bool tcg_out_movi_one(TCGContext *s, TCGReg ret, tcg_target_long arg)
50
{
51
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
52
case INDEX_op_bswap32_i64:
53
tcg_out_bswap32(s, args[0], args[1]);
54
break;
55
-
56
case INDEX_op_bswap64_i64:
57
- a0 = args[0], a1 = args[1], a2 = TCG_REG_R0;
58
- if (a0 == a1) {
59
- a0 = TCG_REG_R0;
60
- a2 = a1;
61
- }
62
-
63
- /* a1 = # abcd efgh */
64
- /* a0 = rl32(a1, 8) # 0000 fghe */
65
- tcg_out_rlw(s, RLWINM, a0, a1, 8, 0, 31);
66
- /* a0 = dep(a0, rl32(a1, 24), 0xff000000) # 0000 hghe */
67
- tcg_out_rlw(s, RLWIMI, a0, a1, 24, 0, 7);
68
- /* a0 = dep(a0, rl32(a1, 24), 0x0000ff00) # 0000 hgfe */
69
- tcg_out_rlw(s, RLWIMI, a0, a1, 24, 16, 23);
70
-
71
- /* a0 = rl64(a0, 32) # hgfe 0000 */
72
- /* a2 = rl64(a1, 32) # efgh abcd */
73
- tcg_out_rld(s, RLDICL, a0, a0, 32, 0);
74
- tcg_out_rld(s, RLDICL, a2, a1, 32, 0);
75
-
76
- /* a0 = dep(a0, rl32(a2, 8), 0xffffffff) # hgfe bcda */
77
- tcg_out_rlw(s, RLWIMI, a0, a2, 8, 0, 31);
78
- /* a0 = dep(a0, rl32(a2, 24), 0xff000000) # hgfe dcda */
79
- tcg_out_rlw(s, RLWIMI, a0, a2, 24, 0, 7);
80
- /* a0 = dep(a0, rl32(a2, 24), 0x0000ff00) # hgfe dcba */
81
- tcg_out_rlw(s, RLWIMI, a0, a2, 24, 16, 23);
82
-
83
- if (a0 == 0) {
84
- tcg_out_mov(s, TCG_TYPE_REG, args[0], a0);
85
- }
86
+ tcg_out_bswap64(s, args[0], args[1]);
87
break;
88
89
case INDEX_op_deposit_i32:
90
--
91
2.25.1
92
93
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
From: Leandro Lupori <leandro.lupori@eldorado.org.br>
2
3
PowerPC64 processors handle direct branches better than indirect
4
ones, resulting in less stalled cycles and branch misses.
5
6
However, PPC's tb_target_set_jmp_target() was only using direct
7
branches for 16-bit jumps, while PowerPC64's unconditional branch
8
instructions are able to handle displacements of up to 26 bits.
9
To take advantage of this, now jumps whose displacements fit in
10
between 17 and 26 bits are also converted to direct branches.
11
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Leandro Lupori <leandro.lupori@eldorado.org.br>
14
[rth: Expanded some commentary.]
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
15
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
16
---
4
tcg/ppc/tcg-target.c.inc | 34 ++++++++++++++++++++++++++++++++++
17
tcg/ppc/tcg-target.c.inc | 119 +++++++++++++++++++++++++++++----------
5
1 file changed, 34 insertions(+)
18
1 file changed, 88 insertions(+), 31 deletions(-)
6
19
7
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
20
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
8
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/ppc/tcg-target.c.inc
22
--- a/tcg/ppc/tcg-target.c.inc
10
+++ b/tcg/ppc/tcg-target.c.inc
23
+++ b/tcg/ppc/tcg-target.c.inc
11
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
24
@@ -XXX,XX +XXX,XX @@ static void tcg_out_mb(TCGContext *s, TCGArg a0)
12
#define SRAD XO31(794)
25
tcg_out32(s, insn);
13
#define SRADI XO31(413<<1)
14
15
+#define BRH XO31(219)
16
+#define BRW XO31(155)
17
+#define BRD XO31(187)
18
+
19
#define TW XO31( 4)
20
#define TRAP (TW | TO(31))
21
22
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_ext16s(TCGContext *s, TCGReg dst, TCGReg src)
23
tcg_out32(s, EXTSH | RA(dst) | RS(src));
24
}
26
}
25
27
26
+static inline void tcg_out_ext16u(TCGContext *s, TCGReg dst, TCGReg src)
28
+static inline uint64_t make_pair(tcg_insn_unit i1, tcg_insn_unit i2)
27
+{
29
+{
28
+ tcg_out32(s, ANDI | SAI(src, dst, 0xffff));
30
+ if (HOST_BIG_ENDIAN) {
31
+ return (uint64_t)i1 << 32 | i2;
32
+ }
33
+ return (uint64_t)i2 << 32 | i1;
29
+}
34
+}
30
+
35
+
31
static inline void tcg_out_ext32s(TCGContext *s, TCGReg dst, TCGReg src)
36
+static inline void ppc64_replace2(uintptr_t rx, uintptr_t rw,
37
+ tcg_insn_unit i0, tcg_insn_unit i1)
38
+{
39
+#if TCG_TARGET_REG_BITS == 64
40
+ qatomic_set((uint64_t *)rw, make_pair(i0, i1));
41
+ flush_idcache_range(rx, rw, 8);
42
+#else
43
+ qemu_build_not_reached();
44
+#endif
45
+}
46
+
47
+static inline void ppc64_replace4(uintptr_t rx, uintptr_t rw,
48
+ tcg_insn_unit i0, tcg_insn_unit i1,
49
+ tcg_insn_unit i2, tcg_insn_unit i3)
50
+{
51
+ uint64_t p[2];
52
+
53
+ p[!HOST_BIG_ENDIAN] = make_pair(i0, i1);
54
+ p[HOST_BIG_ENDIAN] = make_pair(i2, i3);
55
+
56
+ /*
57
+ * There's no convenient way to get the compiler to allocate a pair
58
+ * of registers at an even index, so copy into r6/r7 and clobber.
59
+ */
60
+ asm("mr %%r6, %1\n\t"
61
+ "mr %%r7, %2\n\t"
62
+ "stq %%r6, %0"
63
+ : "=Q"(*(__int128 *)rw) : "r"(p[0]), "r"(p[1]) : "r6", "r7");
64
+ flush_idcache_range(rx, rw, 16);
65
+}
66
+
67
void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_rx,
68
uintptr_t jmp_rw, uintptr_t addr)
32
{
69
{
33
tcg_out32(s, EXTSW | RA(dst) | RS(src));
70
- if (TCG_TARGET_REG_BITS == 64) {
34
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap16(TCGContext *s, TCGReg dst, TCGReg src, int flags)
71
- tcg_insn_unit i1, i2;
35
{
72
- intptr_t tb_diff = addr - tc_ptr;
36
TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
73
- intptr_t br_diff = addr - (jmp_rx + 4);
37
74
- uint64_t pair;
38
+ if (have_isa_3_10) {
75
+ tcg_insn_unit i0, i1, i2, i3;
39
+ tcg_out32(s, BRH | RA(dst) | RS(src));
76
+ intptr_t tb_diff = addr - tc_ptr;
40
+ if (flags & TCG_BSWAP_OS) {
77
+ intptr_t br_diff = addr - (jmp_rx + 4);
41
+ tcg_out_ext16s(s, dst, dst);
78
+ intptr_t lo, hi;
42
+ } else if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
79
43
+ tcg_out_ext16u(s, dst, dst);
80
- /* This does not exercise the range of the branch, but we do
44
+ }
81
- still need to be able to load the new value of TCG_REG_TB.
82
- But this does still happen quite often. */
83
- if (tb_diff == (int16_t)tb_diff) {
84
- i1 = ADDI | TAI(TCG_REG_TB, TCG_REG_TB, tb_diff);
85
- i2 = B | (br_diff & 0x3fffffc);
86
- } else {
87
- intptr_t lo = (int16_t)tb_diff;
88
- intptr_t hi = (int32_t)(tb_diff - lo);
89
- assert(tb_diff == hi + lo);
90
- i1 = ADDIS | TAI(TCG_REG_TB, TCG_REG_TB, hi >> 16);
91
- i2 = ADDI | TAI(TCG_REG_TB, TCG_REG_TB, lo);
92
- }
93
-#if HOST_BIG_ENDIAN
94
- pair = (uint64_t)i1 << 32 | i2;
95
-#else
96
- pair = (uint64_t)i2 << 32 | i1;
97
-#endif
98
-
99
- /* As per the enclosing if, this is ppc64. Avoid the _Static_assert
100
- within qatomic_set that would fail to build a ppc32 host. */
101
- qatomic_set__nocheck((uint64_t *)jmp_rw, pair);
102
- flush_idcache_range(jmp_rx, jmp_rw, 8);
103
- } else {
104
+ if (TCG_TARGET_REG_BITS == 32) {
105
intptr_t diff = addr - jmp_rx;
106
tcg_debug_assert(in_range_b(diff));
107
qatomic_set((uint32_t *)jmp_rw, B | (diff & 0x3fffffc));
108
flush_idcache_range(jmp_rx, jmp_rw, 4);
109
+ return;
110
}
111
+
112
+ /*
113
+ * For 16-bit displacements, we can use a single add + branch.
114
+ * This happens quite often.
115
+ */
116
+ if (tb_diff == (int16_t)tb_diff) {
117
+ i0 = ADDI | TAI(TCG_REG_TB, TCG_REG_TB, tb_diff);
118
+ i1 = B | (br_diff & 0x3fffffc);
119
+ ppc64_replace2(jmp_rx, jmp_rw, i0, i1);
45
+ return;
120
+ return;
46
+ }
121
+ }
47
+
122
+
48
/*
123
+ lo = (int16_t)tb_diff;
49
* In the following,
124
+ hi = (int32_t)(tb_diff - lo);
50
* dep(a, b, m) -> (a & ~m) | (b & m)
125
+ assert(tb_diff == hi + lo);
51
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap32(TCGContext *s, TCGReg dst, TCGReg src, int flags)
126
+ i0 = ADDIS | TAI(TCG_REG_TB, TCG_REG_TB, hi >> 16);
52
{
127
+ i1 = ADDI | TAI(TCG_REG_TB, TCG_REG_TB, lo);
53
TCGReg tmp = dst == src ? TCG_REG_R0 : dst;
128
+
54
129
+ /*
55
+ if (have_isa_3_10) {
130
+ * Without stq from 2.07, we can only update two insns,
56
+ tcg_out32(s, BRW | RA(dst) | RS(src));
131
+ * and those must be the ones that load the target address.
57
+ if (flags & TCG_BSWAP_OS) {
132
+ */
58
+ tcg_out_ext32s(s, dst, dst);
133
+ if (!have_isa_2_07) {
59
+ } else if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
134
+ ppc64_replace2(jmp_rx, jmp_rw, i0, i1);
60
+ tcg_out_ext32u(s, dst, dst);
61
+ }
62
+ return;
135
+ return;
63
+ }
136
+ }
64
+
137
+
65
/*
138
+ /*
66
* Stolen from gcc's builtin_bswap32.
139
+ * For 26-bit displacements, we can use a direct branch.
67
* In the following,
140
+ * Otherwise we still need the indirect branch, which we
68
@@ -XXX,XX +XXX,XX @@ static void tcg_out_bswap64(TCGContext *s, TCGReg dst, TCGReg src)
141
+ * must restore after a potential direct branch write.
69
TCGReg t0 = dst == src ? TCG_REG_R0 : dst;
142
+ */
70
TCGReg t1 = dst == src ? dst : TCG_REG_R0;
143
+ br_diff -= 4;
71
144
+ if (in_range_b(br_diff)) {
72
+ if (have_isa_3_10) {
145
+ i2 = B | (br_diff & 0x3fffffc);
73
+ tcg_out32(s, BRD | RA(dst) | RS(src));
146
+ i3 = NOP;
74
+ return;
147
+ } else {
148
+ i2 = MTSPR | RS(TCG_REG_TB) | CTR;
149
+ i3 = BCCTR | BO_ALWAYS;
75
+ }
150
+ }
76
+
151
+ ppc64_replace4(jmp_rx, jmp_rw, i0, i1, i2, i3);
77
/*
152
}
78
* In the following,
153
79
* dep(a, b, m) -> (a & ~m) | (b & m)
154
static void tcg_out_call_int(TCGContext *s, int lk,
155
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
156
if (s->tb_jmp_insn_offset) {
157
/* Direct jump. */
158
if (TCG_TARGET_REG_BITS == 64) {
159
- /* Ensure the next insns are 8-byte aligned. */
160
- if ((uintptr_t)s->code_ptr & 7) {
161
+ /* Ensure the next insns are 8 or 16-byte aligned. */
162
+ while ((uintptr_t)s->code_ptr & (have_isa_2_07 ? 15 : 7)) {
163
tcg_out32(s, NOP);
164
}
165
s->tb_jmp_insn_offset[args[0]] = tcg_current_code_size(s);
80
--
166
--
81
2.25.1
167
2.34.1
82
83
diff view generated by jsdifflib
Deleted patch
1
For INDEX_op_bswap16_i64, use 64-bit instructions so that we can
2
easily provide the extension to 64-bits. Drop the special case,
3
previously used, where the input is already zero-extended -- the
4
minor code size savings is not worth the complication.
5
1
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
tcg/s390/tcg-target.c.inc | 34 ++++++++++++++++++++++++++++------
10
1 file changed, 28 insertions(+), 6 deletions(-)
11
12
diff --git a/tcg/s390/tcg-target.c.inc b/tcg/s390/tcg-target.c.inc
13
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/s390/tcg-target.c.inc
15
+++ b/tcg/s390/tcg-target.c.inc
16
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
17
tgen_ext16u(s, TCG_TYPE_I32, args[0], args[1]);
18
break;
19
20
- OP_32_64(bswap16):
21
- /* The TCG bswap definition requires bits 0-47 already be zero.
22
- Thus we don't need the G-type insns to implement bswap16_i64. */
23
- tcg_out_insn(s, RRE, LRVR, args[0], args[1]);
24
- tcg_out_sh32(s, RS_SRL, args[0], TCG_REG_NONE, 16);
25
+ case INDEX_op_bswap16_i32:
26
+ a0 = args[0], a1 = args[1], a2 = args[2];
27
+ tcg_out_insn(s, RRE, LRVR, a0, a1);
28
+ if (a2 & TCG_BSWAP_OS) {
29
+ tcg_out_sh32(s, RS_SRA, a0, TCG_REG_NONE, 16);
30
+ } else {
31
+ tcg_out_sh32(s, RS_SRL, a0, TCG_REG_NONE, 16);
32
+ }
33
break;
34
- OP_32_64(bswap32):
35
+ case INDEX_op_bswap16_i64:
36
+ a0 = args[0], a1 = args[1], a2 = args[2];
37
+ tcg_out_insn(s, RRE, LRVGR, a0, a1);
38
+ if (a2 & TCG_BSWAP_OS) {
39
+ tcg_out_sh64(s, RSY_SRAG, a0, a0, TCG_REG_NONE, 48);
40
+ } else {
41
+ tcg_out_sh64(s, RSY_SRLG, a0, a0, TCG_REG_NONE, 48);
42
+ }
43
+ break;
44
+
45
+ case INDEX_op_bswap32_i32:
46
tcg_out_insn(s, RRE, LRVR, args[0], args[1]);
47
break;
48
+ case INDEX_op_bswap32_i64:
49
+ a0 = args[0], a1 = args[1], a2 = args[2];
50
+ tcg_out_insn(s, RRE, LRVR, a0, a1);
51
+ if (a2 & TCG_BSWAP_OS) {
52
+ tgen_ext32s(s, a0, a0);
53
+ } else if ((a2 & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
54
+ tgen_ext32u(s, a0, a0);
55
+ }
56
+ break;
57
58
case INDEX_op_add2_i32:
59
if (const_args[4]) {
60
--
61
2.25.1
62
63
diff view generated by jsdifflib
Deleted patch
1
Notice when the input is known to be zero-extended and force
2
the TCG_BSWAP_IZ flag on. Honor the TCG_BSWAP_OS bit during
3
constant folding. Propagate the input to the output mask.
4
1
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/optimize.c | 56 +++++++++++++++++++++++++++++++++++++++++++++-----
9
1 file changed, 51 insertions(+), 5 deletions(-)
10
11
diff --git a/tcg/optimize.c b/tcg/optimize.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/optimize.c
14
+++ b/tcg/optimize.c
15
@@ -XXX,XX +XXX,XX @@ static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y)
16
return (uint16_t)x;
17
18
CASE_OP_32_64(bswap16):
19
- return bswap16(x);
20
+ x = bswap16(x);
21
+ return y & TCG_BSWAP_OS ? (int16_t)x : x;
22
23
CASE_OP_32_64(bswap32):
24
- return bswap32(x);
25
+ x = bswap32(x);
26
+ return y & TCG_BSWAP_OS ? (int32_t)x : x;
27
28
case INDEX_op_bswap64_i64:
29
return bswap64(x);
30
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
31
}
32
break;
33
34
+ CASE_OP_32_64(bswap16):
35
+ mask = arg_info(op->args[1])->mask;
36
+ if (mask <= 0xffff) {
37
+ op->args[2] |= TCG_BSWAP_IZ;
38
+ }
39
+ mask = bswap16(mask);
40
+ switch (op->args[2] & (TCG_BSWAP_OZ | TCG_BSWAP_OS)) {
41
+ case TCG_BSWAP_OZ:
42
+ break;
43
+ case TCG_BSWAP_OS:
44
+ mask = (int16_t)mask;
45
+ break;
46
+ default: /* undefined high bits */
47
+ mask |= MAKE_64BIT_MASK(16, 48);
48
+ break;
49
+ }
50
+ break;
51
+
52
+ case INDEX_op_bswap32_i64:
53
+ mask = arg_info(op->args[1])->mask;
54
+ if (mask <= 0xffffffffu) {
55
+ op->args[2] |= TCG_BSWAP_IZ;
56
+ }
57
+ mask = bswap32(mask);
58
+ switch (op->args[2] & (TCG_BSWAP_OZ | TCG_BSWAP_OS)) {
59
+ case TCG_BSWAP_OZ:
60
+ break;
61
+ case TCG_BSWAP_OS:
62
+ mask = (int32_t)mask;
63
+ break;
64
+ default: /* undefined high bits */
65
+ mask |= MAKE_64BIT_MASK(32, 32);
66
+ break;
67
+ }
68
+ break;
69
+
70
default:
71
break;
72
}
73
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
74
CASE_OP_32_64(ext16s):
75
CASE_OP_32_64(ext16u):
76
CASE_OP_32_64(ctpop):
77
- CASE_OP_32_64(bswap16):
78
- CASE_OP_32_64(bswap32):
79
- case INDEX_op_bswap64_i64:
80
case INDEX_op_ext32s_i64:
81
case INDEX_op_ext32u_i64:
82
case INDEX_op_ext_i32_i64:
83
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
84
}
85
goto do_default;
86
87
+ CASE_OP_32_64(bswap16):
88
+ CASE_OP_32_64(bswap32):
89
+ case INDEX_op_bswap64_i64:
90
+ if (arg_is_const(op->args[1])) {
91
+ tmp = do_constant_folding(opc, arg_info(op->args[1])->val,
92
+ op->args[2]);
93
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
94
+ break;
95
+ }
96
+ goto do_default;
97
+
98
CASE_OP_32_64(add):
99
CASE_OP_32_64(sub):
100
CASE_OP_32_64(mul):
101
--
102
2.25.1
103
104
diff view generated by jsdifflib
1
Implement the new semantics in the fallback expansion.
1
The value previously chosen overlaps GUSA_MASK.
2
Change all callers to supply the flags that keep the
3
semantics unchanged locally.
4
2
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Rename all DELAY_SLOT_* and GUSA_* defines to emphasize
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
that they are included in TB_FLAGs. Add aliases for the
5
FPSCR and SR bits that are included in TB_FLAGS, so that
6
we don't accidentally reassign those bits.
7
8
Fixes: 4da06fb3062 ("target/sh4: Implement prctl_unalign_sigbus")
9
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/856
10
Reviewed-by: Yoshinori Sato <ysato@users.sourceforge.jp>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
12
---
9
include/tcg/tcg-op.h | 8 +--
13
target/sh4/cpu.h | 56 +++++++++++++------------
10
target/arm/translate-a64.c | 12 ++--
14
linux-user/sh4/signal.c | 6 +--
11
target/arm/translate.c | 2 +-
15
target/sh4/cpu.c | 6 +--
12
target/i386/tcg/translate.c | 2 +-
16
target/sh4/helper.c | 6 +--
13
target/mips/tcg/mxu_translate.c | 2 +-
17
target/sh4/translate.c | 90 ++++++++++++++++++++++-------------------
14
target/s390x/translate.c | 4 +-
18
5 files changed, 88 insertions(+), 76 deletions(-)
15
target/sh4/translate.c | 2 +-
16
tcg/tcg-op.c | 121 ++++++++++++++++++++++----------
17
8 files changed, 99 insertions(+), 54 deletions(-)
18
19
19
diff --git a/include/tcg/tcg-op.h b/include/tcg/tcg-op.h
20
diff --git a/target/sh4/cpu.h b/target/sh4/cpu.h
20
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
21
--- a/include/tcg/tcg-op.h
22
--- a/target/sh4/cpu.h
22
+++ b/include/tcg/tcg-op.h
23
+++ b/target/sh4/cpu.h
23
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ext8s_i32(TCGv_i32 ret, TCGv_i32 arg);
24
@@ -XXX,XX +XXX,XX @@
24
void tcg_gen_ext16s_i32(TCGv_i32 ret, TCGv_i32 arg);
25
#define FPSCR_RM_NEAREST (0 << 0)
25
void tcg_gen_ext8u_i32(TCGv_i32 ret, TCGv_i32 arg);
26
#define FPSCR_RM_ZERO (1 << 0)
26
void tcg_gen_ext16u_i32(TCGv_i32 ret, TCGv_i32 arg);
27
27
-void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg);
28
-#define DELAY_SLOT_MASK 0x7
28
+void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg, int flags);
29
-#define DELAY_SLOT (1 << 0)
29
void tcg_gen_bswap32_i32(TCGv_i32 ret, TCGv_i32 arg);
30
-#define DELAY_SLOT_CONDITIONAL (1 << 1)
30
void tcg_gen_smin_i32(TCGv_i32, TCGv_i32 arg1, TCGv_i32 arg2);
31
-#define DELAY_SLOT_RTE (1 << 2)
31
void tcg_gen_smax_i32(TCGv_i32, TCGv_i32 arg1, TCGv_i32 arg2);
32
+#define TB_FLAG_DELAY_SLOT (1 << 0)
32
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ext32s_i64(TCGv_i64 ret, TCGv_i64 arg);
33
+#define TB_FLAG_DELAY_SLOT_COND (1 << 1)
33
void tcg_gen_ext8u_i64(TCGv_i64 ret, TCGv_i64 arg);
34
+#define TB_FLAG_DELAY_SLOT_RTE (1 << 2)
34
void tcg_gen_ext16u_i64(TCGv_i64 ret, TCGv_i64 arg);
35
+#define TB_FLAG_PENDING_MOVCA (1 << 3)
35
void tcg_gen_ext32u_i64(TCGv_i64 ret, TCGv_i64 arg);
36
+#define TB_FLAG_GUSA_SHIFT 4 /* [11:4] */
36
-void tcg_gen_bswap16_i64(TCGv_i64 ret, TCGv_i64 arg);
37
+#define TB_FLAG_GUSA_EXCLUSIVE (1 << 12)
37
-void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg);
38
+#define TB_FLAG_UNALIGN (1 << 13)
38
+void tcg_gen_bswap16_i64(TCGv_i64 ret, TCGv_i64 arg, int flags);
39
+#define TB_FLAG_SR_FD (1 << SR_FD) /* 15 */
39
+void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg, int flags);
40
+#define TB_FLAG_FPSCR_PR FPSCR_PR /* 19 */
40
void tcg_gen_bswap64_i64(TCGv_i64 ret, TCGv_i64 arg);
41
+#define TB_FLAG_FPSCR_SZ FPSCR_SZ /* 20 */
41
void tcg_gen_smin_i64(TCGv_i64, TCGv_i64 arg1, TCGv_i64 arg2);
42
+#define TB_FLAG_FPSCR_FR FPSCR_FR /* 21 */
42
void tcg_gen_smax_i64(TCGv_i64, TCGv_i64 arg1, TCGv_i64 arg2);
43
+#define TB_FLAG_SR_RB (1 << SR_RB) /* 29 */
43
@@ -XXX,XX +XXX,XX @@ void tcg_gen_stl_vec(TCGv_vec r, TCGv_ptr base, TCGArg offset, TCGType t);
44
+#define TB_FLAG_SR_MD (1 << SR_MD) /* 30 */
44
#define tcg_gen_ext32u_tl tcg_gen_mov_i32
45
45
#define tcg_gen_ext32s_tl tcg_gen_mov_i32
46
-#define TB_FLAG_PENDING_MOVCA (1 << 3)
46
#define tcg_gen_bswap16_tl tcg_gen_bswap16_i32
47
-#define TB_FLAG_UNALIGN (1 << 4)
47
-#define tcg_gen_bswap32_tl tcg_gen_bswap32_i32
48
-
48
+#define tcg_gen_bswap32_tl(D, S, F) tcg_gen_bswap32_i32(D, S)
49
-#define GUSA_SHIFT 4
49
#define tcg_gen_bswap_tl tcg_gen_bswap32_i32
50
-#ifdef CONFIG_USER_ONLY
50
#define tcg_gen_concat_tl_i64 tcg_gen_concat_i32_i64
51
-#define GUSA_EXCLUSIVE (1 << 12)
51
#define tcg_gen_extr_i64_tl tcg_gen_extr_i64_i32
52
-#define GUSA_MASK ((0xff << GUSA_SHIFT) | GUSA_EXCLUSIVE)
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
53
-#else
54
-/* Provide dummy versions of the above to allow tests against tbflags
55
- to be elided while avoiding ifdefs. */
56
-#define GUSA_EXCLUSIVE 0
57
-#define GUSA_MASK 0
58
-#endif
59
-
60
-#define TB_FLAG_ENVFLAGS_MASK (DELAY_SLOT_MASK | GUSA_MASK)
61
+#define TB_FLAG_DELAY_SLOT_MASK (TB_FLAG_DELAY_SLOT | \
62
+ TB_FLAG_DELAY_SLOT_COND | \
63
+ TB_FLAG_DELAY_SLOT_RTE)
64
+#define TB_FLAG_GUSA_MASK ((0xff << TB_FLAG_GUSA_SHIFT) | \
65
+ TB_FLAG_GUSA_EXCLUSIVE)
66
+#define TB_FLAG_FPSCR_MASK (TB_FLAG_FPSCR_PR | \
67
+ TB_FLAG_FPSCR_SZ | \
68
+ TB_FLAG_FPSCR_FR)
69
+#define TB_FLAG_SR_MASK (TB_FLAG_SR_FD | \
70
+ TB_FLAG_SR_RB | \
71
+ TB_FLAG_SR_MD)
72
+#define TB_FLAG_ENVFLAGS_MASK (TB_FLAG_DELAY_SLOT_MASK | \
73
+ TB_FLAG_GUSA_MASK)
74
75
typedef struct tlb_t {
76
uint32_t vpn;        /* virtual page number */
77
@@ -XXX,XX +XXX,XX @@ static inline int cpu_mmu_index (CPUSH4State *env, bool ifetch)
78
{
79
/* The instruction in a RTE delay slot is fetched in privileged
80
mode, but executed in user mode. */
81
- if (ifetch && (env->flags & DELAY_SLOT_RTE)) {
82
+ if (ifetch && (env->flags & TB_FLAG_DELAY_SLOT_RTE)) {
83
return 0;
84
} else {
85
return (env->sr & (1u << SR_MD)) == 0 ? 1 : 0;
86
@@ -XXX,XX +XXX,XX @@ static inline void cpu_get_tb_cpu_state(CPUSH4State *env, target_ulong *pc,
87
{
88
*pc = env->pc;
89
/* For a gUSA region, notice the end of the region. */
90
- *cs_base = env->flags & GUSA_MASK ? env->gregs[0] : 0;
91
- *flags = env->flags /* TB_FLAG_ENVFLAGS_MASK: bits 0-2, 4-12 */
92
- | (env->fpscr & (FPSCR_FR | FPSCR_SZ | FPSCR_PR)) /* Bits 19-21 */
93
- | (env->sr & ((1u << SR_MD) | (1u << SR_RB))) /* Bits 29-30 */
94
- | (env->sr & (1u << SR_FD)) /* Bit 15 */
95
+ *cs_base = env->flags & TB_FLAG_GUSA_MASK ? env->gregs[0] : 0;
96
+ *flags = env->flags
97
+ | (env->fpscr & TB_FLAG_FPSCR_MASK)
98
+ | (env->sr & TB_FLAG_SR_MASK)
99
| (env->movcal_backup ? TB_FLAG_PENDING_MOVCA : 0); /* Bit 3 */
100
#ifdef CONFIG_USER_ONLY
101
*flags |= TB_FLAG_UNALIGN * !env_cpu(env)->prctl_unalign_sigbus;
102
diff --git a/linux-user/sh4/signal.c b/linux-user/sh4/signal.c
53
index XXXXXXX..XXXXXXX 100644
103
index XXXXXXX..XXXXXXX 100644
54
--- a/target/arm/translate-a64.c
104
--- a/linux-user/sh4/signal.c
55
+++ b/target/arm/translate-a64.c
105
+++ b/linux-user/sh4/signal.c
56
@@ -XXX,XX +XXX,XX @@ static void handle_rev32(DisasContext *s, unsigned int sf,
106
@@ -XXX,XX +XXX,XX @@ static void restore_sigcontext(CPUSH4State *regs, struct target_sigcontext *sc)
57
107
__get_user(regs->fpul, &sc->sc_fpul);
58
/* bswap32_i64 requires zero high word */
108
59
tcg_gen_ext32u_i64(tcg_tmp, tcg_rn);
109
regs->tra = -1; /* disable syscall checks */
60
- tcg_gen_bswap32_i64(tcg_rd, tcg_tmp);
110
- regs->flags &= ~(DELAY_SLOT_MASK | GUSA_MASK);
61
+ tcg_gen_bswap32_i64(tcg_rd, tcg_tmp, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
111
+ regs->flags = 0;
62
tcg_gen_shri_i64(tcg_tmp, tcg_rn, 32);
63
- tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp);
64
+ tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
65
tcg_gen_concat32_i64(tcg_rd, tcg_rd, tcg_tmp);
66
67
tcg_temp_free_i64(tcg_tmp);
68
} else {
69
tcg_gen_ext32u_i64(tcg_rd, cpu_reg(s, rn));
70
- tcg_gen_bswap32_i64(tcg_rd, tcg_rd);
71
+ tcg_gen_bswap32_i64(tcg_rd, tcg_rd, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
72
}
73
}
112
}
74
113
75
@@ -XXX,XX +XXX,XX @@ static void handle_rev(DisasContext *s, int opcode, bool u,
114
void setup_frame(int sig, struct target_sigaction *ka,
76
read_vec_element(s, tcg_tmp, rn, i, grp_size);
115
@@ -XXX,XX +XXX,XX @@ void setup_frame(int sig, struct target_sigaction *ka,
77
switch (grp_size) {
116
regs->gregs[5] = 0;
78
case MO_16:
117
regs->gregs[6] = frame_addr += offsetof(typeof(*frame), sc);
79
- tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp);
118
regs->pc = (unsigned long) ka->_sa_handler;
80
+ tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp,
119
- regs->flags &= ~(DELAY_SLOT_MASK | GUSA_MASK);
81
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
120
+ regs->flags &= ~(TB_FLAG_DELAY_SLOT_MASK | TB_FLAG_GUSA_MASK);
82
break;
121
83
case MO_32:
122
unlock_user_struct(frame, frame_addr, 1);
84
- tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp);
123
return;
85
+ tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp,
124
@@ -XXX,XX +XXX,XX @@ void setup_rt_frame(int sig, struct target_sigaction *ka,
86
+ TCG_BSWAP_IZ | TCG_BSWAP_OZ);
125
regs->gregs[5] = frame_addr + offsetof(typeof(*frame), info);
87
break;
126
regs->gregs[6] = frame_addr + offsetof(typeof(*frame), uc);
88
case MO_64:
127
regs->pc = (unsigned long) ka->_sa_handler;
89
tcg_gen_bswap64_i64(tcg_tmp, tcg_tmp);
128
- regs->flags &= ~(DELAY_SLOT_MASK | GUSA_MASK);
90
diff --git a/target/arm/translate.c b/target/arm/translate.c
129
+ regs->flags &= ~(TB_FLAG_DELAY_SLOT_MASK | TB_FLAG_GUSA_MASK);
130
131
unlock_user_struct(frame, frame_addr, 1);
132
return;
133
diff --git a/target/sh4/cpu.c b/target/sh4/cpu.c
91
index XXXXXXX..XXXXXXX 100644
134
index XXXXXXX..XXXXXXX 100644
92
--- a/target/arm/translate.c
135
--- a/target/sh4/cpu.c
93
+++ b/target/arm/translate.c
136
+++ b/target/sh4/cpu.c
94
@@ -XXX,XX +XXX,XX @@ void gen_rev16(TCGv_i32 dest, TCGv_i32 var)
137
@@ -XXX,XX +XXX,XX @@ static void superh_cpu_synchronize_from_tb(CPUState *cs,
95
static void gen_revsh(TCGv_i32 dest, TCGv_i32 var)
138
SuperHCPU *cpu = SUPERH_CPU(cs);
96
{
139
97
tcg_gen_ext16u_i32(var, var);
140
cpu->env.pc = tb_pc(tb);
98
- tcg_gen_bswap16_i32(var, var);
141
- cpu->env.flags = tb->flags & TB_FLAG_ENVFLAGS_MASK;
99
+ tcg_gen_bswap16_i32(var, var, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
142
+ cpu->env.flags = tb->flags;
100
tcg_gen_ext16s_i32(dest, var);
101
}
143
}
102
144
103
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
145
#ifndef CONFIG_USER_ONLY
146
@@ -XXX,XX +XXX,XX @@ static bool superh_io_recompile_replay_branch(CPUState *cs,
147
SuperHCPU *cpu = SUPERH_CPU(cs);
148
CPUSH4State *env = &cpu->env;
149
150
- if ((env->flags & ((DELAY_SLOT | DELAY_SLOT_CONDITIONAL))) != 0
151
+ if ((env->flags & (TB_FLAG_DELAY_SLOT | TB_FLAG_DELAY_SLOT_COND))
152
&& env->pc != tb_pc(tb)) {
153
env->pc -= 2;
154
- env->flags &= ~(DELAY_SLOT | DELAY_SLOT_CONDITIONAL);
155
+ env->flags &= ~(TB_FLAG_DELAY_SLOT | TB_FLAG_DELAY_SLOT_COND);
156
return true;
157
}
158
return false;
159
diff --git a/target/sh4/helper.c b/target/sh4/helper.c
104
index XXXXXXX..XXXXXXX 100644
160
index XXXXXXX..XXXXXXX 100644
105
--- a/target/i386/tcg/translate.c
161
--- a/target/sh4/helper.c
106
+++ b/target/i386/tcg/translate.c
162
+++ b/target/sh4/helper.c
107
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
163
@@ -XXX,XX +XXX,XX @@ void superh_cpu_do_interrupt(CPUState *cs)
108
{
164
env->sr |= (1u << SR_BL) | (1u << SR_MD) | (1u << SR_RB);
109
gen_op_mov_v_reg(s, MO_32, s->T0, reg);
165
env->lock_addr = -1;
110
tcg_gen_ext32u_tl(s->T0, s->T0);
166
111
- tcg_gen_bswap32_tl(s->T0, s->T0);
167
- if (env->flags & DELAY_SLOT_MASK) {
112
+ tcg_gen_bswap32_tl(s->T0, s->T0, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
168
+ if (env->flags & TB_FLAG_DELAY_SLOT_MASK) {
113
gen_op_mov_reg_v(s, MO_32, reg, s->T0);
169
/* Branch instruction should be executed again before delay slot. */
114
}
170
    env->spc -= 2;
115
break;
171
    /* Clear flags for exception/interrupt routine. */
116
diff --git a/target/mips/tcg/mxu_translate.c b/target/mips/tcg/mxu_translate.c
172
- env->flags &= ~DELAY_SLOT_MASK;
117
index XXXXXXX..XXXXXXX 100644
173
+ env->flags &= ~TB_FLAG_DELAY_SLOT_MASK;
118
--- a/target/mips/tcg/mxu_translate.c
174
}
119
+++ b/target/mips/tcg/mxu_translate.c
175
120
@@ -XXX,XX +XXX,XX @@ static void gen_mxu_s32ldd_s32lddr(DisasContext *ctx)
176
if (do_exp) {
121
177
@@ -XXX,XX +XXX,XX @@ bool superh_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
122
if (sel == 1) {
178
CPUSH4State *env = &cpu->env;
123
/* S32LDDR */
179
124
- tcg_gen_bswap32_tl(t1, t1);
180
/* Delay slots are indivisible, ignore interrupts */
125
+ tcg_gen_bswap32_tl(t1, t1, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
181
- if (env->flags & DELAY_SLOT_MASK) {
126
}
182
+ if (env->flags & TB_FLAG_DELAY_SLOT_MASK) {
127
gen_store_mxu_gpr(t1, XRa);
183
return false;
128
184
} else {
129
diff --git a/target/s390x/translate.c b/target/s390x/translate.c
185
superh_cpu_do_interrupt(cs);
130
index XXXXXXX..XXXXXXX 100644
131
--- a/target/s390x/translate.c
132
+++ b/target/s390x/translate.c
133
@@ -XXX,XX +XXX,XX @@ static DisasJumpType op_rosbg(DisasContext *s, DisasOps *o)
134
135
static DisasJumpType op_rev16(DisasContext *s, DisasOps *o)
136
{
137
- tcg_gen_bswap16_i64(o->out, o->in2);
138
+ tcg_gen_bswap16_i64(o->out, o->in2, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
139
return DISAS_NEXT;
140
}
141
142
static DisasJumpType op_rev32(DisasContext *s, DisasOps *o)
143
{
144
- tcg_gen_bswap32_i64(o->out, o->in2);
145
+ tcg_gen_bswap32_i64(o->out, o->in2, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
146
return DISAS_NEXT;
147
}
148
149
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
186
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
150
index XXXXXXX..XXXXXXX 100644
187
index XXXXXXX..XXXXXXX 100644
151
--- a/target/sh4/translate.c
188
--- a/target/sh4/translate.c
152
+++ b/target/sh4/translate.c
189
+++ b/target/sh4/translate.c
153
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
190
@@ -XXX,XX +XXX,XX @@ void superh_cpu_dump_state(CPUState *cs, FILE *f, int flags)
154
    {
191
         i, env->gregs[i], i + 1, env->gregs[i + 1],
155
TCGv low = tcg_temp_new();
192
         i + 2, env->gregs[i + 2], i + 3, env->gregs[i + 3]);
156
     tcg_gen_ext16u_i32(low, REG(B7_4));
193
}
157
-     tcg_gen_bswap16_i32(low, low);
194
- if (env->flags & DELAY_SLOT) {
158
+     tcg_gen_bswap16_i32(low, low, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
195
+ if (env->flags & TB_FLAG_DELAY_SLOT) {
159
tcg_gen_deposit_i32(REG(B11_8), REG(B7_4), low, 0, 16);
196
qemu_printf("in delay slot (delayed_pc=0x%08x)\n",
160
     tcg_temp_free(low);
197
         env->delayed_pc);
161
    }
198
- } else if (env->flags & DELAY_SLOT_CONDITIONAL) {
162
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
199
+ } else if (env->flags & TB_FLAG_DELAY_SLOT_COND) {
163
index XXXXXXX..XXXXXXX 100644
200
qemu_printf("in conditional delay slot (delayed_pc=0x%08x)\n",
164
--- a/tcg/tcg-op.c
201
         env->delayed_pc);
165
+++ b/tcg/tcg-op.c
202
- } else if (env->flags & DELAY_SLOT_RTE) {
166
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ext16u_i32(TCGv_i32 ret, TCGv_i32 arg)
203
+ } else if (env->flags & TB_FLAG_DELAY_SLOT_RTE) {
167
}
204
qemu_fprintf(f, "in rte delay slot (delayed_pc=0x%08x)\n",
205
env->delayed_pc);
206
}
207
@@ -XXX,XX +XXX,XX @@ static inline void gen_save_cpu_state(DisasContext *ctx, bool save_pc)
208
209
static inline bool use_exit_tb(DisasContext *ctx)
210
{
211
- return (ctx->tbflags & GUSA_EXCLUSIVE) != 0;
212
+ return (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) != 0;
168
}
213
}
169
214
170
-/* Note: we assume the two high bytes are set to zero */
215
static bool use_goto_tb(DisasContext *ctx, target_ulong dest)
171
-void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg)
216
@@ -XXX,XX +XXX,XX @@ static void gen_conditional_jump(DisasContext *ctx, target_ulong dest,
172
+void tcg_gen_bswap16_i32(TCGv_i32 ret, TCGv_i32 arg, int flags)
217
TCGLabel *l1 = gen_new_label();
218
TCGCond cond_not_taken = jump_if_true ? TCG_COND_EQ : TCG_COND_NE;
219
220
- if (ctx->tbflags & GUSA_EXCLUSIVE) {
221
+ if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) {
222
/* When in an exclusive region, we must continue to the end.
223
Therefore, exit the region on a taken branch, but otherwise
224
fall through to the next instruction. */
225
tcg_gen_brcondi_i32(cond_not_taken, cpu_sr_t, 0, l1);
226
- tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~GUSA_MASK);
227
+ tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK);
228
/* Note that this won't actually use a goto_tb opcode because we
229
disallow it in use_goto_tb, but it handles exit + singlestep. */
230
gen_goto_tb(ctx, 0, dest);
231
@@ -XXX,XX +XXX,XX @@ static void gen_delayed_conditional_jump(DisasContext * ctx)
232
tcg_gen_mov_i32(ds, cpu_delayed_cond);
233
tcg_gen_discard_i32(cpu_delayed_cond);
234
235
- if (ctx->tbflags & GUSA_EXCLUSIVE) {
236
+ if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) {
237
/* When in an exclusive region, we must continue to the end.
238
Therefore, exit the region on a taken branch, but otherwise
239
fall through to the next instruction. */
240
tcg_gen_brcondi_i32(TCG_COND_EQ, ds, 0, l1);
241
242
/* Leave the gUSA region. */
243
- tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~GUSA_MASK);
244
+ tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK);
245
gen_jump(ctx);
246
247
gen_set_label(l1);
248
@@ -XXX,XX +XXX,XX @@ static inline void gen_store_fpr64(DisasContext *ctx, TCGv_i64 t, int reg)
249
#define XHACK(x) ((((x) & 1 ) << 4) | ((x) & 0xe))
250
251
#define CHECK_NOT_DELAY_SLOT \
252
- if (ctx->envflags & DELAY_SLOT_MASK) { \
253
- goto do_illegal_slot; \
254
+ if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { \
255
+ goto do_illegal_slot; \
256
}
257
258
#define CHECK_PRIVILEGED \
259
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
260
case 0x000b:        /* rts */
261
    CHECK_NOT_DELAY_SLOT
262
    tcg_gen_mov_i32(cpu_delayed_pc, cpu_pr);
263
- ctx->envflags |= DELAY_SLOT;
264
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
265
    ctx->delayed_pc = (uint32_t) - 1;
266
    return;
267
case 0x0028:        /* clrmac */
268
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
269
    CHECK_NOT_DELAY_SLOT
270
gen_write_sr(cpu_ssr);
271
    tcg_gen_mov_i32(cpu_delayed_pc, cpu_spc);
272
- ctx->envflags |= DELAY_SLOT_RTE;
273
+ ctx->envflags |= TB_FLAG_DELAY_SLOT_RTE;
274
    ctx->delayed_pc = (uint32_t) - 1;
275
ctx->base.is_jmp = DISAS_STOP;
276
    return;
277
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
278
    return;
279
case 0xe000:        /* mov #imm,Rn */
280
#ifdef CONFIG_USER_ONLY
281
- /* Detect the start of a gUSA region. If so, update envflags
282
- and end the TB. This will allow us to see the end of the
283
- region (stored in R0) in the next TB. */
284
+ /*
285
+ * Detect the start of a gUSA region (mov #-n, r15).
286
+ * If so, update envflags and end the TB. This will allow us
287
+ * to see the end of the region (stored in R0) in the next TB.
288
+ */
289
if (B11_8 == 15 && B7_0s < 0 &&
290
(tb_cflags(ctx->base.tb) & CF_PARALLEL)) {
291
- ctx->envflags = deposit32(ctx->envflags, GUSA_SHIFT, 8, B7_0s);
292
+ ctx->envflags =
293
+ deposit32(ctx->envflags, TB_FLAG_GUSA_SHIFT, 8, B7_0s);
294
ctx->base.is_jmp = DISAS_STOP;
295
}
296
#endif
297
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
298
case 0xa000:        /* bra disp */
299
    CHECK_NOT_DELAY_SLOT
300
ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2;
301
- ctx->envflags |= DELAY_SLOT;
302
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
303
    return;
304
case 0xb000:        /* bsr disp */
305
    CHECK_NOT_DELAY_SLOT
306
tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4);
307
ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2;
308
- ctx->envflags |= DELAY_SLOT;
309
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
310
    return;
311
}
312
313
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
314
    CHECK_NOT_DELAY_SLOT
315
tcg_gen_xori_i32(cpu_delayed_cond, cpu_sr_t, 1);
316
ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2;
317
- ctx->envflags |= DELAY_SLOT_CONDITIONAL;
318
+ ctx->envflags |= TB_FLAG_DELAY_SLOT_COND;
319
    return;
320
case 0x8900:        /* bt label */
321
    CHECK_NOT_DELAY_SLOT
322
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
323
    CHECK_NOT_DELAY_SLOT
324
tcg_gen_mov_i32(cpu_delayed_cond, cpu_sr_t);
325
ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2;
326
- ctx->envflags |= DELAY_SLOT_CONDITIONAL;
327
+ ctx->envflags |= TB_FLAG_DELAY_SLOT_COND;
328
    return;
329
case 0x8800:        /* cmp/eq #imm,R0 */
330
tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, REG(0), B7_0s);
331
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
332
case 0x0023:        /* braf Rn */
333
    CHECK_NOT_DELAY_SLOT
334
tcg_gen_addi_i32(cpu_delayed_pc, REG(B11_8), ctx->base.pc_next + 4);
335
- ctx->envflags |= DELAY_SLOT;
336
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
337
    ctx->delayed_pc = (uint32_t) - 1;
338
    return;
339
case 0x0003:        /* bsrf Rn */
340
    CHECK_NOT_DELAY_SLOT
341
tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4);
342
    tcg_gen_add_i32(cpu_delayed_pc, REG(B11_8), cpu_pr);
343
- ctx->envflags |= DELAY_SLOT;
344
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
345
    ctx->delayed_pc = (uint32_t) - 1;
346
    return;
347
case 0x4015:        /* cmp/pl Rn */
348
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
349
case 0x402b:        /* jmp @Rn */
350
    CHECK_NOT_DELAY_SLOT
351
    tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8));
352
- ctx->envflags |= DELAY_SLOT;
353
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
354
    ctx->delayed_pc = (uint32_t) - 1;
355
    return;
356
case 0x400b:        /* jsr @Rn */
357
    CHECK_NOT_DELAY_SLOT
358
tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4);
359
    tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8));
360
- ctx->envflags |= DELAY_SLOT;
361
+ ctx->envflags |= TB_FLAG_DELAY_SLOT;
362
    ctx->delayed_pc = (uint32_t) - 1;
363
    return;
364
case 0x400e:        /* ldc Rm,SR */
365
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
366
fflush(stderr);
367
#endif
368
do_illegal:
369
- if (ctx->envflags & DELAY_SLOT_MASK) {
370
+ if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) {
371
do_illegal_slot:
372
gen_save_cpu_state(ctx, true);
373
gen_helper_raise_slot_illegal_instruction(cpu_env);
374
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
375
376
do_fpu_disabled:
377
gen_save_cpu_state(ctx, true);
378
- if (ctx->envflags & DELAY_SLOT_MASK) {
379
+ if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) {
380
gen_helper_raise_slot_fpu_disable(cpu_env);
381
} else {
382
gen_helper_raise_fpu_disable(cpu_env);
383
@@ -XXX,XX +XXX,XX @@ static void decode_opc(DisasContext * ctx)
384
385
_decode_opc(ctx);
386
387
- if (old_flags & DELAY_SLOT_MASK) {
388
+ if (old_flags & TB_FLAG_DELAY_SLOT_MASK) {
389
/* go out of the delay slot */
390
- ctx->envflags &= ~DELAY_SLOT_MASK;
391
+ ctx->envflags &= ~TB_FLAG_DELAY_SLOT_MASK;
392
393
/* When in an exclusive region, we must continue to the end
394
for conditional branches. */
395
- if (ctx->tbflags & GUSA_EXCLUSIVE
396
- && old_flags & DELAY_SLOT_CONDITIONAL) {
397
+ if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE
398
+ && old_flags & TB_FLAG_DELAY_SLOT_COND) {
399
gen_delayed_conditional_jump(ctx);
400
return;
401
}
402
/* Otherwise this is probably an invalid gUSA region.
403
Drop the GUSA bits so the next TB doesn't see them. */
404
- ctx->envflags &= ~GUSA_MASK;
405
+ ctx->envflags &= ~TB_FLAG_GUSA_MASK;
406
407
tcg_gen_movi_i32(cpu_flags, ctx->envflags);
408
- if (old_flags & DELAY_SLOT_CONDITIONAL) {
409
+ if (old_flags & TB_FLAG_DELAY_SLOT_COND) {
410
     gen_delayed_conditional_jump(ctx);
411
} else {
412
gen_jump(ctx);
413
@@ -XXX,XX +XXX,XX @@ static void decode_gusa(DisasContext *ctx, CPUSH4State *env)
414
}
415
416
/* The entire region has been translated. */
417
- ctx->envflags &= ~GUSA_MASK;
418
+ ctx->envflags &= ~TB_FLAG_GUSA_MASK;
419
ctx->base.pc_next = pc_end;
420
ctx->base.num_insns += max_insns - 1;
421
return;
422
@@ -XXX,XX +XXX,XX @@ static void decode_gusa(DisasContext *ctx, CPUSH4State *env)
423
424
/* Restart with the EXCLUSIVE bit set, within a TB run via
425
cpu_exec_step_atomic holding the exclusive lock. */
426
- ctx->envflags |= GUSA_EXCLUSIVE;
427
+ ctx->envflags |= TB_FLAG_GUSA_EXCLUSIVE;
428
gen_save_cpu_state(ctx, false);
429
gen_helper_exclusive(cpu_env);
430
ctx->base.is_jmp = DISAS_NORETURN;
431
@@ -XXX,XX +XXX,XX @@ static void sh4_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
432
(tbflags & (1 << SR_RB))) * 0x10;
433
ctx->fbank = tbflags & FPSCR_FR ? 0x10 : 0;
434
435
- if (tbflags & GUSA_MASK) {
436
+#ifdef CONFIG_USER_ONLY
437
+ if (tbflags & TB_FLAG_GUSA_MASK) {
438
+ /* In gUSA exclusive region. */
439
uint32_t pc = ctx->base.pc_next;
440
uint32_t pc_end = ctx->base.tb->cs_base;
441
- int backup = sextract32(ctx->tbflags, GUSA_SHIFT, 8);
442
+ int backup = sextract32(ctx->tbflags, TB_FLAG_GUSA_SHIFT, 8);
443
int max_insns = (pc_end - pc) / 2;
444
445
if (pc != pc_end + backup || max_insns < 2) {
446
/* This is a malformed gUSA region. Don't do anything special,
447
since the interpreter is likely to get confused. */
448
- ctx->envflags &= ~GUSA_MASK;
449
- } else if (tbflags & GUSA_EXCLUSIVE) {
450
+ ctx->envflags &= ~TB_FLAG_GUSA_MASK;
451
+ } else if (tbflags & TB_FLAG_GUSA_EXCLUSIVE) {
452
/* Regardless of single-stepping or the end of the page,
453
we must complete execution of the gUSA region while
454
holding the exclusive lock. */
455
@@ -XXX,XX +XXX,XX @@ static void sh4_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
456
return;
457
}
458
}
459
+#endif
460
461
/* Since the ISA is fixed-width, we can bound by the number
462
of instructions remaining on the page. */
463
@@ -XXX,XX +XXX,XX @@ static void sh4_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
464
DisasContext *ctx = container_of(dcbase, DisasContext, base);
465
466
#ifdef CONFIG_USER_ONLY
467
- if (unlikely(ctx->envflags & GUSA_MASK)
468
- && !(ctx->envflags & GUSA_EXCLUSIVE)) {
469
+ if (unlikely(ctx->envflags & TB_FLAG_GUSA_MASK)
470
+ && !(ctx->envflags & TB_FLAG_GUSA_EXCLUSIVE)) {
471
/* We're in an gUSA region, and we have not already fallen
472
back on using an exclusive region. Attempt to parse the
473
region into a single supported atomic operation. Failure
474
@@ -XXX,XX +XXX,XX @@ static void sh4_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
173
{
475
{
174
+ /* Only one extension flag may be present. */
476
DisasContext *ctx = container_of(dcbase, DisasContext, base);
175
+ tcg_debug_assert(!(flags & TCG_BSWAP_OS) || !(flags & TCG_BSWAP_OZ));
477
176
+
478
- if (ctx->tbflags & GUSA_EXCLUSIVE) {
177
if (TCG_TARGET_HAS_bswap16_i32) {
479
+ if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) {
178
- tcg_gen_op3i_i32(INDEX_op_bswap16_i32, ret, arg,
480
/* Ending the region of exclusivity. Clear the bits. */
179
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
481
- ctx->envflags &= ~GUSA_MASK;
180
+ tcg_gen_op3i_i32(INDEX_op_bswap16_i32, ret, arg, flags);
482
+ ctx->envflags &= ~TB_FLAG_GUSA_MASK;
181
} else {
483
}
182
TCGv_i32 t0 = tcg_temp_new_i32();
484
183
+ TCGv_i32 t1 = tcg_temp_new_i32();
485
switch (ctx->base.is_jmp) {
184
185
- tcg_gen_ext8u_i32(t0, arg);
186
- tcg_gen_shli_i32(t0, t0, 8);
187
- tcg_gen_shri_i32(ret, arg, 8);
188
- tcg_gen_or_i32(ret, ret, t0);
189
+ tcg_gen_shri_i32(t0, arg, 8);
190
+ if (!(flags & TCG_BSWAP_IZ)) {
191
+ tcg_gen_ext8u_i32(t0, t0);
192
+ }
193
+
194
+ if (flags & TCG_BSWAP_OS) {
195
+ tcg_gen_shli_i32(t1, arg, 24);
196
+ tcg_gen_sari_i32(t1, t1, 16);
197
+ } else if (flags & TCG_BSWAP_OZ) {
198
+ tcg_gen_ext8u_i32(t1, arg);
199
+ tcg_gen_shli_i32(t1, t1, 8);
200
+ } else {
201
+ tcg_gen_shli_i32(t1, arg, 8);
202
+ }
203
+
204
+ tcg_gen_or_i32(ret, t0, t1);
205
tcg_temp_free_i32(t0);
206
+ tcg_temp_free_i32(t1);
207
}
208
}
209
210
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ext32u_i64(TCGv_i64 ret, TCGv_i64 arg)
211
}
212
}
213
214
-/* Note: we assume the six high bytes are set to zero */
215
-void tcg_gen_bswap16_i64(TCGv_i64 ret, TCGv_i64 arg)
216
+void tcg_gen_bswap16_i64(TCGv_i64 ret, TCGv_i64 arg, int flags)
217
{
218
+ /* Only one extension flag may be present. */
219
+ tcg_debug_assert(!(flags & TCG_BSWAP_OS) || !(flags & TCG_BSWAP_OZ));
220
+
221
if (TCG_TARGET_REG_BITS == 32) {
222
- tcg_gen_bswap16_i32(TCGV_LOW(ret), TCGV_LOW(arg));
223
- tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
224
+ tcg_gen_bswap16_i32(TCGV_LOW(ret), TCGV_LOW(arg), flags);
225
+ if (flags & TCG_BSWAP_OS) {
226
+ tcg_gen_sari_i32(TCGV_HIGH(ret), TCGV_LOW(ret), 31);
227
+ } else {
228
+ tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
229
+ }
230
} else if (TCG_TARGET_HAS_bswap16_i64) {
231
- tcg_gen_op3i_i64(INDEX_op_bswap16_i64, ret, arg,
232
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
233
+ tcg_gen_op3i_i64(INDEX_op_bswap16_i64, ret, arg, flags);
234
} else {
235
TCGv_i64 t0 = tcg_temp_new_i64();
236
+ TCGv_i64 t1 = tcg_temp_new_i64();
237
238
- tcg_gen_ext8u_i64(t0, arg);
239
- tcg_gen_shli_i64(t0, t0, 8);
240
- tcg_gen_shri_i64(ret, arg, 8);
241
- tcg_gen_or_i64(ret, ret, t0);
242
+ tcg_gen_shri_i64(t0, arg, 8);
243
+ if (!(flags & TCG_BSWAP_IZ)) {
244
+ tcg_gen_ext8u_i64(t0, t0);
245
+ }
246
+
247
+ if (flags & TCG_BSWAP_OS) {
248
+ tcg_gen_shli_i64(t1, arg, 56);
249
+ tcg_gen_sari_i64(t1, t1, 48);
250
+ } else if (flags & TCG_BSWAP_OZ) {
251
+ tcg_gen_ext8u_i64(t1, arg);
252
+ tcg_gen_shli_i64(t1, t1, 8);
253
+ } else {
254
+ tcg_gen_shli_i64(t1, arg, 8);
255
+ }
256
+
257
+ tcg_gen_or_i64(ret, t0, t1);
258
tcg_temp_free_i64(t0);
259
+ tcg_temp_free_i64(t1);
260
}
261
}
262
263
-/* Note: we assume the four high bytes are set to zero */
264
-void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg)
265
+void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg, int flags)
266
{
267
+ /* Only one extension flag may be present. */
268
+ tcg_debug_assert(!(flags & TCG_BSWAP_OS) || !(flags & TCG_BSWAP_OZ));
269
+
270
if (TCG_TARGET_REG_BITS == 32) {
271
tcg_gen_bswap32_i32(TCGV_LOW(ret), TCGV_LOW(arg));
272
- tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
273
+ if (flags & TCG_BSWAP_OS) {
274
+ tcg_gen_sari_i32(TCGV_HIGH(ret), TCGV_LOW(ret), 31);
275
+ } else {
276
+ tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
277
+ }
278
} else if (TCG_TARGET_HAS_bswap32_i64) {
279
- tcg_gen_op3i_i64(INDEX_op_bswap32_i64, ret, arg,
280
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
281
+ tcg_gen_op3i_i64(INDEX_op_bswap32_i64, ret, arg, flags);
282
} else {
283
TCGv_i64 t0 = tcg_temp_new_i64();
284
TCGv_i64 t1 = tcg_temp_new_i64();
285
TCGv_i64 t2 = tcg_constant_i64(0x00ff00ff);
286
287
- /* arg = ....abcd */
288
- tcg_gen_shri_i64(t0, arg, 8); /* t0 = .....abc */
289
- tcg_gen_and_i64(t1, arg, t2); /* t1 = .....b.d */
290
- tcg_gen_and_i64(t0, t0, t2); /* t0 = .....a.c */
291
- tcg_gen_shli_i64(t1, t1, 8); /* t1 = ....b.d. */
292
- tcg_gen_or_i64(ret, t0, t1); /* ret = ....badc */
293
+ /* arg = xxxxabcd */
294
+ tcg_gen_shri_i64(t0, arg, 8); /* t0 = .xxxxabc */
295
+ tcg_gen_and_i64(t1, arg, t2); /* t1 = .....b.d */
296
+ tcg_gen_and_i64(t0, t0, t2); /* t0 = .....a.c */
297
+ tcg_gen_shli_i64(t1, t1, 8); /* t1 = ....b.d. */
298
+ tcg_gen_or_i64(ret, t0, t1); /* ret = ....badc */
299
300
- tcg_gen_shli_i64(t1, ret, 48); /* t1 = dc...... */
301
- tcg_gen_shri_i64(t0, ret, 16); /* t0 = ......ba */
302
- tcg_gen_shri_i64(t1, t1, 32); /* t1 = ....dc.. */
303
- tcg_gen_or_i64(ret, t0, t1); /* ret = ....dcba */
304
+ tcg_gen_shli_i64(t1, ret, 48); /* t1 = dc...... */
305
+ tcg_gen_shri_i64(t0, ret, 16); /* t0 = ......ba */
306
+ if (flags & TCG_BSWAP_OS) {
307
+ tcg_gen_sari_i64(t1, t1, 32); /* t1 = ssssdc.. */
308
+ } else {
309
+ tcg_gen_shri_i64(t1, t1, 32); /* t1 = ....dc.. */
310
+ }
311
+ tcg_gen_or_i64(ret, t0, t1); /* ret = ssssdcba */
312
313
tcg_temp_free_i64(t0);
314
tcg_temp_free_i64(t1);
315
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
316
if ((orig_memop ^ memop) & MO_BSWAP) {
317
switch (orig_memop & MO_SIZE) {
318
case MO_16:
319
- tcg_gen_bswap16_i32(val, val);
320
+ tcg_gen_bswap16_i32(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
321
if (orig_memop & MO_SIGN) {
322
tcg_gen_ext16s_i32(val, val);
323
}
324
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_st_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
325
switch (memop & MO_SIZE) {
326
case MO_16:
327
tcg_gen_ext16u_i32(swap, val);
328
- tcg_gen_bswap16_i32(swap, swap);
329
+ tcg_gen_bswap16_i32(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
330
break;
331
case MO_32:
332
tcg_gen_bswap32_i32(swap, val);
333
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
334
if ((orig_memop ^ memop) & MO_BSWAP) {
335
switch (orig_memop & MO_SIZE) {
336
case MO_16:
337
- tcg_gen_bswap16_i64(val, val);
338
+ tcg_gen_bswap16_i64(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
339
if (orig_memop & MO_SIGN) {
340
tcg_gen_ext16s_i64(val, val);
341
}
342
break;
343
case MO_32:
344
- tcg_gen_bswap32_i64(val, val);
345
+ tcg_gen_bswap32_i64(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
346
if (orig_memop & MO_SIGN) {
347
tcg_gen_ext32s_i64(val, val);
348
}
349
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_st_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
350
switch (memop & MO_SIZE) {
351
case MO_16:
352
tcg_gen_ext16u_i64(swap, val);
353
- tcg_gen_bswap16_i64(swap, swap);
354
+ tcg_gen_bswap16_i64(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
355
break;
356
case MO_32:
357
tcg_gen_ext32u_i64(swap, val);
358
- tcg_gen_bswap32_i64(swap, swap);
359
+ tcg_gen_bswap32_i64(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
360
break;
361
case MO_64:
362
tcg_gen_bswap64_i64(swap, val);
363
--
486
--
364
2.25.1
487
2.34.1
365
366
diff view generated by jsdifflib
Deleted patch
1
We can perform any required sign-extension via TCG_BSWAP_OS.
2
1
3
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/tcg-op.c | 24 ++++++++++--------------
8
1 file changed, 10 insertions(+), 14 deletions(-)
9
10
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/tcg-op.c
13
+++ b/tcg/tcg-op.c
14
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
15
orig_memop = memop;
16
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
17
memop &= ~MO_BSWAP;
18
- /* The bswap primitive requires zero-extended input. */
19
+ /* The bswap primitive benefits from zero-extended input. */
20
if ((memop & MO_SSIZE) == MO_SW) {
21
memop &= ~MO_SIGN;
22
}
23
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
24
if ((orig_memop ^ memop) & MO_BSWAP) {
25
switch (orig_memop & MO_SIZE) {
26
case MO_16:
27
- tcg_gen_bswap16_i32(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
28
- if (orig_memop & MO_SIGN) {
29
- tcg_gen_ext16s_i32(val, val);
30
- }
31
+ tcg_gen_bswap16_i32(val, val, (orig_memop & MO_SIGN
32
+ ? TCG_BSWAP_IZ | TCG_BSWAP_OS
33
+ : TCG_BSWAP_IZ | TCG_BSWAP_OZ));
34
break;
35
case MO_32:
36
tcg_gen_bswap32_i32(val, val);
37
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
38
orig_memop = memop;
39
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
40
memop &= ~MO_BSWAP;
41
- /* The bswap primitive requires zero-extended input. */
42
+ /* The bswap primitive benefits from zero-extended input. */
43
if ((memop & MO_SIGN) && (memop & MO_SIZE) < MO_64) {
44
memop &= ~MO_SIGN;
45
}
46
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_ld_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
47
plugin_gen_mem_callbacks(addr, info);
48
49
if ((orig_memop ^ memop) & MO_BSWAP) {
50
+ int flags = (orig_memop & MO_SIGN
51
+ ? TCG_BSWAP_IZ | TCG_BSWAP_OS
52
+ : TCG_BSWAP_IZ | TCG_BSWAP_OZ);
53
switch (orig_memop & MO_SIZE) {
54
case MO_16:
55
- tcg_gen_bswap16_i64(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
56
- if (orig_memop & MO_SIGN) {
57
- tcg_gen_ext16s_i64(val, val);
58
- }
59
+ tcg_gen_bswap16_i64(val, val, flags);
60
break;
61
case MO_32:
62
- tcg_gen_bswap32_i64(val, val, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
63
- if (orig_memop & MO_SIGN) {
64
- tcg_gen_ext32s_i64(val, val);
65
- }
66
+ tcg_gen_bswap32_i64(val, val, flags);
67
break;
68
case MO_64:
69
tcg_gen_bswap64_i64(val, val);
70
--
71
2.25.1
72
73
diff view generated by jsdifflib
Deleted patch
1
By removing TCG_BSWAP_IZ we indicate that the input is
2
not zero-extended, and thus can remove an explicit extend.
3
By removing TCG_BSWAP_OZ, we allow the implementation to
4
leave high bits set, which will be ignored by the store.
5
1
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
tcg/tcg-op.c | 9 +++------
10
1 file changed, 3 insertions(+), 6 deletions(-)
11
12
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/tcg-op.c
15
+++ b/tcg/tcg-op.c
16
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_st_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
17
swap = tcg_temp_new_i32();
18
switch (memop & MO_SIZE) {
19
case MO_16:
20
- tcg_gen_ext16u_i32(swap, val);
21
- tcg_gen_bswap16_i32(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
22
+ tcg_gen_bswap16_i32(swap, val, 0);
23
break;
24
case MO_32:
25
tcg_gen_bswap32_i32(swap, val);
26
@@ -XXX,XX +XXX,XX @@ void tcg_gen_qemu_st_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
27
swap = tcg_temp_new_i64();
28
switch (memop & MO_SIZE) {
29
case MO_16:
30
- tcg_gen_ext16u_i64(swap, val);
31
- tcg_gen_bswap16_i64(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
32
+ tcg_gen_bswap16_i64(swap, val, 0);
33
break;
34
case MO_32:
35
- tcg_gen_ext32u_i64(swap, val);
36
- tcg_gen_bswap32_i64(swap, swap, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
37
+ tcg_gen_bswap32_i64(swap, val, 0);
38
break;
39
case MO_64:
40
tcg_gen_bswap64_i64(swap, val);
41
--
42
2.25.1
43
44
diff view generated by jsdifflib
Deleted patch
1
We can eliminate the requirement for a zero-extended output,
2
because the following store will ignore any garbage high bits.
3
1
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/arm/translate-a64.c | 6 ++----
9
1 file changed, 2 insertions(+), 4 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void handle_rev(DisasContext *s, int opcode, bool u,
16
read_vec_element(s, tcg_tmp, rn, i, grp_size);
17
switch (grp_size) {
18
case MO_16:
19
- tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp,
20
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
21
+ tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp, TCG_BSWAP_IZ);
22
break;
23
case MO_32:
24
- tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp,
25
- TCG_BSWAP_IZ | TCG_BSWAP_OZ);
26
+ tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp, TCG_BSWAP_IZ);
27
break;
28
case MO_64:
29
tcg_gen_bswap64_i64(tcg_tmp, tcg_tmp);
30
--
31
2.25.1
32
33
diff view generated by jsdifflib
Deleted patch
1
Use a break instead of an ifdefed else.
2
There's no need to move the values through s->T0.
3
Remove TCG_BSWAP_IZ and the preceding zero-extension.
4
1
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
target/i386/tcg/translate.c | 14 ++++----------
9
1 file changed, 4 insertions(+), 10 deletions(-)
10
11
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/i386/tcg/translate.c
14
+++ b/target/i386/tcg/translate.c
15
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
16
reg = (b & 7) | REX_B(s);
17
#ifdef TARGET_X86_64
18
if (dflag == MO_64) {
19
- gen_op_mov_v_reg(s, MO_64, s->T0, reg);
20
- tcg_gen_bswap64_i64(s->T0, s->T0);
21
- gen_op_mov_reg_v(s, MO_64, reg, s->T0);
22
- } else
23
-#endif
24
- {
25
- gen_op_mov_v_reg(s, MO_32, s->T0, reg);
26
- tcg_gen_ext32u_tl(s->T0, s->T0);
27
- tcg_gen_bswap32_tl(s->T0, s->T0, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
28
- gen_op_mov_reg_v(s, MO_32, reg, s->T0);
29
+ tcg_gen_bswap64_i64(cpu_regs[reg], cpu_regs[reg]);
30
+ break;
31
}
32
+#endif
33
+ tcg_gen_bswap32_tl(cpu_regs[reg], cpu_regs[reg], TCG_BSWAP_OZ);
34
break;
35
case 0xd6: /* salc */
36
if (CODE64(s))
37
--
38
2.25.1
39
40
diff view generated by jsdifflib
Deleted patch
1
Remove TCG_BSWAP_IZ and the preceding zero-extension.
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
6
target/sh4/translate.c | 3 +--
7
1 file changed, 1 insertion(+), 2 deletions(-)
8
9
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
10
index XXXXXXX..XXXXXXX 100644
11
--- a/target/sh4/translate.c
12
+++ b/target/sh4/translate.c
13
@@ -XXX,XX +XXX,XX @@ static void _decode_opc(DisasContext * ctx)
14
case 0x6008:        /* swap.b Rm,Rn */
15
    {
16
TCGv low = tcg_temp_new();
17
-     tcg_gen_ext16u_i32(low, REG(B7_4));
18
-     tcg_gen_bswap16_i32(low, low, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
19
+ tcg_gen_bswap16_i32(low, REG(B7_4), 0);
20
tcg_gen_deposit_i32(REG(B11_8), REG(B7_4), low, 0, 16);
21
     tcg_temp_free(low);
22
    }
23
--
24
2.25.1
25
26
diff view generated by jsdifflib
Deleted patch
1
There were two bugs here: (1) the required endianness was
2
not present in the MemOp, and (2) we were not providing a
3
zero-extended input to the bswap as semantics required.
4
1
5
The best fix is to fold the bswap into the memory operation,
6
producing the desired result directly.
7
8
Acked-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
11
target/mips/tcg/mxu_translate.c | 6 +-----
12
1 file changed, 1 insertion(+), 5 deletions(-)
13
14
diff --git a/target/mips/tcg/mxu_translate.c b/target/mips/tcg/mxu_translate.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/mips/tcg/mxu_translate.c
17
+++ b/target/mips/tcg/mxu_translate.c
18
@@ -XXX,XX +XXX,XX @@ static void gen_mxu_s32ldd_s32lddr(DisasContext *ctx)
19
tcg_gen_ori_tl(t1, t1, 0xFFFFF000);
20
}
21
tcg_gen_add_tl(t1, t0, t1);
22
- tcg_gen_qemu_ld_tl(t1, t1, ctx->mem_idx, MO_SL);
23
+ tcg_gen_qemu_ld_tl(t1, t1, ctx->mem_idx, MO_TESL ^ (sel * MO_BSWAP));
24
25
- if (sel == 1) {
26
- /* S32LDDR */
27
- tcg_gen_bswap32_tl(t1, t1, TCG_BSWAP_IZ | TCG_BSWAP_OZ);
28
- }
29
gen_store_mxu_gpr(t1, XRa);
30
31
tcg_temp_free(t0);
32
--
33
2.25.1
34
35
diff view generated by jsdifflib