1
The following changes since commit 45240eed4f064576d589ea60ebadf3c11d7ab891:
1
The following changes since commit e93ded1bf6c94ab95015b33e188bc8b0b0c32670:
2
2
3
Merge remote-tracking branch 'remotes/armbru/tags/pull-yank-2021-01-13' into staging (2021-01-13 14:19:24 +0000)
3
Merge tag 'testing-pull-request-2022-08-30' of https://gitlab.com/thuth/qemu into staging (2022-08-31 18:19:03 -0400)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20210113
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220901
8
8
9
for you to fetch changes up to 4cacecaaa2bbf8af0967bd3eee43297fada475a9:
9
for you to fetch changes up to 20011be2e30b8aa8ef1fc258485f00c688703deb:
10
10
11
decodetree: Open files with encoding='utf-8' (2021-01-13 08:39:08 -1000)
11
target/riscv: Make translator stop before the end of a page (2022-09-01 07:43:08 +0100)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Improvements to tcg constant handling.
14
Respect PROT_EXEC in user-only mode.
15
Force utf8 for decodetree.
15
Fix s390x, i386 and riscv for translations crossing a page.
16
16
17
----------------------------------------------------------------
17
----------------------------------------------------------------
18
Philippe Mathieu-Daudé (1):
18
Ilya Leoshkevich (4):
19
decodetree: Open files with encoding='utf-8'
19
linux-user: Clear translations on mprotect()
20
accel/tcg: Introduce is_same_page()
21
target/s390x: Make translator stop before the end of a page
22
target/i386: Make translator stop before the end of a page
20
23
21
Richard Henderson (23):
24
Richard Henderson (16):
22
tcg: Use tcg_out_dupi_vec from temp_load
25
linux-user/arm: Mark the commpage executable
23
tcg: Increase tcg_out_dupi_vec immediate to int64_t
26
linux-user/hppa: Allocate page zero as a commpage
24
tcg: Consolidate 3 bits into enum TCGTempKind
27
linux-user/x86_64: Allocate vsyscall page as a commpage
25
tcg: Add temp_readonly
28
linux-user: Honor PT_GNU_STACK
26
tcg: Expand TCGTemp.val to 64-bits
29
tests/tcg/i386: Move smc_code2 to an executable section
27
tcg: Rename struct tcg_temp_info to TempOptInfo
30
accel/tcg: Properly implement get_page_addr_code for user-only
28
tcg: Expand TempOptInfo to 64-bits
31
accel/tcg: Unlock mmap_lock after longjmp
29
tcg: Introduce TYPE_CONST temporaries
32
accel/tcg: Make tb_htable_lookup static
30
tcg/optimize: Improve find_better_copy
33
accel/tcg: Move qemu_ram_addr_from_host_nofail to physmem.c
31
tcg/optimize: Adjust TempOptInfo allocation
34
accel/tcg: Use probe_access_internal for softmmu get_page_addr_code_hostp
32
tcg/optimize: Use tcg_constant_internal with constant folding
35
accel/tcg: Document the faulting lookup in tb_lookup_cmp
33
tcg: Convert tcg_gen_dupi_vec to TCG_CONST
36
accel/tcg: Remove translator_ldsw
34
tcg: Use tcg_constant_i32 with icount expander
37
accel/tcg: Add pc and host_pc params to gen_intermediate_code
35
tcg: Use tcg_constant_{i32,i64} with tcg int expanders
38
accel/tcg: Add fast path for translator_ld*
36
tcg: Use tcg_constant_{i32,i64} with tcg plugins
39
target/riscv: Add MAX_INSN_LEN and insn_len
37
tcg: Use tcg_constant_{i32,i64,vec} with gvec expanders
40
target/riscv: Make translator stop before the end of a page
38
tcg/tci: Add special tci_movi_{i32,i64} opcodes
39
tcg: Remove movi and dupi opcodes
40
tcg: Add tcg_reg_alloc_dup2
41
tcg/i386: Use tcg_constant_vec with tcg vec expanders
42
tcg: Remove tcg_gen_dup{8,16,32,64}i_vec
43
tcg/ppc: Use tcg_constant_vec with tcg vec expanders
44
tcg/aarch64: Use tcg_constant_vec with tcg vec expanders
45
41
46
include/exec/gen-icount.h | 25 +--
42
include/elf.h | 1 +
47
include/tcg/tcg-op.h | 17 +-
43
include/exec/cpu-common.h | 1 +
48
include/tcg/tcg-opc.h | 11 +-
44
include/exec/exec-all.h | 89 ++++++++----------------
49
include/tcg/tcg.h | 50 ++++-
45
include/exec/translator.h | 96 ++++++++++++++++---------
50
accel/tcg/plugin-gen.c | 49 ++---
46
linux-user/arm/target_cpu.h | 4 +-
51
tcg/optimize.c | 249 +++++++++++-----------
47
linux-user/qemu.h | 1 +
52
tcg/tcg-op-gvec.c | 129 +++++-------
48
accel/tcg/cpu-exec.c | 143 ++++++++++++++++++++------------------
53
tcg/tcg-op-vec.c | 52 +----
49
accel/tcg/cputlb.c | 93 +++++++------------------
54
tcg/tcg-op.c | 227 ++++++++++----------
50
accel/tcg/translate-all.c | 29 ++++----
55
tcg/tcg.c | 488 +++++++++++++++++++++++++++++++++----------
51
accel/tcg/translator.c | 135 ++++++++++++++++++++++++++---------
56
tcg/tci.c | 4 +-
52
accel/tcg/user-exec.c | 17 ++++-
57
tcg/aarch64/tcg-target.c.inc | 32 +--
53
linux-user/elfload.c | 82 ++++++++++++++++++++--
58
tcg/arm/tcg-target.c.inc | 1 -
54
linux-user/mmap.c | 6 +-
59
tcg/i386/tcg-target.c.inc | 112 ++++++----
55
softmmu/physmem.c | 12 ++++
60
tcg/mips/tcg-target.c.inc | 2 -
56
target/alpha/translate.c | 5 +-
61
tcg/ppc/tcg-target.c.inc | 90 ++++----
57
target/arm/translate.c | 5 +-
62
tcg/riscv/tcg-target.c.inc | 2 -
58
target/avr/translate.c | 5 +-
63
tcg/s390/tcg-target.c.inc | 2 -
59
target/cris/translate.c | 5 +-
64
tcg/sparc/tcg-target.c.inc | 2 -
60
target/hexagon/translate.c | 6 +-
65
tcg/tci/tcg-target.c.inc | 6 +-
61
target/hppa/translate.c | 5 +-
66
scripts/decodetree.py | 9 +-
62
target/i386/tcg/translate.c | 71 +++++++++++--------
67
21 files changed, 890 insertions(+), 669 deletions(-)
63
target/loongarch/translate.c | 6 +-
68
64
target/m68k/translate.c | 5 +-
65
target/microblaze/translate.c | 5 +-
66
target/mips/tcg/translate.c | 5 +-
67
target/nios2/translate.c | 5 +-
68
target/openrisc/translate.c | 6 +-
69
target/ppc/translate.c | 5 +-
70
target/riscv/translate.c | 32 +++++++--
71
target/rx/translate.c | 5 +-
72
target/s390x/tcg/translate.c | 20 ++++--
73
target/sh4/translate.c | 5 +-
74
target/sparc/translate.c | 5 +-
75
target/tricore/translate.c | 6 +-
76
target/xtensa/translate.c | 6 +-
77
tests/tcg/i386/test-i386.c | 2 +-
78
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++
79
tests/tcg/s390x/noexec.c | 106 ++++++++++++++++++++++++++++
80
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++
81
tests/tcg/multiarch/noexec.c.inc | 139 ++++++++++++++++++++++++++++++++++++
82
tests/tcg/riscv64/Makefile.target | 1 +
83
tests/tcg/s390x/Makefile.target | 1 +
84
tests/tcg/x86_64/Makefile.target | 3 +-
85
43 files changed, 966 insertions(+), 367 deletions(-)
86
create mode 100644 tests/tcg/riscv64/noexec.c
87
create mode 100644 tests/tcg/s390x/noexec.c
88
create mode 100644 tests/tcg/x86_64/noexec.c
89
create mode 100644 tests/tcg/multiarch/noexec.c.inc
diff view generated by jsdifflib
1
This propagates the extended value of TCGTemp.val that we did before.
1
We're about to start validating PAGE_EXEC, which means
2
In addition, it will be required for vector constants.
2
that we've got to mark the commpage executable. We had
3
been placing the commpage outside of reserved_va, which
4
was incorrect and lead to an abort.
3
5
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
9
---
6
tcg/optimize.c | 40 +++++++++++++++++++++-------------------
10
linux-user/arm/target_cpu.h | 4 ++--
7
1 file changed, 21 insertions(+), 19 deletions(-)
11
linux-user/elfload.c | 6 +++++-
12
2 files changed, 7 insertions(+), 3 deletions(-)
8
13
9
diff --git a/tcg/optimize.c b/tcg/optimize.c
14
diff --git a/linux-user/arm/target_cpu.h b/linux-user/arm/target_cpu.h
10
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
11
--- a/tcg/optimize.c
16
--- a/linux-user/arm/target_cpu.h
12
+++ b/tcg/optimize.c
17
+++ b/linux-user/arm/target_cpu.h
13
@@ -XXX,XX +XXX,XX @@ typedef struct TempOptInfo {
18
@@ -XXX,XX +XXX,XX @@ static inline unsigned long arm_max_reserved_va(CPUState *cs)
14
bool is_const;
19
} else {
15
TCGTemp *prev_copy;
20
/*
16
TCGTemp *next_copy;
21
* We need to be able to map the commpage.
17
- tcg_target_ulong val;
22
- * See validate_guest_space in linux-user/elfload.c.
18
- tcg_target_ulong mask;
23
+ * See init_guest_commpage in linux-user/elfload.c.
19
+ uint64_t val;
24
*/
20
+ uint64_t mask;
25
- return 0xffff0000ul;
21
} TempOptInfo;
26
+ return 0xfffffffful;
22
23
static inline TempOptInfo *ts_info(TCGTemp *ts)
24
@@ -XXX,XX +XXX,XX @@ static bool args_are_copies(TCGArg arg1, TCGArg arg2)
25
return ts_are_copies(arg_temp(arg1), arg_temp(arg2));
26
}
27
28
-static void tcg_opt_gen_movi(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg val)
29
+static void tcg_opt_gen_movi(TCGContext *s, TCGOp *op, TCGArg dst, uint64_t val)
30
{
31
const TCGOpDef *def;
32
TCGOpcode new_op;
33
- tcg_target_ulong mask;
34
+ uint64_t mask;
35
TempOptInfo *di = arg_info(dst);
36
37
def = &tcg_op_defs[op->opc];
38
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
39
const TCGOpDef *def;
40
TempOptInfo *di;
41
TempOptInfo *si;
42
- tcg_target_ulong mask;
43
+ uint64_t mask;
44
TCGOpcode new_op;
45
46
if (ts_are_copies(dst_ts, src_ts)) {
47
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
48
}
27
}
49
}
28
}
50
29
#define MAX_RESERVED_VA arm_max_reserved_va
51
-static TCGArg do_constant_folding_2(TCGOpcode op, TCGArg x, TCGArg y)
30
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
52
+static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y)
31
index XXXXXXX..XXXXXXX 100644
32
--- a/linux-user/elfload.c
33
+++ b/linux-user/elfload.c
34
@@ -XXX,XX +XXX,XX @@ enum {
35
36
static bool init_guest_commpage(void)
53
{
37
{
54
uint64_t l64, h64;
38
- void *want = g2h_untagged(HI_COMMPAGE & -qemu_host_page_size);
55
39
+ abi_ptr commpage = HI_COMMPAGE & -qemu_host_page_size;
56
@@ -XXX,XX +XXX,XX @@ static TCGArg do_constant_folding_2(TCGOpcode op, TCGArg x, TCGArg y)
40
+ void *want = g2h_untagged(commpage);
41
void *addr = mmap(want, qemu_host_page_size, PROT_READ | PROT_WRITE,
42
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
43
44
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
45
perror("Protecting guest commpage");
46
exit(EXIT_FAILURE);
57
}
47
}
48
+
49
+ page_set_flags(commpage, commpage + qemu_host_page_size,
50
+ PAGE_READ | PAGE_EXEC | PAGE_VALID);
51
return true;
58
}
52
}
59
53
60
-static TCGArg do_constant_folding(TCGOpcode op, TCGArg x, TCGArg y)
61
+static uint64_t do_constant_folding(TCGOpcode op, uint64_t x, uint64_t y)
62
{
63
const TCGOpDef *def = &tcg_op_defs[op];
64
- TCGArg res = do_constant_folding_2(op, x, y);
65
+ uint64_t res = do_constant_folding_2(op, x, y);
66
if (!(def->flags & TCG_OPF_64BIT)) {
67
res = (int32_t)res;
68
}
69
@@ -XXX,XX +XXX,XX @@ static bool do_constant_folding_cond_eq(TCGCond c)
70
static TCGArg do_constant_folding_cond(TCGOpcode op, TCGArg x,
71
TCGArg y, TCGCond c)
72
{
73
- tcg_target_ulong xv = arg_info(x)->val;
74
- tcg_target_ulong yv = arg_info(y)->val;
75
+ uint64_t xv = arg_info(x)->val;
76
+ uint64_t yv = arg_info(y)->val;
77
+
78
if (arg_is_const(x) && arg_is_const(y)) {
79
const TCGOpDef *def = &tcg_op_defs[op];
80
tcg_debug_assert(!(def->flags & TCG_OPF_VECTOR));
81
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
82
infos = tcg_malloc(sizeof(TempOptInfo) * nb_temps);
83
84
QTAILQ_FOREACH_SAFE(op, &s->ops, link, op_next) {
85
- tcg_target_ulong mask, partmask, affected;
86
+ uint64_t mask, partmask, affected, tmp;
87
int nb_oargs, nb_iargs, i;
88
- TCGArg tmp;
89
TCGOpcode opc = op->opc;
90
const TCGOpDef *def = &tcg_op_defs[opc];
91
92
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
93
94
CASE_OP_32_64(extract2):
95
if (arg_is_const(op->args[1]) && arg_is_const(op->args[2])) {
96
- TCGArg v1 = arg_info(op->args[1])->val;
97
- TCGArg v2 = arg_info(op->args[2])->val;
98
+ uint64_t v1 = arg_info(op->args[1])->val;
99
+ uint64_t v2 = arg_info(op->args[2])->val;
100
+ int shr = op->args[3];
101
102
if (opc == INDEX_op_extract2_i64) {
103
- tmp = (v1 >> op->args[3]) | (v2 << (64 - op->args[3]));
104
+ tmp = (v1 >> shr) | (v2 << (64 - shr));
105
} else {
106
- tmp = (int32_t)(((uint32_t)v1 >> op->args[3]) |
107
- ((uint32_t)v2 << (32 - op->args[3])));
108
+ tmp = (int32_t)(((uint32_t)v1 >> shr) |
109
+ ((uint32_t)v2 << (32 - shr)));
110
}
111
tcg_opt_gen_movi(s, op, op->args[0], tmp);
112
break;
113
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
114
break;
115
}
116
if (arg_is_const(op->args[3]) && arg_is_const(op->args[4])) {
117
- tcg_target_ulong tv = arg_info(op->args[3])->val;
118
- tcg_target_ulong fv = arg_info(op->args[4])->val;
119
+ uint64_t tv = arg_info(op->args[3])->val;
120
+ uint64_t fv = arg_info(op->args[4])->val;
121
TCGCond cond = op->args[5];
122
+
123
if (fv == 1 && tv == 0) {
124
cond = tcg_invert_cond(cond);
125
} else if (!(tv == 1 && fv == 0)) {
126
--
54
--
127
2.25.1
55
2.34.1
128
129
diff view generated by jsdifflib
1
Fix this name vs our coding style.
1
While there are no target-specific nonfaulting probes,
2
generic code may grow some uses at some point.
2
3
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Note that the attrs argument was incorrect -- it should have
5
been MEMTXATTRS_UNSPECIFIED. Just use the simpler interface.
6
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
9
---
7
tcg/optimize.c | 32 ++++++++++++++++----------------
10
target/avr/helper.c | 46 ++++++++++++++++++++++++++++-----------------
8
1 file changed, 16 insertions(+), 16 deletions(-)
11
1 file changed, 29 insertions(+), 17 deletions(-)
9
12
10
diff --git a/tcg/optimize.c b/tcg/optimize.c
13
diff --git a/target/avr/helper.c b/target/avr/helper.c
11
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/optimize.c
15
--- a/target/avr/helper.c
13
+++ b/tcg/optimize.c
16
+++ b/target/avr/helper.c
14
@@ -XXX,XX +XXX,XX @@
17
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
15
glue(glue(case INDEX_op_, x), _i64): \
18
MMUAccessType access_type, int mmu_idx,
16
glue(glue(case INDEX_op_, x), _vec)
19
bool probe, uintptr_t retaddr)
17
18
-struct tcg_temp_info {
19
+typedef struct TempOptInfo {
20
bool is_const;
21
TCGTemp *prev_copy;
22
TCGTemp *next_copy;
23
tcg_target_ulong val;
24
tcg_target_ulong mask;
25
-};
26
+} TempOptInfo;
27
28
-static inline struct tcg_temp_info *ts_info(TCGTemp *ts)
29
+static inline TempOptInfo *ts_info(TCGTemp *ts)
30
{
20
{
31
return ts->state_ptr;
21
- int prot = 0;
22
- MemTxAttrs attrs = {};
23
+ int prot, page_size = TARGET_PAGE_SIZE;
24
uint32_t paddr;
25
26
address &= TARGET_PAGE_MASK;
27
28
if (mmu_idx == MMU_CODE_IDX) {
29
- /* access to code in flash */
30
+ /* Access to code in flash. */
31
paddr = OFFSET_CODE + address;
32
prot = PAGE_READ | PAGE_EXEC;
33
- if (paddr + TARGET_PAGE_SIZE > OFFSET_DATA) {
34
+ if (paddr >= OFFSET_DATA) {
35
+ /*
36
+ * This should not be possible via any architectural operations.
37
+ * There is certainly not an exception that we can deliver.
38
+ * Accept probing that might come from generic code.
39
+ */
40
+ if (probe) {
41
+ return false;
42
+ }
43
error_report("execution left flash memory");
44
abort();
45
}
46
- } else if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
47
- /*
48
- * access to CPU registers, exit and rebuilt this TB to use full access
49
- * incase it touches specially handled registers like SREG or SP
50
- */
51
- AVRCPU *cpu = AVR_CPU(cs);
52
- CPUAVRState *env = &cpu->env;
53
- env->fullacc = 1;
54
- cpu_loop_exit_restore(cs, retaddr);
55
} else {
56
- /* access to memory. nothing special */
57
+ /* Access to memory. */
58
paddr = OFFSET_DATA + address;
59
prot = PAGE_READ | PAGE_WRITE;
60
+ if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
61
+ /*
62
+ * Access to CPU registers, exit and rebuilt this TB to use
63
+ * full access in case it touches specially handled registers
64
+ * like SREG or SP. For probing, set page_size = 1, in order
65
+ * to force tlb_fill to be called for the next access.
66
+ */
67
+ if (probe) {
68
+ page_size = 1;
69
+ } else {
70
+ AVRCPU *cpu = AVR_CPU(cs);
71
+ CPUAVRState *env = &cpu->env;
72
+ env->fullacc = 1;
73
+ cpu_loop_exit_restore(cs, retaddr);
74
+ }
75
+ }
76
}
77
78
- tlb_set_page_with_attrs(cs, address, paddr, attrs, prot,
79
- mmu_idx, TARGET_PAGE_SIZE);
80
-
81
+ tlb_set_page(cs, address, paddr, prot, mmu_idx, page_size);
82
return true;
32
}
83
}
33
84
34
-static inline struct tcg_temp_info *arg_info(TCGArg arg)
35
+static inline TempOptInfo *arg_info(TCGArg arg)
36
{
37
return ts_info(arg_temp(arg));
38
}
39
@@ -XXX,XX +XXX,XX @@ static inline bool ts_is_copy(TCGTemp *ts)
40
/* Reset TEMP's state, possibly removing the temp for the list of copies. */
41
static void reset_ts(TCGTemp *ts)
42
{
43
- struct tcg_temp_info *ti = ts_info(ts);
44
- struct tcg_temp_info *pi = ts_info(ti->prev_copy);
45
- struct tcg_temp_info *ni = ts_info(ti->next_copy);
46
+ TempOptInfo *ti = ts_info(ts);
47
+ TempOptInfo *pi = ts_info(ti->prev_copy);
48
+ TempOptInfo *ni = ts_info(ti->next_copy);
49
50
ni->prev_copy = ti->prev_copy;
51
pi->next_copy = ti->next_copy;
52
@@ -XXX,XX +XXX,XX @@ static void reset_temp(TCGArg arg)
53
}
54
55
/* Initialize and activate a temporary. */
56
-static void init_ts_info(struct tcg_temp_info *infos,
57
+static void init_ts_info(TempOptInfo *infos,
58
TCGTempSet *temps_used, TCGTemp *ts)
59
{
60
size_t idx = temp_idx(ts);
61
if (!test_bit(idx, temps_used->l)) {
62
- struct tcg_temp_info *ti = &infos[idx];
63
+ TempOptInfo *ti = &infos[idx];
64
65
ts->state_ptr = ti;
66
ti->next_copy = ts;
67
@@ -XXX,XX +XXX,XX @@ static void init_ts_info(struct tcg_temp_info *infos,
68
}
69
}
70
71
-static void init_arg_info(struct tcg_temp_info *infos,
72
+static void init_arg_info(TempOptInfo *infos,
73
TCGTempSet *temps_used, TCGArg arg)
74
{
75
init_ts_info(infos, temps_used, arg_temp(arg));
76
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_movi(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg val)
77
const TCGOpDef *def;
78
TCGOpcode new_op;
79
tcg_target_ulong mask;
80
- struct tcg_temp_info *di = arg_info(dst);
81
+ TempOptInfo *di = arg_info(dst);
82
83
def = &tcg_op_defs[op->opc];
84
if (def->flags & TCG_OPF_VECTOR) {
85
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
86
TCGTemp *dst_ts = arg_temp(dst);
87
TCGTemp *src_ts = arg_temp(src);
88
const TCGOpDef *def;
89
- struct tcg_temp_info *di;
90
- struct tcg_temp_info *si;
91
+ TempOptInfo *di;
92
+ TempOptInfo *si;
93
tcg_target_ulong mask;
94
TCGOpcode new_op;
95
96
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
97
di->mask = mask;
98
99
if (src_ts->type == dst_ts->type) {
100
- struct tcg_temp_info *ni = ts_info(si->next_copy);
101
+ TempOptInfo *ni = ts_info(si->next_copy);
102
103
di->next_copy = si->next_copy;
104
di->prev_copy = src_ts;
105
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
106
{
107
int nb_temps, nb_globals;
108
TCGOp *op, *op_next, *prev_mb = NULL;
109
- struct tcg_temp_info *infos;
110
+ TempOptInfo *infos;
111
TCGTempSet temps_used;
112
113
/* Array VALS has an element for each temp.
114
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
115
nb_temps = s->nb_temps;
116
nb_globals = s->nb_globals;
117
bitmap_zero(temps_used.l, nb_temps);
118
- infos = tcg_malloc(sizeof(struct tcg_temp_info) * nb_temps);
119
+ infos = tcg_malloc(sizeof(TempOptInfo) * nb_temps);
120
121
QTAILQ_FOREACH_SAFE(op, &s->ops, link, op_next) {
122
tcg_target_ulong mask, partmask, affected;
123
--
85
--
124
2.25.1
86
2.34.1
125
87
126
88
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
There is no need to go through cc->tcg_ops when
2
we know what value that must have.
2
3
3
When decodetree.py was added in commit 568ae7efae7, QEMU was
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
4
using Python 2 which happily reads UTF-8 files in text mode.
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Python 3 requires either UTF-8 locale or an explicit encoding
6
passed to open(). Now that Python 3 is required, explicit
7
UTF-8 encoding for decodetree source files.
8
9
To avoid further problems with the user locale, also explicit
10
UTF-8 encoding for the generated C files.
11
12
Explicit both input/output are plain text by using the 't' mode.
13
14
This fixes:
15
16
$ /usr/bin/python3 scripts/decodetree.py test.decode
17
Traceback (most recent call last):
18
File "scripts/decodetree.py", line 1397, in <module>
19
main()
20
File "scripts/decodetree.py", line 1308, in main
21
parse_file(f, toppat)
22
File "scripts/decodetree.py", line 994, in parse_file
23
for line in f:
24
File "/usr/lib/python3.6/encodings/ascii.py", line 26, in decode
25
return codecs.ascii_decode(input, self.errors)[0]
26
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 80:
27
ordinal not in range(128)
28
29
Reported-by: Peter Maydell <peter.maydell@linaro.org>
30
Suggested-by: Yonggang Luo <luoyonggang@gmail.com>
31
Reviewed-by: Eduardo Habkost <ehabkost@redhat.com>
32
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
33
Message-Id: <20210110000240.761122-1-f4bug@amsat.org>
34
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
35
---
7
---
36
scripts/decodetree.py | 9 ++++++---
8
target/avr/helper.c | 5 ++---
37
1 file changed, 6 insertions(+), 3 deletions(-)
9
1 file changed, 2 insertions(+), 3 deletions(-)
38
10
39
diff --git a/scripts/decodetree.py b/scripts/decodetree.py
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
40
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
41
--- a/scripts/decodetree.py
13
--- a/target/avr/helper.c
42
+++ b/scripts/decodetree.py
14
+++ b/target/avr/helper.c
43
@@ -XXX,XX +XXX,XX @@
15
@@ -XXX,XX +XXX,XX @@
44
# See the syntax and semantics in docs/devel/decodetree.rst.
16
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
45
#
17
{
46
18
bool ret = false;
47
+import io
19
- CPUClass *cc = CPU_GET_CLASS(cs);
48
import os
20
AVRCPU *cpu = AVR_CPU(cs);
49
import re
21
CPUAVRState *env = &cpu->env;
50
import sys
22
51
@@ -XXX,XX +XXX,XX @@ def main():
23
if (interrupt_request & CPU_INTERRUPT_RESET) {
52
24
if (cpu_interrupts_enabled(env)) {
53
for filename in args:
25
cs->exception_index = EXCP_RESET;
54
input_file = filename
26
- cc->tcg_ops->do_interrupt(cs);
55
- f = open(filename, 'r')
27
+ avr_cpu_do_interrupt(cs);
56
+ f = open(filename, 'rt', encoding='utf-8')
28
57
parse_file(f, toppat)
29
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
58
f.close()
30
59
31
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
60
@@ -XXX,XX +XXX,XX @@ def main():
32
if (cpu_interrupts_enabled(env) && env->intsrc != 0) {
61
prop_size(stree)
33
int index = ctz32(env->intsrc);
62
34
cs->exception_index = EXCP_INT(index);
63
if output_file:
35
- cc->tcg_ops->do_interrupt(cs);
64
- output_fd = open(output_file, 'w')
36
+ avr_cpu_do_interrupt(cs);
65
+ output_fd = open(output_file, 'wt', encoding='utf-8')
37
66
else:
38
env->intsrc &= env->intsrc - 1; /* clear the interrupt */
67
- output_fd = sys.stdout
39
if (!env->intsrc) {
68
+ output_fd = io.TextIOWrapper(sys.stdout.buffer,
69
+ encoding=sys.stdout.encoding,
70
+ errors="ignore")
71
72
output_autogen()
73
for n in sorted(arguments.keys()):
74
--
40
--
75
2.25.1
41
2.34.1
76
42
77
43
diff view generated by jsdifflib
1
There are several ways we can expand a vector dup of a 64-bit
1
We're about to start validating PAGE_EXEC, which means that we've
2
element on a 32-bit host.
2
got to mark page zero executable. We had been special casing this
3
entirely within translate.
3
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
8
---
6
tcg/tcg.c | 97 +++++++++++++++++++++++++++++++++++++++++++++++++++++++
9
linux-user/elfload.c | 34 +++++++++++++++++++++++++++++++---
7
1 file changed, 97 insertions(+)
10
1 file changed, 31 insertions(+), 3 deletions(-)
8
11
9
diff --git a/tcg/tcg.c b/tcg/tcg.c
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
10
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
11
--- a/tcg/tcg.c
14
--- a/linux-user/elfload.c
12
+++ b/tcg/tcg.c
15
+++ b/linux-user/elfload.c
13
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
16
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
14
}
17
regs->gr[31] = infop->entry;
15
}
18
}
16
19
17
+static bool tcg_reg_alloc_dup2(TCGContext *s, const TCGOp *op)
20
+#define LO_COMMPAGE 0
21
+
22
+static bool init_guest_commpage(void)
18
+{
23
+{
19
+ const TCGLifeData arg_life = op->life;
24
+ void *want = g2h_untagged(LO_COMMPAGE);
20
+ TCGTemp *ots, *itsl, *itsh;
25
+ void *addr = mmap(want, qemu_host_page_size, PROT_NONE,
21
+ TCGType vtype = TCGOP_VECL(op) + TCG_TYPE_V64;
26
+ MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
22
+
27
+
23
+ /* This opcode is only valid for 32-bit hosts, for 64-bit elements. */
28
+ if (addr == MAP_FAILED) {
24
+ tcg_debug_assert(TCG_TARGET_REG_BITS == 32);
29
+ perror("Allocating guest commpage");
25
+ tcg_debug_assert(TCGOP_VECE(op) == MO_64);
30
+ exit(EXIT_FAILURE);
26
+
31
+ }
27
+ ots = arg_temp(op->args[0]);
32
+ if (addr != want) {
28
+ itsl = arg_temp(op->args[1]);
33
+ return false;
29
+ itsh = arg_temp(op->args[2]);
30
+
31
+ /* ENV should not be modified. */
32
+ tcg_debug_assert(!temp_readonly(ots));
33
+
34
+ /* Allocate the output register now. */
35
+ if (ots->val_type != TEMP_VAL_REG) {
36
+ TCGRegSet allocated_regs = s->reserved_regs;
37
+ TCGRegSet dup_out_regs =
38
+ tcg_op_defs[INDEX_op_dup_vec].args_ct[0].regs;
39
+
40
+ /* Make sure to not spill the input registers. */
41
+ if (!IS_DEAD_ARG(1) && itsl->val_type == TEMP_VAL_REG) {
42
+ tcg_regset_set_reg(allocated_regs, itsl->reg);
43
+ }
44
+ if (!IS_DEAD_ARG(2) && itsh->val_type == TEMP_VAL_REG) {
45
+ tcg_regset_set_reg(allocated_regs, itsh->reg);
46
+ }
47
+
48
+ ots->reg = tcg_reg_alloc(s, dup_out_regs, allocated_regs,
49
+ op->output_pref[0], ots->indirect_base);
50
+ ots->val_type = TEMP_VAL_REG;
51
+ ots->mem_coherent = 0;
52
+ s->reg_to_temp[ots->reg] = ots;
53
+ }
34
+ }
54
+
35
+
55
+ /* Promote dup2 of immediates to dupi_vec. */
36
+ /*
56
+ if (itsl->val_type == TEMP_VAL_CONST && itsh->val_type == TEMP_VAL_CONST) {
37
+ * On Linux, page zero is normally marked execute only + gateway.
57
+ uint64_t val = deposit64(itsl->val, 32, 32, itsh->val);
38
+ * Normal read or write is supposed to fail (thus PROT_NONE above),
58
+ MemOp vece = MO_64;
39
+ * but specific offsets have kernel code mapped to raise permissions
59
+
40
+ * and implement syscalls. Here, simply mark the page executable.
60
+ if (val == dup_const(MO_8, val)) {
41
+ * Special case the entry points during translation (see do_page_zero).
61
+ vece = MO_8;
42
+ */
62
+ } else if (val == dup_const(MO_16, val)) {
43
+ page_set_flags(LO_COMMPAGE, LO_COMMPAGE + TARGET_PAGE_SIZE,
63
+ vece = MO_16;
44
+ PAGE_EXEC | PAGE_VALID);
64
+ } else if (val == dup_const(MO_32, val)) {
65
+ vece = MO_32;
66
+ }
67
+
68
+ tcg_out_dupi_vec(s, vtype, vece, ots->reg, val);
69
+ goto done;
70
+ }
71
+
72
+ /* If the two inputs form one 64-bit value, try dupm_vec. */
73
+ if (itsl + 1 == itsh && itsl->base_type == TCG_TYPE_I64) {
74
+ if (!itsl->mem_coherent) {
75
+ temp_sync(s, itsl, s->reserved_regs, 0, 0);
76
+ }
77
+ if (!itsh->mem_coherent) {
78
+ temp_sync(s, itsh, s->reserved_regs, 0, 0);
79
+ }
80
+#ifdef HOST_WORDS_BIGENDIAN
81
+ TCGTemp *its = itsh;
82
+#else
83
+ TCGTemp *its = itsl;
84
+#endif
85
+ if (tcg_out_dupm_vec(s, vtype, MO_64, ots->reg,
86
+ its->mem_base->reg, its->mem_offset)) {
87
+ goto done;
88
+ }
89
+ }
90
+
91
+ /* Fall back to generic expansion. */
92
+ return false;
93
+
94
+ done:
95
+ if (IS_DEAD_ARG(1)) {
96
+ temp_dead(s, itsl);
97
+ }
98
+ if (IS_DEAD_ARG(2)) {
99
+ temp_dead(s, itsh);
100
+ }
101
+ if (NEED_SYNC_ARG(0)) {
102
+ temp_sync(s, ots, s->reserved_regs, 0, IS_DEAD_ARG(0));
103
+ } else if (IS_DEAD_ARG(0)) {
104
+ temp_dead(s, ots);
105
+ }
106
+ return true;
45
+ return true;
107
+}
46
+}
108
+
47
+
109
#ifdef TCG_TARGET_STACK_GROWSUP
48
#endif /* TARGET_HPPA */
110
#define STACK_DIR(x) (-(x))
49
50
#ifdef TARGET_XTENSA
51
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
52
}
53
54
#if defined(HI_COMMPAGE)
55
-#define LO_COMMPAGE 0
56
+#define LO_COMMPAGE -1
57
#elif defined(LO_COMMPAGE)
58
#define HI_COMMPAGE 0
111
#else
59
#else
112
@@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
60
#define HI_COMMPAGE 0
113
case INDEX_op_call:
61
-#define LO_COMMPAGE 0
114
tcg_reg_alloc_call(s, op);
62
+#define LO_COMMPAGE -1
115
break;
63
#define init_guest_commpage() true
116
+ case INDEX_op_dup2_vec:
64
#endif
117
+ if (tcg_reg_alloc_dup2(s, op)) {
65
118
+ break;
66
@@ -XXX,XX +XXX,XX @@ static void pgb_static(const char *image_name, abi_ulong orig_loaddr,
119
+ }
67
} else {
120
+ /* fall through */
68
offset = -(HI_COMMPAGE & -align);
121
default:
69
}
122
/* Sanity check that we've not introduced any unhandled opcodes. */
70
- } else if (LO_COMMPAGE != 0) {
123
tcg_debug_assert(tcg_op_supported(opc));
71
+ } else if (LO_COMMPAGE != -1) {
72
loaddr = MIN(loaddr, LO_COMMPAGE & -align);
73
}
74
124
--
75
--
125
2.25.1
76
2.34.1
126
127
diff view generated by jsdifflib
1
These interfaces have been replaced by tcg_gen_dupi_vec
1
We're about to start validating PAGE_EXEC, which means that we've
2
and tcg_constant_vec.
2
got to mark the vsyscall page executable. We had been special
3
casing this entirely within translate.
3
4
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
8
---
7
include/tcg/tcg-op.h | 4 ----
9
linux-user/elfload.c | 23 +++++++++++++++++++++++
8
tcg/tcg-op-vec.c | 20 --------------------
10
1 file changed, 23 insertions(+)
9
2 files changed, 24 deletions(-)
10
11
11
diff --git a/include/tcg/tcg-op.h b/include/tcg/tcg-op.h
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
12
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
13
--- a/include/tcg/tcg-op.h
14
--- a/linux-user/elfload.c
14
+++ b/include/tcg/tcg-op.h
15
+++ b/linux-user/elfload.c
15
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mov_vec(TCGv_vec, TCGv_vec);
16
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
16
void tcg_gen_dup_i32_vec(unsigned vece, TCGv_vec, TCGv_i32);
17
(*regs)[26] = tswapreg(env->segs[R_GS].selector & 0xffff);
17
void tcg_gen_dup_i64_vec(unsigned vece, TCGv_vec, TCGv_i64);
18
void tcg_gen_dup_mem_vec(unsigned vece, TCGv_vec, TCGv_ptr, tcg_target_long);
19
-void tcg_gen_dup8i_vec(TCGv_vec, uint32_t);
20
-void tcg_gen_dup16i_vec(TCGv_vec, uint32_t);
21
-void tcg_gen_dup32i_vec(TCGv_vec, uint32_t);
22
-void tcg_gen_dup64i_vec(TCGv_vec, uint64_t);
23
void tcg_gen_dupi_vec(unsigned vece, TCGv_vec, uint64_t);
24
void tcg_gen_add_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b);
25
void tcg_gen_sub_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b);
26
diff --git a/tcg/tcg-op-vec.c b/tcg/tcg-op-vec.c
27
index XXXXXXX..XXXXXXX 100644
28
--- a/tcg/tcg-op-vec.c
29
+++ b/tcg/tcg-op-vec.c
30
@@ -XXX,XX +XXX,XX @@ TCGv_vec tcg_const_ones_vec_matching(TCGv_vec m)
31
return tcg_const_ones_vec(t->base_type);
32
}
18
}
33
19
34
-void tcg_gen_dup64i_vec(TCGv_vec r, uint64_t a)
20
+#if ULONG_MAX >= TARGET_VSYSCALL_PAGE
35
-{
21
+#define INIT_GUEST_COMMPAGE
36
- tcg_gen_dupi_vec(MO_64, r, a);
22
+static bool init_guest_commpage(void)
37
-}
23
+{
38
-
24
+ /*
39
-void tcg_gen_dup32i_vec(TCGv_vec r, uint32_t a)
25
+ * The vsyscall page is at a high negative address aka kernel space,
40
-{
26
+ * which means that we cannot actually allocate it with target_mmap.
41
- tcg_gen_dupi_vec(MO_32, r, a);
27
+ * We still should be able to use page_set_flags, unless the user
42
-}
28
+ * has specified -R reserved_va, which would trigger an assert().
43
-
29
+ */
44
-void tcg_gen_dup16i_vec(TCGv_vec r, uint32_t a)
30
+ if (reserved_va != 0 &&
45
-{
31
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE >= reserved_va) {
46
- tcg_gen_dupi_vec(MO_16, r, a);
32
+ error_report("Cannot allocate vsyscall page");
47
-}
33
+ exit(EXIT_FAILURE);
48
-
34
+ }
49
-void tcg_gen_dup8i_vec(TCGv_vec r, uint32_t a)
35
+ page_set_flags(TARGET_VSYSCALL_PAGE,
50
-{
36
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE,
51
- tcg_gen_dupi_vec(MO_8, r, a);
37
+ PAGE_EXEC | PAGE_VALID);
52
-}
38
+ return true;
53
-
39
+}
54
void tcg_gen_dupi_vec(unsigned vece, TCGv_vec r, uint64_t a)
40
+#endif
41
#else
42
43
#define ELF_START_MMAP 0x80000000
44
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
45
#else
46
#define HI_COMMPAGE 0
47
#define LO_COMMPAGE -1
48
+#ifndef INIT_GUEST_COMMPAGE
49
#define init_guest_commpage() true
50
#endif
51
+#endif
52
53
static void pgb_fail_in_use(const char *image_name)
55
{
54
{
56
TCGTemp *rt = tcgv_vec_temp(r);
57
--
55
--
58
2.25.1
56
2.34.1
59
60
diff view generated by jsdifflib
1
The temp_fixed, temp_global, temp_local bits are all related.
1
We cannot deliver two interrupts simultaneously;
2
Combine them into a single enumeration.
2
the first interrupt handler must execute first.
3
3
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
7
---
8
include/tcg/tcg.h | 20 +++++---
8
target/avr/helper.c | 9 +++------
9
tcg/optimize.c | 8 +--
9
1 file changed, 3 insertions(+), 6 deletions(-)
10
tcg/tcg.c | 126 ++++++++++++++++++++++++++++------------------
11
3 files changed, 92 insertions(+), 62 deletions(-)
12
10
13
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
14
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg.h
13
--- a/target/avr/helper.c
16
+++ b/include/tcg/tcg.h
14
+++ b/target/avr/helper.c
17
@@ -XXX,XX +XXX,XX @@ typedef enum TCGTempVal {
15
@@ -XXX,XX +XXX,XX @@
18
TEMP_VAL_CONST,
16
19
} TCGTempVal;
17
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
20
18
{
21
+typedef enum TCGTempKind {
19
- bool ret = false;
22
+ /* Temp is dead at the end of all basic blocks. */
20
AVRCPU *cpu = AVR_CPU(cs);
23
+ TEMP_NORMAL,
21
CPUAVRState *env = &cpu->env;
24
+ /* Temp is saved across basic blocks but dead at the end of TBs. */
22
25
+ TEMP_LOCAL,
23
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
26
+ /* Temp is saved across both basic blocks and translation blocks. */
24
avr_cpu_do_interrupt(cs);
27
+ TEMP_GLOBAL,
25
28
+ /* Temp is in a fixed register. */
26
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
29
+ TEMP_FIXED,
27
-
30
+} TCGTempKind;
28
- ret = true;
31
+
29
+ return true;
32
typedef struct TCGTemp {
33
TCGReg reg:8;
34
TCGTempVal val_type:8;
35
TCGType base_type:8;
36
TCGType type:8;
37
- unsigned int fixed_reg:1;
38
+ TCGTempKind kind:3;
39
unsigned int indirect_reg:1;
40
unsigned int indirect_base:1;
41
unsigned int mem_coherent:1;
42
unsigned int mem_allocated:1;
43
- /* If true, the temp is saved across both basic blocks and
44
- translation blocks. */
45
- unsigned int temp_global:1;
46
- /* If true, the temp is saved across basic blocks but dead
47
- at the end of translation blocks. If false, the temp is
48
- dead at the end of basic blocks. */
49
- unsigned int temp_local:1;
50
unsigned int temp_allocated:1;
51
52
tcg_target_long val;
53
diff --git a/tcg/optimize.c b/tcg/optimize.c
54
index XXXXXXX..XXXXXXX 100644
55
--- a/tcg/optimize.c
56
+++ b/tcg/optimize.c
57
@@ -XXX,XX +XXX,XX @@ static TCGTemp *find_better_copy(TCGContext *s, TCGTemp *ts)
58
TCGTemp *i;
59
60
/* If this is already a global, we can't do better. */
61
- if (ts->temp_global) {
62
+ if (ts->kind >= TEMP_GLOBAL) {
63
return ts;
64
}
65
66
/* Search for a global first. */
67
for (i = ts_info(ts)->next_copy; i != ts; i = ts_info(i)->next_copy) {
68
- if (i->temp_global) {
69
+ if (i->kind >= TEMP_GLOBAL) {
70
return i;
71
}
30
}
72
}
31
}
73
32
if (interrupt_request & CPU_INTERRUPT_HARD) {
74
/* If it is a temp, search for a temp local. */
33
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
75
- if (!ts->temp_local) {
34
if (!env->intsrc) {
76
+ if (ts->kind == TEMP_NORMAL) {
35
cs->interrupt_request &= ~CPU_INTERRUPT_HARD;
77
for (i = ts_info(ts)->next_copy; i != ts; i = ts_info(i)->next_copy) {
78
- if (ts->temp_local) {
79
+ if (i->kind >= TEMP_LOCAL) {
80
return i;
81
}
36
}
82
}
37
-
83
diff --git a/tcg/tcg.c b/tcg/tcg.c
38
- ret = true;
84
index XXXXXXX..XXXXXXX 100644
39
+ return true;
85
--- a/tcg/tcg.c
86
+++ b/tcg/tcg.c
87
@@ -XXX,XX +XXX,XX @@ static inline TCGTemp *tcg_global_alloc(TCGContext *s)
88
tcg_debug_assert(s->nb_globals == s->nb_temps);
89
s->nb_globals++;
90
ts = tcg_temp_alloc(s);
91
- ts->temp_global = 1;
92
+ ts->kind = TEMP_GLOBAL;
93
94
return ts;
95
}
96
@@ -XXX,XX +XXX,XX @@ static TCGTemp *tcg_global_reg_new_internal(TCGContext *s, TCGType type,
97
ts = tcg_global_alloc(s);
98
ts->base_type = type;
99
ts->type = type;
100
- ts->fixed_reg = 1;
101
+ ts->kind = TEMP_FIXED;
102
ts->reg = reg;
103
ts->name = name;
104
tcg_regset_set_reg(s->reserved_regs, reg);
105
@@ -XXX,XX +XXX,XX @@ TCGTemp *tcg_global_mem_new_internal(TCGType type, TCGv_ptr base,
106
bigendian = 1;
107
#endif
108
109
- if (!base_ts->fixed_reg) {
110
+ if (base_ts->kind != TEMP_FIXED) {
111
/* We do not support double-indirect registers. */
112
tcg_debug_assert(!base_ts->indirect_reg);
113
base_ts->indirect_base = 1;
114
@@ -XXX,XX +XXX,XX @@ TCGTemp *tcg_global_mem_new_internal(TCGType type, TCGv_ptr base,
115
TCGTemp *tcg_temp_new_internal(TCGType type, bool temp_local)
116
{
117
TCGContext *s = tcg_ctx;
118
+ TCGTempKind kind = temp_local ? TEMP_LOCAL : TEMP_NORMAL;
119
TCGTemp *ts;
120
int idx, k;
121
122
@@ -XXX,XX +XXX,XX @@ TCGTemp *tcg_temp_new_internal(TCGType type, bool temp_local)
123
ts = &s->temps[idx];
124
ts->temp_allocated = 1;
125
tcg_debug_assert(ts->base_type == type);
126
- tcg_debug_assert(ts->temp_local == temp_local);
127
+ tcg_debug_assert(ts->kind == kind);
128
} else {
129
ts = tcg_temp_alloc(s);
130
if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
131
@@ -XXX,XX +XXX,XX @@ TCGTemp *tcg_temp_new_internal(TCGType type, bool temp_local)
132
ts->base_type = type;
133
ts->type = TCG_TYPE_I32;
134
ts->temp_allocated = 1;
135
- ts->temp_local = temp_local;
136
+ ts->kind = kind;
137
138
tcg_debug_assert(ts2 == ts + 1);
139
ts2->base_type = TCG_TYPE_I64;
140
ts2->type = TCG_TYPE_I32;
141
ts2->temp_allocated = 1;
142
- ts2->temp_local = temp_local;
143
+ ts2->kind = kind;
144
} else {
145
ts->base_type = type;
146
ts->type = type;
147
ts->temp_allocated = 1;
148
- ts->temp_local = temp_local;
149
+ ts->kind = kind;
150
}
40
}
151
}
41
}
152
42
- return ret;
153
@@ -XXX,XX +XXX,XX @@ void tcg_temp_free_internal(TCGTemp *ts)
43
+ return false;
154
}
155
#endif
156
157
- tcg_debug_assert(ts->temp_global == 0);
158
+ tcg_debug_assert(ts->kind < TEMP_GLOBAL);
159
tcg_debug_assert(ts->temp_allocated != 0);
160
ts->temp_allocated = 0;
161
162
idx = temp_idx(ts);
163
- k = ts->base_type + (ts->temp_local ? TCG_TYPE_COUNT : 0);
164
+ k = ts->base_type + (ts->kind == TEMP_NORMAL ? 0 : TCG_TYPE_COUNT);
165
set_bit(idx, s->free_temps[k].l);
166
}
44
}
167
45
168
@@ -XXX,XX +XXX,XX @@ void tcg_gen_callN(void *func, TCGTemp *ret, int nargs, TCGTemp **args)
46
void avr_cpu_do_interrupt(CPUState *cs)
169
static void tcg_reg_alloc_start(TCGContext *s)
170
{
171
int i, n;
172
- TCGTemp *ts;
173
174
- for (i = 0, n = s->nb_globals; i < n; i++) {
175
- ts = &s->temps[i];
176
- ts->val_type = (ts->fixed_reg ? TEMP_VAL_REG : TEMP_VAL_MEM);
177
- }
178
- for (n = s->nb_temps; i < n; i++) {
179
- ts = &s->temps[i];
180
- ts->val_type = (ts->temp_local ? TEMP_VAL_MEM : TEMP_VAL_DEAD);
181
- ts->mem_allocated = 0;
182
- ts->fixed_reg = 0;
183
+ for (i = 0, n = s->nb_temps; i < n; i++) {
184
+ TCGTemp *ts = &s->temps[i];
185
+ TCGTempVal val = TEMP_VAL_MEM;
186
+
187
+ switch (ts->kind) {
188
+ case TEMP_FIXED:
189
+ val = TEMP_VAL_REG;
190
+ break;
191
+ case TEMP_GLOBAL:
192
+ break;
193
+ case TEMP_NORMAL:
194
+ val = TEMP_VAL_DEAD;
195
+ /* fall through */
196
+ case TEMP_LOCAL:
197
+ ts->mem_allocated = 0;
198
+ break;
199
+ default:
200
+ g_assert_not_reached();
201
+ }
202
+ ts->val_type = val;
203
}
204
205
memset(s->reg_to_temp, 0, sizeof(s->reg_to_temp));
206
@@ -XXX,XX +XXX,XX @@ static char *tcg_get_arg_str_ptr(TCGContext *s, char *buf, int buf_size,
207
{
208
int idx = temp_idx(ts);
209
210
- if (ts->temp_global) {
211
+ switch (ts->kind) {
212
+ case TEMP_FIXED:
213
+ case TEMP_GLOBAL:
214
pstrcpy(buf, buf_size, ts->name);
215
- } else if (ts->temp_local) {
216
+ break;
217
+ case TEMP_LOCAL:
218
snprintf(buf, buf_size, "loc%d", idx - s->nb_globals);
219
- } else {
220
+ break;
221
+ case TEMP_NORMAL:
222
snprintf(buf, buf_size, "tmp%d", idx - s->nb_globals);
223
+ break;
224
}
225
return buf;
226
}
227
@@ -XXX,XX +XXX,XX @@ static void la_bb_end(TCGContext *s, int ng, int nt)
228
{
229
int i;
230
231
- for (i = 0; i < ng; ++i) {
232
- s->temps[i].state = TS_DEAD | TS_MEM;
233
- la_reset_pref(&s->temps[i]);
234
- }
235
- for (i = ng; i < nt; ++i) {
236
- s->temps[i].state = (s->temps[i].temp_local
237
- ? TS_DEAD | TS_MEM
238
- : TS_DEAD);
239
- la_reset_pref(&s->temps[i]);
240
+ for (i = 0; i < nt; ++i) {
241
+ TCGTemp *ts = &s->temps[i];
242
+ int state;
243
+
244
+ switch (ts->kind) {
245
+ case TEMP_FIXED:
246
+ case TEMP_GLOBAL:
247
+ case TEMP_LOCAL:
248
+ state = TS_DEAD | TS_MEM;
249
+ break;
250
+ case TEMP_NORMAL:
251
+ state = TS_DEAD;
252
+ break;
253
+ default:
254
+ g_assert_not_reached();
255
+ }
256
+ ts->state = state;
257
+ la_reset_pref(ts);
258
}
259
}
260
261
@@ -XXX,XX +XXX,XX @@ static void la_bb_sync(TCGContext *s, int ng, int nt)
262
la_global_sync(s, ng);
263
264
for (int i = ng; i < nt; ++i) {
265
- if (s->temps[i].temp_local) {
266
+ if (s->temps[i].kind == TEMP_LOCAL) {
267
int state = s->temps[i].state;
268
s->temps[i].state = state | TS_MEM;
269
if (state != TS_DEAD) {
270
@@ -XXX,XX +XXX,XX @@ static void check_regs(TCGContext *s)
271
}
272
for (k = 0; k < s->nb_temps; k++) {
273
ts = &s->temps[k];
274
- if (ts->val_type == TEMP_VAL_REG && !ts->fixed_reg
275
+ if (ts->val_type == TEMP_VAL_REG
276
+ && ts->kind != TEMP_FIXED
277
&& s->reg_to_temp[ts->reg] != ts) {
278
printf("Inconsistency for temp %s:\n",
279
tcg_get_arg_str_ptr(s, buf, sizeof(buf), ts));
280
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *, TCGTemp *, TCGRegSet, TCGRegSet, TCGRegSet);
281
mark it free; otherwise mark it dead. */
282
static void temp_free_or_dead(TCGContext *s, TCGTemp *ts, int free_or_dead)
283
{
284
- if (ts->fixed_reg) {
285
+ if (ts->kind == TEMP_FIXED) {
286
return;
287
}
288
if (ts->val_type == TEMP_VAL_REG) {
289
s->reg_to_temp[ts->reg] = NULL;
290
}
291
ts->val_type = (free_or_dead < 0
292
- || ts->temp_local
293
- || ts->temp_global
294
+ || ts->kind != TEMP_NORMAL
295
? TEMP_VAL_MEM : TEMP_VAL_DEAD);
296
}
297
298
@@ -XXX,XX +XXX,XX @@ static inline void temp_dead(TCGContext *s, TCGTemp *ts)
299
static void temp_sync(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs,
300
TCGRegSet preferred_regs, int free_or_dead)
301
{
302
- if (ts->fixed_reg) {
303
+ if (ts->kind == TEMP_FIXED) {
304
return;
305
}
306
if (!ts->mem_coherent) {
307
@@ -XXX,XX +XXX,XX @@ static void temp_save(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs)
308
{
309
/* The liveness analysis already ensures that globals are back
310
in memory. Keep an tcg_debug_assert for safety. */
311
- tcg_debug_assert(ts->val_type == TEMP_VAL_MEM || ts->fixed_reg);
312
+ tcg_debug_assert(ts->val_type == TEMP_VAL_MEM
313
+ || ts->kind == TEMP_FIXED);
314
}
315
316
/* save globals to their canonical location and assume they can be
317
@@ -XXX,XX +XXX,XX @@ static void sync_globals(TCGContext *s, TCGRegSet allocated_regs)
318
for (i = 0, n = s->nb_globals; i < n; i++) {
319
TCGTemp *ts = &s->temps[i];
320
tcg_debug_assert(ts->val_type != TEMP_VAL_REG
321
- || ts->fixed_reg
322
+ || ts->kind == TEMP_FIXED
323
|| ts->mem_coherent);
324
}
325
}
326
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_bb_end(TCGContext *s, TCGRegSet allocated_regs)
327
328
for (i = s->nb_globals; i < s->nb_temps; i++) {
329
TCGTemp *ts = &s->temps[i];
330
- if (ts->temp_local) {
331
+ if (ts->kind == TEMP_LOCAL) {
332
temp_save(s, ts, allocated_regs);
333
} else {
334
/* The liveness analysis already ensures that temps are dead.
335
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_cbranch(TCGContext *s, TCGRegSet allocated_regs)
336
* The liveness analysis already ensures that temps are dead.
337
* Keep tcg_debug_asserts for safety.
338
*/
339
- if (ts->temp_local) {
340
+ if (ts->kind == TEMP_LOCAL) {
341
tcg_debug_assert(ts->val_type != TEMP_VAL_REG || ts->mem_coherent);
342
} else {
343
tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
344
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
345
TCGRegSet preferred_regs)
346
{
347
/* ENV should not be modified. */
348
- tcg_debug_assert(!ots->fixed_reg);
349
+ tcg_debug_assert(ots->kind != TEMP_FIXED);
350
351
/* The movi is not explicitly generated here. */
352
if (ots->val_type == TEMP_VAL_REG) {
353
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
354
ts = arg_temp(op->args[1]);
355
356
/* ENV should not be modified. */
357
- tcg_debug_assert(!ots->fixed_reg);
358
+ tcg_debug_assert(ots->kind != TEMP_FIXED);
359
360
/* Note that otype != itype for no-op truncation. */
361
otype = ots->type;
362
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
363
}
364
temp_dead(s, ots);
365
} else {
366
- if (IS_DEAD_ARG(1) && !ts->fixed_reg) {
367
+ if (IS_DEAD_ARG(1) && ts->kind != TEMP_FIXED) {
368
/* the mov can be suppressed */
369
if (ots->val_type == TEMP_VAL_REG) {
370
s->reg_to_temp[ots->reg] = NULL;
371
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
372
* Store the source register into the destination slot
373
* and leave the destination temp as TEMP_VAL_MEM.
374
*/
375
- assert(!ots->fixed_reg);
376
+ assert(ots->kind != TEMP_FIXED);
377
if (!ts->mem_allocated) {
378
temp_allocate_frame(s, ots);
379
}
380
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_dup(TCGContext *s, const TCGOp *op)
381
its = arg_temp(op->args[1]);
382
383
/* ENV should not be modified. */
384
- tcg_debug_assert(!ots->fixed_reg);
385
+ tcg_debug_assert(ots->kind != TEMP_FIXED);
386
387
itype = its->type;
388
vece = TCGOP_VECE(op);
389
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
390
i_preferred_regs = o_preferred_regs = 0;
391
if (arg_ct->ialias) {
392
o_preferred_regs = op->output_pref[arg_ct->alias_index];
393
- if (ts->fixed_reg) {
394
+ if (ts->kind == TEMP_FIXED) {
395
/* if fixed register, we must allocate a new register
396
if the alias is not the same register */
397
if (arg != op->args[arg_ct->alias_index]) {
398
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
399
ts = arg_temp(arg);
400
401
/* ENV should not be modified. */
402
- tcg_debug_assert(!ts->fixed_reg);
403
+ tcg_debug_assert(ts->kind != TEMP_FIXED);
404
405
if (arg_ct->oalias && !const_args[arg_ct->alias_index]) {
406
reg = new_args[arg_ct->alias_index];
407
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
408
ts = arg_temp(op->args[i]);
409
410
/* ENV should not be modified. */
411
- tcg_debug_assert(!ts->fixed_reg);
412
+ tcg_debug_assert(ts->kind != TEMP_FIXED);
413
414
if (NEED_SYNC_ARG(i)) {
415
temp_sync(s, ts, o_allocated_regs, 0, IS_DEAD_ARG(i));
416
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_call(TCGContext *s, TCGOp *op)
417
ts = arg_temp(arg);
418
419
/* ENV should not be modified. */
420
- tcg_debug_assert(!ts->fixed_reg);
421
+ tcg_debug_assert(ts->kind != TEMP_FIXED);
422
423
reg = tcg_target_call_oarg_regs[i];
424
tcg_debug_assert(s->reg_to_temp[reg] == NULL);
425
--
47
--
426
2.25.1
48
2.34.1
427
49
428
50
diff view generated by jsdifflib
1
While we don't store more than tcg_target_long in TCGTemp,
1
This bit is not saved across interrupts, so we must
2
we shouldn't be limited to that for code generation. We will
2
delay delivering the interrupt until the skip has
3
be able to use this for INDEX_op_dup2_vec with 2 constants.
3
been processed.
4
4
5
Also pass along the minimal vece that may be said to apply
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1118
6
to the constant. This allows some simplification in the
6
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
7
various backends.
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
9
---
11
tcg/tcg.c | 31 +++++++++++++++++++++++++-----
10
target/avr/helper.c | 9 +++++++++
12
tcg/aarch64/tcg-target.c.inc | 12 ++++++------
11
target/avr/translate.c | 26 ++++++++++++++++++++++----
13
tcg/i386/tcg-target.c.inc | 22 ++++++++++++---------
12
2 files changed, 31 insertions(+), 4 deletions(-)
14
tcg/ppc/tcg-target.c.inc | 37 +++++++++++++++++++++++-------------
15
4 files changed, 69 insertions(+), 33 deletions(-)
16
13
17
diff --git a/tcg/tcg.c b/tcg/tcg.c
14
diff --git a/target/avr/helper.c b/target/avr/helper.c
18
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
19
--- a/tcg/tcg.c
16
--- a/target/avr/helper.c
20
+++ b/tcg/tcg.c
17
+++ b/target/avr/helper.c
21
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece,
18
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
22
TCGReg dst, TCGReg src);
19
AVRCPU *cpu = AVR_CPU(cs);
23
static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
20
CPUAVRState *env = &cpu->env;
24
TCGReg dst, TCGReg base, intptr_t offset);
21
25
-static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
22
+ /*
26
- TCGReg dst, tcg_target_long arg);
23
+ * We cannot separate a skip from the next instruction,
27
+static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
24
+ * as the skip would not be preserved across the interrupt.
28
+ TCGReg dst, int64_t arg);
25
+ * Separating the two insn normally only happens at page boundaries.
29
static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, unsigned vecl,
26
+ */
30
unsigned vece, const TCGArg *args,
27
+ if (env->skip) {
31
const int *const_args);
28
+ return false;
32
@@ -XXX,XX +XXX,XX @@ static inline bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
29
+ }
33
{
34
g_assert_not_reached();
35
}
36
-static inline void tcg_out_dupi_vec(TCGContext *s, TCGType type,
37
- TCGReg dst, tcg_target_long arg)
38
+static inline void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
39
+ TCGReg dst, int64_t arg)
40
{
41
g_assert_not_reached();
42
}
43
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
44
if (ts->type <= TCG_TYPE_I64) {
45
tcg_out_movi(s, ts->type, reg, ts->val);
46
} else {
47
- tcg_out_dupi_vec(s, ts->type, reg, ts->val);
48
+ uint64_t val = ts->val;
49
+ MemOp vece = MO_64;
50
+
30
+
51
+ /*
31
if (interrupt_request & CPU_INTERRUPT_RESET) {
52
+ * Find the minimal vector element that matches the constant.
32
if (cpu_interrupts_enabled(env)) {
53
+ * The targets will, in general, have to do this search anyway,
33
cs->exception_index = EXCP_RESET;
54
+ * do this generically.
34
diff --git a/target/avr/translate.c b/target/avr/translate.c
55
+ */
35
index XXXXXXX..XXXXXXX 100644
56
+ if (TCG_TARGET_REG_BITS == 32) {
36
--- a/target/avr/translate.c
57
+ val = dup_const(MO_32, val);
37
+++ b/target/avr/translate.c
58
+ vece = MO_32;
38
@@ -XXX,XX +XXX,XX @@ static void avr_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
39
if (skip_label) {
40
canonicalize_skip(ctx);
41
gen_set_label(skip_label);
42
- if (ctx->base.is_jmp == DISAS_NORETURN) {
43
+
44
+ switch (ctx->base.is_jmp) {
45
+ case DISAS_NORETURN:
46
ctx->base.is_jmp = DISAS_CHAIN;
47
+ break;
48
+ case DISAS_NEXT:
49
+ if (ctx->base.tb->flags & TB_FLAGS_SKIP) {
50
+ ctx->base.is_jmp = DISAS_TOO_MANY;
59
+ }
51
+ }
60
+ if (val == dup_const(MO_8, val)) {
52
+ break;
61
+ vece = MO_8;
53
+ default:
62
+ } else if (val == dup_const(MO_16, val)) {
54
+ break;
63
+ vece = MO_16;
64
+ } else if (TCG_TARGET_REG_BITS == 64 &&
65
+ val == dup_const(MO_32, val)) {
66
+ vece = MO_32;
67
+ }
68
+
69
+ tcg_out_dupi_vec(s, ts->type, vece, reg, ts->val);
70
}
71
ts->mem_coherent = 0;
72
break;
73
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
74
index XXXXXXX..XXXXXXX 100644
75
--- a/tcg/aarch64/tcg-target.c.inc
76
+++ b/tcg/aarch64/tcg-target.c.inc
77
@@ -XXX,XX +XXX,XX @@ static void tcg_out_logicali(TCGContext *s, AArch64Insn insn, TCGType ext,
78
tcg_out_insn_3404(s, insn, ext, rd, rn, ext, r, c);
79
}
80
81
-static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
82
- TCGReg rd, tcg_target_long v64)
83
+static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
84
+ TCGReg rd, int64_t v64)
85
{
86
bool q = type == TCG_TYPE_V128;
87
int cmode, imm8, i;
88
89
/* Test all bytes equal first. */
90
- if (v64 == dup_const(MO_8, v64)) {
91
+ if (vece == MO_8) {
92
imm8 = (uint8_t)v64;
93
tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0xe, imm8);
94
return;
95
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
96
* cannot find an expansion there's no point checking a larger
97
* width because we already know by replication it cannot match.
98
*/
99
- if (v64 == dup_const(MO_16, v64)) {
100
+ if (vece == MO_16) {
101
uint16_t v16 = v64;
102
103
if (is_shimm16(v16, &cmode, &imm8)) {
104
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
105
tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0x8, v16 & 0xff);
106
tcg_out_insn(s, 3606, ORR, q, rd, 0, 0xa, v16 >> 8);
107
return;
108
- } else if (v64 == dup_const(MO_32, v64)) {
109
+ } else if (vece == MO_32) {
110
uint32_t v32 = v64;
111
uint32_t n32 = ~v32;
112
113
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
114
tcg_out_insn_3617(s, insn, is_q, vece, a0, a1);
115
break;
116
}
117
- tcg_out_dupi_vec(s, type, TCG_VEC_TMP, 0);
118
+ tcg_out_dupi_vec(s, type, MO_8, TCG_VEC_TMP, 0);
119
a2 = TCG_VEC_TMP;
120
}
121
insn = cmp_insn[cond];
122
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
123
index XXXXXXX..XXXXXXX 100644
124
--- a/tcg/i386/tcg-target.c.inc
125
+++ b/tcg/i386/tcg-target.c.inc
126
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
127
return true;
128
}
129
130
-static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
131
- TCGReg ret, tcg_target_long arg)
132
+static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
133
+ TCGReg ret, int64_t arg)
134
{
135
int vex_l = (type == TCG_TYPE_V256 ? P_VEXL : 0);
136
137
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
138
return;
139
}
140
141
- if (TCG_TARGET_REG_BITS == 64) {
142
+ if (TCG_TARGET_REG_BITS == 32 && vece < MO_64) {
143
+ if (have_avx2) {
144
+ tcg_out_vex_modrm_pool(s, OPC_VPBROADCASTD + vex_l, ret);
145
+ } else {
146
+ tcg_out_vex_modrm_pool(s, OPC_VBROADCASTSS, ret);
147
+ }
148
+ new_pool_label(s, arg, R_386_32, s->code_ptr - 4, 0);
149
+ } else {
150
if (type == TCG_TYPE_V64) {
151
tcg_out_vex_modrm_pool(s, OPC_MOVQ_VqWq, ret);
152
} else if (have_avx2) {
153
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
154
} else {
155
tcg_out_vex_modrm_pool(s, OPC_MOVDDUP, ret);
156
}
157
- new_pool_label(s, arg, R_386_PC32, s->code_ptr - 4, -4);
158
- } else {
159
- if (have_avx2) {
160
- tcg_out_vex_modrm_pool(s, OPC_VPBROADCASTD + vex_l, ret);
161
+ if (TCG_TARGET_REG_BITS == 64) {
162
+ new_pool_label(s, arg, R_386_PC32, s->code_ptr - 4, -4);
163
} else {
164
- tcg_out_vex_modrm_pool(s, OPC_VBROADCASTSS, ret);
165
+ new_pool_l2(s, R_386_32, s->code_ptr - 4, 0, arg, arg >> 32);
166
}
167
- new_pool_label(s, arg, R_386_32, s->code_ptr - 4, 0);
168
}
169
}
170
171
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
172
index XXXXXXX..XXXXXXX 100644
173
--- a/tcg/ppc/tcg-target.c.inc
174
+++ b/tcg/ppc/tcg-target.c.inc
175
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
176
}
177
}
178
179
-static void tcg_out_dupi_vec(TCGContext *s, TCGType type, TCGReg ret,
180
- tcg_target_long val)
181
+static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
182
+ TCGReg ret, int64_t val)
183
{
184
uint32_t load_insn;
185
int rel, low;
186
intptr_t add;
187
188
- low = (int8_t)val;
189
- if (low >= -16 && low < 16) {
190
- if (val == (tcg_target_long)dup_const(MO_8, low)) {
191
+ switch (vece) {
192
+ case MO_8:
193
+ low = (int8_t)val;
194
+ if (low >= -16 && low < 16) {
195
tcg_out32(s, VSPLTISB | VRT(ret) | ((val & 31) << 16));
196
return;
197
}
198
- if (val == (tcg_target_long)dup_const(MO_16, low)) {
199
+ if (have_isa_3_00) {
200
+ tcg_out32(s, XXSPLTIB | VRT(ret) | ((val & 0xff) << 11));
201
+ return;
202
+ }
203
+ break;
204
+
205
+ case MO_16:
206
+ low = (int16_t)val;
207
+ if (low >= -16 && low < 16) {
208
tcg_out32(s, VSPLTISH | VRT(ret) | ((val & 31) << 16));
209
return;
210
}
211
- if (val == (tcg_target_long)dup_const(MO_32, low)) {
212
+ break;
213
+
214
+ case MO_32:
215
+ low = (int32_t)val;
216
+ if (low >= -16 && low < 16) {
217
tcg_out32(s, VSPLTISW | VRT(ret) | ((val & 31) << 16));
218
return;
219
}
220
- }
221
- if (have_isa_3_00 && val == (tcg_target_long)dup_const(MO_8, val)) {
222
- tcg_out32(s, XXSPLTIB | VRT(ret) | ((val & 0xff) << 11));
223
- return;
224
+ break;
225
}
226
227
/*
228
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type, TCGReg ret,
229
if (TCG_TARGET_REG_BITS == 64) {
230
new_pool_label(s, val, rel, s->code_ptr, add);
231
} else {
232
- new_pool_l2(s, rel, s->code_ptr, add, val, val);
233
+ new_pool_l2(s, rel, s->code_ptr, add, val >> 32, val);
234
}
235
} else {
236
load_insn = LVX | VRT(ret) | RB(TCG_REG_TMP1);
237
if (TCG_TARGET_REG_BITS == 64) {
238
new_pool_l2(s, rel, s->code_ptr, add, val, val);
239
} else {
240
- new_pool_l4(s, rel, s->code_ptr, add, val, val, val, val);
241
+ new_pool_l4(s, rel, s->code_ptr, add,
242
+ val >> 32, val, val >> 32, val);
243
}
55
}
244
}
56
}
245
57
58
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
59
{
60
DisasContext *ctx = container_of(dcbase, DisasContext, base);
61
bool nonconst_skip = canonicalize_skip(ctx);
62
+ /*
63
+ * Because we disable interrupts while env->skip is set,
64
+ * we must return to the main loop to re-evaluate afterward.
65
+ */
66
+ bool force_exit = ctx->base.tb->flags & TB_FLAGS_SKIP;
67
68
switch (ctx->base.is_jmp) {
69
case DISAS_NORETURN:
70
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
71
case DISAS_NEXT:
72
case DISAS_TOO_MANY:
73
case DISAS_CHAIN:
74
- if (!nonconst_skip) {
75
+ if (!nonconst_skip && !force_exit) {
76
/* Note gen_goto_tb checks singlestep. */
77
gen_goto_tb(ctx, 1, ctx->npc);
78
break;
79
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
80
tcg_gen_movi_tl(cpu_pc, ctx->npc);
81
/* fall through */
82
case DISAS_LOOKUP:
83
- tcg_gen_lookup_and_goto_ptr();
84
- break;
85
+ if (!force_exit) {
86
+ tcg_gen_lookup_and_goto_ptr();
87
+ break;
88
+ }
89
+ /* fall through */
90
case DISAS_EXIT:
91
tcg_gen_exit_tb(NULL, 0);
92
break;
246
--
93
--
247
2.25.1
94
2.34.1
248
95
249
96
diff view generated by jsdifflib
1
Prefer TEMP_CONST over anything else.
1
Map the stack executable if required by default or on demand.
2
2
3
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
---
6
---
5
tcg/optimize.c | 27 ++++++++++++---------------
7
include/elf.h | 1 +
6
1 file changed, 12 insertions(+), 15 deletions(-)
8
linux-user/qemu.h | 1 +
9
linux-user/elfload.c | 19 ++++++++++++++++++-
10
3 files changed, 20 insertions(+), 1 deletion(-)
7
11
8
diff --git a/tcg/optimize.c b/tcg/optimize.c
12
diff --git a/include/elf.h b/include/elf.h
9
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/optimize.c
14
--- a/include/elf.h
11
+++ b/tcg/optimize.c
15
+++ b/include/elf.h
12
@@ -XXX,XX +XXX,XX @@ static void init_arg_info(TempOptInfo *infos,
16
@@ -XXX,XX +XXX,XX @@ typedef int64_t Elf64_Sxword;
13
17
#define PT_LOPROC 0x70000000
14
static TCGTemp *find_better_copy(TCGContext *s, TCGTemp *ts)
18
#define PT_HIPROC 0x7fffffff
19
20
+#define PT_GNU_STACK (PT_LOOS + 0x474e551)
21
#define PT_GNU_PROPERTY (PT_LOOS + 0x474e553)
22
23
#define PT_MIPS_REGINFO 0x70000000
24
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
25
index XXXXXXX..XXXXXXX 100644
26
--- a/linux-user/qemu.h
27
+++ b/linux-user/qemu.h
28
@@ -XXX,XX +XXX,XX @@ struct image_info {
29
uint32_t elf_flags;
30
int personality;
31
abi_ulong alignment;
32
+ bool exec_stack;
33
34
/* Generic semihosting knows about these pointers. */
35
abi_ulong arg_strings; /* strings for argv */
36
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/linux-user/elfload.c
39
+++ b/linux-user/elfload.c
40
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
41
#define ELF_ARCH EM_386
42
43
#define ELF_PLATFORM get_elf_platform()
44
+#define EXSTACK_DEFAULT true
45
46
static const char *get_elf_platform(void)
15
{
47
{
16
- TCGTemp *i;
48
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
17
+ TCGTemp *i, *g, *l;
49
18
50
#define ELF_ARCH EM_ARM
19
- /* If this is already a global, we can't do better. */
51
#define ELF_CLASS ELFCLASS32
20
- if (ts->kind >= TEMP_GLOBAL) {
52
+#define EXSTACK_DEFAULT true
21
+ /* If this is already readonly, we can't do better. */
53
22
+ if (temp_readonly(ts)) {
54
static inline void init_thread(struct target_pt_regs *regs,
23
return ts;
55
struct image_info *infop)
56
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
57
#else
58
59
#define ELF_CLASS ELFCLASS32
60
+#define EXSTACK_DEFAULT true
61
62
#endif
63
64
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUPPCState *en
65
66
#define ELF_CLASS ELFCLASS64
67
#define ELF_ARCH EM_LOONGARCH
68
+#define EXSTACK_DEFAULT true
69
70
#define elf_check_arch(x) ((x) == EM_LOONGARCH)
71
72
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
73
#define ELF_CLASS ELFCLASS32
74
#endif
75
#define ELF_ARCH EM_MIPS
76
+#define EXSTACK_DEFAULT true
77
78
#ifdef TARGET_ABI_MIPSN32
79
#define elf_check_abi(x) ((x) & EF_MIPS_ABI2)
80
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
81
#define bswaptls(ptr) bswap32s(ptr)
82
#endif
83
84
+#ifndef EXSTACK_DEFAULT
85
+#define EXSTACK_DEFAULT false
86
+#endif
87
+
88
#include "elf.h"
89
90
/* We must delay the following stanzas until after "elf.h". */
91
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
92
struct image_info *info)
93
{
94
abi_ulong size, error, guard;
95
+ int prot;
96
97
size = guest_stack_size;
98
if (size < STACK_LOWER_LIMIT) {
99
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
100
guard = qemu_real_host_page_size();
24
}
101
}
25
102
26
- /* Search for a global first. */
103
- error = target_mmap(0, size + guard, PROT_READ | PROT_WRITE,
27
+ g = l = NULL;
104
+ prot = PROT_READ | PROT_WRITE;
28
for (i = ts_info(ts)->next_copy; i != ts; i = ts_info(i)->next_copy) {
105
+ if (info->exec_stack) {
29
- if (i->kind >= TEMP_GLOBAL) {
106
+ prot |= PROT_EXEC;
30
+ if (temp_readonly(i)) {
107
+ }
31
return i;
108
+ error = target_mmap(0, size + guard, prot,
32
- }
109
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
33
- }
110
if (error == -1) {
34
-
111
perror("mmap stack");
35
- /* If it is a temp, search for a temp local. */
112
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
36
- if (ts->kind == TEMP_NORMAL) {
113
*/
37
- for (i = ts_info(ts)->next_copy; i != ts; i = ts_info(i)->next_copy) {
114
loaddr = -1, hiaddr = 0;
38
- if (i->kind >= TEMP_LOCAL) {
115
info->alignment = 0;
39
- return i;
116
+ info->exec_stack = EXSTACK_DEFAULT;
40
+ } else if (i->kind > ts->kind) {
117
for (i = 0; i < ehdr->e_phnum; ++i) {
41
+ if (i->kind == TEMP_GLOBAL) {
118
struct elf_phdr *eppnt = phdr + i;
42
+ g = i;
119
if (eppnt->p_type == PT_LOAD) {
43
+ } else if (i->kind == TEMP_LOCAL) {
120
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
44
+ l = i;
121
if (!parse_elf_properties(image_fd, info, eppnt, bprm_buf, &err)) {
122
goto exit_errmsg;
45
}
123
}
124
+ } else if (eppnt->p_type == PT_GNU_STACK) {
125
+ info->exec_stack = eppnt->p_flags & PF_X;
46
}
126
}
47
}
127
}
48
128
49
- /* Failure to find a better representation, return the same temp. */
50
- return ts;
51
+ /* If we didn't find a better representation, return the same temp. */
52
+ return g ? g : l ? l : ts;
53
}
54
55
static bool ts_are_copies(TCGTemp *ts1, TCGTemp *ts2)
56
--
129
--
57
2.25.1
130
2.34.1
58
59
diff view generated by jsdifflib
1
Improve rotrv_vec to reduce "t1 = -v2, t2 = t1 + c" to
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
"t1 = -v2, t2 = c - v2". This avoids a serial dependency
3
between t1 and t2.
4
2
3
Currently it's possible to execute pages that do not have PAGE_EXEC
4
if there is an existing translation block. Fix by invalidating TBs
5
that touch the affected pages.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Message-Id: <20220817150506.592862-2-iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
10
---
7
tcg/aarch64/tcg-target.c.inc | 10 +++++-----
11
linux-user/mmap.c | 6 ++++--
8
1 file changed, 5 insertions(+), 5 deletions(-)
12
1 file changed, 4 insertions(+), 2 deletions(-)
9
13
10
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
14
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
11
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/aarch64/tcg-target.c.inc
16
--- a/linux-user/mmap.c
13
+++ b/tcg/aarch64/tcg-target.c.inc
17
+++ b/linux-user/mmap.c
14
@@ -XXX,XX +XXX,XX @@ void tcg_expand_vec_op(TCGOpcode opc, TCGType type, unsigned vece,
18
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
15
TCGArg a0, ...)
19
goto error;
16
{
20
}
17
va_list va;
21
}
18
- TCGv_vec v0, v1, v2, t1, t2;
22
+
19
+ TCGv_vec v0, v1, v2, t1, t2, c1;
23
page_set_flags(start, start + len, page_flags);
20
TCGArg a2;
24
- mmap_unlock();
21
25
- return 0;
22
va_start(va, a0);
26
+ tb_invalidate_phys_range(start, start + len);
23
@@ -XXX,XX +XXX,XX @@ void tcg_expand_vec_op(TCGOpcode opc, TCGType type, unsigned vece,
27
+ ret = 0;
24
28
+
25
case INDEX_op_rotlv_vec:
29
error:
26
t1 = tcg_temp_new_vec(type);
30
mmap_unlock();
27
- tcg_gen_dupi_vec(vece, t1, 8 << vece);
31
return ret;
28
- tcg_gen_sub_vec(vece, t1, v2, t1);
29
+ c1 = tcg_constant_vec(type, vece, 8 << vece);
30
+ tcg_gen_sub_vec(vece, t1, v2, c1);
31
/* Right shifts are negative left shifts for AArch64. */
32
vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(t1),
33
tcgv_vec_arg(v1), tcgv_vec_arg(t1));
34
@@ -XXX,XX +XXX,XX @@ void tcg_expand_vec_op(TCGOpcode opc, TCGType type, unsigned vece,
35
case INDEX_op_rotrv_vec:
36
t1 = tcg_temp_new_vec(type);
37
t2 = tcg_temp_new_vec(type);
38
+ c1 = tcg_constant_vec(type, vece, 8 << vece);
39
tcg_gen_neg_vec(vece, t1, v2);
40
- tcg_gen_dupi_vec(vece, t2, 8 << vece);
41
- tcg_gen_add_vec(vece, t2, t1, t2);
42
+ tcg_gen_sub_vec(vece, t2, c1, v2);
43
/* Right shifts are negative left shifts for AArch64. */
44
vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(t1),
45
tcgv_vec_arg(v1), tcgv_vec_arg(t1));
46
--
32
--
47
2.25.1
33
2.34.1
48
49
diff view generated by jsdifflib
1
Improve expand_vec_shi to use sign-extraction for MO_32.
1
We're about to start validating PAGE_EXEC, which means
2
This allows a single VSPLTISB instruction to load all of
2
that we've got to put this code into a section that is
3
the valid shift constants.
3
both writable and executable.
4
4
5
Note that this test did not run on hardware beforehand either.
6
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
10
---
7
tcg/ppc/tcg-target.c.inc | 44 ++++++++++++++++++++++++----------------
11
tests/tcg/i386/test-i386.c | 2 +-
8
1 file changed, 27 insertions(+), 17 deletions(-)
12
1 file changed, 1 insertion(+), 1 deletion(-)
9
13
10
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
14
diff --git a/tests/tcg/i386/test-i386.c b/tests/tcg/i386/test-i386.c
11
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/ppc/tcg-target.c.inc
16
--- a/tests/tcg/i386/test-i386.c
13
+++ b/tcg/ppc/tcg-target.c.inc
17
+++ b/tests/tcg/i386/test-i386.c
14
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
18
@@ -XXX,XX +XXX,XX @@ uint8_t code[] = {
15
static void expand_vec_shi(TCGType type, unsigned vece, TCGv_vec v0,
19
0xc3, /* ret */
16
TCGv_vec v1, TCGArg imm, TCGOpcode opci)
20
};
17
{
21
18
- TCGv_vec t1 = tcg_temp_new_vec(type);
22
-asm(".section \".data\"\n"
19
+ TCGv_vec t1;
23
+asm(".section \".data_x\",\"awx\"\n"
20
24
"smc_code2:\n"
21
- /* Splat w/bytes for xxspltib. */
25
"movl 4(%esp), %eax\n"
22
- tcg_gen_dupi_vec(MO_8, t1, imm & ((8 << vece) - 1));
26
"movl %eax, smc_patch_addr2 + 1\n"
23
+ if (vece == MO_32) {
24
+ /*
25
+ * Only 5 bits are significant, and VSPLTISB can represent -16..15.
26
+ * So using negative numbers gets us the 4th bit easily.
27
+ */
28
+ imm = sextract32(imm, 0, 5);
29
+ } else {
30
+ imm &= (8 << vece) - 1;
31
+ }
32
+
33
+ /* Splat w/bytes for xxspltib when 2.07 allows MO_64. */
34
+ t1 = tcg_constant_vec(type, MO_8, imm);
35
vec_gen_3(opci, type, vece, tcgv_vec_arg(v0),
36
tcgv_vec_arg(v1), tcgv_vec_arg(t1));
37
- tcg_temp_free_vec(t1);
38
}
39
40
static void expand_vec_cmp(TCGType type, unsigned vece, TCGv_vec v0,
41
@@ -XXX,XX +XXX,XX @@ static void expand_vec_mul(TCGType type, unsigned vece, TCGv_vec v0,
42
{
43
TCGv_vec t1 = tcg_temp_new_vec(type);
44
TCGv_vec t2 = tcg_temp_new_vec(type);
45
- TCGv_vec t3, t4;
46
+ TCGv_vec c0, c16;
47
48
switch (vece) {
49
case MO_8:
50
@@ -XXX,XX +XXX,XX @@ static void expand_vec_mul(TCGType type, unsigned vece, TCGv_vec v0,
51
52
case MO_32:
53
tcg_debug_assert(!have_isa_2_07);
54
- t3 = tcg_temp_new_vec(type);
55
- t4 = tcg_temp_new_vec(type);
56
- tcg_gen_dupi_vec(MO_8, t4, -16);
57
+ /*
58
+ * Only 5 bits are significant, and VSPLTISB can represent -16..15.
59
+ * So using -16 is a quick way to represent 16.
60
+ */
61
+ c16 = tcg_constant_vec(type, MO_8, -16);
62
+ c0 = tcg_constant_vec(type, MO_8, 0);
63
+
64
vec_gen_3(INDEX_op_rotlv_vec, type, MO_32, tcgv_vec_arg(t1),
65
- tcgv_vec_arg(v2), tcgv_vec_arg(t4));
66
+ tcgv_vec_arg(v2), tcgv_vec_arg(c16));
67
vec_gen_3(INDEX_op_ppc_mulou_vec, type, MO_16, tcgv_vec_arg(t2),
68
tcgv_vec_arg(v1), tcgv_vec_arg(v2));
69
- tcg_gen_dupi_vec(MO_8, t3, 0);
70
- vec_gen_4(INDEX_op_ppc_msum_vec, type, MO_16, tcgv_vec_arg(t3),
71
- tcgv_vec_arg(v1), tcgv_vec_arg(t1), tcgv_vec_arg(t3));
72
- vec_gen_3(INDEX_op_shlv_vec, type, MO_32, tcgv_vec_arg(t3),
73
- tcgv_vec_arg(t3), tcgv_vec_arg(t4));
74
- tcg_gen_add_vec(MO_32, v0, t2, t3);
75
- tcg_temp_free_vec(t3);
76
- tcg_temp_free_vec(t4);
77
+ vec_gen_4(INDEX_op_ppc_msum_vec, type, MO_16, tcgv_vec_arg(t1),
78
+ tcgv_vec_arg(v1), tcgv_vec_arg(t1), tcgv_vec_arg(c0));
79
+ vec_gen_3(INDEX_op_shlv_vec, type, MO_32, tcgv_vec_arg(t1),
80
+ tcgv_vec_arg(t1), tcgv_vec_arg(c16));
81
+ tcg_gen_add_vec(MO_32, v0, t1, t2);
82
break;
83
84
default:
85
--
27
--
86
2.25.1
28
2.34.1
87
88
diff view generated by jsdifflib
1
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
3
Introduce a function that checks whether a given address is on the same
4
page as where disassembly started. Having it improves readability of
5
the following patches.
6
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Message-Id: <20220811095534.241224-3-iii@linux.ibm.com>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
[rth: Make the DisasContextBase parameter const.]
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
13
---
4
tcg/i386/tcg-target.c.inc | 26 +++++++++++++-------------
14
include/exec/translator.h | 10 ++++++++++
5
1 file changed, 13 insertions(+), 13 deletions(-)
15
1 file changed, 10 insertions(+)
6
16
7
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
17
diff --git a/include/exec/translator.h b/include/exec/translator.h
8
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/i386/tcg-target.c.inc
19
--- a/include/exec/translator.h
10
+++ b/tcg/i386/tcg-target.c.inc
20
+++ b/include/exec/translator.h
11
@@ -XXX,XX +XXX,XX @@ static void expand_vec_rotv(TCGType type, unsigned vece, TCGv_vec v0,
21
@@ -XXX,XX +XXX,XX @@ FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
12
static void expand_vec_mul(TCGType type, unsigned vece,
22
13
TCGv_vec v0, TCGv_vec v1, TCGv_vec v2)
23
#undef GEN_TRANSLATOR_LD
14
{
24
15
- TCGv_vec t1, t2, t3, t4;
25
+/*
16
+ TCGv_vec t1, t2, t3, t4, zero;
26
+ * Return whether addr is on the same page as where disassembly started.
17
27
+ * Translators can use this to enforce the rule that only single-insn
18
tcg_debug_assert(vece == MO_8);
28
+ * translation blocks are allowed to cross page boundaries.
19
29
+ */
20
@@ -XXX,XX +XXX,XX @@ static void expand_vec_mul(TCGType type, unsigned vece,
30
+static inline bool is_same_page(const DisasContextBase *db, target_ulong addr)
21
case TCG_TYPE_V64:
31
+{
22
t1 = tcg_temp_new_vec(TCG_TYPE_V128);
32
+ return ((addr ^ db->pc_first) & TARGET_PAGE_MASK) == 0;
23
t2 = tcg_temp_new_vec(TCG_TYPE_V128);
33
+}
24
- tcg_gen_dup16i_vec(t2, 0);
34
+
25
+ zero = tcg_constant_vec(TCG_TYPE_V128, MO_8, 0);
35
#endif /* EXEC__TRANSLATOR_H */
26
vec_gen_3(INDEX_op_x86_punpckl_vec, TCG_TYPE_V128, MO_8,
27
- tcgv_vec_arg(t1), tcgv_vec_arg(v1), tcgv_vec_arg(t2));
28
+ tcgv_vec_arg(t1), tcgv_vec_arg(v1), tcgv_vec_arg(zero));
29
vec_gen_3(INDEX_op_x86_punpckl_vec, TCG_TYPE_V128, MO_8,
30
- tcgv_vec_arg(t2), tcgv_vec_arg(t2), tcgv_vec_arg(v2));
31
+ tcgv_vec_arg(t2), tcgv_vec_arg(zero), tcgv_vec_arg(v2));
32
tcg_gen_mul_vec(MO_16, t1, t1, t2);
33
tcg_gen_shri_vec(MO_16, t1, t1, 8);
34
vec_gen_3(INDEX_op_x86_packus_vec, TCG_TYPE_V128, MO_8,
35
@@ -XXX,XX +XXX,XX @@ static void expand_vec_mul(TCGType type, unsigned vece,
36
t2 = tcg_temp_new_vec(type);
37
t3 = tcg_temp_new_vec(type);
38
t4 = tcg_temp_new_vec(type);
39
- tcg_gen_dup16i_vec(t4, 0);
40
+ zero = tcg_constant_vec(TCG_TYPE_V128, MO_8, 0);
41
vec_gen_3(INDEX_op_x86_punpckl_vec, type, MO_8,
42
- tcgv_vec_arg(t1), tcgv_vec_arg(v1), tcgv_vec_arg(t4));
43
+ tcgv_vec_arg(t1), tcgv_vec_arg(v1), tcgv_vec_arg(zero));
44
vec_gen_3(INDEX_op_x86_punpckl_vec, type, MO_8,
45
- tcgv_vec_arg(t2), tcgv_vec_arg(t4), tcgv_vec_arg(v2));
46
+ tcgv_vec_arg(t2), tcgv_vec_arg(zero), tcgv_vec_arg(v2));
47
vec_gen_3(INDEX_op_x86_punpckh_vec, type, MO_8,
48
- tcgv_vec_arg(t3), tcgv_vec_arg(v1), tcgv_vec_arg(t4));
49
+ tcgv_vec_arg(t3), tcgv_vec_arg(v1), tcgv_vec_arg(zero));
50
vec_gen_3(INDEX_op_x86_punpckh_vec, type, MO_8,
51
- tcgv_vec_arg(t4), tcgv_vec_arg(t4), tcgv_vec_arg(v2));
52
+ tcgv_vec_arg(t4), tcgv_vec_arg(zero), tcgv_vec_arg(v2));
53
tcg_gen_mul_vec(MO_16, t1, t1, t2);
54
tcg_gen_mul_vec(MO_16, t3, t3, t4);
55
tcg_gen_shri_vec(MO_16, t1, t1, 8);
56
@@ -XXX,XX +XXX,XX @@ static bool expand_vec_cmp_noinv(TCGType type, unsigned vece, TCGv_vec v0,
57
NEED_UMIN = 8,
58
NEED_UMAX = 16,
59
};
60
- TCGv_vec t1, t2;
61
+ TCGv_vec t1, t2, t3;
62
uint8_t fixup;
63
64
switch (cond) {
65
@@ -XXX,XX +XXX,XX @@ static bool expand_vec_cmp_noinv(TCGType type, unsigned vece, TCGv_vec v0,
66
} else if (fixup & NEED_BIAS) {
67
t1 = tcg_temp_new_vec(type);
68
t2 = tcg_temp_new_vec(type);
69
- tcg_gen_dupi_vec(vece, t2, 1ull << ((8 << vece) - 1));
70
- tcg_gen_sub_vec(vece, t1, v1, t2);
71
- tcg_gen_sub_vec(vece, t2, v2, t2);
72
+ t3 = tcg_constant_vec(type, vece, 1ull << ((8 << vece) - 1));
73
+ tcg_gen_sub_vec(vece, t1, v1, t3);
74
+ tcg_gen_sub_vec(vece, t2, v2, t3);
75
v1 = t1;
76
v2 = t2;
77
cond = tcg_signed_cond(cond);
78
--
36
--
79
2.25.1
37
2.34.1
80
81
diff view generated by jsdifflib
1
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
1
The current implementation is a no-op, simply returning addr.
2
This is incorrect, because we ought to be checking the page
3
permissions for execution.
4
5
Make get_page_addr_code inline for both implementations.
6
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Acked-by: Alistair Francis <alistair.francis@wdc.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
11
---
4
accel/tcg/plugin-gen.c | 49 +++++++++++++++++++-----------------------
12
include/exec/exec-all.h | 85 ++++++++++++++---------------------------
5
1 file changed, 22 insertions(+), 27 deletions(-)
13
accel/tcg/cputlb.c | 5 ---
14
accel/tcg/user-exec.c | 14 +++++++
15
3 files changed, 42 insertions(+), 62 deletions(-)
6
16
7
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
17
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
8
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
9
--- a/accel/tcg/plugin-gen.c
19
--- a/include/exec/exec-all.h
10
+++ b/accel/tcg/plugin-gen.c
20
+++ b/include/exec/exec-all.h
11
@@ -XXX,XX +XXX,XX @@ static TCGOp *copy_extu_i32_i64(TCGOp **begin_op, TCGOp *op)
21
@@ -XXX,XX +XXX,XX @@ struct MemoryRegionSection *iotlb_to_section(CPUState *cpu,
12
if (TCG_TARGET_REG_BITS == 32) {
22
hwaddr index, MemTxAttrs attrs);
13
/* mov_i32 */
23
#endif
14
op = copy_op(begin_op, op, INDEX_op_mov_i32);
24
15
- /* movi_i32 */
25
-#if defined(CONFIG_USER_ONLY)
16
- op = copy_op(begin_op, op, INDEX_op_movi_i32);
26
-void mmap_lock(void);
17
+ /* mov_i32 w/ $0 */
27
-void mmap_unlock(void);
18
+ op = copy_op(begin_op, op, INDEX_op_mov_i32);
28
-bool have_mmap_lock(void);
19
} else {
29
-
20
/* extu_i32_i64 */
30
/**
21
op = copy_op(begin_op, op, INDEX_op_extu_i32_i64);
31
- * get_page_addr_code() - user-mode version
22
@@ -XXX,XX +XXX,XX @@ static TCGOp *copy_mov_i64(TCGOp **begin_op, TCGOp *op)
32
+ * get_page_addr_code_hostp()
23
return op;
33
* @env: CPUArchState
34
* @addr: guest virtual address of guest code
35
*
36
- * Returns @addr.
37
+ * See get_page_addr_code() (full-system version) for documentation on the
38
+ * return value.
39
+ *
40
+ * Sets *@hostp (when @hostp is non-NULL) as follows.
41
+ * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
42
+ * to the host address where @addr's content is kept.
43
+ *
44
+ * Note: this function can trigger an exception.
45
+ */
46
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
47
+ void **hostp);
48
+
49
+/**
50
+ * get_page_addr_code()
51
+ * @env: CPUArchState
52
+ * @addr: guest virtual address of guest code
53
+ *
54
+ * If we cannot translate and execute from the entire RAM page, or if
55
+ * the region is not backed by RAM, returns -1. Otherwise, returns the
56
+ * ram_addr_t corresponding to the guest code at @addr.
57
+ *
58
+ * Note: this function can trigger an exception.
59
*/
60
static inline tb_page_addr_t get_page_addr_code(CPUArchState *env,
61
target_ulong addr)
62
{
63
- return addr;
64
+ return get_page_addr_code_hostp(env, addr, NULL);
24
}
65
}
25
66
26
-static TCGOp *copy_movi_i64(TCGOp **begin_op, TCGOp *op, uint64_t v)
67
-/**
68
- * get_page_addr_code_hostp() - user-mode version
69
- * @env: CPUArchState
70
- * @addr: guest virtual address of guest code
71
- *
72
- * Returns @addr.
73
- *
74
- * If @hostp is non-NULL, sets *@hostp to the host address where @addr's content
75
- * is kept.
76
- */
77
-static inline tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env,
78
- target_ulong addr,
79
- void **hostp)
27
-{
80
-{
28
- if (TCG_TARGET_REG_BITS == 32) {
81
- if (hostp) {
29
- /* 2x movi_i32 */
82
- *hostp = g2h_untagged(addr);
30
- op = copy_op(begin_op, op, INDEX_op_movi_i32);
83
- }
31
- op->args[1] = v;
84
- return addr;
85
-}
86
+#if defined(CONFIG_USER_ONLY)
87
+void mmap_lock(void);
88
+void mmap_unlock(void);
89
+bool have_mmap_lock(void);
90
91
/**
92
* adjust_signal_pc:
93
@@ -XXX,XX +XXX,XX @@ G_NORETURN void cpu_loop_exit_sigbus(CPUState *cpu, target_ulong addr,
94
static inline void mmap_lock(void) {}
95
static inline void mmap_unlock(void) {}
96
97
-/**
98
- * get_page_addr_code() - full-system version
99
- * @env: CPUArchState
100
- * @addr: guest virtual address of guest code
101
- *
102
- * If we cannot translate and execute from the entire RAM page, or if
103
- * the region is not backed by RAM, returns -1. Otherwise, returns the
104
- * ram_addr_t corresponding to the guest code at @addr.
105
- *
106
- * Note: this function can trigger an exception.
107
- */
108
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr);
32
-
109
-
33
- op = copy_op(begin_op, op, INDEX_op_movi_i32);
110
-/**
34
- op->args[1] = v >> 32;
111
- * get_page_addr_code_hostp() - full-system version
35
- } else {
112
- * @env: CPUArchState
36
- /* movi_i64 */
113
- * @addr: guest virtual address of guest code
37
- op = copy_op(begin_op, op, INDEX_op_movi_i64);
114
- *
38
- op->args[1] = v;
115
- * See get_page_addr_code() (full-system version) for documentation on the
39
- }
116
- * return value.
40
- return op;
117
- *
118
- * Sets *@hostp (when @hostp is non-NULL) as follows.
119
- * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
120
- * to the host address where @addr's content is kept.
121
- *
122
- * Note: this function can trigger an exception.
123
- */
124
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
125
- void **hostp);
126
-
127
void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length);
128
void tlb_set_dirty(CPUState *cpu, target_ulong vaddr);
129
130
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
131
index XXXXXXX..XXXXXXX 100644
132
--- a/accel/tcg/cputlb.c
133
+++ b/accel/tcg/cputlb.c
134
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
135
return qemu_ram_addr_from_host_nofail(p);
136
}
137
138
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr)
139
-{
140
- return get_page_addr_code_hostp(env, addr, NULL);
41
-}
141
-}
42
-
142
-
43
static TCGOp *copy_const_ptr(TCGOp **begin_op, TCGOp *op, void *ptr)
143
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
144
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
44
{
145
{
45
if (UINTPTR_MAX == UINT32_MAX) {
146
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
46
- /* movi_i32 */
147
index XXXXXXX..XXXXXXX 100644
47
- op = copy_op(begin_op, op, INDEX_op_movi_i32);
148
--- a/accel/tcg/user-exec.c
48
- op->args[1] = (uintptr_t)ptr;
149
+++ b/accel/tcg/user-exec.c
49
+ /* mov_i32 */
150
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
50
+ op = copy_op(begin_op, op, INDEX_op_mov_i32);
151
return size ? g2h(env_cpu(env), addr) : NULL;
51
+ op->args[1] = tcgv_i32_arg(tcg_constant_i32((uintptr_t)ptr));
52
} else {
53
- /* movi_i64 */
54
- op = copy_movi_i64(begin_op, op, (uint64_t)(uintptr_t)ptr);
55
+ /* mov_i64 */
56
+ op = copy_op(begin_op, op, INDEX_op_mov_i64);
57
+ op->args[1] = tcgv_i64_arg(tcg_constant_i64((uintptr_t)ptr));
58
}
59
return op;
60
}
152
}
61
153
62
static TCGOp *copy_const_i64(TCGOp **begin_op, TCGOp *op, uint64_t v)
154
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
63
{
155
+ void **hostp)
64
- return copy_movi_i64(begin_op, op, v);
156
+{
65
+ if (TCG_TARGET_REG_BITS == 32) {
157
+ int flags;
66
+ /* 2x mov_i32 */
158
+
67
+ op = copy_op(begin_op, op, INDEX_op_mov_i32);
159
+ flags = probe_access_internal(env, addr, 1, MMU_INST_FETCH, false, 0);
68
+ op->args[1] = tcgv_i32_arg(tcg_constant_i32(v));
160
+ g_assert(flags == 0);
69
+ op = copy_op(begin_op, op, INDEX_op_mov_i32);
161
+
70
+ op->args[1] = tcgv_i32_arg(tcg_constant_i32(v >> 32));
162
+ if (hostp) {
71
+ } else {
163
+ *hostp = g2h_untagged(addr);
72
+ /* mov_i64 */
73
+ op = copy_op(begin_op, op, INDEX_op_mov_i64);
74
+ op->args[1] = tcgv_i64_arg(tcg_constant_i64(v));
75
+ }
164
+ }
76
+ return op;
165
+ return addr;
77
}
166
+}
78
167
+
79
static TCGOp *copy_extu_tl_i64(TCGOp **begin_op, TCGOp *op)
168
/* The softmmu versions of these helpers are in cputlb.c. */
80
@@ -XXX,XX +XXX,XX @@ static TCGOp *append_mem_cb(const struct qemu_plugin_dyn_cb *cb,
169
81
170
/*
82
tcg_debug_assert(type == PLUGIN_GEN_CB_MEM);
83
84
- /* const_i32 == movi_i32 ("info", so it remains as is) */
85
- op = copy_op(&begin_op, op, INDEX_op_movi_i32);
86
+ /* const_i32 == mov_i32 ("info", so it remains as is) */
87
+ op = copy_op(&begin_op, op, INDEX_op_mov_i32);
88
89
/* const_ptr */
90
op = copy_const_ptr(&begin_op, op, cb->userp);
91
--
171
--
92
2.25.1
172
2.34.1
93
94
diff view generated by jsdifflib
1
We must do this before we adjust tcg_out_movi_i32, lest the
1
The mmap_lock is held around tb_gen_code. While the comment
2
under-the-hood poking that we do for icount be broken.
2
is correct that the lock is dropped when tb_gen_code runs out
3
of memory, the lock is *not* dropped when an exception is
4
raised reading code for translation.
3
5
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Acked-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
10
---
7
include/exec/gen-icount.h | 25 +++++++++++++------------
11
accel/tcg/cpu-exec.c | 12 ++++++------
8
1 file changed, 13 insertions(+), 12 deletions(-)
12
accel/tcg/user-exec.c | 3 ---
13
2 files changed, 6 insertions(+), 9 deletions(-)
9
14
10
diff --git a/include/exec/gen-icount.h b/include/exec/gen-icount.h
15
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
11
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
12
--- a/include/exec/gen-icount.h
17
--- a/accel/tcg/cpu-exec.c
13
+++ b/include/exec/gen-icount.h
18
+++ b/accel/tcg/cpu-exec.c
14
@@ -XXX,XX +XXX,XX @@ static inline void gen_io_end(void)
19
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
15
20
cpu_tb_exec(cpu, tb, &tb_exit);
16
static inline void gen_tb_start(const TranslationBlock *tb)
21
cpu_exec_exit(cpu);
17
{
22
} else {
18
- TCGv_i32 count, imm;
23
- /*
19
+ TCGv_i32 count;
24
- * The mmap_lock is dropped by tb_gen_code if it runs out of
20
25
- * memory.
21
tcg_ctx->exitreq_label = gen_new_label();
26
- */
22
if (tb_cflags(tb) & CF_USE_ICOUNT) {
27
#ifndef CONFIG_SOFTMMU
23
@@ -XXX,XX +XXX,XX @@ static inline void gen_tb_start(const TranslationBlock *tb)
28
clear_helper_retaddr();
24
offsetof(ArchCPU, env));
29
- tcg_debug_assert(!have_mmap_lock());
25
30
+ if (have_mmap_lock()) {
26
if (tb_cflags(tb) & CF_USE_ICOUNT) {
31
+ mmap_unlock();
27
- imm = tcg_temp_new_i32();
32
+ }
28
- /* We emit a movi with a dummy immediate argument. Keep the insn index
33
#endif
29
- * of the movi so that we later (when we know the actual insn count)
34
if (qemu_mutex_iothread_locked()) {
30
- * can update the immediate argument with the actual insn count. */
35
qemu_mutex_unlock_iothread();
31
- tcg_gen_movi_i32(imm, 0xdeadbeef);
36
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
32
+ /*
37
33
+ * We emit a sub with a dummy immediate argument. Keep the insn index
38
#ifndef CONFIG_SOFTMMU
34
+ * of the sub so that we later (when we know the actual insn count)
39
clear_helper_retaddr();
35
+ * can update the argument with the actual insn count.
40
- tcg_debug_assert(!have_mmap_lock());
36
+ */
41
+ if (have_mmap_lock()) {
37
+ tcg_gen_sub_i32(count, count, tcg_constant_i32(0));
42
+ mmap_unlock();
38
icount_start_insn = tcg_last_op();
43
+ }
39
-
44
#endif
40
- tcg_gen_sub_i32(count, count, imm);
45
if (qemu_mutex_iothread_locked()) {
41
- tcg_temp_free_i32(imm);
46
qemu_mutex_unlock_iothread();
47
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/accel/tcg/user-exec.c
50
+++ b/accel/tcg/user-exec.c
51
@@ -XXX,XX +XXX,XX @@ MMUAccessType adjust_signal_pc(uintptr_t *pc, bool is_write)
52
* (and if the translator doesn't handle page boundaries correctly
53
* there's little we can do about that here). Therefore, do not
54
* trigger the unwinder.
55
- *
56
- * Like tb_gen_code, release the memory lock before cpu_loop_exit.
57
*/
58
- mmap_unlock();
59
*pc = 0;
60
return MMU_INST_FETCH;
42
}
61
}
43
44
tcg_gen_brcondi_i32(TCG_COND_LT, count, 0, tcg_ctx->exitreq_label);
45
@@ -XXX,XX +XXX,XX @@ static inline void gen_tb_start(const TranslationBlock *tb)
46
static inline void gen_tb_end(const TranslationBlock *tb, int num_insns)
47
{
48
if (tb_cflags(tb) & CF_USE_ICOUNT) {
49
- /* Update the num_insn immediate parameter now that we know
50
- * the actual insn count. */
51
- tcg_set_insn_param(icount_start_insn, 1, num_insns);
52
+ /*
53
+ * Update the num_insn immediate parameter now that we know
54
+ * the actual insn count.
55
+ */
56
+ tcg_set_insn_param(icount_start_insn, 2,
57
+ tcgv_i32_arg(tcg_constant_i32(num_insns)));
58
}
59
60
gen_set_label(tcg_ctx->exitreq_label);
61
--
62
--
62
2.25.1
63
2.34.1
63
64
diff view generated by jsdifflib
1
These are now completely covered by mov from a
1
The function is not used outside of cpu-exec.c. Move it and
2
TYPE_CONST temporary.
2
its subroutines up in the file, before the first use.
3
3
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Reviewed-by: Aleksandar Markovic <aleksandar.qemu.devel@gmail.com>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
---
8
include/tcg/tcg-opc.h | 3 ---
9
include/exec/exec-all.h | 3 -
9
tcg/optimize.c | 4 ----
10
accel/tcg/cpu-exec.c | 122 ++++++++++++++++++++--------------------
10
tcg/tcg-op-vec.c | 1 -
11
2 files changed, 61 insertions(+), 64 deletions(-)
11
tcg/tcg.c | 18 +-----------------
12
tcg/aarch64/tcg-target.c.inc | 3 ---
13
tcg/arm/tcg-target.c.inc | 1 -
14
tcg/i386/tcg-target.c.inc | 3 ---
15
tcg/mips/tcg-target.c.inc | 2 --
16
tcg/ppc/tcg-target.c.inc | 3 ---
17
tcg/riscv/tcg-target.c.inc | 2 --
18
tcg/s390/tcg-target.c.inc | 2 --
19
tcg/sparc/tcg-target.c.inc | 2 --
20
tcg/tci/tcg-target.c.inc | 2 --
21
13 files changed, 1 insertion(+), 45 deletions(-)
22
12
23
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
13
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
24
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
25
--- a/include/tcg/tcg-opc.h
15
--- a/include/exec/exec-all.h
26
+++ b/include/tcg/tcg-opc.h
16
+++ b/include/exec/exec-all.h
27
@@ -XXX,XX +XXX,XX @@ DEF(br, 0, 0, 1, TCG_OPF_BB_END)
17
@@ -XXX,XX +XXX,XX @@ void tb_invalidate_phys_addr(AddressSpace *as, hwaddr addr, MemTxAttrs attrs);
28
DEF(mb, 0, 0, 1, 0)
18
#endif
29
19
void tb_flush(CPUState *cpu);
30
DEF(mov_i32, 1, 1, 0, TCG_OPF_NOT_PRESENT)
20
void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr);
31
-DEF(movi_i32, 1, 0, 1, TCG_OPF_NOT_PRESENT)
21
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
32
DEF(setcond_i32, 1, 2, 1, 0)
22
- target_ulong cs_base, uint32_t flags,
33
DEF(movcond_i32, 1, 4, 1, IMPL(TCG_TARGET_HAS_movcond_i32))
23
- uint32_t cflags);
34
/* load/store */
24
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr);
35
@@ -XXX,XX +XXX,XX @@ DEF(ctz_i32, 1, 2, 0, IMPL(TCG_TARGET_HAS_ctz_i32))
25
36
DEF(ctpop_i32, 1, 1, 0, IMPL(TCG_TARGET_HAS_ctpop_i32))
26
/* GETPC is the true target of the return instruction that we'll execute. */
37
27
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
38
DEF(mov_i64, 1, 1, 0, TCG_OPF_64BIT | TCG_OPF_NOT_PRESENT)
39
-DEF(movi_i64, 1, 0, 1, TCG_OPF_64BIT | TCG_OPF_NOT_PRESENT)
40
DEF(setcond_i64, 1, 2, 1, IMPL64)
41
DEF(movcond_i64, 1, 4, 1, IMPL64 | IMPL(TCG_TARGET_HAS_movcond_i64))
42
/* load/store */
43
@@ -XXX,XX +XXX,XX @@ DEF(qemu_st8_i32, 0, TLADDR_ARGS + 1, 1,
44
#define IMPLVEC TCG_OPF_VECTOR | IMPL(TCG_TARGET_MAYBE_vec)
45
46
DEF(mov_vec, 1, 1, 0, TCG_OPF_VECTOR | TCG_OPF_NOT_PRESENT)
47
-DEF(dupi_vec, 1, 0, 1, TCG_OPF_VECTOR | TCG_OPF_NOT_PRESENT)
48
49
DEF(dup_vec, 1, 1, 0, IMPLVEC)
50
DEF(dup2_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_REG_BITS == 32))
51
diff --git a/tcg/optimize.c b/tcg/optimize.c
52
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
53
--- a/tcg/optimize.c
29
--- a/accel/tcg/cpu-exec.c
54
+++ b/tcg/optimize.c
30
+++ b/accel/tcg/cpu-exec.c
55
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
31
@@ -XXX,XX +XXX,XX @@ uint32_t curr_cflags(CPUState *cpu)
56
CASE_OP_32_64_VEC(mov):
32
return cflags;
57
tcg_opt_gen_mov(s, op, op->args[0], op->args[1]);
58
break;
59
- CASE_OP_32_64(movi):
60
- case INDEX_op_dupi_vec:
61
- tcg_opt_gen_movi(s, &temps_used, op, op->args[0], op->args[1]);
62
- break;
63
64
case INDEX_op_dup_vec:
65
if (arg_is_const(op->args[1])) {
66
diff --git a/tcg/tcg-op-vec.c b/tcg/tcg-op-vec.c
67
index XXXXXXX..XXXXXXX 100644
68
--- a/tcg/tcg-op-vec.c
69
+++ b/tcg/tcg-op-vec.c
70
@@ -XXX,XX +XXX,XX @@ bool tcg_can_emit_vecop_list(const TCGOpcode *list,
71
case INDEX_op_xor_vec:
72
case INDEX_op_mov_vec:
73
case INDEX_op_dup_vec:
74
- case INDEX_op_dupi_vec:
75
case INDEX_op_dup2_vec:
76
case INDEX_op_ld_vec:
77
case INDEX_op_st_vec:
78
diff --git a/tcg/tcg.c b/tcg/tcg.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/tcg/tcg.c
81
+++ b/tcg/tcg.c
82
@@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op)
83
return TCG_TARGET_HAS_goto_ptr;
84
85
case INDEX_op_mov_i32:
86
- case INDEX_op_movi_i32:
87
case INDEX_op_setcond_i32:
88
case INDEX_op_brcond_i32:
89
case INDEX_op_ld8u_i32:
90
@@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op)
91
return TCG_TARGET_REG_BITS == 32;
92
93
case INDEX_op_mov_i64:
94
- case INDEX_op_movi_i64:
95
case INDEX_op_setcond_i64:
96
case INDEX_op_brcond_i64:
97
case INDEX_op_ld8u_i64:
98
@@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op)
99
100
case INDEX_op_mov_vec:
101
case INDEX_op_dup_vec:
102
- case INDEX_op_dupi_vec:
103
case INDEX_op_dupm_vec:
104
case INDEX_op_ld_vec:
105
case INDEX_op_st_vec:
106
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_cbranch(TCGContext *s, TCGRegSet allocated_regs)
107
}
33
}
108
34
109
/*
35
+struct tb_desc {
110
- * Specialized code generation for INDEX_op_movi_*.
36
+ target_ulong pc;
111
+ * Specialized code generation for INDEX_op_mov_* with a constant.
37
+ target_ulong cs_base;
112
*/
38
+ CPUArchState *env;
113
static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
39
+ tb_page_addr_t phys_page1;
114
tcg_target_ulong val, TCGLifeData arg_life,
40
+ uint32_t flags;
115
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
41
+ uint32_t cflags;
116
}
42
+ uint32_t trace_vcpu_dstate;
43
+};
44
+
45
+static bool tb_lookup_cmp(const void *p, const void *d)
46
+{
47
+ const TranslationBlock *tb = p;
48
+ const struct tb_desc *desc = d;
49
+
50
+ if (tb->pc == desc->pc &&
51
+ tb->page_addr[0] == desc->phys_page1 &&
52
+ tb->cs_base == desc->cs_base &&
53
+ tb->flags == desc->flags &&
54
+ tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
55
+ tb_cflags(tb) == desc->cflags) {
56
+ /* check next page if needed */
57
+ if (tb->page_addr[1] == -1) {
58
+ return true;
59
+ } else {
60
+ tb_page_addr_t phys_page2;
61
+ target_ulong virt_page2;
62
+
63
+ virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
64
+ phys_page2 = get_page_addr_code(desc->env, virt_page2);
65
+ if (tb->page_addr[1] == phys_page2) {
66
+ return true;
67
+ }
68
+ }
69
+ }
70
+ return false;
71
+}
72
+
73
+static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
74
+ target_ulong cs_base, uint32_t flags,
75
+ uint32_t cflags)
76
+{
77
+ tb_page_addr_t phys_pc;
78
+ struct tb_desc desc;
79
+ uint32_t h;
80
+
81
+ desc.env = cpu->env_ptr;
82
+ desc.cs_base = cs_base;
83
+ desc.flags = flags;
84
+ desc.cflags = cflags;
85
+ desc.trace_vcpu_dstate = *cpu->trace_dstate;
86
+ desc.pc = pc;
87
+ phys_pc = get_page_addr_code(desc.env, pc);
88
+ if (phys_pc == -1) {
89
+ return NULL;
90
+ }
91
+ desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
92
+ h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
93
+ return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
94
+}
95
+
96
/* Might cause an exception, so have a longjmp destination ready */
97
static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
98
target_ulong cs_base,
99
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
100
end_exclusive();
117
}
101
}
118
102
119
-static void tcg_reg_alloc_movi(TCGContext *s, const TCGOp *op)
103
-struct tb_desc {
104
- target_ulong pc;
105
- target_ulong cs_base;
106
- CPUArchState *env;
107
- tb_page_addr_t phys_page1;
108
- uint32_t flags;
109
- uint32_t cflags;
110
- uint32_t trace_vcpu_dstate;
111
-};
112
-
113
-static bool tb_lookup_cmp(const void *p, const void *d)
120
-{
114
-{
121
- TCGTemp *ots = arg_temp(op->args[0]);
115
- const TranslationBlock *tb = p;
122
- tcg_target_ulong val = op->args[1];
116
- const struct tb_desc *desc = d;
123
-
117
-
124
- tcg_reg_alloc_do_movi(s, ots, val, op->life, op->output_pref[0]);
118
- if (tb->pc == desc->pc &&
119
- tb->page_addr[0] == desc->phys_page1 &&
120
- tb->cs_base == desc->cs_base &&
121
- tb->flags == desc->flags &&
122
- tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
123
- tb_cflags(tb) == desc->cflags) {
124
- /* check next page if needed */
125
- if (tb->page_addr[1] == -1) {
126
- return true;
127
- } else {
128
- tb_page_addr_t phys_page2;
129
- target_ulong virt_page2;
130
-
131
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
132
- phys_page2 = get_page_addr_code(desc->env, virt_page2);
133
- if (tb->page_addr[1] == phys_page2) {
134
- return true;
135
- }
136
- }
137
- }
138
- return false;
125
-}
139
-}
126
-
140
-
127
/*
141
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
128
* Specialized code generation for INDEX_op_mov_*.
142
- target_ulong cs_base, uint32_t flags,
129
*/
143
- uint32_t cflags)
130
@@ -XXX,XX +XXX,XX @@ int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
144
-{
131
case INDEX_op_mov_vec:
145
- tb_page_addr_t phys_pc;
132
tcg_reg_alloc_mov(s, op);
146
- struct tb_desc desc;
133
break;
147
- uint32_t h;
134
- case INDEX_op_movi_i32:
148
-
135
- case INDEX_op_movi_i64:
149
- desc.env = cpu->env_ptr;
136
- case INDEX_op_dupi_vec:
150
- desc.cs_base = cs_base;
137
- tcg_reg_alloc_movi(s, op);
151
- desc.flags = flags;
138
- break;
152
- desc.cflags = cflags;
139
case INDEX_op_dup_vec:
153
- desc.trace_vcpu_dstate = *cpu->trace_dstate;
140
tcg_reg_alloc_dup(s, op);
154
- desc.pc = pc;
141
break;
155
- phys_pc = get_page_addr_code(desc.env, pc);
142
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
156
- if (phys_pc == -1) {
143
index XXXXXXX..XXXXXXX 100644
157
- return NULL;
144
--- a/tcg/aarch64/tcg-target.c.inc
158
- }
145
+++ b/tcg/aarch64/tcg-target.c.inc
159
- desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
146
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
160
- h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
147
161
- return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
148
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
162
-}
149
case INDEX_op_mov_i64:
163
-
150
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
164
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr)
151
- case INDEX_op_movi_i64:
165
{
152
case INDEX_op_call: /* Always emitted via tcg_out_call. */
166
if (TCG_TARGET_HAS_direct_jump) {
153
default:
154
g_assert_not_reached();
155
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
156
break;
157
158
case INDEX_op_mov_vec: /* Always emitted via tcg_out_mov. */
159
- case INDEX_op_dupi_vec: /* Always emitted via tcg_out_movi. */
160
case INDEX_op_dup_vec: /* Always emitted via tcg_out_dup_vec. */
161
default:
162
g_assert_not_reached();
163
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
164
index XXXXXXX..XXXXXXX 100644
165
--- a/tcg/arm/tcg-target.c.inc
166
+++ b/tcg/arm/tcg-target.c.inc
167
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
168
break;
169
170
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
171
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
172
case INDEX_op_call: /* Always emitted via tcg_out_call. */
173
default:
174
tcg_abort();
175
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
176
index XXXXXXX..XXXXXXX 100644
177
--- a/tcg/i386/tcg-target.c.inc
178
+++ b/tcg/i386/tcg-target.c.inc
179
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
180
break;
181
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
182
case INDEX_op_mov_i64:
183
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
184
- case INDEX_op_movi_i64:
185
case INDEX_op_call: /* Always emitted via tcg_out_call. */
186
default:
187
tcg_abort();
188
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
189
break;
190
191
case INDEX_op_mov_vec: /* Always emitted via tcg_out_mov. */
192
- case INDEX_op_dupi_vec: /* Always emitted via tcg_out_movi. */
193
case INDEX_op_dup_vec: /* Always emitted via tcg_out_dup_vec. */
194
default:
195
g_assert_not_reached();
196
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
197
index XXXXXXX..XXXXXXX 100644
198
--- a/tcg/mips/tcg-target.c.inc
199
+++ b/tcg/mips/tcg-target.c.inc
200
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
201
break;
202
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
203
case INDEX_op_mov_i64:
204
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
205
- case INDEX_op_movi_i64:
206
case INDEX_op_call: /* Always emitted via tcg_out_call. */
207
default:
208
tcg_abort();
209
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
210
index XXXXXXX..XXXXXXX 100644
211
--- a/tcg/ppc/tcg-target.c.inc
212
+++ b/tcg/ppc/tcg-target.c.inc
213
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
214
215
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
216
case INDEX_op_mov_i64:
217
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
218
- case INDEX_op_movi_i64:
219
case INDEX_op_call: /* Always emitted via tcg_out_call. */
220
default:
221
tcg_abort();
222
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
223
return;
224
225
case INDEX_op_mov_vec: /* Always emitted via tcg_out_mov. */
226
- case INDEX_op_dupi_vec: /* Always emitted via tcg_out_movi. */
227
case INDEX_op_dup_vec: /* Always emitted via tcg_out_dup_vec. */
228
default:
229
g_assert_not_reached();
230
diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc
231
index XXXXXXX..XXXXXXX 100644
232
--- a/tcg/riscv/tcg-target.c.inc
233
+++ b/tcg/riscv/tcg-target.c.inc
234
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
235
236
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
237
case INDEX_op_mov_i64:
238
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
239
- case INDEX_op_movi_i64:
240
case INDEX_op_call: /* Always emitted via tcg_out_call. */
241
default:
242
g_assert_not_reached();
243
diff --git a/tcg/s390/tcg-target.c.inc b/tcg/s390/tcg-target.c.inc
244
index XXXXXXX..XXXXXXX 100644
245
--- a/tcg/s390/tcg-target.c.inc
246
+++ b/tcg/s390/tcg-target.c.inc
247
@@ -XXX,XX +XXX,XX @@ static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
248
249
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
250
case INDEX_op_mov_i64:
251
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
252
- case INDEX_op_movi_i64:
253
case INDEX_op_call: /* Always emitted via tcg_out_call. */
254
default:
255
tcg_abort();
256
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
257
index XXXXXXX..XXXXXXX 100644
258
--- a/tcg/sparc/tcg-target.c.inc
259
+++ b/tcg/sparc/tcg-target.c.inc
260
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
261
262
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
263
case INDEX_op_mov_i64:
264
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
265
- case INDEX_op_movi_i64:
266
case INDEX_op_call: /* Always emitted via tcg_out_call. */
267
default:
268
tcg_abort();
269
diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc
270
index XXXXXXX..XXXXXXX 100644
271
--- a/tcg/tci/tcg-target.c.inc
272
+++ b/tcg/tci/tcg-target.c.inc
273
@@ -XXX,XX +XXX,XX @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
274
break;
275
case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
276
case INDEX_op_mov_i64:
277
- case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
278
- case INDEX_op_movi_i64:
279
case INDEX_op_call: /* Always emitted via tcg_out_call. */
280
default:
281
tcg_abort();
282
--
167
--
283
2.25.1
168
2.34.1
284
285
diff view generated by jsdifflib
1
The base qemu_ram_addr_from_host function is already in
2
softmmu/physmem.c; move the nofail version to be adjacent.
3
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
1
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
---
8
---
3
include/tcg/tcg-op.h | 13 +--
9
include/exec/cpu-common.h | 1 +
4
tcg/tcg-op.c | 227 ++++++++++++++++++++-----------------------
10
accel/tcg/cputlb.c | 12 ------------
5
2 files changed, 109 insertions(+), 131 deletions(-)
11
softmmu/physmem.c | 12 ++++++++++++
12
3 files changed, 13 insertions(+), 12 deletions(-)
6
13
7
diff --git a/include/tcg/tcg-op.h b/include/tcg/tcg-op.h
14
diff --git a/include/exec/cpu-common.h b/include/exec/cpu-common.h
8
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
9
--- a/include/tcg/tcg-op.h
16
--- a/include/exec/cpu-common.h
10
+++ b/include/tcg/tcg-op.h
17
+++ b/include/exec/cpu-common.h
11
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mb(TCGBar);
18
@@ -XXX,XX +XXX,XX @@ typedef uintptr_t ram_addr_t;
12
19
void qemu_ram_remap(ram_addr_t addr, ram_addr_t length);
13
/* 32 bit ops */
20
/* This should not be used by devices. */
14
21
ram_addr_t qemu_ram_addr_from_host(void *ptr);
15
+void tcg_gen_movi_i32(TCGv_i32 ret, int32_t arg);
22
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr);
16
void tcg_gen_addi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2);
23
RAMBlock *qemu_ram_block_by_name(const char *name);
17
void tcg_gen_subfi_i32(TCGv_i32 ret, int32_t arg1, TCGv_i32 arg2);
24
RAMBlock *qemu_ram_block_from_host(void *ptr, bool round_offset,
18
void tcg_gen_subi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2);
25
ram_addr_t *offset);
19
@@ -XXX,XX +XXX,XX @@ static inline void tcg_gen_mov_i32(TCGv_i32 ret, TCGv_i32 arg)
26
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
20
}
27
index XXXXXXX..XXXXXXX 100644
28
--- a/accel/tcg/cputlb.c
29
+++ b/accel/tcg/cputlb.c
30
@@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr,
31
prot, mmu_idx, size);
21
}
32
}
22
33
23
-static inline void tcg_gen_movi_i32(TCGv_i32 ret, int32_t arg)
34
-static inline ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
24
-{
35
-{
25
- tcg_gen_op2i_i32(INDEX_op_movi_i32, ret, arg);
36
- ram_addr_t ram_addr;
37
-
38
- ram_addr = qemu_ram_addr_from_host(ptr);
39
- if (ram_addr == RAM_ADDR_INVALID) {
40
- error_report("Bad ram pointer %p", ptr);
41
- abort();
42
- }
43
- return ram_addr;
26
-}
44
-}
27
-
45
-
28
static inline void tcg_gen_ld8u_i32(TCGv_i32 ret, TCGv_ptr arg2,
46
/*
29
tcg_target_long offset)
47
* Note: tlb_fill() can trigger a resize of the TLB. This means that all of the
30
{
48
* caller's prior references to the TLB table (e.g. CPUTLBEntry pointers) must
31
@@ -XXX,XX +XXX,XX @@ static inline void tcg_gen_not_i32(TCGv_i32 ret, TCGv_i32 arg)
49
diff --git a/softmmu/physmem.c b/softmmu/physmem.c
32
50
index XXXXXXX..XXXXXXX 100644
33
/* 64 bit ops */
51
--- a/softmmu/physmem.c
34
52
+++ b/softmmu/physmem.c
35
+void tcg_gen_movi_i64(TCGv_i64 ret, int64_t arg);
53
@@ -XXX,XX +XXX,XX @@ ram_addr_t qemu_ram_addr_from_host(void *ptr)
36
void tcg_gen_addi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2);
54
return block->offset + offset;
37
void tcg_gen_subfi_i64(TCGv_i64 ret, int64_t arg1, TCGv_i64 arg2);
38
void tcg_gen_subi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2);
39
@@ -XXX,XX +XXX,XX @@ static inline void tcg_gen_mov_i64(TCGv_i64 ret, TCGv_i64 arg)
40
}
41
}
55
}
42
56
43
-static inline void tcg_gen_movi_i64(TCGv_i64 ret, int64_t arg)
57
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
44
-{
45
- tcg_gen_op2i_i64(INDEX_op_movi_i64, ret, arg);
46
-}
47
-
48
static inline void tcg_gen_ld8u_i64(TCGv_i64 ret, TCGv_ptr arg2,
49
tcg_target_long offset)
50
{
51
@@ -XXX,XX +XXX,XX @@ static inline void tcg_gen_sub_i64(TCGv_i64 ret, TCGv_i64 arg1, TCGv_i64 arg2)
52
53
void tcg_gen_discard_i64(TCGv_i64 arg);
54
void tcg_gen_mov_i64(TCGv_i64 ret, TCGv_i64 arg);
55
-void tcg_gen_movi_i64(TCGv_i64 ret, int64_t arg);
56
void tcg_gen_ld8u_i64(TCGv_i64 ret, TCGv_ptr arg2, tcg_target_long offset);
57
void tcg_gen_ld8s_i64(TCGv_i64 ret, TCGv_ptr arg2, tcg_target_long offset);
58
void tcg_gen_ld16u_i64(TCGv_i64 ret, TCGv_ptr arg2, tcg_target_long offset);
59
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/tcg/tcg-op.c
62
+++ b/tcg/tcg-op.c
63
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mb(TCGBar mb_type)
64
65
/* 32 bit ops */
66
67
+void tcg_gen_movi_i32(TCGv_i32 ret, int32_t arg)
68
+{
58
+{
69
+ tcg_gen_mov_i32(ret, tcg_constant_i32(arg));
59
+ ram_addr_t ram_addr;
60
+
61
+ ram_addr = qemu_ram_addr_from_host(ptr);
62
+ if (ram_addr == RAM_ADDR_INVALID) {
63
+ error_report("Bad ram pointer %p", ptr);
64
+ abort();
65
+ }
66
+ return ram_addr;
70
+}
67
+}
71
+
68
+
72
void tcg_gen_addi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
69
static MemTxResult flatview_read(FlatView *fv, hwaddr addr,
73
{
70
MemTxAttrs attrs, void *buf, hwaddr len);
74
/* some cases can be optimized here */
71
static MemTxResult flatview_write(FlatView *fv, hwaddr addr, MemTxAttrs attrs,
75
if (arg2 == 0) {
76
tcg_gen_mov_i32(ret, arg1);
77
} else {
78
- TCGv_i32 t0 = tcg_const_i32(arg2);
79
- tcg_gen_add_i32(ret, arg1, t0);
80
- tcg_temp_free_i32(t0);
81
+ tcg_gen_add_i32(ret, arg1, tcg_constant_i32(arg2));
82
}
83
}
84
85
@@ -XXX,XX +XXX,XX @@ void tcg_gen_subfi_i32(TCGv_i32 ret, int32_t arg1, TCGv_i32 arg2)
86
/* Don't recurse with tcg_gen_neg_i32. */
87
tcg_gen_op2_i32(INDEX_op_neg_i32, ret, arg2);
88
} else {
89
- TCGv_i32 t0 = tcg_const_i32(arg1);
90
- tcg_gen_sub_i32(ret, t0, arg2);
91
- tcg_temp_free_i32(t0);
92
+ tcg_gen_sub_i32(ret, tcg_constant_i32(arg1), arg2);
93
}
94
}
95
96
@@ -XXX,XX +XXX,XX @@ void tcg_gen_subi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
97
if (arg2 == 0) {
98
tcg_gen_mov_i32(ret, arg1);
99
} else {
100
- TCGv_i32 t0 = tcg_const_i32(arg2);
101
- tcg_gen_sub_i32(ret, arg1, t0);
102
- tcg_temp_free_i32(t0);
103
+ tcg_gen_sub_i32(ret, arg1, tcg_constant_i32(arg2));
104
}
105
}
106
107
void tcg_gen_andi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
108
{
109
- TCGv_i32 t0;
110
/* Some cases can be optimized here. */
111
switch (arg2) {
112
case 0:
113
@@ -XXX,XX +XXX,XX @@ void tcg_gen_andi_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
114
}
115
break;
116
}
117
- t0 = tcg_const_i32(arg2);
118
- tcg_gen_and_i32(ret, arg1, t0);
119
- tcg_temp_free_i32(t0);
120
+
121
+ tcg_gen_and_i32(ret, arg1, tcg_constant_i32(arg2));
122
}
123
124
void tcg_gen_ori_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
125
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ori_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
126
} else if (arg2 == 0) {
127
tcg_gen_mov_i32(ret, arg1);
128
} else {
129
- TCGv_i32 t0 = tcg_const_i32(arg2);
130
- tcg_gen_or_i32(ret, arg1, t0);
131
- tcg_temp_free_i32(t0);
132
+ tcg_gen_or_i32(ret, arg1, tcg_constant_i32(arg2));
133
}
134
}
135
136
@@ -XXX,XX +XXX,XX @@ void tcg_gen_xori_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
137
/* Don't recurse with tcg_gen_not_i32. */
138
tcg_gen_op2_i32(INDEX_op_not_i32, ret, arg1);
139
} else {
140
- TCGv_i32 t0 = tcg_const_i32(arg2);
141
- tcg_gen_xor_i32(ret, arg1, t0);
142
- tcg_temp_free_i32(t0);
143
+ tcg_gen_xor_i32(ret, arg1, tcg_constant_i32(arg2));
144
}
145
}
146
147
@@ -XXX,XX +XXX,XX @@ void tcg_gen_shli_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
148
if (arg2 == 0) {
149
tcg_gen_mov_i32(ret, arg1);
150
} else {
151
- TCGv_i32 t0 = tcg_const_i32(arg2);
152
- tcg_gen_shl_i32(ret, arg1, t0);
153
- tcg_temp_free_i32(t0);
154
+ tcg_gen_shl_i32(ret, arg1, tcg_constant_i32(arg2));
155
}
156
}
157
158
@@ -XXX,XX +XXX,XX @@ void tcg_gen_shri_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
159
if (arg2 == 0) {
160
tcg_gen_mov_i32(ret, arg1);
161
} else {
162
- TCGv_i32 t0 = tcg_const_i32(arg2);
163
- tcg_gen_shr_i32(ret, arg1, t0);
164
- tcg_temp_free_i32(t0);
165
+ tcg_gen_shr_i32(ret, arg1, tcg_constant_i32(arg2));
166
}
167
}
168
169
@@ -XXX,XX +XXX,XX @@ void tcg_gen_sari_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
170
if (arg2 == 0) {
171
tcg_gen_mov_i32(ret, arg1);
172
} else {
173
- TCGv_i32 t0 = tcg_const_i32(arg2);
174
- tcg_gen_sar_i32(ret, arg1, t0);
175
- tcg_temp_free_i32(t0);
176
+ tcg_gen_sar_i32(ret, arg1, tcg_constant_i32(arg2));
177
}
178
}
179
180
@@ -XXX,XX +XXX,XX @@ void tcg_gen_brcondi_i32(TCGCond cond, TCGv_i32 arg1, int32_t arg2, TCGLabel *l)
181
if (cond == TCG_COND_ALWAYS) {
182
tcg_gen_br(l);
183
} else if (cond != TCG_COND_NEVER) {
184
- TCGv_i32 t0 = tcg_const_i32(arg2);
185
- tcg_gen_brcond_i32(cond, arg1, t0, l);
186
- tcg_temp_free_i32(t0);
187
+ tcg_gen_brcond_i32(cond, arg1, tcg_constant_i32(arg2), l);
188
}
189
}
190
191
@@ -XXX,XX +XXX,XX @@ void tcg_gen_setcond_i32(TCGCond cond, TCGv_i32 ret,
192
void tcg_gen_setcondi_i32(TCGCond cond, TCGv_i32 ret,
193
TCGv_i32 arg1, int32_t arg2)
194
{
195
- TCGv_i32 t0 = tcg_const_i32(arg2);
196
- tcg_gen_setcond_i32(cond, ret, arg1, t0);
197
- tcg_temp_free_i32(t0);
198
+ tcg_gen_setcond_i32(cond, ret, arg1, tcg_constant_i32(arg2));
199
}
200
201
void tcg_gen_muli_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
202
@@ -XXX,XX +XXX,XX @@ void tcg_gen_muli_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
203
} else if (is_power_of_2(arg2)) {
204
tcg_gen_shli_i32(ret, arg1, ctz32(arg2));
205
} else {
206
- TCGv_i32 t0 = tcg_const_i32(arg2);
207
- tcg_gen_mul_i32(ret, arg1, t0);
208
- tcg_temp_free_i32(t0);
209
+ tcg_gen_mul_i32(ret, arg1, tcg_constant_i32(arg2));
210
}
211
}
212
213
@@ -XXX,XX +XXX,XX @@ void tcg_gen_clz_i32(TCGv_i32 ret, TCGv_i32 arg1, TCGv_i32 arg2)
214
215
void tcg_gen_clzi_i32(TCGv_i32 ret, TCGv_i32 arg1, uint32_t arg2)
216
{
217
- TCGv_i32 t = tcg_const_i32(arg2);
218
- tcg_gen_clz_i32(ret, arg1, t);
219
- tcg_temp_free_i32(t);
220
+ tcg_gen_clz_i32(ret, arg1, tcg_constant_i32(arg2));
221
}
222
223
void tcg_gen_ctz_i32(TCGv_i32 ret, TCGv_i32 arg1, TCGv_i32 arg2)
224
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ctz_i32(TCGv_i32 ret, TCGv_i32 arg1, TCGv_i32 arg2)
225
tcg_gen_clzi_i32(t, t, 32);
226
tcg_gen_xori_i32(t, t, 31);
227
}
228
- z = tcg_const_i32(0);
229
+ z = tcg_constant_i32(0);
230
tcg_gen_movcond_i32(TCG_COND_EQ, ret, arg1, z, arg2, t);
231
tcg_temp_free_i32(t);
232
- tcg_temp_free_i32(z);
233
} else {
234
gen_helper_ctz_i32(ret, arg1, arg2);
235
}
236
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ctzi_i32(TCGv_i32 ret, TCGv_i32 arg1, uint32_t arg2)
237
tcg_gen_ctpop_i32(ret, t);
238
tcg_temp_free_i32(t);
239
} else {
240
- TCGv_i32 t = tcg_const_i32(arg2);
241
- tcg_gen_ctz_i32(ret, arg1, t);
242
- tcg_temp_free_i32(t);
243
+ tcg_gen_ctz_i32(ret, arg1, tcg_constant_i32(arg2));
244
}
245
}
246
247
@@ -XXX,XX +XXX,XX @@ void tcg_gen_rotli_i32(TCGv_i32 ret, TCGv_i32 arg1, int32_t arg2)
248
if (arg2 == 0) {
249
tcg_gen_mov_i32(ret, arg1);
250
} else if (TCG_TARGET_HAS_rot_i32) {
251
- TCGv_i32 t0 = tcg_const_i32(arg2);
252
- tcg_gen_rotl_i32(ret, arg1, t0);
253
- tcg_temp_free_i32(t0);
254
+ tcg_gen_rotl_i32(ret, arg1, tcg_constant_i32(arg2));
255
} else {
256
TCGv_i32 t0, t1;
257
t0 = tcg_temp_new_i32();
258
@@ -XXX,XX +XXX,XX @@ void tcg_gen_deposit_z_i32(TCGv_i32 ret, TCGv_i32 arg,
259
tcg_gen_andi_i32(ret, arg, (1u << len) - 1);
260
} else if (TCG_TARGET_HAS_deposit_i32
261
&& TCG_TARGET_deposit_i32_valid(ofs, len)) {
262
- TCGv_i32 zero = tcg_const_i32(0);
263
+ TCGv_i32 zero = tcg_constant_i32(0);
264
tcg_gen_op5ii_i32(INDEX_op_deposit_i32, ret, zero, arg, ofs, len);
265
- tcg_temp_free_i32(zero);
266
} else {
267
/* To help two-operand hosts we prefer to zero-extend first,
268
which allows ARG to stay live. */
269
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap32_i32(TCGv_i32 ret, TCGv_i32 arg)
270
} else {
271
TCGv_i32 t0 = tcg_temp_new_i32();
272
TCGv_i32 t1 = tcg_temp_new_i32();
273
- TCGv_i32 t2 = tcg_const_i32(0x00ff00ff);
274
+ TCGv_i32 t2 = tcg_constant_i32(0x00ff00ff);
275
276
/* arg = abcd */
277
tcg_gen_shri_i32(t0, arg, 8); /* t0 = .abc */
278
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap32_i32(TCGv_i32 ret, TCGv_i32 arg)
279
280
tcg_temp_free_i32(t0);
281
tcg_temp_free_i32(t1);
282
- tcg_temp_free_i32(t2);
283
}
284
}
285
286
@@ -XXX,XX +XXX,XX @@ void tcg_gen_discard_i64(TCGv_i64 arg)
287
288
void tcg_gen_mov_i64(TCGv_i64 ret, TCGv_i64 arg)
289
{
290
- tcg_gen_mov_i32(TCGV_LOW(ret), TCGV_LOW(arg));
291
- tcg_gen_mov_i32(TCGV_HIGH(ret), TCGV_HIGH(arg));
292
+ TCGTemp *ts = tcgv_i64_temp(arg);
293
+
294
+ /* Canonicalize TCGv_i64 TEMP_CONST into TCGv_i32 TEMP_CONST. */
295
+ if (ts->kind == TEMP_CONST) {
296
+ tcg_gen_movi_i64(ret, ts->val);
297
+ } else {
298
+ tcg_gen_mov_i32(TCGV_LOW(ret), TCGV_LOW(arg));
299
+ tcg_gen_mov_i32(TCGV_HIGH(ret), TCGV_HIGH(arg));
300
+ }
301
}
302
303
void tcg_gen_movi_i64(TCGv_i64 ret, int64_t arg)
304
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mul_i64(TCGv_i64 ret, TCGv_i64 arg1, TCGv_i64 arg2)
305
tcg_temp_free_i64(t0);
306
tcg_temp_free_i32(t1);
307
}
308
+
309
+#else
310
+
311
+void tcg_gen_movi_i64(TCGv_i64 ret, int64_t arg)
312
+{
313
+ tcg_gen_mov_i64(ret, tcg_constant_i64(arg));
314
+}
315
+
316
#endif /* TCG_TARGET_REG_SIZE == 32 */
317
318
void tcg_gen_addi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
319
@@ -XXX,XX +XXX,XX @@ void tcg_gen_addi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
320
/* some cases can be optimized here */
321
if (arg2 == 0) {
322
tcg_gen_mov_i64(ret, arg1);
323
+ } else if (TCG_TARGET_REG_BITS == 64) {
324
+ tcg_gen_add_i64(ret, arg1, tcg_constant_i64(arg2));
325
} else {
326
- TCGv_i64 t0 = tcg_const_i64(arg2);
327
- tcg_gen_add_i64(ret, arg1, t0);
328
- tcg_temp_free_i64(t0);
329
+ tcg_gen_add2_i32(TCGV_LOW(ret), TCGV_HIGH(ret),
330
+ TCGV_LOW(arg1), TCGV_HIGH(arg1),
331
+ tcg_constant_i32(arg2), tcg_constant_i32(arg2 >> 32));
332
}
333
}
334
335
@@ -XXX,XX +XXX,XX @@ void tcg_gen_subfi_i64(TCGv_i64 ret, int64_t arg1, TCGv_i64 arg2)
336
if (arg1 == 0 && TCG_TARGET_HAS_neg_i64) {
337
/* Don't recurse with tcg_gen_neg_i64. */
338
tcg_gen_op2_i64(INDEX_op_neg_i64, ret, arg2);
339
+ } else if (TCG_TARGET_REG_BITS == 64) {
340
+ tcg_gen_sub_i64(ret, tcg_constant_i64(arg1), arg2);
341
} else {
342
- TCGv_i64 t0 = tcg_const_i64(arg1);
343
- tcg_gen_sub_i64(ret, t0, arg2);
344
- tcg_temp_free_i64(t0);
345
+ tcg_gen_sub2_i32(TCGV_LOW(ret), TCGV_HIGH(ret),
346
+ tcg_constant_i32(arg1), tcg_constant_i32(arg1 >> 32),
347
+ TCGV_LOW(arg2), TCGV_HIGH(arg2));
348
}
349
}
350
351
@@ -XXX,XX +XXX,XX @@ void tcg_gen_subi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
352
/* some cases can be optimized here */
353
if (arg2 == 0) {
354
tcg_gen_mov_i64(ret, arg1);
355
+ } else if (TCG_TARGET_REG_BITS == 64) {
356
+ tcg_gen_sub_i64(ret, arg1, tcg_constant_i64(arg2));
357
} else {
358
- TCGv_i64 t0 = tcg_const_i64(arg2);
359
- tcg_gen_sub_i64(ret, arg1, t0);
360
- tcg_temp_free_i64(t0);
361
+ tcg_gen_sub2_i32(TCGV_LOW(ret), TCGV_HIGH(ret),
362
+ TCGV_LOW(arg1), TCGV_HIGH(arg1),
363
+ tcg_constant_i32(arg2), tcg_constant_i32(arg2 >> 32));
364
}
365
}
366
367
void tcg_gen_andi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
368
{
369
- TCGv_i64 t0;
370
-
371
if (TCG_TARGET_REG_BITS == 32) {
372
tcg_gen_andi_i32(TCGV_LOW(ret), TCGV_LOW(arg1), arg2);
373
tcg_gen_andi_i32(TCGV_HIGH(ret), TCGV_HIGH(arg1), arg2 >> 32);
374
@@ -XXX,XX +XXX,XX @@ void tcg_gen_andi_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
375
}
376
break;
377
}
378
- t0 = tcg_const_i64(arg2);
379
- tcg_gen_and_i64(ret, arg1, t0);
380
- tcg_temp_free_i64(t0);
381
+
382
+ tcg_gen_and_i64(ret, arg1, tcg_constant_i64(arg2));
383
}
384
385
void tcg_gen_ori_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
386
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ori_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
387
} else if (arg2 == 0) {
388
tcg_gen_mov_i64(ret, arg1);
389
} else {
390
- TCGv_i64 t0 = tcg_const_i64(arg2);
391
- tcg_gen_or_i64(ret, arg1, t0);
392
- tcg_temp_free_i64(t0);
393
+ tcg_gen_or_i64(ret, arg1, tcg_constant_i64(arg2));
394
}
395
}
396
397
@@ -XXX,XX +XXX,XX @@ void tcg_gen_xori_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
398
/* Don't recurse with tcg_gen_not_i64. */
399
tcg_gen_op2_i64(INDEX_op_not_i64, ret, arg1);
400
} else {
401
- TCGv_i64 t0 = tcg_const_i64(arg2);
402
- tcg_gen_xor_i64(ret, arg1, t0);
403
- tcg_temp_free_i64(t0);
404
+ tcg_gen_xor_i64(ret, arg1, tcg_constant_i64(arg2));
405
}
406
}
407
408
@@ -XXX,XX +XXX,XX @@ void tcg_gen_shli_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
409
} else if (arg2 == 0) {
410
tcg_gen_mov_i64(ret, arg1);
411
} else {
412
- TCGv_i64 t0 = tcg_const_i64(arg2);
413
- tcg_gen_shl_i64(ret, arg1, t0);
414
- tcg_temp_free_i64(t0);
415
+ tcg_gen_shl_i64(ret, arg1, tcg_constant_i64(arg2));
416
}
417
}
418
419
@@ -XXX,XX +XXX,XX @@ void tcg_gen_shri_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
420
} else if (arg2 == 0) {
421
tcg_gen_mov_i64(ret, arg1);
422
} else {
423
- TCGv_i64 t0 = tcg_const_i64(arg2);
424
- tcg_gen_shr_i64(ret, arg1, t0);
425
- tcg_temp_free_i64(t0);
426
+ tcg_gen_shr_i64(ret, arg1, tcg_constant_i64(arg2));
427
}
428
}
429
430
@@ -XXX,XX +XXX,XX @@ void tcg_gen_sari_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
431
} else if (arg2 == 0) {
432
tcg_gen_mov_i64(ret, arg1);
433
} else {
434
- TCGv_i64 t0 = tcg_const_i64(arg2);
435
- tcg_gen_sar_i64(ret, arg1, t0);
436
- tcg_temp_free_i64(t0);
437
+ tcg_gen_sar_i64(ret, arg1, tcg_constant_i64(arg2));
438
}
439
}
440
441
@@ -XXX,XX +XXX,XX @@ void tcg_gen_brcond_i64(TCGCond cond, TCGv_i64 arg1, TCGv_i64 arg2, TCGLabel *l)
442
443
void tcg_gen_brcondi_i64(TCGCond cond, TCGv_i64 arg1, int64_t arg2, TCGLabel *l)
444
{
445
- if (cond == TCG_COND_ALWAYS) {
446
+ if (TCG_TARGET_REG_BITS == 64) {
447
+ tcg_gen_brcond_i64(cond, arg1, tcg_constant_i64(arg2), l);
448
+ } else if (cond == TCG_COND_ALWAYS) {
449
tcg_gen_br(l);
450
} else if (cond != TCG_COND_NEVER) {
451
- TCGv_i64 t0 = tcg_const_i64(arg2);
452
- tcg_gen_brcond_i64(cond, arg1, t0, l);
453
- tcg_temp_free_i64(t0);
454
+ l->refs++;
455
+ tcg_gen_op6ii_i32(INDEX_op_brcond2_i32,
456
+ TCGV_LOW(arg1), TCGV_HIGH(arg1),
457
+ tcg_constant_i32(arg2),
458
+ tcg_constant_i32(arg2 >> 32),
459
+ cond, label_arg(l));
460
}
461
}
462
463
@@ -XXX,XX +XXX,XX @@ void tcg_gen_setcond_i64(TCGCond cond, TCGv_i64 ret,
464
void tcg_gen_setcondi_i64(TCGCond cond, TCGv_i64 ret,
465
TCGv_i64 arg1, int64_t arg2)
466
{
467
- TCGv_i64 t0 = tcg_const_i64(arg2);
468
- tcg_gen_setcond_i64(cond, ret, arg1, t0);
469
- tcg_temp_free_i64(t0);
470
+ if (TCG_TARGET_REG_BITS == 64) {
471
+ tcg_gen_setcond_i64(cond, ret, arg1, tcg_constant_i64(arg2));
472
+ } else if (cond == TCG_COND_ALWAYS) {
473
+ tcg_gen_movi_i64(ret, 1);
474
+ } else if (cond == TCG_COND_NEVER) {
475
+ tcg_gen_movi_i64(ret, 0);
476
+ } else {
477
+ tcg_gen_op6i_i32(INDEX_op_setcond2_i32, TCGV_LOW(ret),
478
+ TCGV_LOW(arg1), TCGV_HIGH(arg1),
479
+ tcg_constant_i32(arg2),
480
+ tcg_constant_i32(arg2 >> 32), cond);
481
+ tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
482
+ }
483
}
484
485
void tcg_gen_muli_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
486
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg)
487
} else {
488
TCGv_i64 t0 = tcg_temp_new_i64();
489
TCGv_i64 t1 = tcg_temp_new_i64();
490
- TCGv_i64 t2 = tcg_const_i64(0x00ff00ff);
491
+ TCGv_i64 t2 = tcg_constant_i64(0x00ff00ff);
492
493
/* arg = ....abcd */
494
tcg_gen_shri_i64(t0, arg, 8); /* t0 = .....abc */
495
@@ -XXX,XX +XXX,XX @@ void tcg_gen_bswap32_i64(TCGv_i64 ret, TCGv_i64 arg)
496
497
tcg_temp_free_i64(t0);
498
tcg_temp_free_i64(t1);
499
- tcg_temp_free_i64(t2);
500
}
501
}
502
503
@@ -XXX,XX +XXX,XX @@ void tcg_gen_clzi_i64(TCGv_i64 ret, TCGv_i64 arg1, uint64_t arg2)
504
if (TCG_TARGET_REG_BITS == 32
505
&& TCG_TARGET_HAS_clz_i32
506
&& arg2 <= 0xffffffffu) {
507
- TCGv_i32 t = tcg_const_i32((uint32_t)arg2 - 32);
508
- tcg_gen_clz_i32(t, TCGV_LOW(arg1), t);
509
+ TCGv_i32 t = tcg_temp_new_i32();
510
+ tcg_gen_clzi_i32(t, TCGV_LOW(arg1), arg2 - 32);
511
tcg_gen_addi_i32(t, t, 32);
512
tcg_gen_clz_i32(TCGV_LOW(ret), TCGV_HIGH(arg1), t);
513
tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
514
tcg_temp_free_i32(t);
515
} else {
516
- TCGv_i64 t = tcg_const_i64(arg2);
517
- tcg_gen_clz_i64(ret, arg1, t);
518
- tcg_temp_free_i64(t);
519
+ TCGv_i64 t0 = tcg_const_i64(arg2);
520
+ tcg_gen_clz_i64(ret, arg1, t0);
521
+ tcg_temp_free_i64(t0);
522
}
523
}
524
525
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ctz_i64(TCGv_i64 ret, TCGv_i64 arg1, TCGv_i64 arg2)
526
tcg_gen_clzi_i64(t, t, 64);
527
tcg_gen_xori_i64(t, t, 63);
528
}
529
- z = tcg_const_i64(0);
530
+ z = tcg_constant_i64(0);
531
tcg_gen_movcond_i64(TCG_COND_EQ, ret, arg1, z, arg2, t);
532
tcg_temp_free_i64(t);
533
tcg_temp_free_i64(z);
534
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ctzi_i64(TCGv_i64 ret, TCGv_i64 arg1, uint64_t arg2)
535
if (TCG_TARGET_REG_BITS == 32
536
&& TCG_TARGET_HAS_ctz_i32
537
&& arg2 <= 0xffffffffu) {
538
- TCGv_i32 t32 = tcg_const_i32((uint32_t)arg2 - 32);
539
- tcg_gen_ctz_i32(t32, TCGV_HIGH(arg1), t32);
540
+ TCGv_i32 t32 = tcg_temp_new_i32();
541
+ tcg_gen_ctzi_i32(t32, TCGV_HIGH(arg1), arg2 - 32);
542
tcg_gen_addi_i32(t32, t32, 32);
543
tcg_gen_ctz_i32(TCGV_LOW(ret), TCGV_LOW(arg1), t32);
544
tcg_gen_movi_i32(TCGV_HIGH(ret), 0);
545
@@ -XXX,XX +XXX,XX @@ void tcg_gen_ctzi_i64(TCGv_i64 ret, TCGv_i64 arg1, uint64_t arg2)
546
tcg_gen_ctpop_i64(ret, t);
547
tcg_temp_free_i64(t);
548
} else {
549
- TCGv_i64 t64 = tcg_const_i64(arg2);
550
- tcg_gen_ctz_i64(ret, arg1, t64);
551
- tcg_temp_free_i64(t64);
552
+ TCGv_i64 t0 = tcg_const_i64(arg2);
553
+ tcg_gen_ctz_i64(ret, arg1, t0);
554
+ tcg_temp_free_i64(t0);
555
}
556
}
557
558
@@ -XXX,XX +XXX,XX @@ void tcg_gen_rotli_i64(TCGv_i64 ret, TCGv_i64 arg1, int64_t arg2)
559
if (arg2 == 0) {
560
tcg_gen_mov_i64(ret, arg1);
561
} else if (TCG_TARGET_HAS_rot_i64) {
562
- TCGv_i64 t0 = tcg_const_i64(arg2);
563
- tcg_gen_rotl_i64(ret, arg1, t0);
564
- tcg_temp_free_i64(t0);
565
+ tcg_gen_rotl_i64(ret, arg1, tcg_constant_i64(arg2));
566
} else {
567
TCGv_i64 t0, t1;
568
t0 = tcg_temp_new_i64();
569
@@ -XXX,XX +XXX,XX @@ void tcg_gen_deposit_z_i64(TCGv_i64 ret, TCGv_i64 arg,
570
tcg_gen_andi_i64(ret, arg, (1ull << len) - 1);
571
} else if (TCG_TARGET_HAS_deposit_i64
572
&& TCG_TARGET_deposit_i64_valid(ofs, len)) {
573
- TCGv_i64 zero = tcg_const_i64(0);
574
+ TCGv_i64 zero = tcg_constant_i64(0);
575
tcg_gen_op5ii_i64(INDEX_op_deposit_i64, ret, zero, arg, ofs, len);
576
- tcg_temp_free_i64(zero);
577
} else {
578
if (TCG_TARGET_REG_BITS == 32) {
579
if (ofs >= 32) {
580
@@ -XXX,XX +XXX,XX @@ void tcg_gen_atomic_cmpxchg_i32(TCGv_i32 retv, TCGv addr, TCGv_i32 cmpv,
581
582
#ifdef CONFIG_SOFTMMU
583
{
584
- TCGv_i32 oi = tcg_const_i32(make_memop_idx(memop & ~MO_SIGN, idx));
585
- gen(retv, cpu_env, addr, cmpv, newv, oi);
586
- tcg_temp_free_i32(oi);
587
+ TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
588
+ gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
589
}
590
#else
591
gen(retv, cpu_env, addr, cmpv, newv);
592
@@ -XXX,XX +XXX,XX @@ void tcg_gen_atomic_cmpxchg_i64(TCGv_i64 retv, TCGv addr, TCGv_i64 cmpv,
593
594
#ifdef CONFIG_SOFTMMU
595
{
596
- TCGv_i32 oi = tcg_const_i32(make_memop_idx(memop, idx));
597
- gen(retv, cpu_env, addr, cmpv, newv, oi);
598
- tcg_temp_free_i32(oi);
599
+ TCGMemOpIdx oi = make_memop_idx(memop, idx);
600
+ gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
601
}
602
#else
603
gen(retv, cpu_env, addr, cmpv, newv);
604
@@ -XXX,XX +XXX,XX @@ static void do_atomic_op_i32(TCGv_i32 ret, TCGv addr, TCGv_i32 val,
605
606
#ifdef CONFIG_SOFTMMU
607
{
608
- TCGv_i32 oi = tcg_const_i32(make_memop_idx(memop & ~MO_SIGN, idx));
609
- gen(ret, cpu_env, addr, val, oi);
610
- tcg_temp_free_i32(oi);
611
+ TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
612
+ gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
613
}
614
#else
615
gen(ret, cpu_env, addr, val);
616
@@ -XXX,XX +XXX,XX @@ static void do_atomic_op_i64(TCGv_i64 ret, TCGv addr, TCGv_i64 val,
617
618
#ifdef CONFIG_SOFTMMU
619
{
620
- TCGv_i32 oi = tcg_const_i32(make_memop_idx(memop & ~MO_SIGN, idx));
621
- gen(ret, cpu_env, addr, val, oi);
622
- tcg_temp_free_i32(oi);
623
+ TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
624
+ gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
625
}
626
#else
627
gen(ret, cpu_env, addr, val);
628
--
72
--
629
2.25.1
73
2.34.1
630
631
diff view generated by jsdifflib
1
Simplify the implementation of get_page_addr_code_hostp
2
by reusing the existing probe_access infrastructure.
3
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
1
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
---
7
---
3
tcg/optimize.c | 108 ++++++++++++++++++++++---------------------------
8
accel/tcg/cputlb.c | 76 ++++++++++++++++------------------------------
4
1 file changed, 49 insertions(+), 59 deletions(-)
9
1 file changed, 26 insertions(+), 50 deletions(-)
5
10
6
diff --git a/tcg/optimize.c b/tcg/optimize.c
11
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
7
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
8
--- a/tcg/optimize.c
13
--- a/accel/tcg/cputlb.c
9
+++ b/tcg/optimize.c
14
+++ b/accel/tcg/cputlb.c
10
@@ -XXX,XX +XXX,XX @@ static bool args_are_copies(TCGArg arg1, TCGArg arg2)
15
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
11
return ts_are_copies(arg_temp(arg1), arg_temp(arg2));
16
victim_tlb_hit(env, mmu_idx, index, offsetof(CPUTLBEntry, TY), \
12
}
17
(ADDR) & TARGET_PAGE_MASK)
13
18
14
-static void tcg_opt_gen_movi(TCGContext *s, TCGOp *op, TCGArg dst, uint64_t val)
19
-/*
20
- * Return a ram_addr_t for the virtual address for execution.
21
- *
22
- * Return -1 if we can't translate and execute from an entire page
23
- * of RAM. This will force us to execute by loading and translating
24
- * one insn at a time, without caching.
25
- *
26
- * NOTE: This function will trigger an exception if the page is
27
- * not executable.
28
- */
29
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
30
- void **hostp)
15
-{
31
-{
16
- const TCGOpDef *def;
32
- uintptr_t mmu_idx = cpu_mmu_index(env, true);
17
- TCGOpcode new_op;
33
- uintptr_t index = tlb_index(env, mmu_idx, addr);
18
- uint64_t mask;
34
- CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr);
19
- TempOptInfo *di = arg_info(dst);
35
- void *p;
20
-
36
-
21
- def = &tcg_op_defs[op->opc];
37
- if (unlikely(!tlb_hit(entry->addr_code, addr))) {
22
- if (def->flags & TCG_OPF_VECTOR) {
38
- if (!VICTIM_TLB_HIT(addr_code, addr)) {
23
- new_op = INDEX_op_dupi_vec;
39
- tlb_fill(env_cpu(env), addr, 0, MMU_INST_FETCH, mmu_idx, 0);
24
- } else if (def->flags & TCG_OPF_64BIT) {
40
- index = tlb_index(env, mmu_idx, addr);
25
- new_op = INDEX_op_movi_i64;
41
- entry = tlb_entry(env, mmu_idx, addr);
26
- } else {
42
-
27
- new_op = INDEX_op_movi_i32;
43
- if (unlikely(entry->addr_code & TLB_INVALID_MASK)) {
44
- /*
45
- * The MMU protection covers a smaller range than a target
46
- * page, so we must redo the MMU check for every insn.
47
- */
48
- return -1;
49
- }
50
- }
51
- assert(tlb_hit(entry->addr_code, addr));
28
- }
52
- }
29
- op->opc = new_op;
30
- /* TCGOP_VECL and TCGOP_VECE remain unchanged. */
31
- op->args[0] = dst;
32
- op->args[1] = val;
33
-
53
-
34
- reset_temp(dst);
54
- if (unlikely(entry->addr_code & TLB_MMIO)) {
35
- di->is_const = true;
55
- /* The region is not backed by RAM. */
36
- di->val = val;
56
- if (hostp) {
37
- mask = val;
57
- *hostp = NULL;
38
- if (TCG_TARGET_REG_BITS > 32 && new_op == INDEX_op_movi_i32) {
58
- }
39
- /* High bits of the destination are now garbage. */
59
- return -1;
40
- mask |= ~0xffffffffull;
41
- }
60
- }
42
- di->mask = mask;
61
-
62
- p = (void *)((uintptr_t)addr + entry->addend);
63
- if (hostp) {
64
- *hostp = p;
65
- }
66
- return qemu_ram_addr_from_host_nofail(p);
43
-}
67
-}
44
-
68
-
45
static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
69
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
70
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
46
{
71
{
47
TCGTemp *dst_ts = arg_temp(dst);
72
@@ -XXX,XX +XXX,XX @@ void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
48
@@ -XXX,XX +XXX,XX @@ static void tcg_opt_gen_mov(TCGContext *s, TCGOp *op, TCGArg dst, TCGArg src)
73
return flags ? NULL : host;
49
}
50
}
74
}
51
75
52
+static void tcg_opt_gen_movi(TCGContext *s, TCGTempSet *temps_used,
76
+/*
53
+ TCGOp *op, TCGArg dst, uint64_t val)
77
+ * Return a ram_addr_t for the virtual address for execution.
78
+ *
79
+ * Return -1 if we can't translate and execute from an entire page
80
+ * of RAM. This will force us to execute by loading and translating
81
+ * one insn at a time, without caching.
82
+ *
83
+ * NOTE: This function will trigger an exception if the page is
84
+ * not executable.
85
+ */
86
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
87
+ void **hostp)
54
+{
88
+{
55
+ const TCGOpDef *def = &tcg_op_defs[op->opc];
89
+ void *p;
56
+ TCGType type;
57
+ TCGTemp *tv;
58
+
90
+
59
+ if (def->flags & TCG_OPF_VECTOR) {
91
+ (void)probe_access_internal(env, addr, 1, MMU_INST_FETCH,
60
+ type = TCGOP_VECL(op) + TCG_TYPE_V64;
92
+ cpu_mmu_index(env, true), false, &p, 0);
61
+ } else if (def->flags & TCG_OPF_64BIT) {
93
+ if (p == NULL) {
62
+ type = TCG_TYPE_I64;
94
+ return -1;
63
+ } else {
64
+ type = TCG_TYPE_I32;
65
+ }
95
+ }
66
+
96
+ if (hostp) {
67
+ /* Convert movi to mov with constant temp. */
97
+ *hostp = p;
68
+ tv = tcg_constant_internal(type, val);
98
+ }
69
+ init_ts_info(temps_used, tv);
99
+ return qemu_ram_addr_from_host_nofail(p);
70
+ tcg_opt_gen_mov(s, op, dst, temp_arg(tv));
71
+}
100
+}
72
+
101
+
73
static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y)
102
#ifdef CONFIG_PLUGIN
74
{
103
/*
75
uint64_t l64, h64;
104
* Perform a TLB lookup and populate the qemu_plugin_hwaddr structure.
76
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
77
nb_temps = s->nb_temps;
78
nb_globals = s->nb_globals;
79
80
- bitmap_zero(temps_used.l, nb_temps);
81
+ memset(&temps_used, 0, sizeof(temps_used));
82
for (i = 0; i < nb_temps; ++i) {
83
s->temps[i].state_ptr = NULL;
84
}
85
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
86
CASE_OP_32_64(rotr):
87
if (arg_is_const(op->args[1])
88
&& arg_info(op->args[1])->val == 0) {
89
- tcg_opt_gen_movi(s, op, op->args[0], 0);
90
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], 0);
91
continue;
92
}
93
break;
94
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
95
96
if (partmask == 0) {
97
tcg_debug_assert(nb_oargs == 1);
98
- tcg_opt_gen_movi(s, op, op->args[0], 0);
99
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], 0);
100
continue;
101
}
102
if (affected == 0) {
103
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
104
CASE_OP_32_64(mulsh):
105
if (arg_is_const(op->args[2])
106
&& arg_info(op->args[2])->val == 0) {
107
- tcg_opt_gen_movi(s, op, op->args[0], 0);
108
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], 0);
109
continue;
110
}
111
break;
112
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
113
CASE_OP_32_64_VEC(sub):
114
CASE_OP_32_64_VEC(xor):
115
if (args_are_copies(op->args[1], op->args[2])) {
116
- tcg_opt_gen_movi(s, op, op->args[0], 0);
117
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], 0);
118
continue;
119
}
120
break;
121
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
122
break;
123
CASE_OP_32_64(movi):
124
case INDEX_op_dupi_vec:
125
- tcg_opt_gen_movi(s, op, op->args[0], op->args[1]);
126
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], op->args[1]);
127
break;
128
129
case INDEX_op_dup_vec:
130
if (arg_is_const(op->args[1])) {
131
tmp = arg_info(op->args[1])->val;
132
tmp = dup_const(TCGOP_VECE(op), tmp);
133
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
134
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
135
break;
136
}
137
goto do_default;
138
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
139
if (arg_is_const(op->args[1]) && arg_is_const(op->args[2])) {
140
tmp = arg_info(op->args[1])->val;
141
if (tmp == arg_info(op->args[2])->val) {
142
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
143
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
144
break;
145
}
146
} else if (args_are_copies(op->args[1], op->args[2])) {
147
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
148
case INDEX_op_extrh_i64_i32:
149
if (arg_is_const(op->args[1])) {
150
tmp = do_constant_folding(opc, arg_info(op->args[1])->val, 0);
151
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
152
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
153
break;
154
}
155
goto do_default;
156
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
157
if (arg_is_const(op->args[1]) && arg_is_const(op->args[2])) {
158
tmp = do_constant_folding(opc, arg_info(op->args[1])->val,
159
arg_info(op->args[2])->val);
160
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
161
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
162
break;
163
}
164
goto do_default;
165
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
166
TCGArg v = arg_info(op->args[1])->val;
167
if (v != 0) {
168
tmp = do_constant_folding(opc, v, 0);
169
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
170
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
171
} else {
172
tcg_opt_gen_mov(s, op, op->args[0], op->args[2]);
173
}
174
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
175
tmp = deposit64(arg_info(op->args[1])->val,
176
op->args[3], op->args[4],
177
arg_info(op->args[2])->val);
178
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
179
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
180
break;
181
}
182
goto do_default;
183
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
184
if (arg_is_const(op->args[1])) {
185
tmp = extract64(arg_info(op->args[1])->val,
186
op->args[2], op->args[3]);
187
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
188
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
189
break;
190
}
191
goto do_default;
192
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
193
if (arg_is_const(op->args[1])) {
194
tmp = sextract64(arg_info(op->args[1])->val,
195
op->args[2], op->args[3]);
196
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
197
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
198
break;
199
}
200
goto do_default;
201
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
202
tmp = (int32_t)(((uint32_t)v1 >> shr) |
203
((uint32_t)v2 << (32 - shr)));
204
}
205
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
206
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
207
break;
208
}
209
goto do_default;
210
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
211
tmp = do_constant_folding_cond(opc, op->args[1],
212
op->args[2], op->args[3]);
213
if (tmp != 2) {
214
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
215
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
216
break;
217
}
218
goto do_default;
219
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
220
op->args[1], op->args[2]);
221
if (tmp != 2) {
222
if (tmp) {
223
- bitmap_zero(temps_used.l, nb_temps);
224
+ memset(&temps_used, 0, sizeof(temps_used));
225
op->opc = INDEX_op_br;
226
op->args[0] = op->args[3];
227
} else {
228
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
229
uint64_t a = ((uint64_t)ah << 32) | al;
230
uint64_t b = ((uint64_t)bh << 32) | bl;
231
TCGArg rl, rh;
232
- TCGOp *op2 = tcg_op_insert_before(s, op, INDEX_op_movi_i32);
233
+ TCGOp *op2 = tcg_op_insert_before(s, op, INDEX_op_mov_i32);
234
235
if (opc == INDEX_op_add2_i32) {
236
a += b;
237
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
238
239
rl = op->args[0];
240
rh = op->args[1];
241
- tcg_opt_gen_movi(s, op, rl, (int32_t)a);
242
- tcg_opt_gen_movi(s, op2, rh, (int32_t)(a >> 32));
243
+ tcg_opt_gen_movi(s, &temps_used, op, rl, (int32_t)a);
244
+ tcg_opt_gen_movi(s, &temps_used, op2, rh, (int32_t)(a >> 32));
245
break;
246
}
247
goto do_default;
248
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
249
uint32_t b = arg_info(op->args[3])->val;
250
uint64_t r = (uint64_t)a * b;
251
TCGArg rl, rh;
252
- TCGOp *op2 = tcg_op_insert_before(s, op, INDEX_op_movi_i32);
253
+ TCGOp *op2 = tcg_op_insert_before(s, op, INDEX_op_mov_i32);
254
255
rl = op->args[0];
256
rh = op->args[1];
257
- tcg_opt_gen_movi(s, op, rl, (int32_t)r);
258
- tcg_opt_gen_movi(s, op2, rh, (int32_t)(r >> 32));
259
+ tcg_opt_gen_movi(s, &temps_used, op, rl, (int32_t)r);
260
+ tcg_opt_gen_movi(s, &temps_used, op2, rh, (int32_t)(r >> 32));
261
break;
262
}
263
goto do_default;
264
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
265
if (tmp != 2) {
266
if (tmp) {
267
do_brcond_true:
268
- bitmap_zero(temps_used.l, nb_temps);
269
+ memset(&temps_used, 0, sizeof(temps_used));
270
op->opc = INDEX_op_br;
271
op->args[0] = op->args[5];
272
} else {
273
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
274
/* Simplify LT/GE comparisons vs zero to a single compare
275
vs the high word of the input. */
276
do_brcond_high:
277
- bitmap_zero(temps_used.l, nb_temps);
278
+ memset(&temps_used, 0, sizeof(temps_used));
279
op->opc = INDEX_op_brcond_i32;
280
op->args[0] = op->args[1];
281
op->args[1] = op->args[3];
282
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
283
goto do_default;
284
}
285
do_brcond_low:
286
- bitmap_zero(temps_used.l, nb_temps);
287
+ memset(&temps_used, 0, sizeof(temps_used));
288
op->opc = INDEX_op_brcond_i32;
289
op->args[1] = op->args[2];
290
op->args[2] = op->args[4];
291
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
292
op->args[5]);
293
if (tmp != 2) {
294
do_setcond_const:
295
- tcg_opt_gen_movi(s, op, op->args[0], tmp);
296
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
297
} else if ((op->args[5] == TCG_COND_LT
298
|| op->args[5] == TCG_COND_GE)
299
&& arg_is_const(op->args[3])
300
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
301
block, otherwise we only trash the output args. "mask" is
302
the non-zero bits mask for the first output arg. */
303
if (def->flags & TCG_OPF_BB_END) {
304
- bitmap_zero(temps_used.l, nb_temps);
305
+ memset(&temps_used, 0, sizeof(temps_used));
306
} else {
307
do_reset_output:
308
for (i = 0; i < nb_oargs; i++) {
309
--
105
--
310
2.25.1
106
2.34.1
311
312
diff view generated by jsdifflib
1
This will reduce the differences between 32-bit and 64-bit hosts,
1
It was non-obvious to me why we can raise an exception in
2
allowing full 64-bit constants to be created with the same interface.
2
the middle of a comparison function, but it works.
3
While nearby, use TARGET_PAGE_ALIGN instead of open-coding.
3
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
7
---
6
include/tcg/tcg.h | 2 +-
8
accel/tcg/cpu-exec.c | 11 ++++++++++-
7
tcg/tcg.c | 2 +-
9
1 file changed, 10 insertions(+), 1 deletion(-)
8
2 files changed, 2 insertions(+), 2 deletions(-)
9
10
10
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
11
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
11
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
12
--- a/include/tcg/tcg.h
13
--- a/accel/tcg/cpu-exec.c
13
+++ b/include/tcg/tcg.h
14
+++ b/accel/tcg/cpu-exec.c
14
@@ -XXX,XX +XXX,XX @@ typedef struct TCGTemp {
15
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
15
unsigned int mem_allocated:1;
16
tb_page_addr_t phys_page2;
16
unsigned int temp_allocated:1;
17
target_ulong virt_page2;
17
18
18
- tcg_target_long val;
19
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
19
+ int64_t val;
20
+ /*
20
struct TCGTemp *mem_base;
21
+ * We know that the first page matched, and an otherwise valid TB
21
intptr_t mem_offset;
22
+ * encountered an incomplete instruction at the end of that page,
22
const char *name;
23
+ * therefore we know that generating a new TB from the current PC
23
diff --git a/tcg/tcg.c b/tcg/tcg.c
24
+ * must also require reading from the next page -- even if the
24
index XXXXXXX..XXXXXXX 100644
25
+ * second pages do not match, and therefore the resulting insn
25
--- a/tcg/tcg.c
26
+ * is different for the new TB. Therefore any exception raised
26
+++ b/tcg/tcg.c
27
+ * here by the faulting lookup is not premature.
27
@@ -XXX,XX +XXX,XX @@ static void dump_regs(TCGContext *s)
28
+ */
28
tcg_target_reg_names[ts->mem_base->reg]);
29
+ virt_page2 = TARGET_PAGE_ALIGN(desc->pc);
29
break;
30
phys_page2 = get_page_addr_code(desc->env, virt_page2);
30
case TEMP_VAL_CONST:
31
if (tb->page_addr[1] == phys_page2) {
31
- printf("$0x%" TCG_PRIlx, ts->val);
32
return true;
32
+ printf("$0x%" PRIx64, ts->val);
33
break;
34
case TEMP_VAL_DEAD:
35
printf("D");
36
--
33
--
37
2.25.1
34
2.34.1
38
39
diff view generated by jsdifflib
1
Do not allocate a large block for indexing. Instead, allocate
1
The only user can easily use translator_lduw and
2
for each temporary as they are seen.
2
adjust the type to signed during the return.
3
3
4
In general, this will use less memory, if we consider that most
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
TBs do not touch every target register. This also allows us to
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
allocate TempOptInfo for new temps created during optimization.
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
8
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
8
---
11
tcg/optimize.c | 60 ++++++++++++++++++++++++++++----------------------
9
include/exec/translator.h | 1 -
12
1 file changed, 34 insertions(+), 26 deletions(-)
10
target/i386/tcg/translate.c | 2 +-
11
2 files changed, 1 insertion(+), 2 deletions(-)
13
12
14
diff --git a/tcg/optimize.c b/tcg/optimize.c
13
diff --git a/include/exec/translator.h b/include/exec/translator.h
15
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
16
--- a/tcg/optimize.c
15
--- a/include/exec/translator.h
17
+++ b/tcg/optimize.c
16
+++ b/include/exec/translator.h
18
@@ -XXX,XX +XXX,XX @@ static void reset_temp(TCGArg arg)
17
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
18
19
#define FOR_EACH_TRANSLATOR_LD(F) \
20
F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
21
- F(translator_ldsw, int16_t, cpu_ldsw_code, bswap16) \
22
F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
23
F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
24
F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
25
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/i386/tcg/translate.c
28
+++ b/target/i386/tcg/translate.c
29
@@ -XXX,XX +XXX,XX @@ static inline uint8_t x86_ldub_code(CPUX86State *env, DisasContext *s)
30
31
static inline int16_t x86_ldsw_code(CPUX86State *env, DisasContext *s)
32
{
33
- return translator_ldsw(env, &s->base, advance_pc(env, s, 2));
34
+ return translator_lduw(env, &s->base, advance_pc(env, s, 2));
19
}
35
}
20
36
21
/* Initialize and activate a temporary. */
37
static inline uint16_t x86_lduw_code(CPUX86State *env, DisasContext *s)
22
-static void init_ts_info(TempOptInfo *infos,
23
- TCGTempSet *temps_used, TCGTemp *ts)
24
+static void init_ts_info(TCGTempSet *temps_used, TCGTemp *ts)
25
{
26
size_t idx = temp_idx(ts);
27
- if (!test_bit(idx, temps_used->l)) {
28
- TempOptInfo *ti = &infos[idx];
29
+ TempOptInfo *ti;
30
31
+ if (test_bit(idx, temps_used->l)) {
32
+ return;
33
+ }
34
+ set_bit(idx, temps_used->l);
35
+
36
+ ti = ts->state_ptr;
37
+ if (ti == NULL) {
38
+ ti = tcg_malloc(sizeof(TempOptInfo));
39
ts->state_ptr = ti;
40
- ti->next_copy = ts;
41
- ti->prev_copy = ts;
42
- if (ts->kind == TEMP_CONST) {
43
- ti->is_const = true;
44
- ti->val = ti->mask = ts->val;
45
- if (TCG_TARGET_REG_BITS > 32 && ts->type == TCG_TYPE_I32) {
46
- /* High bits of a 32-bit quantity are garbage. */
47
- ti->mask |= ~0xffffffffull;
48
- }
49
- } else {
50
- ti->is_const = false;
51
- ti->mask = -1;
52
+ }
53
+
54
+ ti->next_copy = ts;
55
+ ti->prev_copy = ts;
56
+ if (ts->kind == TEMP_CONST) {
57
+ ti->is_const = true;
58
+ ti->val = ts->val;
59
+ ti->mask = ts->val;
60
+ if (TCG_TARGET_REG_BITS > 32 && ts->type == TCG_TYPE_I32) {
61
+ /* High bits of a 32-bit quantity are garbage. */
62
+ ti->mask |= ~0xffffffffull;
63
}
64
- set_bit(idx, temps_used->l);
65
+ } else {
66
+ ti->is_const = false;
67
+ ti->mask = -1;
68
}
69
}
70
71
-static void init_arg_info(TempOptInfo *infos,
72
- TCGTempSet *temps_used, TCGArg arg)
73
+static void init_arg_info(TCGTempSet *temps_used, TCGArg arg)
74
{
75
- init_ts_info(infos, temps_used, arg_temp(arg));
76
+ init_ts_info(temps_used, arg_temp(arg));
77
}
78
79
static TCGTemp *find_better_copy(TCGContext *s, TCGTemp *ts)
80
@@ -XXX,XX +XXX,XX @@ static bool swap_commutative2(TCGArg *p1, TCGArg *p2)
81
/* Propagate constants and copies, fold constant expressions. */
82
void tcg_optimize(TCGContext *s)
83
{
84
- int nb_temps, nb_globals;
85
+ int nb_temps, nb_globals, i;
86
TCGOp *op, *op_next, *prev_mb = NULL;
87
- TempOptInfo *infos;
88
TCGTempSet temps_used;
89
90
/* Array VALS has an element for each temp.
91
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
92
93
nb_temps = s->nb_temps;
94
nb_globals = s->nb_globals;
95
+
96
bitmap_zero(temps_used.l, nb_temps);
97
- infos = tcg_malloc(sizeof(TempOptInfo) * nb_temps);
98
+ for (i = 0; i < nb_temps; ++i) {
99
+ s->temps[i].state_ptr = NULL;
100
+ }
101
102
QTAILQ_FOREACH_SAFE(op, &s->ops, link, op_next) {
103
uint64_t mask, partmask, affected, tmp;
104
- int nb_oargs, nb_iargs, i;
105
+ int nb_oargs, nb_iargs;
106
TCGOpcode opc = op->opc;
107
const TCGOpDef *def = &tcg_op_defs[opc];
108
109
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
110
for (i = 0; i < nb_oargs + nb_iargs; i++) {
111
TCGTemp *ts = arg_temp(op->args[i]);
112
if (ts) {
113
- init_ts_info(infos, &temps_used, ts);
114
+ init_ts_info(&temps_used, ts);
115
}
116
}
117
} else {
118
nb_oargs = def->nb_oargs;
119
nb_iargs = def->nb_iargs;
120
for (i = 0; i < nb_oargs + nb_iargs; i++) {
121
- init_arg_info(infos, &temps_used, op->args[i]);
122
+ init_arg_info(&temps_used, op->args[i]);
123
}
124
}
125
126
--
38
--
127
2.25.1
39
2.34.1
128
129
diff view generated by jsdifflib
1
The normal movi opcodes are going away. We need something
1
Pass these along to translator_loop -- pc may be used instead
2
for TCI to use internally.
2
of tb->pc, and host_pc is currently unused. Adjust all targets
3
at one time.
3
4
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Alistair Francis <alistair.francis@wdc.com>
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
9
---
7
include/tcg/tcg-opc.h | 8 ++++++++
10
include/exec/exec-all.h | 1 -
8
tcg/tci.c | 4 ++--
11
include/exec/translator.h | 24 ++++++++++++++++++++----
9
tcg/tci/tcg-target.c.inc | 4 ++--
12
accel/tcg/translate-all.c | 6 ++++--
10
3 files changed, 12 insertions(+), 4 deletions(-)
13
accel/tcg/translator.c | 9 +++++----
14
target/alpha/translate.c | 5 +++--
15
target/arm/translate.c | 5 +++--
16
target/avr/translate.c | 5 +++--
17
target/cris/translate.c | 5 +++--
18
target/hexagon/translate.c | 6 ++++--
19
target/hppa/translate.c | 5 +++--
20
target/i386/tcg/translate.c | 5 +++--
21
target/loongarch/translate.c | 6 ++++--
22
target/m68k/translate.c | 5 +++--
23
target/microblaze/translate.c | 5 +++--
24
target/mips/tcg/translate.c | 5 +++--
25
target/nios2/translate.c | 5 +++--
26
target/openrisc/translate.c | 6 ++++--
27
target/ppc/translate.c | 5 +++--
28
target/riscv/translate.c | 5 +++--
29
target/rx/translate.c | 5 +++--
30
target/s390x/tcg/translate.c | 5 +++--
31
target/sh4/translate.c | 5 +++--
32
target/sparc/translate.c | 5 +++--
33
target/tricore/translate.c | 6 ++++--
34
target/xtensa/translate.c | 6 ++++--
35
25 files changed, 97 insertions(+), 53 deletions(-)
11
36
12
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
37
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
13
index XXXXXXX..XXXXXXX 100644
38
index XXXXXXX..XXXXXXX 100644
14
--- a/include/tcg/tcg-opc.h
39
--- a/include/exec/exec-all.h
15
+++ b/include/tcg/tcg-opc.h
40
+++ b/include/exec/exec-all.h
16
@@ -XXX,XX +XXX,XX @@ DEF(last_generic, 0, 0, 0, TCG_OPF_NOT_PRESENT)
41
@@ -XXX,XX +XXX,XX @@ typedef ram_addr_t tb_page_addr_t;
17
#include "tcg-target.opc.h"
42
#define TB_PAGE_ADDR_FMT RAM_ADDR_FMT
18
#endif
43
#endif
19
44
20
+#ifdef TCG_TARGET_INTERPRETER
45
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns);
21
+/* These opcodes are only for use between the tci generator and interpreter. */
46
void restore_state_to_opc(CPUArchState *env, TranslationBlock *tb,
22
+DEF(tci_movi_i32, 1, 0, 1, TCG_OPF_NOT_PRESENT)
47
target_ulong *data);
23
+#if TCG_TARGET_REG_BITS == 64
48
24
+DEF(tci_movi_i64, 1, 0, 1, TCG_OPF_64BIT | TCG_OPF_NOT_PRESENT)
49
diff --git a/include/exec/translator.h b/include/exec/translator.h
25
+#endif
50
index XXXXXXX..XXXXXXX 100644
26
+#endif
51
--- a/include/exec/translator.h
27
+
52
+++ b/include/exec/translator.h
28
#undef TLADDR_ARGS
53
@@ -XXX,XX +XXX,XX @@
29
#undef DATA64_ARGS
54
#include "exec/translate-all.h"
30
#undef IMPL
55
#include "tcg/tcg.h"
31
diff --git a/tcg/tci.c b/tcg/tci.c
56
32
index XXXXXXX..XXXXXXX 100644
57
+/**
33
--- a/tcg/tci.c
58
+ * gen_intermediate_code
34
+++ b/tcg/tci.c
59
+ * @cpu: cpu context
35
@@ -XXX,XX +XXX,XX @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState *env,
60
+ * @tb: translation block
36
t1 = tci_read_r32(regs, &tb_ptr);
61
+ * @max_insns: max number of instructions to translate
37
tci_write_reg32(regs, t0, t1);
62
+ * @pc: guest virtual program counter address
38
break;
63
+ * @host_pc: host physical program counter address
39
- case INDEX_op_movi_i32:
64
+ *
40
+ case INDEX_op_tci_movi_i32:
65
+ * This function must be provided by the target, which should create
41
t0 = *tb_ptr++;
66
+ * the target-specific DisasContext, and then invoke translator_loop.
42
t1 = tci_read_i32(&tb_ptr);
67
+ */
43
tci_write_reg32(regs, t0, t1);
68
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
44
@@ -XXX,XX +XXX,XX @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState *env,
69
+ target_ulong pc, void *host_pc);
45
t1 = tci_read_r64(regs, &tb_ptr);
70
46
tci_write_reg64(regs, t0, t1);
71
/**
47
break;
72
* DisasJumpType:
48
- case INDEX_op_movi_i64:
73
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
49
+ case INDEX_op_tci_movi_i64:
74
50
t0 = *tb_ptr++;
75
/**
51
t1 = tci_read_i64(&tb_ptr);
76
* translator_loop:
52
tci_write_reg64(regs, t0, t1);
77
- * @ops: Target-specific operations.
53
diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc
78
- * @db: Disassembly context.
54
index XXXXXXX..XXXXXXX 100644
79
* @cpu: Target vCPU.
55
--- a/tcg/tci/tcg-target.c.inc
80
* @tb: Translation block.
56
+++ b/tcg/tci/tcg-target.c.inc
81
* @max_insns: Maximum number of insns to translate.
57
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi(TCGContext *s, TCGType type,
82
+ * @pc: guest virtual program counter address
58
uint8_t *old_code_ptr = s->code_ptr;
83
+ * @host_pc: host physical program counter address
59
uint32_t arg32 = arg;
84
+ * @ops: Target-specific operations.
60
if (type == TCG_TYPE_I32 || arg == arg32) {
85
+ * @db: Disassembly context.
61
- tcg_out_op_t(s, INDEX_op_movi_i32);
86
*
62
+ tcg_out_op_t(s, INDEX_op_tci_movi_i32);
87
* Generic translator loop.
63
tcg_out_r(s, t0);
88
*
64
tcg_out32(s, arg32);
89
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
65
} else {
90
* - When single-stepping is enabled (system-wide or on the current vCPU).
66
tcg_debug_assert(type == TCG_TYPE_I64);
91
* - When too many instructions have been translated.
67
#if TCG_TARGET_REG_BITS == 64
92
*/
68
- tcg_out_op_t(s, INDEX_op_movi_i64);
93
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
69
+ tcg_out_op_t(s, INDEX_op_tci_movi_i64);
94
- CPUState *cpu, TranslationBlock *tb, int max_insns);
70
tcg_out_r(s, t0);
95
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
71
tcg_out64(s, arg);
96
+ target_ulong pc, void *host_pc,
72
#else
97
+ const TranslatorOps *ops, DisasContextBase *db);
98
99
void translator_loop_temp_check(DisasContextBase *db);
100
101
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
102
index XXXXXXX..XXXXXXX 100644
103
--- a/accel/tcg/translate-all.c
104
+++ b/accel/tcg/translate-all.c
105
@@ -XXX,XX +XXX,XX @@
106
107
#include "exec/cputlb.h"
108
#include "exec/translate-all.h"
109
+#include "exec/translator.h"
110
#include "qemu/bitmap.h"
111
#include "qemu/qemu-print.h"
112
#include "qemu/timer.h"
113
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
114
TCGProfile *prof = &tcg_ctx->prof;
115
int64_t ti;
116
#endif
117
+ void *host_pc;
118
119
assert_memory_lock();
120
qemu_thread_jit_write();
121
122
- phys_pc = get_page_addr_code(env, pc);
123
+ phys_pc = get_page_addr_code_hostp(env, pc, &host_pc);
124
125
if (phys_pc == -1) {
126
/* Generate a one-shot TB with 1 insn in it */
127
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
128
tcg_func_start(tcg_ctx);
129
130
tcg_ctx->cpu = env_cpu(env);
131
- gen_intermediate_code(cpu, tb, max_insns);
132
+ gen_intermediate_code(cpu, tb, max_insns, pc, host_pc);
133
assert(tb->size != 0);
134
tcg_ctx->cpu = NULL;
135
max_insns = tb->icount;
136
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
137
index XXXXXXX..XXXXXXX 100644
138
--- a/accel/tcg/translator.c
139
+++ b/accel/tcg/translator.c
140
@@ -XXX,XX +XXX,XX @@ static inline void translator_page_protect(DisasContextBase *dcbase,
141
#endif
142
}
143
144
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
145
- CPUState *cpu, TranslationBlock *tb, int max_insns)
146
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
147
+ target_ulong pc, void *host_pc,
148
+ const TranslatorOps *ops, DisasContextBase *db)
149
{
150
uint32_t cflags = tb_cflags(tb);
151
bool plugin_enabled;
152
153
/* Initialize DisasContext */
154
db->tb = tb;
155
- db->pc_first = tb->pc;
156
- db->pc_next = db->pc_first;
157
+ db->pc_first = pc;
158
+ db->pc_next = pc;
159
db->is_jmp = DISAS_NEXT;
160
db->num_insns = 0;
161
db->max_insns = max_insns;
162
diff --git a/target/alpha/translate.c b/target/alpha/translate.c
163
index XXXXXXX..XXXXXXX 100644
164
--- a/target/alpha/translate.c
165
+++ b/target/alpha/translate.c
166
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps alpha_tr_ops = {
167
.disas_log = alpha_tr_disas_log,
168
};
169
170
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
171
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
172
+ target_ulong pc, void *host_pc)
173
{
174
DisasContext dc;
175
- translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
176
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &alpha_tr_ops, &dc.base);
177
}
178
179
void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
180
diff --git a/target/arm/translate.c b/target/arm/translate.c
181
index XXXXXXX..XXXXXXX 100644
182
--- a/target/arm/translate.c
183
+++ b/target/arm/translate.c
184
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps thumb_translator_ops = {
185
};
186
187
/* generate intermediate code for basic block 'tb'. */
188
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
189
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
190
+ target_ulong pc, void *host_pc)
191
{
192
DisasContext dc = { };
193
const TranslatorOps *ops = &arm_translator_ops;
194
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
195
}
196
#endif
197
198
- translator_loop(ops, &dc.base, cpu, tb, max_insns);
199
+ translator_loop(cpu, tb, max_insns, pc, host_pc, ops, &dc.base);
200
}
201
202
void restore_state_to_opc(CPUARMState *env, TranslationBlock *tb,
203
diff --git a/target/avr/translate.c b/target/avr/translate.c
204
index XXXXXXX..XXXXXXX 100644
205
--- a/target/avr/translate.c
206
+++ b/target/avr/translate.c
207
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps avr_tr_ops = {
208
.disas_log = avr_tr_disas_log,
209
};
210
211
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
212
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
213
+ target_ulong pc, void *host_pc)
214
{
215
DisasContext dc = { };
216
- translator_loop(&avr_tr_ops, &dc.base, cs, tb, max_insns);
217
+ translator_loop(cs, tb, max_insns, pc, host_pc, &avr_tr_ops, &dc.base);
218
}
219
220
void restore_state_to_opc(CPUAVRState *env, TranslationBlock *tb,
221
diff --git a/target/cris/translate.c b/target/cris/translate.c
222
index XXXXXXX..XXXXXXX 100644
223
--- a/target/cris/translate.c
224
+++ b/target/cris/translate.c
225
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps cris_tr_ops = {
226
.disas_log = cris_tr_disas_log,
227
};
228
229
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
230
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
231
+ target_ulong pc, void *host_pc)
232
{
233
DisasContext dc;
234
- translator_loop(&cris_tr_ops, &dc.base, cs, tb, max_insns);
235
+ translator_loop(cs, tb, max_insns, pc, host_pc, &cris_tr_ops, &dc.base);
236
}
237
238
void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
239
diff --git a/target/hexagon/translate.c b/target/hexagon/translate.c
240
index XXXXXXX..XXXXXXX 100644
241
--- a/target/hexagon/translate.c
242
+++ b/target/hexagon/translate.c
243
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hexagon_tr_ops = {
244
.disas_log = hexagon_tr_disas_log,
245
};
246
247
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
248
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
249
+ target_ulong pc, void *host_pc)
250
{
251
DisasContext ctx;
252
253
- translator_loop(&hexagon_tr_ops, &ctx.base, cs, tb, max_insns);
254
+ translator_loop(cs, tb, max_insns, pc, host_pc,
255
+ &hexagon_tr_ops, &ctx.base);
256
}
257
258
#define NAME_LEN 64
259
diff --git a/target/hppa/translate.c b/target/hppa/translate.c
260
index XXXXXXX..XXXXXXX 100644
261
--- a/target/hppa/translate.c
262
+++ b/target/hppa/translate.c
263
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hppa_tr_ops = {
264
.disas_log = hppa_tr_disas_log,
265
};
266
267
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
268
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
269
+ target_ulong pc, void *host_pc)
270
{
271
DisasContext ctx;
272
- translator_loop(&hppa_tr_ops, &ctx.base, cs, tb, max_insns);
273
+ translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base);
274
}
275
276
void restore_state_to_opc(CPUHPPAState *env, TranslationBlock *tb,
277
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
278
index XXXXXXX..XXXXXXX 100644
279
--- a/target/i386/tcg/translate.c
280
+++ b/target/i386/tcg/translate.c
281
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps i386_tr_ops = {
282
};
283
284
/* generate intermediate code for basic block 'tb'. */
285
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
286
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
287
+ target_ulong pc, void *host_pc)
288
{
289
DisasContext dc;
290
291
- translator_loop(&i386_tr_ops, &dc.base, cpu, tb, max_insns);
292
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &i386_tr_ops, &dc.base);
293
}
294
295
void restore_state_to_opc(CPUX86State *env, TranslationBlock *tb,
296
diff --git a/target/loongarch/translate.c b/target/loongarch/translate.c
297
index XXXXXXX..XXXXXXX 100644
298
--- a/target/loongarch/translate.c
299
+++ b/target/loongarch/translate.c
300
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps loongarch_tr_ops = {
301
.disas_log = loongarch_tr_disas_log,
302
};
303
304
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
305
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
306
+ target_ulong pc, void *host_pc)
307
{
308
DisasContext ctx;
309
310
- translator_loop(&loongarch_tr_ops, &ctx.base, cs, tb, max_insns);
311
+ translator_loop(cs, tb, max_insns, pc, host_pc,
312
+ &loongarch_tr_ops, &ctx.base);
313
}
314
315
void loongarch_translate_init(void)
316
diff --git a/target/m68k/translate.c b/target/m68k/translate.c
317
index XXXXXXX..XXXXXXX 100644
318
--- a/target/m68k/translate.c
319
+++ b/target/m68k/translate.c
320
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps m68k_tr_ops = {
321
.disas_log = m68k_tr_disas_log,
322
};
323
324
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
325
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
326
+ target_ulong pc, void *host_pc)
327
{
328
DisasContext dc;
329
- translator_loop(&m68k_tr_ops, &dc.base, cpu, tb, max_insns);
330
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &m68k_tr_ops, &dc.base);
331
}
332
333
static double floatx80_to_double(CPUM68KState *env, uint16_t high, uint64_t low)
334
diff --git a/target/microblaze/translate.c b/target/microblaze/translate.c
335
index XXXXXXX..XXXXXXX 100644
336
--- a/target/microblaze/translate.c
337
+++ b/target/microblaze/translate.c
338
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mb_tr_ops = {
339
.disas_log = mb_tr_disas_log,
340
};
341
342
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
343
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
344
+ target_ulong pc, void *host_pc)
345
{
346
DisasContext dc;
347
- translator_loop(&mb_tr_ops, &dc.base, cpu, tb, max_insns);
348
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &mb_tr_ops, &dc.base);
349
}
350
351
void mb_cpu_dump_state(CPUState *cs, FILE *f, int flags)
352
diff --git a/target/mips/tcg/translate.c b/target/mips/tcg/translate.c
353
index XXXXXXX..XXXXXXX 100644
354
--- a/target/mips/tcg/translate.c
355
+++ b/target/mips/tcg/translate.c
356
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mips_tr_ops = {
357
.disas_log = mips_tr_disas_log,
358
};
359
360
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
361
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
362
+ target_ulong pc, void *host_pc)
363
{
364
DisasContext ctx;
365
366
- translator_loop(&mips_tr_ops, &ctx.base, cs, tb, max_insns);
367
+ translator_loop(cs, tb, max_insns, pc, host_pc, &mips_tr_ops, &ctx.base);
368
}
369
370
void mips_tcg_init(void)
371
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
372
index XXXXXXX..XXXXXXX 100644
373
--- a/target/nios2/translate.c
374
+++ b/target/nios2/translate.c
375
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps nios2_tr_ops = {
376
.disas_log = nios2_tr_disas_log,
377
};
378
379
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
380
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
381
+ target_ulong pc, void *host_pc)
382
{
383
DisasContext dc;
384
- translator_loop(&nios2_tr_ops, &dc.base, cs, tb, max_insns);
385
+ translator_loop(cs, tb, max_insns, pc, host_pc, &nios2_tr_ops, &dc.base);
386
}
387
388
void nios2_cpu_dump_state(CPUState *cs, FILE *f, int flags)
389
diff --git a/target/openrisc/translate.c b/target/openrisc/translate.c
390
index XXXXXXX..XXXXXXX 100644
391
--- a/target/openrisc/translate.c
392
+++ b/target/openrisc/translate.c
393
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps openrisc_tr_ops = {
394
.disas_log = openrisc_tr_disas_log,
395
};
396
397
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
398
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
399
+ target_ulong pc, void *host_pc)
400
{
401
DisasContext ctx;
402
403
- translator_loop(&openrisc_tr_ops, &ctx.base, cs, tb, max_insns);
404
+ translator_loop(cs, tb, max_insns, pc, host_pc,
405
+ &openrisc_tr_ops, &ctx.base);
406
}
407
408
void openrisc_cpu_dump_state(CPUState *cs, FILE *f, int flags)
409
diff --git a/target/ppc/translate.c b/target/ppc/translate.c
410
index XXXXXXX..XXXXXXX 100644
411
--- a/target/ppc/translate.c
412
+++ b/target/ppc/translate.c
413
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps ppc_tr_ops = {
414
.disas_log = ppc_tr_disas_log,
415
};
416
417
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
418
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
419
+ target_ulong pc, void *host_pc)
420
{
421
DisasContext ctx;
422
423
- translator_loop(&ppc_tr_ops, &ctx.base, cs, tb, max_insns);
424
+ translator_loop(cs, tb, max_insns, pc, host_pc, &ppc_tr_ops, &ctx.base);
425
}
426
427
void restore_state_to_opc(CPUPPCState *env, TranslationBlock *tb,
428
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
429
index XXXXXXX..XXXXXXX 100644
430
--- a/target/riscv/translate.c
431
+++ b/target/riscv/translate.c
432
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps riscv_tr_ops = {
433
.disas_log = riscv_tr_disas_log,
434
};
435
436
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
437
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
438
+ target_ulong pc, void *host_pc)
439
{
440
DisasContext ctx;
441
442
- translator_loop(&riscv_tr_ops, &ctx.base, cs, tb, max_insns);
443
+ translator_loop(cs, tb, max_insns, pc, host_pc, &riscv_tr_ops, &ctx.base);
444
}
445
446
void riscv_translate_init(void)
447
diff --git a/target/rx/translate.c b/target/rx/translate.c
448
index XXXXXXX..XXXXXXX 100644
449
--- a/target/rx/translate.c
450
+++ b/target/rx/translate.c
451
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps rx_tr_ops = {
452
.disas_log = rx_tr_disas_log,
453
};
454
455
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
456
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
457
+ target_ulong pc, void *host_pc)
458
{
459
DisasContext dc;
460
461
- translator_loop(&rx_tr_ops, &dc.base, cs, tb, max_insns);
462
+ translator_loop(cs, tb, max_insns, pc, host_pc, &rx_tr_ops, &dc.base);
463
}
464
465
void restore_state_to_opc(CPURXState *env, TranslationBlock *tb,
466
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
467
index XXXXXXX..XXXXXXX 100644
468
--- a/target/s390x/tcg/translate.c
469
+++ b/target/s390x/tcg/translate.c
470
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps s390x_tr_ops = {
471
.disas_log = s390x_tr_disas_log,
472
};
473
474
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
475
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
476
+ target_ulong pc, void *host_pc)
477
{
478
DisasContext dc;
479
480
- translator_loop(&s390x_tr_ops, &dc.base, cs, tb, max_insns);
481
+ translator_loop(cs, tb, max_insns, pc, host_pc, &s390x_tr_ops, &dc.base);
482
}
483
484
void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb,
485
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
486
index XXXXXXX..XXXXXXX 100644
487
--- a/target/sh4/translate.c
488
+++ b/target/sh4/translate.c
489
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sh4_tr_ops = {
490
.disas_log = sh4_tr_disas_log,
491
};
492
493
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
494
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
495
+ target_ulong pc, void *host_pc)
496
{
497
DisasContext ctx;
498
499
- translator_loop(&sh4_tr_ops, &ctx.base, cs, tb, max_insns);
500
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base);
501
}
502
503
void restore_state_to_opc(CPUSH4State *env, TranslationBlock *tb,
504
diff --git a/target/sparc/translate.c b/target/sparc/translate.c
505
index XXXXXXX..XXXXXXX 100644
506
--- a/target/sparc/translate.c
507
+++ b/target/sparc/translate.c
508
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sparc_tr_ops = {
509
.disas_log = sparc_tr_disas_log,
510
};
511
512
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
513
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
514
+ target_ulong pc, void *host_pc)
515
{
516
DisasContext dc = {};
517
518
- translator_loop(&sparc_tr_ops, &dc.base, cs, tb, max_insns);
519
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sparc_tr_ops, &dc.base);
520
}
521
522
void sparc_tcg_init(void)
523
diff --git a/target/tricore/translate.c b/target/tricore/translate.c
524
index XXXXXXX..XXXXXXX 100644
525
--- a/target/tricore/translate.c
526
+++ b/target/tricore/translate.c
527
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps tricore_tr_ops = {
528
};
529
530
531
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
532
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
533
+ target_ulong pc, void *host_pc)
534
{
535
DisasContext ctx;
536
- translator_loop(&tricore_tr_ops, &ctx.base, cs, tb, max_insns);
537
+ translator_loop(cs, tb, max_insns, pc, host_pc,
538
+ &tricore_tr_ops, &ctx.base);
539
}
540
541
void
542
diff --git a/target/xtensa/translate.c b/target/xtensa/translate.c
543
index XXXXXXX..XXXXXXX 100644
544
--- a/target/xtensa/translate.c
545
+++ b/target/xtensa/translate.c
546
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps xtensa_translator_ops = {
547
.disas_log = xtensa_tr_disas_log,
548
};
549
550
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
551
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
552
+ target_ulong pc, void *host_pc)
553
{
554
DisasContext dc = {};
555
- translator_loop(&xtensa_translator_ops, &dc.base, cpu, tb, max_insns);
556
+ translator_loop(cpu, tb, max_insns, pc, host_pc,
557
+ &xtensa_translator_ops, &dc.base);
558
}
559
560
void xtensa_cpu_dump_state(CPUState *cs, FILE *f, int flags)
73
--
561
--
74
2.25.1
562
2.34.1
75
76
diff view generated by jsdifflib
1
Because we now store uint64_t in TCGTemp, we can now always
1
Cache the translation from guest to host address, so we may
2
store the full 64-bit duplicate immediate. So remove the
2
use direct loads when we hit on the primary translation page.
3
difference between 32- and 64-bit hosts.
4
3
4
Look up the second translation page only once, during translation.
5
This obviates another lookup of the second page within tb_gen_code
6
after translation.
7
8
Fixes a bug in that plugin_insn_append should be passed the bytes
9
in the original memory order, not bswapped by pieces.
10
11
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
12
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
14
---
7
tcg/optimize.c | 9 ++++-----
15
include/exec/translator.h | 63 +++++++++++--------
8
tcg/tcg-op-vec.c | 39 ++++++++++-----------------------------
16
accel/tcg/translate-all.c | 23 +++----
9
tcg/tcg.c | 7 +------
17
accel/tcg/translator.c | 126 +++++++++++++++++++++++++++++---------
10
3 files changed, 15 insertions(+), 40 deletions(-)
18
3 files changed, 141 insertions(+), 71 deletions(-)
11
19
12
diff --git a/tcg/optimize.c b/tcg/optimize.c
20
diff --git a/include/exec/translator.h b/include/exec/translator.h
13
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/optimize.c
22
--- a/include/exec/translator.h
15
+++ b/tcg/optimize.c
23
+++ b/include/exec/translator.h
16
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
24
@@ -XXX,XX +XXX,XX @@ typedef enum DisasJumpType {
17
case INDEX_op_dup2_vec:
25
* Architecture-agnostic disassembly context.
18
assert(TCG_TARGET_REG_BITS == 32);
26
*/
19
if (arg_is_const(op->args[1]) && arg_is_const(op->args[2])) {
27
typedef struct DisasContextBase {
20
- tmp = arg_info(op->args[1])->val;
28
- const TranslationBlock *tb;
21
- if (tmp == arg_info(op->args[2])->val) {
29
+ TranslationBlock *tb;
22
- tcg_opt_gen_movi(s, &temps_used, op, op->args[0], tmp);
30
target_ulong pc_first;
23
- break;
31
target_ulong pc_next;
24
- }
32
DisasJumpType is_jmp;
25
+ tcg_opt_gen_movi(s, &temps_used, op, op->args[0],
33
int num_insns;
26
+ deposit64(arg_info(op->args[1])->val, 32, 32,
34
int max_insns;
27
+ arg_info(op->args[2])->val));
35
bool singlestep_enabled;
28
+ break;
36
-#ifdef CONFIG_USER_ONLY
29
} else if (args_are_copies(op->args[1], op->args[2])) {
37
- /*
30
op->opc = INDEX_op_dup_vec;
38
- * Guest address of the last byte of the last protected page.
31
TCGOP_VECE(op) = MO_32;
39
- *
32
diff --git a/tcg/tcg-op-vec.c b/tcg/tcg-op-vec.c
40
- * Pages containing the translated instructions are made non-writable in
41
- * order to achieve consistency in case another thread is modifying the
42
- * code while translate_insn() fetches the instruction bytes piecemeal.
43
- * Such writer threads are blocked on mmap_lock() in page_unprotect().
44
- */
45
- target_ulong page_protect_end;
46
-#endif
47
+ void *host_addr[2];
48
} DisasContextBase;
49
50
/**
51
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
52
* the relevant information at translation time.
53
*/
54
55
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
56
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
57
- abi_ptr pc, bool do_swap); \
58
- static inline type fullname(CPUArchState *env, \
59
- DisasContextBase *dcbase, abi_ptr pc) \
60
- { \
61
- return fullname ## _swap(env, dcbase, pc, false); \
62
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
63
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
64
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
65
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
66
+
67
+static inline uint16_t
68
+translator_lduw_swap(CPUArchState *env, DisasContextBase *db,
69
+ abi_ptr pc, bool do_swap)
70
+{
71
+ uint16_t ret = translator_lduw(env, db, pc);
72
+ if (do_swap) {
73
+ ret = bswap16(ret);
74
}
75
+ return ret;
76
+}
77
78
-#define FOR_EACH_TRANSLATOR_LD(F) \
79
- F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
80
- F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
81
- F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
82
- F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
83
+static inline uint32_t
84
+translator_ldl_swap(CPUArchState *env, DisasContextBase *db,
85
+ abi_ptr pc, bool do_swap)
86
+{
87
+ uint32_t ret = translator_ldl(env, db, pc);
88
+ if (do_swap) {
89
+ ret = bswap32(ret);
90
+ }
91
+ return ret;
92
+}
93
94
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
95
-
96
-#undef GEN_TRANSLATOR_LD
97
+static inline uint64_t
98
+translator_ldq_swap(CPUArchState *env, DisasContextBase *db,
99
+ abi_ptr pc, bool do_swap)
100
+{
101
+ uint64_t ret = translator_ldq_swap(env, db, pc, false);
102
+ if (do_swap) {
103
+ ret = bswap64(ret);
104
+ }
105
+ return ret;
106
+}
107
108
/*
109
* Return whether addr is on the same page as where disassembly started.
110
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
33
index XXXXXXX..XXXXXXX 100644
111
index XXXXXXX..XXXXXXX 100644
34
--- a/tcg/tcg-op-vec.c
112
--- a/accel/tcg/translate-all.c
35
+++ b/tcg/tcg-op-vec.c
113
+++ b/accel/tcg/translate-all.c
36
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mov_vec(TCGv_vec r, TCGv_vec a)
114
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
37
}
115
{
116
CPUArchState *env = cpu->env_ptr;
117
TranslationBlock *tb, *existing_tb;
118
- tb_page_addr_t phys_pc, phys_page2;
119
- target_ulong virt_page2;
120
+ tb_page_addr_t phys_pc;
121
tcg_insn_unit *gen_code_buf;
122
int gen_code_size, search_size, max_insns;
123
#ifdef CONFIG_PROFILER
124
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
125
tb->flags = flags;
126
tb->cflags = cflags;
127
tb->trace_vcpu_dstate = *cpu->trace_dstate;
128
+ tb->page_addr[0] = phys_pc;
129
+ tb->page_addr[1] = -1;
130
tcg_ctx->tb_cflags = cflags;
131
tb_overflow:
132
133
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
134
}
135
136
/*
137
- * If the TB is not associated with a physical RAM page then
138
- * it must be a temporary one-insn TB, and we have nothing to do
139
- * except fill in the page_addr[] fields. Return early before
140
- * attempting to link to other TBs or add to the lookup table.
141
+ * If the TB is not associated with a physical RAM page then it must be
142
+ * a temporary one-insn TB, and we have nothing left to do. Return early
143
+ * before attempting to link to other TBs or add to the lookup table.
144
*/
145
- if (phys_pc == -1) {
146
- tb->page_addr[0] = tb->page_addr[1] = -1;
147
+ if (tb->page_addr[0] == -1) {
148
return tb;
149
}
150
151
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
152
*/
153
tcg_tb_insert(tb);
154
155
- /* check next page if needed */
156
- virt_page2 = (pc + tb->size - 1) & TARGET_PAGE_MASK;
157
- phys_page2 = -1;
158
- if ((pc & TARGET_PAGE_MASK) != virt_page2) {
159
- phys_page2 = get_page_addr_code(env, virt_page2);
160
- }
161
/*
162
* No explicit memory barrier is required -- tb_link_page() makes the
163
* TB visible in a consistent state.
164
*/
165
- existing_tb = tb_link_page(tb, phys_pc, phys_page2);
166
+ existing_tb = tb_link_page(tb, tb->page_addr[0], tb->page_addr[1]);
167
/* if the TB already exists, discard what we just translated */
168
if (unlikely(existing_tb != tb)) {
169
uintptr_t orig_aligned = (uintptr_t)gen_code_buf;
170
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
171
index XXXXXXX..XXXXXXX 100644
172
--- a/accel/tcg/translator.c
173
+++ b/accel/tcg/translator.c
174
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest)
175
return ((db->pc_first ^ dest) & TARGET_PAGE_MASK) == 0;
38
}
176
}
39
177
40
-#define MO_REG (TCG_TARGET_REG_BITS == 64 ? MO_64 : MO_32)
178
-static inline void translator_page_protect(DisasContextBase *dcbase,
41
-
179
- target_ulong pc)
42
-static void do_dupi_vec(TCGv_vec r, unsigned vece, TCGArg a)
43
-{
180
-{
44
- TCGTemp *rt = tcgv_vec_temp(r);
181
-#ifdef CONFIG_USER_ONLY
45
- vec_gen_2(INDEX_op_dupi_vec, rt->base_type, vece, temp_arg(rt), a);
182
- dcbase->page_protect_end = pc | ~TARGET_PAGE_MASK;
183
- page_protect(pc);
184
-#endif
46
-}
185
-}
47
-
186
-
48
TCGv_vec tcg_const_zeros_vec(TCGType type)
187
void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
188
target_ulong pc, void *host_pc,
189
const TranslatorOps *ops, DisasContextBase *db)
190
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
191
db->num_insns = 0;
192
db->max_insns = max_insns;
193
db->singlestep_enabled = cflags & CF_SINGLE_STEP;
194
- translator_page_protect(db, db->pc_next);
195
+ db->host_addr[0] = host_pc;
196
+ db->host_addr[1] = NULL;
197
+
198
+#ifdef CONFIG_USER_ONLY
199
+ page_protect(pc);
200
+#endif
201
202
ops->init_disas_context(db, cpu);
203
tcg_debug_assert(db->is_jmp == DISAS_NEXT); /* no early exit */
204
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
205
#endif
206
}
207
208
-static inline void translator_maybe_page_protect(DisasContextBase *dcbase,
209
- target_ulong pc, size_t len)
210
+static void *translator_access(CPUArchState *env, DisasContextBase *db,
211
+ target_ulong pc, size_t len)
49
{
212
{
50
TCGv_vec ret = tcg_temp_new_vec(type);
213
-#ifdef CONFIG_USER_ONLY
51
- do_dupi_vec(ret, MO_REG, 0);
214
- target_ulong end = pc + len - 1;
52
+ tcg_gen_dupi_vec(MO_64, ret, 0);
215
+ void *host;
53
return ret;
216
+ target_ulong base, end;
217
+ TranslationBlock *tb;
218
219
- if (end > dcbase->page_protect_end) {
220
- translator_page_protect(dcbase, end);
221
+ tb = db->tb;
222
+
223
+ /* Use slow path if first page is MMIO. */
224
+ if (unlikely(tb->page_addr[0] == -1)) {
225
+ return NULL;
226
}
227
+
228
+ end = pc + len - 1;
229
+ if (likely(is_same_page(db, end))) {
230
+ host = db->host_addr[0];
231
+ base = db->pc_first;
232
+ } else {
233
+ host = db->host_addr[1];
234
+ base = TARGET_PAGE_ALIGN(db->pc_first);
235
+ if (host == NULL) {
236
+ tb->page_addr[1] =
237
+ get_page_addr_code_hostp(env, base, &db->host_addr[1]);
238
+#ifdef CONFIG_USER_ONLY
239
+ page_protect(end);
240
#endif
241
+ /* We cannot handle MMIO as second page. */
242
+ assert(tb->page_addr[1] != -1);
243
+ host = db->host_addr[1];
244
+ }
245
+
246
+ /* Use slow path when crossing pages. */
247
+ if (is_same_page(db, pc)) {
248
+ return NULL;
249
+ }
250
+ }
251
+
252
+ tcg_debug_assert(pc >= base);
253
+ return host + (pc - base);
54
}
254
}
55
255
56
TCGv_vec tcg_const_ones_vec(TCGType type)
256
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
57
{
257
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
58
TCGv_vec ret = tcg_temp_new_vec(type);
258
- abi_ptr pc, bool do_swap) \
59
- do_dupi_vec(ret, MO_REG, -1);
259
- { \
60
+ tcg_gen_dupi_vec(MO_64, ret, -1);
260
- translator_maybe_page_protect(dcbase, pc, sizeof(type)); \
61
return ret;
261
- type ret = load_fn(env, pc); \
62
}
262
- if (do_swap) { \
63
263
- ret = swap_fn(ret); \
64
@@ -XXX,XX +XXX,XX @@ TCGv_vec tcg_const_ones_vec_matching(TCGv_vec m)
264
- } \
65
265
- plugin_insn_append(pc, &ret, sizeof(ret)); \
66
void tcg_gen_dup64i_vec(TCGv_vec r, uint64_t a)
266
- return ret; \
67
{
267
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
68
- if (TCG_TARGET_REG_BITS == 64) {
268
+{
69
- do_dupi_vec(r, MO_64, a);
269
+ uint8_t ret;
70
- } else if (a == dup_const(MO_32, a)) {
270
+ void *p = translator_access(env, db, pc, sizeof(ret));
71
- do_dupi_vec(r, MO_32, a);
271
+
72
- } else {
272
+ if (p) {
73
- TCGv_i64 c = tcg_const_i64(a);
273
+ plugin_insn_append(pc, p, sizeof(ret));
74
- tcg_gen_dup_i64_vec(MO_64, r, c);
274
+ return ldub_p(p);
75
- tcg_temp_free_i64(c);
275
}
76
- }
276
+ ret = cpu_ldub_code(env, pc);
77
+ tcg_gen_dupi_vec(MO_64, r, a);
277
+ plugin_insn_append(pc, &ret, sizeof(ret));
78
}
278
+ return ret;
79
279
+}
80
void tcg_gen_dup32i_vec(TCGv_vec r, uint32_t a)
280
81
{
281
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
82
- do_dupi_vec(r, MO_REG, dup_const(MO_32, a));
282
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
83
+ tcg_gen_dupi_vec(MO_32, r, a);
283
+{
84
}
284
+ uint16_t ret, plug;
85
285
+ void *p = translator_access(env, db, pc, sizeof(ret));
86
void tcg_gen_dup16i_vec(TCGv_vec r, uint32_t a)
286
87
{
287
-#undef GEN_TRANSLATOR_LD
88
- do_dupi_vec(r, MO_REG, dup_const(MO_16, a));
288
+ if (p) {
89
+ tcg_gen_dupi_vec(MO_16, r, a);
289
+ plugin_insn_append(pc, p, sizeof(ret));
90
}
290
+ return lduw_p(p);
91
291
+ }
92
void tcg_gen_dup8i_vec(TCGv_vec r, uint32_t a)
292
+ ret = cpu_lduw_code(env, pc);
93
{
293
+ plug = tswap16(ret);
94
- do_dupi_vec(r, MO_REG, dup_const(MO_8, a));
294
+ plugin_insn_append(pc, &plug, sizeof(ret));
95
+ tcg_gen_dupi_vec(MO_8, r, a);
295
+ return ret;
96
}
296
+}
97
297
+
98
void tcg_gen_dupi_vec(unsigned vece, TCGv_vec r, uint64_t a)
298
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
99
{
299
+{
100
- if (vece == MO_64) {
300
+ uint32_t ret, plug;
101
- tcg_gen_dup64i_vec(r, a);
301
+ void *p = translator_access(env, db, pc, sizeof(ret));
102
- } else {
302
+
103
- do_dupi_vec(r, MO_REG, dup_const(vece, a));
303
+ if (p) {
104
- }
304
+ plugin_insn_append(pc, p, sizeof(ret));
105
+ TCGTemp *rt = tcgv_vec_temp(r);
305
+ return ldl_p(p);
106
+ tcg_gen_mov_vec(r, tcg_constant_vec(rt->base_type, vece, a));
306
+ }
107
}
307
+ ret = cpu_ldl_code(env, pc);
108
308
+ plug = tswap32(ret);
109
void tcg_gen_dup_i64_vec(unsigned vece, TCGv_vec r, TCGv_i64 a)
309
+ plugin_insn_append(pc, &plug, sizeof(ret));
110
@@ -XXX,XX +XXX,XX @@ void tcg_gen_abs_vec(unsigned vece, TCGv_vec r, TCGv_vec a)
310
+ return ret;
111
if (tcg_can_emit_vec_op(INDEX_op_sari_vec, type, vece) > 0) {
311
+}
112
tcg_gen_sari_vec(vece, t, a, (8 << vece) - 1);
312
+
113
} else {
313
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
114
- do_dupi_vec(t, MO_REG, 0);
314
+{
115
- tcg_gen_cmp_vec(TCG_COND_LT, vece, t, a, t);
315
+ uint64_t ret, plug;
116
+ tcg_gen_cmp_vec(TCG_COND_LT, vece, t, a,
316
+ void *p = translator_access(env, db, pc, sizeof(ret));
117
+ tcg_constant_vec(type, vece, 0));
317
+
118
}
318
+ if (p) {
119
tcg_gen_xor_vec(vece, r, a, t);
319
+ plugin_insn_append(pc, p, sizeof(ret));
120
tcg_gen_sub_vec(vece, r, r, t);
320
+ return ldq_p(p);
121
diff --git a/tcg/tcg.c b/tcg/tcg.c
321
+ }
122
index XXXXXXX..XXXXXXX 100644
322
+ ret = cpu_ldq_code(env, pc);
123
--- a/tcg/tcg.c
323
+ plug = tswap64(ret);
124
+++ b/tcg/tcg.c
324
+ plugin_insn_append(pc, &plug, sizeof(ret));
125
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
325
+ return ret;
126
* The targets will, in general, have to do this search anyway,
326
+}
127
* do this generically.
128
*/
129
- if (TCG_TARGET_REG_BITS == 32) {
130
- val = dup_const(MO_32, val);
131
- vece = MO_32;
132
- }
133
if (val == dup_const(MO_8, val)) {
134
vece = MO_8;
135
} else if (val == dup_const(MO_16, val)) {
136
vece = MO_16;
137
- } else if (TCG_TARGET_REG_BITS == 64 &&
138
- val == dup_const(MO_32, val)) {
139
+ } else if (val == dup_const(MO_32, val)) {
140
vece = MO_32;
141
}
142
143
--
327
--
144
2.25.1
328
2.34.1
145
146
diff view generated by jsdifflib
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
3
Right now translator stops right *after* the end of a page, which
4
breaks reporting of fault locations when the last instruction of a
5
multi-insn translation block crosses a page boundary.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-Id: <20220817150506.592862-3-iii@linux.ibm.com>
1
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
---
11
---
3
include/tcg/tcg.h | 1 +
12
target/s390x/tcg/translate.c | 15 +++-
4
tcg/tcg-op-gvec.c | 129 ++++++++++++++++++----------------------------
13
tests/tcg/s390x/noexec.c | 106 +++++++++++++++++++++++
5
tcg/tcg.c | 8 +++
14
tests/tcg/multiarch/noexec.c.inc | 139 +++++++++++++++++++++++++++++++
6
3 files changed, 60 insertions(+), 78 deletions(-)
15
tests/tcg/s390x/Makefile.target | 1 +
16
4 files changed, 257 insertions(+), 4 deletions(-)
17
create mode 100644 tests/tcg/s390x/noexec.c
18
create mode 100644 tests/tcg/multiarch/noexec.c.inc
7
19
8
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
20
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
9
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
10
--- a/include/tcg/tcg.h
22
--- a/target/s390x/tcg/translate.c
11
+++ b/include/tcg/tcg.h
23
+++ b/target/s390x/tcg/translate.c
12
@@ -XXX,XX +XXX,XX @@ static inline TCGv_i64 tcg_constant_i64(int64_t val)
24
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
25
dc->insn_start = tcg_last_op();
13
}
26
}
14
27
15
TCGv_vec tcg_constant_vec(TCGType type, unsigned vece, int64_t val);
28
+static target_ulong get_next_pc(CPUS390XState *env, DisasContext *s,
16
+TCGv_vec tcg_constant_vec_matching(TCGv_vec match, unsigned vece, int64_t val);
29
+ uint64_t pc)
17
30
+{
18
#if UINTPTR_MAX == UINT32_MAX
31
+ uint64_t insn = ld_code2(env, s, pc);
19
# define tcg_const_ptr(x) ((TCGv_ptr)tcg_const_i32((intptr_t)(x)))
32
+
20
diff --git a/tcg/tcg-op-gvec.c b/tcg/tcg-op-gvec.c
33
+ return pc + get_ilen((insn >> 8) & 0xff);
21
index XXXXXXX..XXXXXXX 100644
34
+}
22
--- a/tcg/tcg-op-gvec.c
35
+
23
+++ b/tcg/tcg-op-gvec.c
36
static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
24
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2_ool(uint32_t dofs, uint32_t aofs,
25
gen_helper_gvec_2 *fn)
26
{
37
{
27
TCGv_ptr a0, a1;
38
CPUS390XState *env = cs->env_ptr;
28
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
39
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
29
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
40
30
41
dc->base.is_jmp = translate_one(env, dc);
31
a0 = tcg_temp_new_ptr();
42
if (dc->base.is_jmp == DISAS_NEXT) {
32
a1 = tcg_temp_new_ptr();
43
- uint64_t page_start;
33
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2_ool(uint32_t dofs, uint32_t aofs,
44
-
34
45
- page_start = dc->base.pc_first & TARGET_PAGE_MASK;
35
tcg_temp_free_ptr(a0);
46
- if (dc->base.pc_next - page_start >= TARGET_PAGE_SIZE || dc->ex_value) {
36
tcg_temp_free_ptr(a1);
47
+ if (!is_same_page(dcbase, dc->base.pc_next) ||
37
- tcg_temp_free_i32(desc);
48
+ !is_same_page(dcbase, get_next_pc(env, dc, dc->base.pc_next)) ||
38
}
49
+ dc->ex_value) {
39
50
dc->base.is_jmp = DISAS_TOO_MANY;
40
/* Generate a call to a gvec-style helper with two vector operands
41
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2i_ool(uint32_t dofs, uint32_t aofs, TCGv_i64 c,
42
gen_helper_gvec_2i *fn)
43
{
44
TCGv_ptr a0, a1;
45
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
46
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
47
48
a0 = tcg_temp_new_ptr();
49
a1 = tcg_temp_new_ptr();
50
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2i_ool(uint32_t dofs, uint32_t aofs, TCGv_i64 c,
51
52
tcg_temp_free_ptr(a0);
53
tcg_temp_free_ptr(a1);
54
- tcg_temp_free_i32(desc);
55
}
56
57
/* Generate a call to a gvec-style helper with three vector operands. */
58
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_3_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
59
gen_helper_gvec_3 *fn)
60
{
61
TCGv_ptr a0, a1, a2;
62
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
63
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
64
65
a0 = tcg_temp_new_ptr();
66
a1 = tcg_temp_new_ptr();
67
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_3_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
68
tcg_temp_free_ptr(a0);
69
tcg_temp_free_ptr(a1);
70
tcg_temp_free_ptr(a2);
71
- tcg_temp_free_i32(desc);
72
}
73
74
/* Generate a call to a gvec-style helper with four vector operands. */
75
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_4_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
76
int32_t data, gen_helper_gvec_4 *fn)
77
{
78
TCGv_ptr a0, a1, a2, a3;
79
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
80
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
81
82
a0 = tcg_temp_new_ptr();
83
a1 = tcg_temp_new_ptr();
84
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_4_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
85
tcg_temp_free_ptr(a1);
86
tcg_temp_free_ptr(a2);
87
tcg_temp_free_ptr(a3);
88
- tcg_temp_free_i32(desc);
89
}
90
91
/* Generate a call to a gvec-style helper with five vector operands. */
92
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_5_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
93
uint32_t maxsz, int32_t data, gen_helper_gvec_5 *fn)
94
{
95
TCGv_ptr a0, a1, a2, a3, a4;
96
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
97
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
98
99
a0 = tcg_temp_new_ptr();
100
a1 = tcg_temp_new_ptr();
101
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_5_ool(uint32_t dofs, uint32_t aofs, uint32_t bofs,
102
tcg_temp_free_ptr(a2);
103
tcg_temp_free_ptr(a3);
104
tcg_temp_free_ptr(a4);
105
- tcg_temp_free_i32(desc);
106
}
107
108
/* Generate a call to a gvec-style helper with three vector operands
109
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2_ptr(uint32_t dofs, uint32_t aofs,
110
int32_t data, gen_helper_gvec_2_ptr *fn)
111
{
112
TCGv_ptr a0, a1;
113
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
114
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
115
116
a0 = tcg_temp_new_ptr();
117
a1 = tcg_temp_new_ptr();
118
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2_ptr(uint32_t dofs, uint32_t aofs,
119
120
tcg_temp_free_ptr(a0);
121
tcg_temp_free_ptr(a1);
122
- tcg_temp_free_i32(desc);
123
}
124
125
/* Generate a call to a gvec-style helper with three vector operands
126
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_3_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
127
int32_t data, gen_helper_gvec_3_ptr *fn)
128
{
129
TCGv_ptr a0, a1, a2;
130
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
131
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
132
133
a0 = tcg_temp_new_ptr();
134
a1 = tcg_temp_new_ptr();
135
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_3_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
136
tcg_temp_free_ptr(a0);
137
tcg_temp_free_ptr(a1);
138
tcg_temp_free_ptr(a2);
139
- tcg_temp_free_i32(desc);
140
}
141
142
/* Generate a call to a gvec-style helper with four vector operands
143
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_4_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
144
gen_helper_gvec_4_ptr *fn)
145
{
146
TCGv_ptr a0, a1, a2, a3;
147
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
148
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
149
150
a0 = tcg_temp_new_ptr();
151
a1 = tcg_temp_new_ptr();
152
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_4_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
153
tcg_temp_free_ptr(a1);
154
tcg_temp_free_ptr(a2);
155
tcg_temp_free_ptr(a3);
156
- tcg_temp_free_i32(desc);
157
}
158
159
/* Generate a call to a gvec-style helper with five vector operands
160
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_5_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
161
gen_helper_gvec_5_ptr *fn)
162
{
163
TCGv_ptr a0, a1, a2, a3, a4;
164
- TCGv_i32 desc = tcg_const_i32(simd_desc(oprsz, maxsz, data));
165
+ TCGv_i32 desc = tcg_constant_i32(simd_desc(oprsz, maxsz, data));
166
167
a0 = tcg_temp_new_ptr();
168
a1 = tcg_temp_new_ptr();
169
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_5_ptr(uint32_t dofs, uint32_t aofs, uint32_t bofs,
170
tcg_temp_free_ptr(a2);
171
tcg_temp_free_ptr(a3);
172
tcg_temp_free_ptr(a4);
173
- tcg_temp_free_i32(desc);
174
}
175
176
/* Return true if we want to implement something of OPRSZ bytes
177
@@ -XXX,XX +XXX,XX @@ static void do_dup(unsigned vece, uint32_t dofs, uint32_t oprsz,
178
|| (TCG_TARGET_REG_BITS == 64
179
&& (in_c == 0 || in_c == -1
180
|| !check_size_impl(oprsz, 4)))) {
181
- t_64 = tcg_const_i64(in_c);
182
+ t_64 = tcg_constant_i64(in_c);
183
} else {
184
- t_32 = tcg_const_i32(in_c);
185
+ t_32 = tcg_constant_i32(in_c);
186
}
187
}
188
189
@@ -XXX,XX +XXX,XX @@ static void do_dup(unsigned vece, uint32_t dofs, uint32_t oprsz,
190
t_val = tcg_temp_new_i32();
191
tcg_gen_extrl_i64_i32(t_val, in_64);
192
} else {
193
- t_val = tcg_const_i32(in_c);
194
+ t_val = tcg_constant_i32(in_c);
195
}
196
gen_helper_memset(t_ptr, t_ptr, t_val, t_size);
197
198
- if (!in_32) {
199
+ if (in_64) {
200
tcg_temp_free_i32(t_val);
201
}
202
tcg_temp_free_ptr(t_size);
203
@@ -XXX,XX +XXX,XX @@ static void do_dup(unsigned vece, uint32_t dofs, uint32_t oprsz,
204
return;
205
}
206
207
- t_desc = tcg_const_i32(simd_desc(oprsz, maxsz, 0));
208
+ t_desc = tcg_constant_i32(simd_desc(oprsz, maxsz, 0));
209
210
if (vece == MO_64) {
211
if (in_64) {
212
gen_helper_gvec_dup64(t_ptr, t_desc, in_64);
213
} else {
214
- t_64 = tcg_const_i64(in_c);
215
+ t_64 = tcg_constant_i64(in_c);
216
gen_helper_gvec_dup64(t_ptr, t_desc, t_64);
217
- tcg_temp_free_i64(t_64);
218
}
219
} else {
220
typedef void dup_fn(TCGv_ptr, TCGv_i32, TCGv_i32);
221
@@ -XXX,XX +XXX,XX @@ static void do_dup(unsigned vece, uint32_t dofs, uint32_t oprsz,
222
223
if (in_32) {
224
fns[vece](t_ptr, t_desc, in_32);
225
- } else {
226
+ } else if (in_64) {
227
t_32 = tcg_temp_new_i32();
228
- if (in_64) {
229
- tcg_gen_extrl_i64_i32(t_32, in_64);
230
- } else if (vece == MO_8) {
231
- tcg_gen_movi_i32(t_32, in_c & 0xff);
232
- } else if (vece == MO_16) {
233
- tcg_gen_movi_i32(t_32, in_c & 0xffff);
234
- } else {
235
- tcg_gen_movi_i32(t_32, in_c);
236
- }
237
+ tcg_gen_extrl_i64_i32(t_32, in_64);
238
fns[vece](t_ptr, t_desc, t_32);
239
tcg_temp_free_i32(t_32);
240
+ } else {
241
+ if (vece == MO_8) {
242
+ in_c &= 0xff;
243
+ } else if (vece == MO_16) {
244
+ in_c &= 0xffff;
245
+ }
246
+ t_32 = tcg_constant_i32(in_c);
247
+ fns[vece](t_ptr, t_desc, t_32);
248
}
51
}
249
}
52
}
250
53
diff --git a/tests/tcg/s390x/noexec.c b/tests/tcg/s390x/noexec.c
251
tcg_temp_free_ptr(t_ptr);
54
new file mode 100644
252
- tcg_temp_free_i32(t_desc);
55
index XXXXXXX..XXXXXXX
253
return;
56
--- /dev/null
254
57
+++ b/tests/tcg/s390x/noexec.c
255
done:
58
@@ -XXX,XX +XXX,XX @@
256
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_2i(uint32_t dofs, uint32_t aofs, uint32_t oprsz,
59
+#include "../multiarch/noexec.c.inc"
257
if (g->fno) {
60
+
258
tcg_gen_gvec_2_ool(dofs, aofs, oprsz, maxsz, c, g->fno);
61
+static void *arch_mcontext_pc(const mcontext_t *ctx)
259
} else {
62
+{
260
- TCGv_i64 tcg_c = tcg_const_i64(c);
63
+ return (void *)ctx->psw.addr;
261
+ TCGv_i64 tcg_c = tcg_constant_i64(c);
64
+}
262
tcg_gen_gvec_2i_ool(dofs, aofs, tcg_c, oprsz,
65
+
263
maxsz, c, g->fnoi);
66
+static int arch_mcontext_arg(const mcontext_t *ctx)
264
- tcg_temp_free_i64(tcg_c);
67
+{
265
}
68
+ return ctx->gregs[2];
266
oprsz = maxsz;
69
+}
267
}
70
+
268
@@ -XXX,XX +XXX,XX @@ static void gen_addv_mask(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b, TCGv_i64 m)
71
+static void arch_flush(void *p, int len)
269
72
+{
270
void tcg_gen_vec_add8_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
73
+}
271
{
74
+
272
- TCGv_i64 m = tcg_const_i64(dup_const(MO_8, 0x80));
75
+extern char noexec_1[];
273
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_8, 0x80));
76
+extern char noexec_2[];
274
gen_addv_mask(d, a, b, m);
77
+extern char noexec_end[];
275
- tcg_temp_free_i64(m);
78
+
276
}
79
+asm("noexec_1:\n"
277
80
+ " lgfi %r2,1\n" /* %r2 is 0 on entry, set 1. */
278
void tcg_gen_vec_add16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
81
+ "noexec_2:\n"
279
{
82
+ " lgfi %r2,2\n" /* %r2 is 0/1; set 2. */
280
- TCGv_i64 m = tcg_const_i64(dup_const(MO_16, 0x8000));
83
+ " br %r14\n" /* return */
281
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_16, 0x8000));
84
+ "noexec_end:");
282
gen_addv_mask(d, a, b, m);
85
+
283
- tcg_temp_free_i64(m);
86
+extern char exrl_1[];
284
}
87
+extern char exrl_2[];
285
88
+extern char exrl_end[];
286
void tcg_gen_vec_add32_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
89
+
287
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_adds(unsigned vece, uint32_t dofs, uint32_t aofs,
90
+asm("exrl_1:\n"
288
void tcg_gen_gvec_addi(unsigned vece, uint32_t dofs, uint32_t aofs,
91
+ " exrl %r0, exrl_2\n"
289
int64_t c, uint32_t oprsz, uint32_t maxsz)
92
+ " br %r14\n"
290
{
93
+ "exrl_2:\n"
291
- TCGv_i64 tmp = tcg_const_i64(c);
94
+ " lgfi %r2,2\n"
292
+ TCGv_i64 tmp = tcg_constant_i64(c);
95
+ "exrl_end:");
293
tcg_gen_gvec_adds(vece, dofs, aofs, tmp, oprsz, maxsz);
96
+
294
- tcg_temp_free_i64(tmp);
97
+int main(void)
295
}
98
+{
296
99
+ struct noexec_test noexec_tests[] = {
297
static const TCGOpcode vecop_list_sub[] = { INDEX_op_sub_vec, 0 };
100
+ {
298
@@ -XXX,XX +XXX,XX @@ static void gen_subv_mask(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b, TCGv_i64 m)
101
+ .name = "fallthrough",
299
102
+ .test_code = noexec_1,
300
void tcg_gen_vec_sub8_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
103
+ .test_len = noexec_end - noexec_1,
301
{
104
+ .page_ofs = noexec_1 - noexec_2,
302
- TCGv_i64 m = tcg_const_i64(dup_const(MO_8, 0x80));
105
+ .entry_ofs = noexec_1 - noexec_2,
303
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_8, 0x80));
106
+ .expected_si_ofs = 0,
304
gen_subv_mask(d, a, b, m);
107
+ .expected_pc_ofs = 0,
305
- tcg_temp_free_i64(m);
108
+ .expected_arg = 1,
306
}
109
+ },
307
110
+ {
308
void tcg_gen_vec_sub16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
111
+ .name = "jump",
309
{
112
+ .test_code = noexec_1,
310
- TCGv_i64 m = tcg_const_i64(dup_const(MO_16, 0x8000));
113
+ .test_len = noexec_end - noexec_1,
311
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_16, 0x8000));
114
+ .page_ofs = noexec_1 - noexec_2,
312
gen_subv_mask(d, a, b, m);
115
+ .entry_ofs = 0,
313
- tcg_temp_free_i64(m);
116
+ .expected_si_ofs = 0,
314
}
117
+ .expected_pc_ofs = 0,
315
118
+ .expected_arg = 0,
316
void tcg_gen_vec_sub32_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
119
+ },
317
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_muls(unsigned vece, uint32_t dofs, uint32_t aofs,
120
+ {
318
void tcg_gen_gvec_muli(unsigned vece, uint32_t dofs, uint32_t aofs,
121
+ .name = "exrl",
319
int64_t c, uint32_t oprsz, uint32_t maxsz)
122
+ .test_code = exrl_1,
320
{
123
+ .test_len = exrl_end - exrl_1,
321
- TCGv_i64 tmp = tcg_const_i64(c);
124
+ .page_ofs = exrl_1 - exrl_2,
322
+ TCGv_i64 tmp = tcg_constant_i64(c);
125
+ .entry_ofs = exrl_1 - exrl_2,
323
tcg_gen_gvec_muls(vece, dofs, aofs, tmp, oprsz, maxsz);
126
+ .expected_si_ofs = 0,
324
- tcg_temp_free_i64(tmp);
127
+ .expected_pc_ofs = exrl_1 - exrl_2,
325
}
128
+ .expected_arg = 0,
326
129
+ },
327
void tcg_gen_gvec_ssadd(unsigned vece, uint32_t dofs, uint32_t aofs,
130
+ {
328
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_sssub(unsigned vece, uint32_t dofs, uint32_t aofs,
131
+ .name = "fallthrough [cross]",
329
132
+ .test_code = noexec_1,
330
static void tcg_gen_usadd_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
133
+ .test_len = noexec_end - noexec_1,
331
{
134
+ .page_ofs = noexec_1 - noexec_2 - 2,
332
- TCGv_i32 max = tcg_const_i32(-1);
135
+ .entry_ofs = noexec_1 - noexec_2 - 2,
333
+ TCGv_i32 max = tcg_constant_i32(-1);
136
+ .expected_si_ofs = 0,
334
tcg_gen_add_i32(d, a, b);
137
+ .expected_pc_ofs = -2,
335
tcg_gen_movcond_i32(TCG_COND_LTU, d, d, a, max, d);
138
+ .expected_arg = 1,
336
- tcg_temp_free_i32(max);
139
+ },
337
}
140
+ {
338
141
+ .name = "jump [cross]",
339
static void tcg_gen_usadd_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
142
+ .test_code = noexec_1,
340
{
143
+ .test_len = noexec_end - noexec_1,
341
- TCGv_i64 max = tcg_const_i64(-1);
144
+ .page_ofs = noexec_1 - noexec_2 - 2,
342
+ TCGv_i64 max = tcg_constant_i64(-1);
145
+ .entry_ofs = -2,
343
tcg_gen_add_i64(d, a, b);
146
+ .expected_si_ofs = 0,
344
tcg_gen_movcond_i64(TCG_COND_LTU, d, d, a, max, d);
147
+ .expected_pc_ofs = -2,
345
- tcg_temp_free_i64(max);
148
+ .expected_arg = 0,
346
}
149
+ },
347
150
+ {
348
void tcg_gen_gvec_usadd(unsigned vece, uint32_t dofs, uint32_t aofs,
151
+ .name = "exrl [cross]",
349
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_usadd(unsigned vece, uint32_t dofs, uint32_t aofs,
152
+ .test_code = exrl_1,
350
153
+ .test_len = exrl_end - exrl_1,
351
static void tcg_gen_ussub_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
154
+ .page_ofs = exrl_1 - exrl_2 - 2,
352
{
155
+ .entry_ofs = exrl_1 - exrl_2 - 2,
353
- TCGv_i32 min = tcg_const_i32(0);
156
+ .expected_si_ofs = 0,
354
+ TCGv_i32 min = tcg_constant_i32(0);
157
+ .expected_pc_ofs = exrl_1 - exrl_2 - 2,
355
tcg_gen_sub_i32(d, a, b);
158
+ .expected_arg = 0,
356
tcg_gen_movcond_i32(TCG_COND_LTU, d, a, b, min, d);
159
+ },
357
- tcg_temp_free_i32(min);
160
+ };
358
}
161
+
359
162
+ return test_noexec(noexec_tests,
360
static void tcg_gen_ussub_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
163
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
361
{
164
+}
362
- TCGv_i64 min = tcg_const_i64(0);
165
diff --git a/tests/tcg/multiarch/noexec.c.inc b/tests/tcg/multiarch/noexec.c.inc
363
+ TCGv_i64 min = tcg_constant_i64(0);
166
new file mode 100644
364
tcg_gen_sub_i64(d, a, b);
167
index XXXXXXX..XXXXXXX
365
tcg_gen_movcond_i64(TCG_COND_LTU, d, a, b, min, d);
168
--- /dev/null
366
- tcg_temp_free_i64(min);
169
+++ b/tests/tcg/multiarch/noexec.c.inc
367
}
170
@@ -XXX,XX +XXX,XX @@
368
171
+/*
369
void tcg_gen_gvec_ussub(unsigned vece, uint32_t dofs, uint32_t aofs,
172
+ * Common code for arch-specific MMU_INST_FETCH fault testing.
370
@@ -XXX,XX +XXX,XX @@ static void gen_negv_mask(TCGv_i64 d, TCGv_i64 b, TCGv_i64 m)
173
+ */
371
174
+
372
void tcg_gen_vec_neg8_i64(TCGv_i64 d, TCGv_i64 b)
175
+#define _GNU_SOURCE
373
{
176
+
374
- TCGv_i64 m = tcg_const_i64(dup_const(MO_8, 0x80));
177
+#include <assert.h>
375
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_8, 0x80));
178
+#include <signal.h>
376
gen_negv_mask(d, b, m);
179
+#include <stdio.h>
377
- tcg_temp_free_i64(m);
180
+#include <stdlib.h>
378
}
181
+#include <string.h>
379
182
+#include <errno.h>
380
void tcg_gen_vec_neg16_i64(TCGv_i64 d, TCGv_i64 b)
183
+#include <unistd.h>
381
{
184
+#include <sys/mman.h>
382
- TCGv_i64 m = tcg_const_i64(dup_const(MO_16, 0x8000));
185
+#include <sys/ucontext.h>
383
+ TCGv_i64 m = tcg_constant_i64(dup_const(MO_16, 0x8000));
186
+
384
gen_negv_mask(d, b, m);
187
+/* Forward declarations. */
385
- tcg_temp_free_i64(m);
188
+
386
}
189
+static void *arch_mcontext_pc(const mcontext_t *ctx);
387
190
+static int arch_mcontext_arg(const mcontext_t *ctx);
388
void tcg_gen_vec_neg32_i64(TCGv_i64 d, TCGv_i64 b)
191
+static void arch_flush(void *p, int len);
389
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_ands(unsigned vece, uint32_t dofs, uint32_t aofs,
192
+
390
void tcg_gen_gvec_andi(unsigned vece, uint32_t dofs, uint32_t aofs,
193
+/* Testing infrastructure. */
391
int64_t c, uint32_t oprsz, uint32_t maxsz)
194
+
392
{
195
+struct noexec_test {
393
- TCGv_i64 tmp = tcg_const_i64(dup_const(vece, c));
196
+ const char *name;
394
+ TCGv_i64 tmp = tcg_constant_i64(dup_const(vece, c));
197
+ const char *test_code;
395
tcg_gen_gvec_2s(dofs, aofs, oprsz, maxsz, tmp, &gop_ands);
198
+ int test_len;
396
- tcg_temp_free_i64(tmp);
199
+ int page_ofs;
397
}
200
+ int entry_ofs;
398
201
+ int expected_si_ofs;
399
static const GVecGen2s gop_xors = {
202
+ int expected_pc_ofs;
400
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_xors(unsigned vece, uint32_t dofs, uint32_t aofs,
203
+ int expected_arg;
401
void tcg_gen_gvec_xori(unsigned vece, uint32_t dofs, uint32_t aofs,
204
+};
402
int64_t c, uint32_t oprsz, uint32_t maxsz)
205
+
403
{
206
+static void *page_base;
404
- TCGv_i64 tmp = tcg_const_i64(dup_const(vece, c));
207
+static int page_size;
405
+ TCGv_i64 tmp = tcg_constant_i64(dup_const(vece, c));
208
+static const struct noexec_test *current_noexec_test;
406
tcg_gen_gvec_2s(dofs, aofs, oprsz, maxsz, tmp, &gop_xors);
209
+
407
- tcg_temp_free_i64(tmp);
210
+static void handle_err(const char *syscall)
408
}
211
+{
409
212
+ printf("[ FAILED ] %s: %s\n", syscall, strerror(errno));
410
static const GVecGen2s gop_ors = {
213
+ exit(EXIT_FAILURE);
411
@@ -XXX,XX +XXX,XX @@ void tcg_gen_gvec_ors(unsigned vece, uint32_t dofs, uint32_t aofs,
214
+}
412
void tcg_gen_gvec_ori(unsigned vece, uint32_t dofs, uint32_t aofs,
215
+
413
int64_t c, uint32_t oprsz, uint32_t maxsz)
216
+static void handle_segv(int sig, siginfo_t *info, void *ucontext)
414
{
217
+{
415
- TCGv_i64 tmp = tcg_const_i64(dup_const(vece, c));
218
+ const struct noexec_test *test = current_noexec_test;
416
+ TCGv_i64 tmp = tcg_constant_i64(dup_const(vece, c));
219
+ const mcontext_t *mc = &((ucontext_t *)ucontext)->uc_mcontext;
417
tcg_gen_gvec_2s(dofs, aofs, oprsz, maxsz, tmp, &gop_ors);
220
+ void *expected_si;
418
- tcg_temp_free_i64(tmp);
221
+ void *expected_pc;
419
}
222
+ void *pc;
420
223
+ int arg;
421
void tcg_gen_vec_shl8i_i64(TCGv_i64 d, TCGv_i64 a, int64_t c)
224
+
422
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_shlv_mod_vec(unsigned vece, TCGv_vec d,
225
+ if (test == NULL) {
423
TCGv_vec a, TCGv_vec b)
226
+ printf("[ FAILED ] unexpected SEGV\n");
424
{
227
+ exit(EXIT_FAILURE);
425
TCGv_vec t = tcg_temp_new_vec_matching(d);
228
+ }
426
+ TCGv_vec m = tcg_constant_vec_matching(d, vece, (8 << vece) - 1);
229
+ current_noexec_test = NULL;
427
230
+
428
- tcg_gen_dupi_vec(vece, t, (8 << vece) - 1);
231
+ expected_si = page_base + test->expected_si_ofs;
429
- tcg_gen_and_vec(vece, t, t, b);
232
+ if (info->si_addr != expected_si) {
430
+ tcg_gen_and_vec(vece, t, b, m);
233
+ printf("[ FAILED ] wrong si_addr (%p != %p)\n",
431
tcg_gen_shlv_vec(vece, d, a, t);
234
+ info->si_addr, expected_si);
432
tcg_temp_free_vec(t);
235
+ exit(EXIT_FAILURE);
433
}
236
+ }
434
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_shrv_mod_vec(unsigned vece, TCGv_vec d,
237
+
435
TCGv_vec a, TCGv_vec b)
238
+ pc = arch_mcontext_pc(mc);
436
{
239
+ expected_pc = page_base + test->expected_pc_ofs;
437
TCGv_vec t = tcg_temp_new_vec_matching(d);
240
+ if (pc != expected_pc) {
438
+ TCGv_vec m = tcg_constant_vec_matching(d, vece, (8 << vece) - 1);
241
+ printf("[ FAILED ] wrong pc (%p != %p)\n", pc, expected_pc);
439
242
+ exit(EXIT_FAILURE);
440
- tcg_gen_dupi_vec(vece, t, (8 << vece) - 1);
243
+ }
441
- tcg_gen_and_vec(vece, t, t, b);
244
+
442
+ tcg_gen_and_vec(vece, t, b, m);
245
+ arg = arch_mcontext_arg(mc);
443
tcg_gen_shrv_vec(vece, d, a, t);
246
+ if (arg != test->expected_arg) {
444
tcg_temp_free_vec(t);
247
+ printf("[ FAILED ] wrong arg (%d != %d)\n", arg, test->expected_arg);
445
}
248
+ exit(EXIT_FAILURE);
446
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_sarv_mod_vec(unsigned vece, TCGv_vec d,
249
+ }
447
TCGv_vec a, TCGv_vec b)
250
+
448
{
251
+ if (mprotect(page_base, page_size,
449
TCGv_vec t = tcg_temp_new_vec_matching(d);
252
+ PROT_READ | PROT_WRITE | PROT_EXEC) < 0) {
450
+ TCGv_vec m = tcg_constant_vec_matching(d, vece, (8 << vece) - 1);
253
+ handle_err("mprotect");
451
254
+ }
452
- tcg_gen_dupi_vec(vece, t, (8 << vece) - 1);
255
+}
453
- tcg_gen_and_vec(vece, t, t, b);
256
+
454
+ tcg_gen_and_vec(vece, t, b, m);
257
+static void test_noexec_1(const struct noexec_test *test)
455
tcg_gen_sarv_vec(vece, d, a, t);
258
+{
456
tcg_temp_free_vec(t);
259
+ void *start = page_base + test->page_ofs;
457
}
260
+ void (*fn)(int arg) = page_base + test->entry_ofs;
458
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_rotlv_mod_vec(unsigned vece, TCGv_vec d,
261
+
459
TCGv_vec a, TCGv_vec b)
262
+ memcpy(start, test->test_code, test->test_len);
460
{
263
+ arch_flush(start, test->test_len);
461
TCGv_vec t = tcg_temp_new_vec_matching(d);
264
+
462
+ TCGv_vec m = tcg_constant_vec_matching(d, vece, (8 << vece) - 1);
265
+ /* Trigger TB creation in order to test invalidation. */
463
266
+ fn(0);
464
- tcg_gen_dupi_vec(vece, t, (8 << vece) - 1);
267
+
465
- tcg_gen_and_vec(vece, t, t, b);
268
+ if (mprotect(page_base, page_size, PROT_NONE) < 0) {
466
+ tcg_gen_and_vec(vece, t, b, m);
269
+ handle_err("mprotect");
467
tcg_gen_rotlv_vec(vece, d, a, t);
270
+ }
468
tcg_temp_free_vec(t);
271
+
469
}
272
+ /* Trigger SEGV and check that handle_segv() ran. */
470
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_rotrv_mod_vec(unsigned vece, TCGv_vec d,
273
+ current_noexec_test = test;
471
TCGv_vec a, TCGv_vec b)
274
+ fn(0);
472
{
275
+ assert(current_noexec_test == NULL);
473
TCGv_vec t = tcg_temp_new_vec_matching(d);
276
+}
474
+ TCGv_vec m = tcg_constant_vec_matching(d, vece, (8 << vece) - 1);
277
+
475
278
+static int test_noexec(struct noexec_test *tests, size_t n_tests)
476
- tcg_gen_dupi_vec(vece, t, (8 << vece) - 1);
279
+{
477
- tcg_gen_and_vec(vece, t, t, b);
280
+ struct sigaction act;
478
+ tcg_gen_and_vec(vece, t, b, m);
281
+ size_t i;
479
tcg_gen_rotrv_vec(vece, d, a, t);
282
+
480
tcg_temp_free_vec(t);
283
+ memset(&act, 0, sizeof(act));
481
}
284
+ act.sa_sigaction = handle_segv;
482
diff --git a/tcg/tcg.c b/tcg/tcg.c
285
+ act.sa_flags = SA_SIGINFO;
286
+ if (sigaction(SIGSEGV, &act, NULL) < 0) {
287
+ handle_err("sigaction");
288
+ }
289
+
290
+ page_size = getpagesize();
291
+ page_base = mmap(NULL, 2 * page_size,
292
+ PROT_READ | PROT_WRITE | PROT_EXEC,
293
+ MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
294
+ if (page_base == MAP_FAILED) {
295
+ handle_err("mmap");
296
+ }
297
+ page_base += page_size;
298
+
299
+ for (i = 0; i < n_tests; i++) {
300
+ struct noexec_test *test = &tests[i];
301
+
302
+ printf("[ RUN ] %s\n", test->name);
303
+ test_noexec_1(test);
304
+ printf("[ OK ]\n");
305
+ }
306
+
307
+ printf("[ PASSED ]\n");
308
+ return EXIT_SUCCESS;
309
+}
310
diff --git a/tests/tcg/s390x/Makefile.target b/tests/tcg/s390x/Makefile.target
483
index XXXXXXX..XXXXXXX 100644
311
index XXXXXXX..XXXXXXX 100644
484
--- a/tcg/tcg.c
312
--- a/tests/tcg/s390x/Makefile.target
485
+++ b/tcg/tcg.c
313
+++ b/tests/tcg/s390x/Makefile.target
486
@@ -XXX,XX +XXX,XX @@ TCGv_vec tcg_constant_vec(TCGType type, unsigned vece, int64_t val)
314
@@ -XXX,XX +XXX,XX @@ TESTS+=shift
487
return temp_tcgv_vec(tcg_constant_internal(type, val));
315
TESTS+=trap
488
}
316
TESTS+=signals-s390x
489
317
TESTS+=branch-relative-long
490
+TCGv_vec tcg_constant_vec_matching(TCGv_vec match, unsigned vece, int64_t val)
318
+TESTS+=noexec
491
+{
319
492
+ TCGTemp *t = tcgv_vec_temp(match);
320
Z14_TESTS=vfminmax
493
+
321
vfminmax: LDFLAGS+=-lm
494
+ tcg_debug_assert(t->temp_allocated != 0);
495
+ return tcg_constant_vec(t->base_type, vece, val);
496
+}
497
+
498
TCGv_i32 tcg_const_i32(int32_t val)
499
{
500
TCGv_i32 t0;
501
--
322
--
502
2.25.1
323
2.34.1
503
504
diff view generated by jsdifflib
1
These will hold a single constant for the duration of the TB.
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
They are hashed, so that each value has one temp across the TB.
2
3
3
Right now translator stops right *after* the end of a page, which
4
Not used yet, this is all infrastructure.
4
breaks reporting of fault locations when the last instruction of a
5
5
multi-insn translation block crosses a page boundary.
6
7
An implementation, like the one arm and s390x have, would require an
8
i386 length disassembler, which is burdensome to maintain. Another
9
alternative would be to single-step at the end of a guest page, but
10
this may come with a performance impact.
11
12
Fix by snapshotting disassembly state and restoring it after we figure
13
out we crossed a page boundary. This includes rolling back cc_op
14
updates and emitted ops.
15
16
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1143
19
Message-Id: <20220817150506.592862-4-iii@linux.ibm.com>
20
[rth: Simplify end-of-insn cross-page checks.]
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
21
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
22
---
8
include/tcg/tcg.h | 24 ++++-
23
target/i386/tcg/translate.c | 64 ++++++++++++++++-----------
9
tcg/optimize.c | 13 ++-
24
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++++++++++++++
10
tcg/tcg.c | 224 ++++++++++++++++++++++++++++++++++++----------
25
tests/tcg/x86_64/Makefile.target | 3 +-
11
3 files changed, 211 insertions(+), 50 deletions(-)
26
3 files changed, 116 insertions(+), 26 deletions(-)
12
27
create mode 100644 tests/tcg/x86_64/noexec.c
13
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
28
29
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
14
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
15
--- a/include/tcg/tcg.h
31
--- a/target/i386/tcg/translate.c
16
+++ b/include/tcg/tcg.h
32
+++ b/target/i386/tcg/translate.c
17
@@ -XXX,XX +XXX,XX @@ typedef enum TCGTempKind {
33
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
18
TEMP_GLOBAL,
34
TCGv_i64 tmp1_i64;
19
/* Temp is in a fixed register. */
35
20
TEMP_FIXED,
36
sigjmp_buf jmpbuf;
21
+ /* Temp is a fixed constant. */
37
+ TCGOp *prev_insn_end;
22
+ TEMP_CONST,
38
} DisasContext;
23
} TCGTempKind;
39
24
40
/* The environment in which user-only runs is constrained. */
25
typedef struct TCGTemp {
41
@@ -XXX,XX +XXX,XX @@ static uint64_t advance_pc(CPUX86State *env, DisasContext *s, int num_bytes)
26
@@ -XXX,XX +XXX,XX @@ struct TCGContext {
27
QSIMPLEQ_HEAD(, TCGOp) plugin_ops;
28
#endif
29
30
+ GHashTable *const_table[TCG_TYPE_COUNT];
31
TCGTempSet free_temps[TCG_TYPE_COUNT * 2];
32
TCGTemp temps[TCG_MAX_TEMPS]; /* globals first, temps after */
33
34
@@ -XXX,XX +XXX,XX @@ struct TCGContext {
35
36
static inline bool temp_readonly(TCGTemp *ts)
37
{
42
{
38
- return ts->kind == TEMP_FIXED;
43
uint64_t pc = s->pc;
39
+ return ts->kind >= TEMP_FIXED;
44
40
}
45
+ /* This is a subsequent insn that crosses a page boundary. */
41
46
+ if (s->base.num_insns > 1 &&
42
extern TCGContext tcg_init_ctx;
47
+ !is_same_page(&s->base, s->pc + num_bytes - 1)) {
43
@@ -XXX,XX +XXX,XX @@ TCGOp *tcg_op_insert_after(TCGContext *s, TCGOp *op, TCGOpcode opc);
48
+ siglongjmp(s->jmpbuf, 2);
44
45
void tcg_optimize(TCGContext *s);
46
47
+/* Allocate a new temporary and initialize it with a constant. */
48
TCGv_i32 tcg_const_i32(int32_t val);
49
TCGv_i64 tcg_const_i64(int64_t val);
50
TCGv_i32 tcg_const_local_i32(int32_t val);
51
@@ -XXX,XX +XXX,XX @@ TCGv_vec tcg_const_ones_vec(TCGType);
52
TCGv_vec tcg_const_zeros_vec_matching(TCGv_vec);
53
TCGv_vec tcg_const_ones_vec_matching(TCGv_vec);
54
55
+/*
56
+ * Locate or create a read-only temporary that is a constant.
57
+ * This kind of temporary need not and should not be freed.
58
+ */
59
+TCGTemp *tcg_constant_internal(TCGType type, int64_t val);
60
+
61
+static inline TCGv_i32 tcg_constant_i32(int32_t val)
62
+{
63
+ return temp_tcgv_i32(tcg_constant_internal(TCG_TYPE_I32, val));
64
+}
65
+
66
+static inline TCGv_i64 tcg_constant_i64(int64_t val)
67
+{
68
+ return temp_tcgv_i64(tcg_constant_internal(TCG_TYPE_I64, val));
69
+}
70
+
71
+TCGv_vec tcg_constant_vec(TCGType type, unsigned vece, int64_t val);
72
+
73
#if UINTPTR_MAX == UINT32_MAX
74
# define tcg_const_ptr(x) ((TCGv_ptr)tcg_const_i32((intptr_t)(x)))
75
# define tcg_const_local_ptr(x) ((TCGv_ptr)tcg_const_local_i32((intptr_t)(x)))
76
diff --git a/tcg/optimize.c b/tcg/optimize.c
77
index XXXXXXX..XXXXXXX 100644
78
--- a/tcg/optimize.c
79
+++ b/tcg/optimize.c
80
@@ -XXX,XX +XXX,XX @@ static void init_ts_info(TempOptInfo *infos,
81
ts->state_ptr = ti;
82
ti->next_copy = ts;
83
ti->prev_copy = ts;
84
- ti->is_const = false;
85
- ti->mask = -1;
86
+ if (ts->kind == TEMP_CONST) {
87
+ ti->is_const = true;
88
+ ti->val = ti->mask = ts->val;
89
+ if (TCG_TARGET_REG_BITS > 32 && ts->type == TCG_TYPE_I32) {
90
+ /* High bits of a 32-bit quantity are garbage. */
91
+ ti->mask |= ~0xffffffffull;
92
+ }
93
+ } else {
94
+ ti->is_const = false;
95
+ ti->mask = -1;
96
+ }
97
set_bit(idx, temps_used->l);
98
}
99
}
100
diff --git a/tcg/tcg.c b/tcg/tcg.c
101
index XXXXXXX..XXXXXXX 100644
102
--- a/tcg/tcg.c
103
+++ b/tcg/tcg.c
104
@@ -XXX,XX +XXX,XX @@ void tcg_func_start(TCGContext *s)
105
/* No temps have been previously allocated for size or locality. */
106
memset(s->free_temps, 0, sizeof(s->free_temps));
107
108
+ /* No constant temps have been previously allocated. */
109
+ for (int i = 0; i < TCG_TYPE_COUNT; ++i) {
110
+ if (s->const_table[i]) {
111
+ g_hash_table_remove_all(s->const_table[i]);
112
+ }
113
+ }
49
+ }
114
+
50
+
115
s->nb_ops = 0;
51
s->pc += num_bytes;
116
s->nb_labels = 0;
52
if (unlikely(s->pc - s->pc_start > X86_MAX_INSN_LENGTH)) {
117
s->current_frame_offset = s->frame_start;
53
/* If the instruction's 16th byte is on a different page than the 1st, a
118
@@ -XXX,XX +XXX,XX @@ TCGTemp *tcg_global_mem_new_internal(TCGType type, TCGv_ptr base,
54
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
119
bigendian = 1;
55
int modrm, reg, rm, mod, op, opreg, val;
120
#endif
56
target_ulong next_eip, tval;
121
57
target_ulong pc_start = s->base.pc_next;
122
- if (base_ts->kind != TEMP_FIXED) {
58
+ bool orig_cc_op_dirty = s->cc_op_dirty;
123
+ switch (base_ts->kind) {
59
+ CCOp orig_cc_op = s->cc_op;
124
+ case TEMP_FIXED:
60
61
s->pc_start = s->pc = pc_start;
62
s->override = -1;
63
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
64
s->rip_offset = 0; /* for relative ip address */
65
s->vex_l = 0;
66
s->vex_v = 0;
67
- if (sigsetjmp(s->jmpbuf, 0) != 0) {
68
+ switch (sigsetjmp(s->jmpbuf, 0)) {
69
+ case 0:
125
+ break;
70
+ break;
126
+ case TEMP_GLOBAL:
71
+ case 1:
127
/* We do not support double-indirect registers. */
72
gen_exception_gpf(s);
128
tcg_debug_assert(!base_ts->indirect_reg);
73
return s->pc;
129
base_ts->indirect_base = 1;
74
+ case 2:
130
s->nb_indirects += (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64
75
+ /* Restore state that may affect the next instruction. */
131
? 2 : 1);
76
+ s->cc_op_dirty = orig_cc_op_dirty;
132
indirect_reg = 1;
77
+ s->cc_op = orig_cc_op;
133
+ break;
78
+ s->base.num_insns--;
79
+ tcg_remove_ops_after(s->prev_insn_end);
80
+ s->base.is_jmp = DISAS_TOO_MANY;
81
+ return pc_start;
134
+ default:
82
+ default:
135
+ g_assert_not_reached();
83
+ g_assert_not_reached();
136
}
84
}
137
85
138
if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
86
prefixes = 0;
139
@@ -XXX,XX +XXX,XX @@ void tcg_temp_free_internal(TCGTemp *ts)
87
@@ -XXX,XX +XXX,XX @@ static void i386_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
140
TCGContext *s = tcg_ctx;
88
{
141
int k, idx;
89
DisasContext *dc = container_of(dcbase, DisasContext, base);
142
90
143
+ /* In order to simplify users of tcg_constant_*, silently ignore free. */
91
+ dc->prev_insn_end = tcg_last_op();
144
+ if (ts->kind == TEMP_CONST) {
92
tcg_gen_insn_start(dc->base.pc_next, dc->cc_op);
145
+ return;
93
}
94
95
@@ -XXX,XX +XXX,XX @@ static void i386_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
96
#endif
97
98
pc_next = disas_insn(dc, cpu);
99
-
100
- if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
101
- /* if single step mode, we generate only one instruction and
102
- generate an exception */
103
- /* if irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
104
- the flag and abort the translation to give the irqs a
105
- chance to happen */
106
- dc->base.is_jmp = DISAS_TOO_MANY;
107
- } else if ((tb_cflags(dc->base.tb) & CF_USE_ICOUNT)
108
- && ((pc_next & TARGET_PAGE_MASK)
109
- != ((pc_next + TARGET_MAX_INSN_SIZE - 1)
110
- & TARGET_PAGE_MASK)
111
- || (pc_next & ~TARGET_PAGE_MASK) == 0)) {
112
- /* Do not cross the boundary of the pages in icount mode,
113
- it can cause an exception. Do it only when boundary is
114
- crossed by the first instruction in the block.
115
- If current instruction already crossed the bound - it's ok,
116
- because an exception hasn't stopped this code.
117
- */
118
- dc->base.is_jmp = DISAS_TOO_MANY;
119
- } else if ((pc_next - dc->base.pc_first) >= (TARGET_PAGE_SIZE - 32)) {
120
- dc->base.is_jmp = DISAS_TOO_MANY;
121
- }
122
-
123
dc->base.pc_next = pc_next;
124
+
125
+ if (dc->base.is_jmp == DISAS_NEXT) {
126
+ if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
127
+ /*
128
+ * If single step mode, we generate only one instruction and
129
+ * generate an exception.
130
+ * If irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
131
+ * the flag and abort the translation to give the irqs a
132
+ * chance to happen.
133
+ */
134
+ dc->base.is_jmp = DISAS_TOO_MANY;
135
+ } else if (!is_same_page(&dc->base, pc_next)) {
136
+ dc->base.is_jmp = DISAS_TOO_MANY;
137
+ }
146
+ }
138
+ }
147
+
148
#if defined(CONFIG_DEBUG_TCG)
149
s->temps_in_use--;
150
if (s->temps_in_use < 0) {
151
@@ -XXX,XX +XXX,XX @@ void tcg_temp_free_internal(TCGTemp *ts)
152
set_bit(idx, s->free_temps[k].l);
153
}
139
}
154
140
155
+TCGTemp *tcg_constant_internal(TCGType type, int64_t val)
141
static void i386_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
156
+{
142
diff --git a/tests/tcg/x86_64/noexec.c b/tests/tcg/x86_64/noexec.c
157
+ TCGContext *s = tcg_ctx;
143
new file mode 100644
158
+ GHashTable *h = s->const_table[type];
144
index XXXXXXX..XXXXXXX
159
+ TCGTemp *ts;
145
--- /dev/null
160
+
146
+++ b/tests/tcg/x86_64/noexec.c
161
+ if (h == NULL) {
147
@@ -XXX,XX +XXX,XX @@
162
+ h = g_hash_table_new(g_int64_hash, g_int64_equal);
148
+#include "../multiarch/noexec.c.inc"
163
+ s->const_table[type] = h;
149
+
164
+ }
150
+static void *arch_mcontext_pc(const mcontext_t *ctx)
165
+
151
+{
166
+ ts = g_hash_table_lookup(h, &val);
152
+ return (void *)ctx->gregs[REG_RIP];
167
+ if (ts == NULL) {
153
+}
168
+ ts = tcg_temp_alloc(s);
154
+
169
+
155
+int arch_mcontext_arg(const mcontext_t *ctx)
170
+ if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
156
+{
171
+ TCGTemp *ts2 = tcg_temp_alloc(s);
157
+ return ctx->gregs[REG_RDI];
172
+
158
+}
173
+ ts->base_type = TCG_TYPE_I64;
159
+
174
+ ts->type = TCG_TYPE_I32;
160
+static void arch_flush(void *p, int len)
175
+ ts->kind = TEMP_CONST;
161
+{
176
+ ts->temp_allocated = 1;
162
+}
177
+ /*
163
+
178
+ * Retain the full value of the 64-bit constant in the low
164
+extern char noexec_1[];
179
+ * part, so that the hash table works. Actual uses will
165
+extern char noexec_2[];
180
+ * truncate the value to the low part.
166
+extern char noexec_end[];
181
+ */
167
+
182
+ ts->val = val;
168
+asm("noexec_1:\n"
183
+
169
+ " movq $1,%rdi\n" /* %rdi is 0 on entry, set 1. */
184
+ tcg_debug_assert(ts2 == ts + 1);
170
+ "noexec_2:\n"
185
+ ts2->base_type = TCG_TYPE_I64;
171
+ " movq $2,%rdi\n" /* %rdi is 0/1; set 2. */
186
+ ts2->type = TCG_TYPE_I32;
172
+ " ret\n"
187
+ ts2->kind = TEMP_CONST;
173
+ "noexec_end:");
188
+ ts2->temp_allocated = 1;
174
+
189
+ ts2->val = val >> 32;
175
+int main(void)
190
+ } else {
176
+{
191
+ ts->base_type = type;
177
+ struct noexec_test noexec_tests[] = {
192
+ ts->type = type;
178
+ {
193
+ ts->kind = TEMP_CONST;
179
+ .name = "fallthrough",
194
+ ts->temp_allocated = 1;
180
+ .test_code = noexec_1,
195
+ ts->val = val;
181
+ .test_len = noexec_end - noexec_1,
196
+ }
182
+ .page_ofs = noexec_1 - noexec_2,
197
+ g_hash_table_insert(h, &ts->val, ts);
183
+ .entry_ofs = noexec_1 - noexec_2,
198
+ }
184
+ .expected_si_ofs = 0,
199
+
185
+ .expected_pc_ofs = 0,
200
+ return ts;
186
+ .expected_arg = 1,
201
+}
187
+ },
202
+
188
+ {
203
+TCGv_vec tcg_constant_vec(TCGType type, unsigned vece, int64_t val)
189
+ .name = "jump",
204
+{
190
+ .test_code = noexec_1,
205
+ val = dup_const(vece, val);
191
+ .test_len = noexec_end - noexec_1,
206
+ return temp_tcgv_vec(tcg_constant_internal(type, val));
192
+ .page_ofs = noexec_1 - noexec_2,
207
+}
193
+ .entry_ofs = 0,
208
+
194
+ .expected_si_ofs = 0,
209
TCGv_i32 tcg_const_i32(int32_t val)
195
+ .expected_pc_ofs = 0,
210
{
196
+ .expected_arg = 0,
211
TCGv_i32 t0;
197
+ },
212
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_start(TCGContext *s)
198
+ {
213
TCGTempVal val = TEMP_VAL_MEM;
199
+ .name = "fallthrough [cross]",
214
200
+ .test_code = noexec_1,
215
switch (ts->kind) {
201
+ .test_len = noexec_end - noexec_1,
216
+ case TEMP_CONST:
202
+ .page_ofs = noexec_1 - noexec_2 - 2,
217
+ val = TEMP_VAL_CONST;
203
+ .entry_ofs = noexec_1 - noexec_2 - 2,
218
+ break;
204
+ .expected_si_ofs = 0,
219
case TEMP_FIXED:
205
+ .expected_pc_ofs = -2,
220
val = TEMP_VAL_REG;
206
+ .expected_arg = 1,
221
break;
207
+ },
222
@@ -XXX,XX +XXX,XX @@ static char *tcg_get_arg_str_ptr(TCGContext *s, char *buf, int buf_size,
208
+ {
223
case TEMP_NORMAL:
209
+ .name = "jump [cross]",
224
snprintf(buf, buf_size, "tmp%d", idx - s->nb_globals);
210
+ .test_code = noexec_1,
225
break;
211
+ .test_len = noexec_end - noexec_1,
226
+ case TEMP_CONST:
212
+ .page_ofs = noexec_1 - noexec_2 - 2,
227
+ switch (ts->type) {
213
+ .entry_ofs = -2,
228
+ case TCG_TYPE_I32:
214
+ .expected_si_ofs = 0,
229
+ snprintf(buf, buf_size, "$0x%x", (int32_t)ts->val);
215
+ .expected_pc_ofs = -2,
230
+ break;
216
+ .expected_arg = 0,
231
+#if TCG_TARGET_REG_BITS > 32
217
+ },
232
+ case TCG_TYPE_I64:
218
+ };
233
+ snprintf(buf, buf_size, "$0x%" PRIx64, ts->val);
219
+
234
+ break;
220
+ return test_noexec(noexec_tests,
235
+#endif
221
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
236
+ case TCG_TYPE_V64:
222
+}
237
+ case TCG_TYPE_V128:
223
diff --git a/tests/tcg/x86_64/Makefile.target b/tests/tcg/x86_64/Makefile.target
238
+ case TCG_TYPE_V256:
224
index XXXXXXX..XXXXXXX 100644
239
+ snprintf(buf, buf_size, "v%d$0x%" PRIx64,
225
--- a/tests/tcg/x86_64/Makefile.target
240
+ 64 << (ts->type - TCG_TYPE_V64), ts->val);
226
+++ b/tests/tcg/x86_64/Makefile.target
241
+ break;
227
@@ -XXX,XX +XXX,XX @@ include $(SRC_PATH)/tests/tcg/i386/Makefile.target
242
+ default:
228
243
+ g_assert_not_reached();
229
ifeq ($(filter %-linux-user, $(TARGET)),$(TARGET))
244
+ }
230
X86_64_TESTS += vsyscall
245
+ break;
231
+X86_64_TESTS += noexec
246
}
232
TESTS=$(MULTIARCH_TESTS) $(X86_64_TESTS) test-x86_64
247
return buf;
233
else
248
}
234
TESTS=$(MULTIARCH_TESTS)
249
@@ -XXX,XX +XXX,XX @@ static void la_bb_end(TCGContext *s, int ng, int nt)
235
@@ -XXX,XX +XXX,XX @@ test-x86_64: LDFLAGS+=-lm -lc
250
state = TS_DEAD | TS_MEM;
236
test-x86_64: test-i386.c test-i386.h test-i386-shift.h test-i386-muldiv.h
251
break;
237
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
252
case TEMP_NORMAL:
238
253
+ case TEMP_CONST:
239
-vsyscall: $(SRC_PATH)/tests/tcg/x86_64/vsyscall.c
254
state = TS_DEAD;
240
+%: $(SRC_PATH)/tests/tcg/x86_64/%.c
255
break;
241
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
256
default:
257
@@ -XXX,XX +XXX,XX @@ static void la_bb_sync(TCGContext *s, int ng, int nt)
258
la_global_sync(s, ng);
259
260
for (int i = ng; i < nt; ++i) {
261
- if (s->temps[i].kind == TEMP_LOCAL) {
262
- int state = s->temps[i].state;
263
- s->temps[i].state = state | TS_MEM;
264
+ TCGTemp *ts = &s->temps[i];
265
+ int state;
266
+
267
+ switch (ts->kind) {
268
+ case TEMP_LOCAL:
269
+ state = ts->state;
270
+ ts->state = state | TS_MEM;
271
if (state != TS_DEAD) {
272
continue;
273
}
274
- } else {
275
+ break;
276
+ case TEMP_NORMAL:
277
s->temps[i].state = TS_DEAD;
278
+ break;
279
+ case TEMP_CONST:
280
+ continue;
281
+ default:
282
+ g_assert_not_reached();
283
}
284
la_reset_pref(&s->temps[i]);
285
}
286
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *, TCGTemp *, TCGRegSet, TCGRegSet, TCGRegSet);
287
mark it free; otherwise mark it dead. */
288
static void temp_free_or_dead(TCGContext *s, TCGTemp *ts, int free_or_dead)
289
{
290
- if (temp_readonly(ts)) {
291
+ TCGTempVal new_type;
292
+
293
+ switch (ts->kind) {
294
+ case TEMP_FIXED:
295
return;
296
+ case TEMP_GLOBAL:
297
+ case TEMP_LOCAL:
298
+ new_type = TEMP_VAL_MEM;
299
+ break;
300
+ case TEMP_NORMAL:
301
+ new_type = free_or_dead < 0 ? TEMP_VAL_MEM : TEMP_VAL_DEAD;
302
+ break;
303
+ case TEMP_CONST:
304
+ new_type = TEMP_VAL_CONST;
305
+ break;
306
+ default:
307
+ g_assert_not_reached();
308
}
309
if (ts->val_type == TEMP_VAL_REG) {
310
s->reg_to_temp[ts->reg] = NULL;
311
}
312
- ts->val_type = (free_or_dead < 0
313
- || ts->kind != TEMP_NORMAL
314
- ? TEMP_VAL_MEM : TEMP_VAL_DEAD);
315
+ ts->val_type = new_type;
316
}
317
318
/* Mark a temporary as dead. */
319
@@ -XXX,XX +XXX,XX @@ static inline void temp_dead(TCGContext *s, TCGTemp *ts)
320
static void temp_sync(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs,
321
TCGRegSet preferred_regs, int free_or_dead)
322
{
323
- if (temp_readonly(ts)) {
324
- return;
325
- }
326
- if (!ts->mem_coherent) {
327
+ if (!temp_readonly(ts) && !ts->mem_coherent) {
328
if (!ts->mem_allocated) {
329
temp_allocate_frame(s, ts);
330
}
331
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_bb_end(TCGContext *s, TCGRegSet allocated_regs)
332
333
for (i = s->nb_globals; i < s->nb_temps; i++) {
334
TCGTemp *ts = &s->temps[i];
335
- if (ts->kind == TEMP_LOCAL) {
336
+
337
+ switch (ts->kind) {
338
+ case TEMP_LOCAL:
339
temp_save(s, ts, allocated_regs);
340
- } else {
341
+ break;
342
+ case TEMP_NORMAL:
343
/* The liveness analysis already ensures that temps are dead.
344
Keep an tcg_debug_assert for safety. */
345
tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
346
+ break;
347
+ case TEMP_CONST:
348
+ /* Similarly, we should have freed any allocated register. */
349
+ tcg_debug_assert(ts->val_type == TEMP_VAL_CONST);
350
+ break;
351
+ default:
352
+ g_assert_not_reached();
353
}
354
}
355
356
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_cbranch(TCGContext *s, TCGRegSet allocated_regs)
357
* The liveness analysis already ensures that temps are dead.
358
* Keep tcg_debug_asserts for safety.
359
*/
360
- if (ts->kind == TEMP_LOCAL) {
361
+ switch (ts->kind) {
362
+ case TEMP_LOCAL:
363
tcg_debug_assert(ts->val_type != TEMP_VAL_REG || ts->mem_coherent);
364
- } else {
365
+ break;
366
+ case TEMP_NORMAL:
367
tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
368
+ break;
369
+ case TEMP_CONST:
370
+ break;
371
+ default:
372
+ g_assert_not_reached();
373
}
374
}
375
}
376
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
377
i_preferred_regs = o_preferred_regs = 0;
378
if (arg_ct->ialias) {
379
o_preferred_regs = op->output_pref[arg_ct->alias_index];
380
- if (ts->kind == TEMP_FIXED) {
381
- /* if fixed register, we must allocate a new register
382
- if the alias is not the same register */
383
- if (arg != op->args[arg_ct->alias_index]) {
384
- goto allocate_in_reg;
385
- }
386
- } else {
387
- /* if the input is aliased to an output and if it is
388
- not dead after the instruction, we must allocate
389
- a new register and move it */
390
- if (!IS_DEAD_ARG(i)) {
391
- goto allocate_in_reg;
392
- }
393
394
- /* check if the current register has already been allocated
395
- for another input aliased to an output */
396
- if (ts->val_type == TEMP_VAL_REG) {
397
- int k2, i2;
398
- reg = ts->reg;
399
- for (k2 = 0 ; k2 < k ; k2++) {
400
- i2 = def->args_ct[nb_oargs + k2].sort_index;
401
- if (def->args_ct[i2].ialias && reg == new_args[i2]) {
402
- goto allocate_in_reg;
403
- }
404
+ /*
405
+ * If the input is readonly, then it cannot also be an
406
+ * output and aliased to itself. If the input is not
407
+ * dead after the instruction, we must allocate a new
408
+ * register and move it.
409
+ */
410
+ if (temp_readonly(ts) || !IS_DEAD_ARG(i)) {
411
+ goto allocate_in_reg;
412
+ }
413
+
414
+ /*
415
+ * Check if the current register has already been allocated
416
+ * for another input aliased to an output.
417
+ */
418
+ if (ts->val_type == TEMP_VAL_REG) {
419
+ reg = ts->reg;
420
+ for (int k2 = 0; k2 < k; k2++) {
421
+ int i2 = def->args_ct[nb_oargs + k2].sort_index;
422
+ if (def->args_ct[i2].ialias && reg == new_args[i2]) {
423
+ goto allocate_in_reg;
424
}
425
}
426
- i_preferred_regs = o_preferred_regs;
427
}
428
+ i_preferred_regs = o_preferred_regs;
429
}
430
431
temp_load(s, ts, arg_ct->regs, i_allocated_regs, i_preferred_regs);
432
reg = ts->reg;
433
434
- if (tcg_regset_test_reg(arg_ct->regs, reg)) {
435
- /* nothing to do : the constraint is satisfied */
436
- } else {
437
- allocate_in_reg:
438
- /* allocate a new register matching the constraint
439
- and move the temporary register into it */
440
+ if (!tcg_regset_test_reg(arg_ct->regs, reg)) {
441
+ allocate_in_reg:
442
+ /*
443
+ * Allocate a new register matching the constraint
444
+ * and move the temporary register into it.
445
+ */
446
temp_load(s, ts, tcg_target_available_regs[ts->type],
447
i_allocated_regs, 0);
448
reg = tcg_reg_alloc(s, arg_ct->regs, i_allocated_regs,
449
--
242
--
450
2.25.1
243
2.34.1
451
452
diff view generated by jsdifflib
1
In most, but not all, places that we check for TEMP_FIXED,
1
These will be useful in properly ending the TB.
2
we are really testing that we do not modify the temporary.
3
2
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
3
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
7
---
8
include/tcg/tcg.h | 5 +++++
8
target/riscv/translate.c | 10 +++++++++-
9
tcg/tcg.c | 21 ++++++++++-----------
9
1 file changed, 9 insertions(+), 1 deletion(-)
10
2 files changed, 15 insertions(+), 11 deletions(-)
11
10
12
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
11
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/include/tcg/tcg.h
13
--- a/target/riscv/translate.c
15
+++ b/include/tcg/tcg.h
14
+++ b/target/riscv/translate.c
16
@@ -XXX,XX +XXX,XX @@ struct TCGContext {
15
@@ -XXX,XX +XXX,XX @@ static uint32_t opcode_at(DisasContextBase *dcbase, target_ulong pc)
17
target_ulong gen_insn_data[TCG_MAX_INSNS][TARGET_INSN_START_WORDS];
16
/* Include decoders for factored-out extensions */
18
};
17
#include "decode-XVentanaCondOps.c.inc"
19
18
20
+static inline bool temp_readonly(TCGTemp *ts)
19
+/* The specification allows for longer insns, but not supported by qemu. */
20
+#define MAX_INSN_LEN 4
21
+
22
+static inline int insn_len(uint16_t first_word)
21
+{
23
+{
22
+ return ts->kind == TEMP_FIXED;
24
+ return (first_word & 3) == 3 ? 4 : 2;
23
+}
25
+}
24
+
26
+
25
extern TCGContext tcg_init_ctx;
27
static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
26
extern __thread TCGContext *tcg_ctx;
27
extern const void *tcg_code_gen_epilogue;
28
diff --git a/tcg/tcg.c b/tcg/tcg.c
29
index XXXXXXX..XXXXXXX 100644
30
--- a/tcg/tcg.c
31
+++ b/tcg/tcg.c
32
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *, TCGTemp *, TCGRegSet, TCGRegSet, TCGRegSet);
33
mark it free; otherwise mark it dead. */
34
static void temp_free_or_dead(TCGContext *s, TCGTemp *ts, int free_or_dead)
35
{
28
{
36
- if (ts->kind == TEMP_FIXED) {
29
/*
37
+ if (temp_readonly(ts)) {
30
@@ -XXX,XX +XXX,XX @@ static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
38
return;
31
};
39
}
32
40
if (ts->val_type == TEMP_VAL_REG) {
33
/* Check for compressed insn */
41
@@ -XXX,XX +XXX,XX @@ static inline void temp_dead(TCGContext *s, TCGTemp *ts)
34
- if (extract16(opcode, 0, 2) != 3) {
42
static void temp_sync(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs,
35
+ if (insn_len(opcode) == 2) {
43
TCGRegSet preferred_regs, int free_or_dead)
36
if (!has_ext(ctx, RVC)) {
44
{
37
gen_exception_illegal(ctx);
45
- if (ts->kind == TEMP_FIXED) {
38
} else {
46
+ if (temp_readonly(ts)) {
47
return;
48
}
49
if (!ts->mem_coherent) {
50
@@ -XXX,XX +XXX,XX @@ static void temp_save(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs)
51
{
52
/* The liveness analysis already ensures that globals are back
53
in memory. Keep an tcg_debug_assert for safety. */
54
- tcg_debug_assert(ts->val_type == TEMP_VAL_MEM
55
- || ts->kind == TEMP_FIXED);
56
+ tcg_debug_assert(ts->val_type == TEMP_VAL_MEM || temp_readonly(ts));
57
}
58
59
/* save globals to their canonical location and assume they can be
60
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
61
TCGRegSet preferred_regs)
62
{
63
/* ENV should not be modified. */
64
- tcg_debug_assert(ots->kind != TEMP_FIXED);
65
+ tcg_debug_assert(!temp_readonly(ots));
66
67
/* The movi is not explicitly generated here. */
68
if (ots->val_type == TEMP_VAL_REG) {
69
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
70
ts = arg_temp(op->args[1]);
71
72
/* ENV should not be modified. */
73
- tcg_debug_assert(ots->kind != TEMP_FIXED);
74
+ tcg_debug_assert(!temp_readonly(ots));
75
76
/* Note that otype != itype for no-op truncation. */
77
otype = ots->type;
78
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
79
* Store the source register into the destination slot
80
* and leave the destination temp as TEMP_VAL_MEM.
81
*/
82
- assert(ots->kind != TEMP_FIXED);
83
+ assert(!temp_readonly(ots));
84
if (!ts->mem_allocated) {
85
temp_allocate_frame(s, ots);
86
}
87
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_dup(TCGContext *s, const TCGOp *op)
88
its = arg_temp(op->args[1]);
89
90
/* ENV should not be modified. */
91
- tcg_debug_assert(ots->kind != TEMP_FIXED);
92
+ tcg_debug_assert(!temp_readonly(ots));
93
94
itype = its->type;
95
vece = TCGOP_VECE(op);
96
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
97
ts = arg_temp(arg);
98
99
/* ENV should not be modified. */
100
- tcg_debug_assert(ts->kind != TEMP_FIXED);
101
+ tcg_debug_assert(!temp_readonly(ts));
102
103
if (arg_ct->oalias && !const_args[arg_ct->alias_index]) {
104
reg = new_args[arg_ct->alias_index];
105
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
106
ts = arg_temp(op->args[i]);
107
108
/* ENV should not be modified. */
109
- tcg_debug_assert(ts->kind != TEMP_FIXED);
110
+ tcg_debug_assert(!temp_readonly(ts));
111
112
if (NEED_SYNC_ARG(i)) {
113
temp_sync(s, ts, o_allocated_regs, 0, IS_DEAD_ARG(i));
114
@@ -XXX,XX +XXX,XX @@ static void tcg_reg_alloc_call(TCGContext *s, TCGOp *op)
115
ts = arg_temp(arg);
116
117
/* ENV should not be modified. */
118
- tcg_debug_assert(ts->kind != TEMP_FIXED);
119
+ tcg_debug_assert(!temp_readonly(ts));
120
121
reg = tcg_target_call_oarg_regs[i];
122
tcg_debug_assert(s->reg_to_temp[reg] == NULL);
123
--
39
--
124
2.25.1
40
2.34.1
125
126
diff view generated by jsdifflib
1
Having dupi pass though movi is confusing and arguably wrong.
1
Right now the translator stops right *after* the end of a page, which
2
breaks reporting of fault locations when the last instruction of a
3
multi-insn translation block crosses a page boundary.
2
4
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1155
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
10
---
6
tcg/tcg.c | 6 +++-
11
target/riscv/translate.c | 17 +++++--
7
tcg/aarch64/tcg-target.c.inc | 7 ----
12
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++++++++++++
8
tcg/i386/tcg-target.c.inc | 63 ++++++++++++++++++++++++------------
13
tests/tcg/riscv64/Makefile.target | 1 +
9
tcg/ppc/tcg-target.c.inc | 6 ----
14
3 files changed, 93 insertions(+), 4 deletions(-)
10
4 files changed, 47 insertions(+), 35 deletions(-)
15
create mode 100644 tests/tcg/riscv64/noexec.c
11
16
12
diff --git a/tcg/tcg.c b/tcg/tcg.c
17
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
13
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/tcg.c
19
--- a/target/riscv/translate.c
15
+++ b/tcg/tcg.c
20
+++ b/target/riscv/translate.c
16
@@ -XXX,XX +XXX,XX @@ static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
21
@@ -XXX,XX +XXX,XX @@ static void riscv_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
17
case TEMP_VAL_CONST:
22
}
18
reg = tcg_reg_alloc(s, desired_regs, allocated_regs,
23
ctx->nftemp = 0;
19
preferred_regs, ts->indirect_base);
24
20
- tcg_out_movi(s, ts->type, reg, ts->val);
25
+ /* Only the first insn within a TB is allowed to cross a page boundary. */
21
+ if (ts->type <= TCG_TYPE_I64) {
26
if (ctx->base.is_jmp == DISAS_NEXT) {
22
+ tcg_out_movi(s, ts->type, reg, ts->val);
27
- target_ulong page_start;
28
-
29
- page_start = ctx->base.pc_first & TARGET_PAGE_MASK;
30
- if (ctx->base.pc_next - page_start >= TARGET_PAGE_SIZE) {
31
+ if (!is_same_page(&ctx->base, ctx->base.pc_next)) {
32
ctx->base.is_jmp = DISAS_TOO_MANY;
23
+ } else {
33
+ } else {
24
+ tcg_out_dupi_vec(s, ts->type, reg, ts->val);
34
+ unsigned page_ofs = ctx->base.pc_next & ~TARGET_PAGE_MASK;
25
+ }
35
+
26
ts->mem_coherent = 0;
36
+ if (page_ofs > TARGET_PAGE_SIZE - MAX_INSN_LEN) {
27
break;
37
+ uint16_t next_insn = cpu_lduw_code(env, ctx->base.pc_next);
28
case TEMP_VAL_MEM:
38
+ int len = insn_len(next_insn);
29
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
39
+
30
index XXXXXXX..XXXXXXX 100644
40
+ if (!is_same_page(&ctx->base, ctx->base.pc_next + len)) {
31
--- a/tcg/aarch64/tcg-target.c.inc
41
+ ctx->base.is_jmp = DISAS_TOO_MANY;
32
+++ b/tcg/aarch64/tcg-target.c.inc
42
+ }
33
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi(TCGContext *s, TCGType type, TCGReg rd,
43
+ }
34
case TCG_TYPE_I64:
44
}
35
tcg_debug_assert(rd < 32);
36
break;
37
-
38
- case TCG_TYPE_V64:
39
- case TCG_TYPE_V128:
40
- tcg_debug_assert(rd >= 32);
41
- tcg_out_dupi_vec(s, type, rd, value);
42
- return;
43
-
44
default:
45
g_assert_not_reached();
46
}
47
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
48
index XXXXXXX..XXXXXXX 100644
49
--- a/tcg/i386/tcg-target.c.inc
50
+++ b/tcg/i386/tcg-target.c.inc
51
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
52
}
45
}
53
}
46
}
54
47
diff --git a/tests/tcg/riscv64/noexec.c b/tests/tcg/riscv64/noexec.c
55
-static void tcg_out_movi(TCGContext *s, TCGType type,
48
new file mode 100644
56
- TCGReg ret, tcg_target_long arg)
49
index XXXXXXX..XXXXXXX
57
+static void tcg_out_movi_vec(TCGContext *s, TCGType type,
50
--- /dev/null
58
+ TCGReg ret, tcg_target_long arg)
51
+++ b/tests/tcg/riscv64/noexec.c
52
@@ -XXX,XX +XXX,XX @@
53
+#include "../multiarch/noexec.c.inc"
54
+
55
+static void *arch_mcontext_pc(const mcontext_t *ctx)
59
+{
56
+{
60
+ if (arg == 0) {
57
+ return (void *)ctx->__gregs[REG_PC];
61
+ tcg_out_vex_modrm(s, OPC_PXOR, ret, ret, ret);
62
+ return;
63
+ }
64
+ if (arg == -1) {
65
+ tcg_out_vex_modrm(s, OPC_PCMPEQB, ret, ret, ret);
66
+ return;
67
+ }
68
+
69
+ int rexw = (type == TCG_TYPE_I32 ? 0 : P_REXW);
70
+ tcg_out_vex_modrm_pool(s, OPC_MOVD_VyEy + rexw, ret);
71
+ if (TCG_TARGET_REG_BITS == 64) {
72
+ new_pool_label(s, arg, R_386_PC32, s->code_ptr - 4, -4);
73
+ } else {
74
+ new_pool_label(s, arg, R_386_32, s->code_ptr - 4, 0);
75
+ }
76
+}
58
+}
77
+
59
+
78
+static void tcg_out_movi_int(TCGContext *s, TCGType type,
60
+static int arch_mcontext_arg(const mcontext_t *ctx)
79
+ TCGReg ret, tcg_target_long arg)
80
{
81
tcg_target_long diff;
82
83
- switch (type) {
84
- case TCG_TYPE_I32:
85
-#if TCG_TARGET_REG_BITS == 64
86
- case TCG_TYPE_I64:
87
-#endif
88
- if (ret < 16) {
89
- break;
90
- }
91
- /* fallthru */
92
- case TCG_TYPE_V64:
93
- case TCG_TYPE_V128:
94
- case TCG_TYPE_V256:
95
- tcg_debug_assert(ret >= 16);
96
- tcg_out_dupi_vec(s, type, ret, arg);
97
- return;
98
- default:
99
- g_assert_not_reached();
100
- }
101
-
102
if (arg == 0) {
103
tgen_arithr(s, ARITH_XOR, ret, ret);
104
return;
105
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi(TCGContext *s, TCGType type,
106
tcg_out64(s, arg);
107
}
108
109
+static void tcg_out_movi(TCGContext *s, TCGType type,
110
+ TCGReg ret, tcg_target_long arg)
111
+{
61
+{
112
+ switch (type) {
62
+ return ctx->__gregs[REG_A0];
113
+ case TCG_TYPE_I32:
114
+#if TCG_TARGET_REG_BITS == 64
115
+ case TCG_TYPE_I64:
116
+#endif
117
+ if (ret < 16) {
118
+ tcg_out_movi_int(s, type, ret, arg);
119
+ } else {
120
+ tcg_out_movi_vec(s, type, ret, arg);
121
+ }
122
+ break;
123
+ default:
124
+ g_assert_not_reached();
125
+ }
126
+}
63
+}
127
+
64
+
128
static inline void tcg_out_pushi(TCGContext *s, tcg_target_long val)
65
+static void arch_flush(void *p, int len)
129
{
66
+{
130
if (val == (int8_t)val) {
67
+ __builtin___clear_cache(p, p + len);
131
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
68
+}
69
+
70
+extern char noexec_1[];
71
+extern char noexec_2[];
72
+extern char noexec_end[];
73
+
74
+asm(".option push\n"
75
+ ".option norvc\n"
76
+ "noexec_1:\n"
77
+ " li a0,1\n" /* a0 is 0 on entry, set 1. */
78
+ "noexec_2:\n"
79
+ " li a0,2\n" /* a0 is 0/1; set 2. */
80
+ " ret\n"
81
+ "noexec_end:\n"
82
+ ".option pop");
83
+
84
+int main(void)
85
+{
86
+ struct noexec_test noexec_tests[] = {
87
+ {
88
+ .name = "fallthrough",
89
+ .test_code = noexec_1,
90
+ .test_len = noexec_end - noexec_1,
91
+ .page_ofs = noexec_1 - noexec_2,
92
+ .entry_ofs = noexec_1 - noexec_2,
93
+ .expected_si_ofs = 0,
94
+ .expected_pc_ofs = 0,
95
+ .expected_arg = 1,
96
+ },
97
+ {
98
+ .name = "jump",
99
+ .test_code = noexec_1,
100
+ .test_len = noexec_end - noexec_1,
101
+ .page_ofs = noexec_1 - noexec_2,
102
+ .entry_ofs = 0,
103
+ .expected_si_ofs = 0,
104
+ .expected_pc_ofs = 0,
105
+ .expected_arg = 0,
106
+ },
107
+ {
108
+ .name = "fallthrough [cross]",
109
+ .test_code = noexec_1,
110
+ .test_len = noexec_end - noexec_1,
111
+ .page_ofs = noexec_1 - noexec_2 - 2,
112
+ .entry_ofs = noexec_1 - noexec_2 - 2,
113
+ .expected_si_ofs = 0,
114
+ .expected_pc_ofs = -2,
115
+ .expected_arg = 1,
116
+ },
117
+ {
118
+ .name = "jump [cross]",
119
+ .test_code = noexec_1,
120
+ .test_len = noexec_end - noexec_1,
121
+ .page_ofs = noexec_1 - noexec_2 - 2,
122
+ .entry_ofs = -2,
123
+ .expected_si_ofs = 0,
124
+ .expected_pc_ofs = -2,
125
+ .expected_arg = 0,
126
+ },
127
+ };
128
+
129
+ return test_noexec(noexec_tests,
130
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
131
+}
132
diff --git a/tests/tcg/riscv64/Makefile.target b/tests/tcg/riscv64/Makefile.target
132
index XXXXXXX..XXXXXXX 100644
133
index XXXXXXX..XXXXXXX 100644
133
--- a/tcg/ppc/tcg-target.c.inc
134
--- a/tests/tcg/riscv64/Makefile.target
134
+++ b/tcg/ppc/tcg-target.c.inc
135
+++ b/tests/tcg/riscv64/Makefile.target
135
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi(TCGContext *s, TCGType type, TCGReg ret,
136
@@ -XXX,XX +XXX,XX @@
136
tcg_out_movi_int(s, type, ret, arg, false);
137
137
break;
138
VPATH += $(SRC_PATH)/tests/tcg/riscv64
138
139
TESTS += test-div
139
- case TCG_TYPE_V64:
140
+TESTS += noexec
140
- case TCG_TYPE_V128:
141
- tcg_debug_assert(ret >= TCG_REG_V0);
142
- tcg_out_dupi_vec(s, type, ret, arg);
143
- break;
144
-
145
default:
146
g_assert_not_reached();
147
}
148
--
141
--
149
2.25.1
142
2.34.1
150
151
diff view generated by jsdifflib