1
The following changes since commit 0a301624c2f4ced3331ffd5bce85b4274fe132af:
1
The following changes since commit e93ded1bf6c94ab95015b33e188bc8b0b0c32670:
2
2
3
Merge remote-tracking branch 'remotes/pmaydell/tags/pull-target-arm-20220208' into staging (2022-02-08 11:40:08 +0000)
3
Merge tag 'testing-pull-request-2022-08-30' of https://gitlab.com/thuth/qemu into staging (2022-08-31 18:19:03 -0400)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220211
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220901
8
8
9
for you to fetch changes up to 5c1a101ef6b85537a4ade93c39ea81cadd5c246e:
9
for you to fetch changes up to 20011be2e30b8aa8ef1fc258485f00c688703deb:
10
10
11
tests/tcg/multiarch: Add sigbus.c (2022-02-09 09:00:01 +1100)
11
target/riscv: Make translator stop before the end of a page (2022-09-01 07:43:08 +0100)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Fix safe_syscall_base for sparc64.
14
Respect PROT_EXEC in user-only mode.
15
Fix host signal handling for sparc64-linux.
15
Fix s390x, i386 and riscv for translations crossing a page.
16
Speedups for jump cache and work list probing.
17
Fix for exception replays.
18
Raise guest SIGBUS for user-only misaligned accesses.
19
16
20
----------------------------------------------------------------
17
----------------------------------------------------------------
21
Idan Horowitz (2):
18
Ilya Leoshkevich (4):
22
accel/tcg: Optimize jump cache flush during tlb range flush
19
linux-user: Clear translations on mprotect()
23
softmmu/cpus: Check if the cpu work list is empty atomically
20
accel/tcg: Introduce is_same_page()
21
target/s390x: Make translator stop before the end of a page
22
target/i386: Make translator stop before the end of a page
24
23
25
Pavel Dovgalyuk (1):
24
Richard Henderson (16):
26
replay: use CF_NOIRQ for special exception-replaying TB
25
linux-user/arm: Mark the commpage executable
26
linux-user/hppa: Allocate page zero as a commpage
27
linux-user/x86_64: Allocate vsyscall page as a commpage
28
linux-user: Honor PT_GNU_STACK
29
tests/tcg/i386: Move smc_code2 to an executable section
30
accel/tcg: Properly implement get_page_addr_code for user-only
31
accel/tcg: Unlock mmap_lock after longjmp
32
accel/tcg: Make tb_htable_lookup static
33
accel/tcg: Move qemu_ram_addr_from_host_nofail to physmem.c
34
accel/tcg: Use probe_access_internal for softmmu get_page_addr_code_hostp
35
accel/tcg: Document the faulting lookup in tb_lookup_cmp
36
accel/tcg: Remove translator_ldsw
37
accel/tcg: Add pc and host_pc params to gen_intermediate_code
38
accel/tcg: Add fast path for translator_ld*
39
target/riscv: Add MAX_INSN_LEN and insn_len
40
target/riscv: Make translator stop before the end of a page
27
41
28
Richard Henderson (29):
42
include/elf.h | 1 +
29
common-user/host/sparc64: Fix safe_syscall_base
43
include/exec/cpu-common.h | 1 +
30
linux-user: Introduce host_signal_mask
44
include/exec/exec-all.h | 89 ++++++++----------------
31
linux-user: Introduce host_sigcontext
45
include/exec/translator.h | 96 ++++++++++++++++---------
32
linux-user: Move sparc/host-signal.h to sparc64/host-signal.h
46
linux-user/arm/target_cpu.h | 4 +-
33
linux-user/include/host/sparc64: Fix host_sigcontext
47
linux-user/qemu.h | 1 +
34
tcg/i386: Support raising sigbus for user-only
48
accel/tcg/cpu-exec.c | 143 ++++++++++++++++++++------------------
35
tcg/aarch64: Support raising sigbus for user-only
49
accel/tcg/cputlb.c | 93 +++++++------------------
36
tcg/ppc: Support raising sigbus for user-only
50
accel/tcg/translate-all.c | 29 ++++----
37
tcg/riscv: Support raising sigbus for user-only
51
accel/tcg/translator.c | 135 ++++++++++++++++++++++++++---------
38
tcg/s390x: Support raising sigbus for user-only
52
accel/tcg/user-exec.c | 17 ++++-
39
tcg/tci: Support raising sigbus for user-only
53
linux-user/elfload.c | 82 ++++++++++++++++++++--
40
tcg/arm: Drop support for armv4 and armv5 hosts
54
linux-user/mmap.c | 6 +-
41
tcg/arm: Remove use_armv5t_instructions
55
softmmu/physmem.c | 12 ++++
42
tcg/arm: Remove use_armv6_instructions
56
target/alpha/translate.c | 5 +-
43
tcg/arm: Check alignment for ldrd and strd
57
target/arm/translate.c | 5 +-
44
tcg/arm: Support unaligned access for softmmu
58
target/avr/translate.c | 5 +-
45
tcg/arm: Reserve a register for guest_base
59
target/cris/translate.c | 5 +-
46
tcg/arm: Support raising sigbus for user-only
60
target/hexagon/translate.c | 6 +-
47
tcg/mips: Support unaligned access for user-only
61
target/hppa/translate.c | 5 +-
48
tcg/mips: Support unaligned access for softmmu
62
target/i386/tcg/translate.c | 71 +++++++++++--------
49
tcg/sparc: Use tcg_out_movi_imm13 in tcg_out_addsub2_i64
63
target/loongarch/translate.c | 6 +-
50
tcg/sparc: Split out tcg_out_movi_imm32
64
target/m68k/translate.c | 5 +-
51
tcg/sparc: Add scratch argument to tcg_out_movi_int
65
target/microblaze/translate.c | 5 +-
52
tcg/sparc: Improve code gen for shifted 32-bit constants
66
target/mips/tcg/translate.c | 5 +-
53
tcg/sparc: Convert patch_reloc to return bool
67
target/nios2/translate.c | 5 +-
54
tcg/sparc: Use the constant pool for 64-bit constants
68
target/openrisc/translate.c | 6 +-
55
tcg/sparc: Add tcg_out_jmpl_const for better tail calls
69
target/ppc/translate.c | 5 +-
56
tcg/sparc: Support unaligned access for user-only
70
target/riscv/translate.c | 32 +++++++--
57
tests/tcg/multiarch: Add sigbus.c
71
target/rx/translate.c | 5 +-
58
72
target/s390x/tcg/translate.c | 20 ++++--
59
WANG Xuerui (2):
73
target/sh4/translate.c | 5 +-
60
tcg/loongarch64: Fix fallout from recent MO_Q renaming
74
target/sparc/translate.c | 5 +-
61
tcg/loongarch64: Support raising sigbus for user-only
75
target/tricore/translate.c | 6 +-
62
76
target/xtensa/translate.c | 6 +-
63
linux-user/include/host/aarch64/host-signal.h | 16 +-
77
tests/tcg/i386/test-i386.c | 2 +-
64
linux-user/include/host/alpha/host-signal.h | 14 +-
78
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++
65
linux-user/include/host/arm/host-signal.h | 14 +-
79
tests/tcg/s390x/noexec.c | 106 ++++++++++++++++++++++++++++
66
linux-user/include/host/i386/host-signal.h | 14 +-
80
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++
67
linux-user/include/host/loongarch64/host-signal.h | 14 +-
81
tests/tcg/multiarch/noexec.c.inc | 139 ++++++++++++++++++++++++++++++++++++
68
linux-user/include/host/mips/host-signal.h | 14 +-
82
tests/tcg/riscv64/Makefile.target | 1 +
69
linux-user/include/host/ppc/host-signal.h | 14 +-
83
tests/tcg/s390x/Makefile.target | 1 +
70
linux-user/include/host/riscv/host-signal.h | 14 +-
84
tests/tcg/x86_64/Makefile.target | 3 +-
71
linux-user/include/host/s390/host-signal.h | 14 +-
85
43 files changed, 966 insertions(+), 367 deletions(-)
72
linux-user/include/host/sparc/host-signal.h | 63 ----
86
create mode 100644 tests/tcg/riscv64/noexec.c
73
linux-user/include/host/sparc64/host-signal.h | 65 +++-
87
create mode 100644 tests/tcg/s390x/noexec.c
74
linux-user/include/host/x86_64/host-signal.h | 14 +-
88
create mode 100644 tests/tcg/x86_64/noexec.c
75
tcg/aarch64/tcg-target.h | 2 -
89
create mode 100644 tests/tcg/multiarch/noexec.c.inc
76
tcg/arm/tcg-target.h | 6 +-
77
tcg/i386/tcg-target.h | 2 -
78
tcg/loongarch64/tcg-target.h | 2 -
79
tcg/mips/tcg-target.h | 2 -
80
tcg/ppc/tcg-target.h | 2 -
81
tcg/riscv/tcg-target.h | 2 -
82
tcg/s390x/tcg-target.h | 2 -
83
accel/tcg/cpu-exec.c | 3 +-
84
accel/tcg/cputlb.c | 9 +
85
linux-user/signal.c | 22 +-
86
softmmu/cpus.c | 7 +-
87
tcg/tci.c | 20 +-
88
tests/tcg/multiarch/sigbus.c | 68 ++++
89
tcg/aarch64/tcg-target.c.inc | 91 ++++-
90
tcg/arm/tcg-target.c.inc | 410 +++++++++-------------
91
tcg/i386/tcg-target.c.inc | 103 +++++-
92
tcg/loongarch64/tcg-target.c.inc | 73 +++-
93
tcg/mips/tcg-target.c.inc | 387 ++++++++++++++++++--
94
tcg/ppc/tcg-target.c.inc | 98 +++++-
95
tcg/riscv/tcg-target.c.inc | 63 +++-
96
tcg/s390x/tcg-target.c.inc | 59 +++-
97
tcg/sparc/tcg-target.c.inc | 348 +++++++++++++++---
98
common-user/host/sparc64/safe-syscall.inc.S | 5 +-
99
36 files changed, 1561 insertions(+), 495 deletions(-)
100
delete mode 100644 linux-user/include/host/sparc/host-signal.h
101
create mode 100644 tests/tcg/multiarch/sigbus.c
102
diff view generated by jsdifflib
Deleted patch
1
Use the "retl" instead of "ret" instruction alias, since we
2
do not allocate a register window in this function.
3
1
4
Fix the offset to the first stacked parameter, which lies
5
beyond the register window save area.
6
7
Fixes: 95c021dac835 ("linux-user/host/sparc64: Add safe-syscall.inc.S")
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
10
common-user/host/sparc64/safe-syscall.inc.S | 5 +++--
11
1 file changed, 3 insertions(+), 2 deletions(-)
12
13
diff --git a/common-user/host/sparc64/safe-syscall.inc.S b/common-user/host/sparc64/safe-syscall.inc.S
14
index XXXXXXX..XXXXXXX 100644
15
--- a/common-user/host/sparc64/safe-syscall.inc.S
16
+++ b/common-user/host/sparc64/safe-syscall.inc.S
17
@@ -XXX,XX +XXX,XX @@
18
.type safe_syscall_end, @function
19
20
#define STACK_BIAS 2047
21
-#define PARAM(N) STACK_BIAS + N*8
22
+#define WINDOW_SIZE 16 * 8
23
+#define PARAM(N) STACK_BIAS + WINDOW_SIZE + N * 8
24
25
/*
26
* This is the entry point for making a system call. The calling
27
@@ -XXX,XX +XXX,XX @@ safe_syscall_end:
28
/* code path for having successfully executed the syscall */
29
bcs,pn %xcc, 1f
30
nop
31
- ret
32
+ retl
33
nop
34
35
/* code path when we didn't execute the syscall */
36
--
37
2.25.1
38
39
diff view generated by jsdifflib
1
This is kinda sorta the opposite of the other tcg hosts, where
1
We're about to start validating PAGE_EXEC, which means
2
we get (normal) alignment checks for free with host SIGBUS and
2
that we've got to mark the commpage executable. We had
3
need to add code to support unaligned accesses.
3
been placing the commpage outside of reserved_va, which
4
was incorrect and lead to an abort.
4
5
5
This inline code expansion is somewhat large, but it takes quite
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
a few instructions to make a function call to a helper anyway.
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
9
---
11
tcg/sparc/tcg-target.c.inc | 219 +++++++++++++++++++++++++++++++++++--
10
linux-user/arm/target_cpu.h | 4 ++--
12
1 file changed, 211 insertions(+), 8 deletions(-)
11
linux-user/elfload.c | 6 +++++-
12
2 files changed, 7 insertions(+), 3 deletions(-)
13
13
14
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
14
diff --git a/linux-user/arm/target_cpu.h b/linux-user/arm/target_cpu.h
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/tcg/sparc/tcg-target.c.inc
16
--- a/linux-user/arm/target_cpu.h
17
+++ b/tcg/sparc/tcg-target.c.inc
17
+++ b/linux-user/arm/target_cpu.h
18
@@ -XXX,XX +XXX,XX @@ static const int tcg_target_call_oarg_regs[] = {
18
@@ -XXX,XX +XXX,XX @@ static inline unsigned long arm_max_reserved_va(CPUState *cs)
19
#define ARITH_ADD (INSN_OP(2) | INSN_OP3(0x00))
19
} else {
20
#define ARITH_ADDCC (INSN_OP(2) | INSN_OP3(0x10))
20
/*
21
#define ARITH_AND (INSN_OP(2) | INSN_OP3(0x01))
21
* We need to be able to map the commpage.
22
+#define ARITH_ANDCC (INSN_OP(2) | INSN_OP3(0x11))
22
- * See validate_guest_space in linux-user/elfload.c.
23
#define ARITH_ANDN (INSN_OP(2) | INSN_OP3(0x05))
23
+ * See init_guest_commpage in linux-user/elfload.c.
24
#define ARITH_OR (INSN_OP(2) | INSN_OP3(0x02))
24
*/
25
#define ARITH_ORCC (INSN_OP(2) | INSN_OP3(0x12))
25
- return 0xffff0000ul;
26
@@ -XXX,XX +XXX,XX @@ static void build_trampolines(TCGContext *s)
26
+ return 0xfffffffful;
27
tcg_out_mov_delay(s, TCG_REG_O0, TCG_AREG0);
28
}
27
}
29
}
28
}
30
+#else
29
#define MAX_RESERVED_VA arm_max_reserved_va
31
+static const tcg_insn_unit *qemu_unalign_ld_trampoline;
30
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
32
+static const tcg_insn_unit *qemu_unalign_st_trampoline;
31
index XXXXXXX..XXXXXXX 100644
32
--- a/linux-user/elfload.c
33
+++ b/linux-user/elfload.c
34
@@ -XXX,XX +XXX,XX @@ enum {
35
36
static bool init_guest_commpage(void)
37
{
38
- void *want = g2h_untagged(HI_COMMPAGE & -qemu_host_page_size);
39
+ abi_ptr commpage = HI_COMMPAGE & -qemu_host_page_size;
40
+ void *want = g2h_untagged(commpage);
41
void *addr = mmap(want, qemu_host_page_size, PROT_READ | PROT_WRITE,
42
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
43
44
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
45
perror("Protecting guest commpage");
46
exit(EXIT_FAILURE);
47
}
33
+
48
+
34
+static void build_trampolines(TCGContext *s)
49
+ page_set_flags(commpage, commpage + qemu_host_page_size,
35
+{
50
+ PAGE_READ | PAGE_EXEC | PAGE_VALID);
36
+ for (int ld = 0; ld < 2; ++ld) {
51
return true;
37
+ void *helper;
38
+
39
+ while ((uintptr_t)s->code_ptr & 15) {
40
+ tcg_out_nop(s);
41
+ }
42
+
43
+ if (ld) {
44
+ helper = helper_unaligned_ld;
45
+ qemu_unalign_ld_trampoline = tcg_splitwx_to_rx(s->code_ptr);
46
+ } else {
47
+ helper = helper_unaligned_st;
48
+ qemu_unalign_st_trampoline = tcg_splitwx_to_rx(s->code_ptr);
49
+ }
50
+
51
+ if (!SPARC64 && TARGET_LONG_BITS == 64) {
52
+ /* Install the high part of the address. */
53
+ tcg_out_arithi(s, TCG_REG_O1, TCG_REG_O2, 32, SHIFT_SRLX);
54
+ }
55
+
56
+ /* Tail call. */
57
+ tcg_out_jmpl_const(s, helper, true, true);
58
+ /* delay slot -- set the env argument */
59
+ tcg_out_mov_delay(s, TCG_REG_O0, TCG_AREG0);
60
+ }
61
+}
62
#endif
63
64
/* Generate global QEMU prologue and epilogue code */
65
@@ -XXX,XX +XXX,XX @@ static void tcg_target_qemu_prologue(TCGContext *s)
66
/* delay slot */
67
tcg_out_movi_imm13(s, TCG_REG_O0, 0);
68
69
-#ifdef CONFIG_SOFTMMU
70
build_trampolines(s);
71
-#endif
72
}
52
}
73
53
74
static void tcg_out_nop_fill(tcg_insn_unit *p, int count)
75
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_load(TCGContext *s, TCGReg addr, int mem_index,
76
static const int qemu_ld_opc[(MO_SSIZE | MO_BSWAP) + 1] = {
77
[MO_UB] = LDUB,
78
[MO_SB] = LDSB,
79
+ [MO_UB | MO_LE] = LDUB,
80
+ [MO_SB | MO_LE] = LDSB,
81
82
[MO_BEUW] = LDUH,
83
[MO_BESW] = LDSH,
84
[MO_BEUL] = LDUW,
85
[MO_BESL] = LDSW,
86
[MO_BEUQ] = LDX,
87
+ [MO_BESQ] = LDX,
88
89
[MO_LEUW] = LDUH_LE,
90
[MO_LESW] = LDSH_LE,
91
[MO_LEUL] = LDUW_LE,
92
[MO_LESL] = LDSW_LE,
93
[MO_LEUQ] = LDX_LE,
94
+ [MO_LESQ] = LDX_LE,
95
};
96
97
static const int qemu_st_opc[(MO_SIZE | MO_BSWAP) + 1] = {
98
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data, TCGReg addr,
99
MemOpIdx oi, bool is_64)
100
{
101
MemOp memop = get_memop(oi);
102
+ tcg_insn_unit *label_ptr;
103
+
104
#ifdef CONFIG_SOFTMMU
105
unsigned memi = get_mmuidx(oi);
106
TCGReg addrz, param;
107
const tcg_insn_unit *func;
108
- tcg_insn_unit *label_ptr;
109
110
addrz = tcg_out_tlb_load(s, addr, memi, memop,
111
offsetof(CPUTLBEntry, addr_read));
112
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data, TCGReg addr,
113
114
*label_ptr |= INSN_OFF19(tcg_ptr_byte_diff(s->code_ptr, label_ptr));
115
#else
116
+ TCGReg index = (guest_base ? TCG_GUEST_BASE_REG : TCG_REG_G0);
117
+ unsigned a_bits = get_alignment_bits(memop);
118
+ unsigned s_bits = memop & MO_SIZE;
119
+ unsigned t_bits;
120
+
121
if (SPARC64 && TARGET_LONG_BITS == 32) {
122
tcg_out_arithi(s, TCG_REG_T1, addr, 0, SHIFT_SRL);
123
addr = TCG_REG_T1;
124
}
125
- tcg_out_ldst_rr(s, data, addr,
126
- (guest_base ? TCG_GUEST_BASE_REG : TCG_REG_G0),
127
+
128
+ /*
129
+ * Normal case: alignment equal to access size.
130
+ */
131
+ if (a_bits == s_bits) {
132
+ tcg_out_ldst_rr(s, data, addr, index,
133
+ qemu_ld_opc[memop & (MO_BSWAP | MO_SSIZE)]);
134
+ return;
135
+ }
136
+
137
+ /*
138
+ * Test for at least natural alignment, and assume most accesses
139
+ * will be aligned -- perform a straight load in the delay slot.
140
+ * This is required to preserve atomicity for aligned accesses.
141
+ */
142
+ t_bits = MAX(a_bits, s_bits);
143
+ tcg_debug_assert(t_bits < 13);
144
+ tcg_out_arithi(s, TCG_REG_G0, addr, (1u << t_bits) - 1, ARITH_ANDCC);
145
+
146
+ /* beq,a,pt %icc, label */
147
+ label_ptr = s->code_ptr;
148
+ tcg_out_bpcc0(s, COND_E, BPCC_A | BPCC_PT | BPCC_ICC, 0);
149
+ /* delay slot */
150
+ tcg_out_ldst_rr(s, data, addr, index,
151
qemu_ld_opc[memop & (MO_BSWAP | MO_SSIZE)]);
152
+
153
+ if (a_bits >= s_bits) {
154
+ /*
155
+ * Overalignment: A successful alignment test will perform the memory
156
+ * operation in the delay slot, and failure need only invoke the
157
+ * handler for SIGBUS.
158
+ */
159
+ TCGReg arg_low = TCG_REG_O1 + (!SPARC64 && TARGET_LONG_BITS == 64);
160
+ tcg_out_call_nodelay(s, qemu_unalign_ld_trampoline, false);
161
+ /* delay slot -- move to low part of argument reg */
162
+ tcg_out_mov_delay(s, arg_low, addr);
163
+ } else {
164
+ /* Underalignment: load by pieces of minimum alignment. */
165
+ int ld_opc, a_size, s_size, i;
166
+
167
+ /*
168
+ * Force full address into T1 early; avoids problems with
169
+ * overlap between @addr and @data.
170
+ */
171
+ tcg_out_arith(s, TCG_REG_T1, addr, index, ARITH_ADD);
172
+
173
+ a_size = 1 << a_bits;
174
+ s_size = 1 << s_bits;
175
+ if ((memop & MO_BSWAP) == MO_BE) {
176
+ ld_opc = qemu_ld_opc[a_bits | MO_BE | (memop & MO_SIGN)];
177
+ tcg_out_ldst(s, data, TCG_REG_T1, 0, ld_opc);
178
+ ld_opc = qemu_ld_opc[a_bits | MO_BE];
179
+ for (i = a_size; i < s_size; i += a_size) {
180
+ tcg_out_ldst(s, TCG_REG_T2, TCG_REG_T1, i, ld_opc);
181
+ tcg_out_arithi(s, data, data, a_size, SHIFT_SLLX);
182
+ tcg_out_arith(s, data, data, TCG_REG_T2, ARITH_OR);
183
+ }
184
+ } else if (a_bits == 0) {
185
+ ld_opc = LDUB;
186
+ tcg_out_ldst(s, data, TCG_REG_T1, 0, ld_opc);
187
+ for (i = a_size; i < s_size; i += a_size) {
188
+ if ((memop & MO_SIGN) && i == s_size - a_size) {
189
+ ld_opc = LDSB;
190
+ }
191
+ tcg_out_ldst(s, TCG_REG_T2, TCG_REG_T1, i, ld_opc);
192
+ tcg_out_arithi(s, TCG_REG_T2, TCG_REG_T2, i * 8, SHIFT_SLLX);
193
+ tcg_out_arith(s, data, data, TCG_REG_T2, ARITH_OR);
194
+ }
195
+ } else {
196
+ ld_opc = qemu_ld_opc[a_bits | MO_LE];
197
+ tcg_out_ldst_rr(s, data, TCG_REG_T1, TCG_REG_G0, ld_opc);
198
+ for (i = a_size; i < s_size; i += a_size) {
199
+ tcg_out_arithi(s, TCG_REG_T1, TCG_REG_T1, a_size, ARITH_ADD);
200
+ if ((memop & MO_SIGN) && i == s_size - a_size) {
201
+ ld_opc = qemu_ld_opc[a_bits | MO_LE | MO_SIGN];
202
+ }
203
+ tcg_out_ldst_rr(s, TCG_REG_T2, TCG_REG_T1, TCG_REG_G0, ld_opc);
204
+ tcg_out_arithi(s, TCG_REG_T2, TCG_REG_T2, i * 8, SHIFT_SLLX);
205
+ tcg_out_arith(s, data, data, TCG_REG_T2, ARITH_OR);
206
+ }
207
+ }
208
+ }
209
+
210
+ *label_ptr |= INSN_OFF19(tcg_ptr_byte_diff(s->code_ptr, label_ptr));
211
#endif /* CONFIG_SOFTMMU */
212
}
213
214
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data, TCGReg addr,
215
MemOpIdx oi)
216
{
217
MemOp memop = get_memop(oi);
218
+ tcg_insn_unit *label_ptr;
219
+
220
#ifdef CONFIG_SOFTMMU
221
unsigned memi = get_mmuidx(oi);
222
TCGReg addrz, param;
223
const tcg_insn_unit *func;
224
- tcg_insn_unit *label_ptr;
225
226
addrz = tcg_out_tlb_load(s, addr, memi, memop,
227
offsetof(CPUTLBEntry, addr_write));
228
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data, TCGReg addr,
229
230
*label_ptr |= INSN_OFF19(tcg_ptr_byte_diff(s->code_ptr, label_ptr));
231
#else
232
+ TCGReg index = (guest_base ? TCG_GUEST_BASE_REG : TCG_REG_G0);
233
+ unsigned a_bits = get_alignment_bits(memop);
234
+ unsigned s_bits = memop & MO_SIZE;
235
+ unsigned t_bits;
236
+
237
if (SPARC64 && TARGET_LONG_BITS == 32) {
238
tcg_out_arithi(s, TCG_REG_T1, addr, 0, SHIFT_SRL);
239
addr = TCG_REG_T1;
240
}
241
- tcg_out_ldst_rr(s, data, addr,
242
- (guest_base ? TCG_GUEST_BASE_REG : TCG_REG_G0),
243
+
244
+ /*
245
+ * Normal case: alignment equal to access size.
246
+ */
247
+ if (a_bits == s_bits) {
248
+ tcg_out_ldst_rr(s, data, addr, index,
249
+ qemu_st_opc[memop & (MO_BSWAP | MO_SIZE)]);
250
+ return;
251
+ }
252
+
253
+ /*
254
+ * Test for at least natural alignment, and assume most accesses
255
+ * will be aligned -- perform a straight store in the delay slot.
256
+ * This is required to preserve atomicity for aligned accesses.
257
+ */
258
+ t_bits = MAX(a_bits, s_bits);
259
+ tcg_debug_assert(t_bits < 13);
260
+ tcg_out_arithi(s, TCG_REG_G0, addr, (1u << t_bits) - 1, ARITH_ANDCC);
261
+
262
+ /* beq,a,pt %icc, label */
263
+ label_ptr = s->code_ptr;
264
+ tcg_out_bpcc0(s, COND_E, BPCC_A | BPCC_PT | BPCC_ICC, 0);
265
+ /* delay slot */
266
+ tcg_out_ldst_rr(s, data, addr, index,
267
qemu_st_opc[memop & (MO_BSWAP | MO_SIZE)]);
268
+
269
+ if (a_bits >= s_bits) {
270
+ /*
271
+ * Overalignment: A successful alignment test will perform the memory
272
+ * operation in the delay slot, and failure need only invoke the
273
+ * handler for SIGBUS.
274
+ */
275
+ TCGReg arg_low = TCG_REG_O1 + (!SPARC64 && TARGET_LONG_BITS == 64);
276
+ tcg_out_call_nodelay(s, qemu_unalign_st_trampoline, false);
277
+ /* delay slot -- move to low part of argument reg */
278
+ tcg_out_mov_delay(s, arg_low, addr);
279
+ } else {
280
+ /* Underalignment: store by pieces of minimum alignment. */
281
+ int st_opc, a_size, s_size, i;
282
+
283
+ /*
284
+ * Force full address into T1 early; avoids problems with
285
+ * overlap between @addr and @data.
286
+ */
287
+ tcg_out_arith(s, TCG_REG_T1, addr, index, ARITH_ADD);
288
+
289
+ a_size = 1 << a_bits;
290
+ s_size = 1 << s_bits;
291
+ if ((memop & MO_BSWAP) == MO_BE) {
292
+ st_opc = qemu_st_opc[a_bits | MO_BE];
293
+ for (i = 0; i < s_size; i += a_size) {
294
+ TCGReg d = data;
295
+ int shift = (s_size - a_size - i) * 8;
296
+ if (shift) {
297
+ d = TCG_REG_T2;
298
+ tcg_out_arithi(s, d, data, shift, SHIFT_SRLX);
299
+ }
300
+ tcg_out_ldst(s, d, TCG_REG_T1, i, st_opc);
301
+ }
302
+ } else if (a_bits == 0) {
303
+ tcg_out_ldst(s, data, TCG_REG_T1, 0, STB);
304
+ for (i = 1; i < s_size; i++) {
305
+ tcg_out_arithi(s, TCG_REG_T2, data, i * 8, SHIFT_SRLX);
306
+ tcg_out_ldst(s, TCG_REG_T2, TCG_REG_T1, i, STB);
307
+ }
308
+ } else {
309
+ /* Note that ST*A with immediate asi must use indexed address. */
310
+ st_opc = qemu_st_opc[a_bits + MO_LE];
311
+ tcg_out_ldst_rr(s, data, TCG_REG_T1, TCG_REG_G0, st_opc);
312
+ for (i = a_size; i < s_size; i += a_size) {
313
+ tcg_out_arithi(s, TCG_REG_T2, data, i * 8, SHIFT_SRLX);
314
+ tcg_out_arithi(s, TCG_REG_T1, TCG_REG_T1, a_size, ARITH_ADD);
315
+ tcg_out_ldst_rr(s, TCG_REG_T2, TCG_REG_T1, TCG_REG_G0, st_opc);
316
+ }
317
+ }
318
+ }
319
+
320
+ *label_ptr |= INSN_OFF19(tcg_ptr_byte_diff(s->code_ptr, label_ptr));
321
#endif /* CONFIG_SOFTMMU */
322
}
323
324
--
54
--
325
2.25.1
55
2.34.1
326
327
diff view generated by jsdifflib
1
We had code for checking for 13 and 21-bit shifted constants,
1
While there are no target-specific nonfaulting probes,
2
but we can do better and allow 32-bit shifted constants.
2
generic code may grow some uses at some point.
3
This is still 2 insns shorter than the full 64-bit sequence.
4
3
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Note that the attrs argument was incorrect -- it should have
5
been MEMTXATTRS_UNSPECIFIED. Just use the simpler interface.
6
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
---
9
tcg/sparc/tcg-target.c.inc | 12 ++++++------
10
target/avr/helper.c | 46 ++++++++++++++++++++++++++++-----------------
10
1 file changed, 6 insertions(+), 6 deletions(-)
11
1 file changed, 29 insertions(+), 17 deletions(-)
11
12
12
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
13
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/sparc/tcg-target.c.inc
15
--- a/target/avr/helper.c
15
+++ b/tcg/sparc/tcg-target.c.inc
16
+++ b/target/avr/helper.c
16
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
17
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
17
return;
18
MMUAccessType access_type, int mmu_idx,
19
bool probe, uintptr_t retaddr)
20
{
21
- int prot = 0;
22
- MemTxAttrs attrs = {};
23
+ int prot, page_size = TARGET_PAGE_SIZE;
24
uint32_t paddr;
25
26
address &= TARGET_PAGE_MASK;
27
28
if (mmu_idx == MMU_CODE_IDX) {
29
- /* access to code in flash */
30
+ /* Access to code in flash. */
31
paddr = OFFSET_CODE + address;
32
prot = PAGE_READ | PAGE_EXEC;
33
- if (paddr + TARGET_PAGE_SIZE > OFFSET_DATA) {
34
+ if (paddr >= OFFSET_DATA) {
35
+ /*
36
+ * This should not be possible via any architectural operations.
37
+ * There is certainly not an exception that we can deliver.
38
+ * Accept probing that might come from generic code.
39
+ */
40
+ if (probe) {
41
+ return false;
42
+ }
43
error_report("execution left flash memory");
44
abort();
45
}
46
- } else if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
47
- /*
48
- * access to CPU registers, exit and rebuilt this TB to use full access
49
- * incase it touches specially handled registers like SREG or SP
50
- */
51
- AVRCPU *cpu = AVR_CPU(cs);
52
- CPUAVRState *env = &cpu->env;
53
- env->fullacc = 1;
54
- cpu_loop_exit_restore(cs, retaddr);
55
} else {
56
- /* access to memory. nothing special */
57
+ /* Access to memory. */
58
paddr = OFFSET_DATA + address;
59
prot = PAGE_READ | PAGE_WRITE;
60
+ if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
61
+ /*
62
+ * Access to CPU registers, exit and rebuilt this TB to use
63
+ * full access in case it touches specially handled registers
64
+ * like SREG or SP. For probing, set page_size = 1, in order
65
+ * to force tlb_fill to be called for the next access.
66
+ */
67
+ if (probe) {
68
+ page_size = 1;
69
+ } else {
70
+ AVRCPU *cpu = AVR_CPU(cs);
71
+ CPUAVRState *env = &cpu->env;
72
+ env->fullacc = 1;
73
+ cpu_loop_exit_restore(cs, retaddr);
74
+ }
75
+ }
18
}
76
}
19
77
20
- /* A 21-bit constant, shifted. */
78
- tlb_set_page_with_attrs(cs, address, paddr, attrs, prot,
21
+ /* A 32-bit constant, shifted. */
79
- mmu_idx, TARGET_PAGE_SIZE);
22
lsb = ctz64(arg);
80
-
23
test = (tcg_target_long)arg >> lsb;
81
+ tlb_set_page(cs, address, paddr, prot, mmu_idx, page_size);
24
- if (check_fit_tl(test, 13)) {
82
return true;
25
- tcg_out_movi_imm13(s, ret, test);
83
}
26
- tcg_out_arithi(s, ret, ret, lsb, SHIFT_SLLX);
84
27
- return;
28
- } else if (lsb > 10 && test == extract64(test, 0, 21)) {
29
+ if (lsb > 10 && test == extract64(test, 0, 21)) {
30
tcg_out_sethi(s, ret, test << 10);
31
tcg_out_arithi(s, ret, ret, lsb - 10, SHIFT_SLLX);
32
return;
33
+ } else if (test == (uint32_t)test || test == (int32_t)test) {
34
+ tcg_out_movi_int(s, TCG_TYPE_I64, ret, test, in_prologue, scratch);
35
+ tcg_out_arithi(s, ret, ret, lsb, SHIFT_SLLX);
36
+ return;
37
}
38
39
/* A 64-bit constant decomposed into 2 32-bit pieces. */
40
--
85
--
41
2.25.1
86
2.34.1
42
87
43
88
diff view generated by jsdifflib
1
There is no need to go through cc->tcg_ops when
2
we know what value that must have.
3
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
1
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
7
---
4
tcg/tci.c | 20 ++++++++++++++------
8
target/avr/helper.c | 5 ++---
5
1 file changed, 14 insertions(+), 6 deletions(-)
9
1 file changed, 2 insertions(+), 3 deletions(-)
6
10
7
diff --git a/tcg/tci.c b/tcg/tci.c
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
8
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/tci.c
13
--- a/target/avr/helper.c
10
+++ b/tcg/tci.c
14
+++ b/target/avr/helper.c
11
@@ -XXX,XX +XXX,XX @@ static bool tci_compare64(uint64_t u0, uint64_t u1, TCGCond condition)
15
@@ -XXX,XX +XXX,XX @@
12
static uint64_t tci_qemu_ld(CPUArchState *env, target_ulong taddr,
16
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
13
MemOpIdx oi, const void *tb_ptr)
14
{
17
{
15
- MemOp mop = get_memop(oi) & (MO_BSWAP | MO_SSIZE);
18
bool ret = false;
16
+ MemOp mop = get_memop(oi);
19
- CPUClass *cc = CPU_GET_CLASS(cs);
17
uintptr_t ra = (uintptr_t)tb_ptr;
20
AVRCPU *cpu = AVR_CPU(cs);
18
21
CPUAVRState *env = &cpu->env;
19
#ifdef CONFIG_SOFTMMU
22
20
- switch (mop) {
23
if (interrupt_request & CPU_INTERRUPT_RESET) {
21
+ switch (mop & (MO_BSWAP | MO_SSIZE)) {
24
if (cpu_interrupts_enabled(env)) {
22
case MO_UB:
25
cs->exception_index = EXCP_RESET;
23
return helper_ret_ldub_mmu(env, taddr, oi, ra);
26
- cc->tcg_ops->do_interrupt(cs);
24
case MO_SB:
27
+ avr_cpu_do_interrupt(cs);
25
@@ -XXX,XX +XXX,XX @@ static uint64_t tci_qemu_ld(CPUArchState *env, target_ulong taddr,
28
26
}
29
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
27
#else
30
28
void *haddr = g2h(env_cpu(env), taddr);
31
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
29
+ unsigned a_mask = (1u << get_alignment_bits(mop)) - 1;
32
if (cpu_interrupts_enabled(env) && env->intsrc != 0) {
30
uint64_t ret;
33
int index = ctz32(env->intsrc);
31
34
cs->exception_index = EXCP_INT(index);
32
set_helper_retaddr(ra);
35
- cc->tcg_ops->do_interrupt(cs);
33
- switch (mop) {
36
+ avr_cpu_do_interrupt(cs);
34
+ if (taddr & a_mask) {
37
35
+ helper_unaligned_ld(env, taddr);
38
env->intsrc &= env->intsrc - 1; /* clear the interrupt */
36
+ }
39
if (!env->intsrc) {
37
+ switch (mop & (MO_BSWAP | MO_SSIZE)) {
38
case MO_UB:
39
ret = ldub_p(haddr);
40
break;
41
@@ -XXX,XX +XXX,XX @@ static uint64_t tci_qemu_ld(CPUArchState *env, target_ulong taddr,
42
static void tci_qemu_st(CPUArchState *env, target_ulong taddr, uint64_t val,
43
MemOpIdx oi, const void *tb_ptr)
44
{
45
- MemOp mop = get_memop(oi) & (MO_BSWAP | MO_SSIZE);
46
+ MemOp mop = get_memop(oi);
47
uintptr_t ra = (uintptr_t)tb_ptr;
48
49
#ifdef CONFIG_SOFTMMU
50
- switch (mop) {
51
+ switch (mop & (MO_BSWAP | MO_SIZE)) {
52
case MO_UB:
53
helper_ret_stb_mmu(env, taddr, val, oi, ra);
54
break;
55
@@ -XXX,XX +XXX,XX @@ static void tci_qemu_st(CPUArchState *env, target_ulong taddr, uint64_t val,
56
}
57
#else
58
void *haddr = g2h(env_cpu(env), taddr);
59
+ unsigned a_mask = (1u << get_alignment_bits(mop)) - 1;
60
61
set_helper_retaddr(ra);
62
- switch (mop) {
63
+ if (taddr & a_mask) {
64
+ helper_unaligned_st(env, taddr);
65
+ }
66
+ switch (mop & (MO_BSWAP | MO_SIZE)) {
67
case MO_UB:
68
stb_p(haddr, val);
69
break;
70
--
40
--
71
2.25.1
41
2.34.1
72
42
73
43
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
We're about to start validating PAGE_EXEC, which means that we've
2
got to mark page zero executable. We had been special casing this
3
entirely within translate.
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
8
---
4
tcg/arm/tcg-target.h | 2 -
9
linux-user/elfload.c | 34 +++++++++++++++++++++++++++++++---
5
tcg/arm/tcg-target.c.inc | 83 +++++++++++++++++++++++++++++++++++++++-
10
1 file changed, 31 insertions(+), 3 deletions(-)
6
2 files changed, 81 insertions(+), 4 deletions(-)
7
11
8
diff --git a/tcg/arm/tcg-target.h b/tcg/arm/tcg-target.h
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
9
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/arm/tcg-target.h
14
--- a/linux-user/elfload.c
11
+++ b/tcg/arm/tcg-target.h
15
+++ b/linux-user/elfload.c
12
@@ -XXX,XX +XXX,XX @@ extern bool use_neon_instructions;
16
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
13
/* not defined -- call should be eliminated at compile time */
17
regs->gr[31] = infop->entry;
14
void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
15
16
-#ifdef CONFIG_SOFTMMU
17
#define TCG_TARGET_NEED_LDST_LABELS
18
-#endif
19
#define TCG_TARGET_NEED_POOL_LABELS
20
21
#endif
22
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
23
index XXXXXXX..XXXXXXX 100644
24
--- a/tcg/arm/tcg-target.c.inc
25
+++ b/tcg/arm/tcg-target.c.inc
26
@@ -XXX,XX +XXX,XX @@
27
*/
28
29
#include "elf.h"
30
+#include "../tcg-ldst.c.inc"
31
#include "../tcg-pool.c.inc"
32
33
int arm_arch = __ARM_ARCH;
34
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vldst(TCGContext *s, ARMInsn insn,
35
}
18
}
36
19
37
#ifdef CONFIG_SOFTMMU
20
+#define LO_COMMPAGE 0
38
-#include "../tcg-ldst.c.inc"
39
-
40
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
41
* int mmu_idx, uintptr_t ra)
42
*/
43
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
44
tcg_out_goto(s, COND_AL, qemu_st_helpers[opc & MO_SIZE]);
45
return true;
46
}
47
+#else
48
+
21
+
49
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addrlo,
22
+static bool init_guest_commpage(void)
50
+ TCGReg addrhi, unsigned a_bits)
51
+{
23
+{
52
+ unsigned a_mask = (1 << a_bits) - 1;
24
+ void *want = g2h_untagged(LO_COMMPAGE);
53
+ TCGLabelQemuLdst *label = new_ldst_label(s);
25
+ void *addr = mmap(want, qemu_host_page_size, PROT_NONE,
26
+ MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
54
+
27
+
55
+ label->is_ld = is_ld;
28
+ if (addr == MAP_FAILED) {
56
+ label->addrlo_reg = addrlo;
29
+ perror("Allocating guest commpage");
57
+ label->addrhi_reg = addrhi;
30
+ exit(EXIT_FAILURE);
58
+
31
+ }
59
+ /* We are expecting a_bits to max out at 7, and can easily support 8. */
32
+ if (addr != want) {
60
+ tcg_debug_assert(a_mask <= 0xff);
61
+ /* tst addr, #mask */
62
+ tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addrlo, a_mask);
63
+
64
+ /* blne slow_path */
65
+ label->label_ptr[0] = s->code_ptr;
66
+ tcg_out_bl_imm(s, COND_NE, 0);
67
+
68
+ label->raddr = tcg_splitwx_to_rx(s->code_ptr);
69
+}
70
+
71
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
72
+{
73
+ if (!reloc_pc24(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
74
+ return false;
33
+ return false;
75
+ }
34
+ }
76
+
35
+
77
+ if (TARGET_LONG_BITS == 64) {
78
+ /* 64-bit target address is aligned into R2:R3. */
79
+ if (l->addrhi_reg != TCG_REG_R2) {
80
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R2, l->addrlo_reg);
81
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R3, l->addrhi_reg);
82
+ } else if (l->addrlo_reg != TCG_REG_R3) {
83
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R3, l->addrhi_reg);
84
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R2, l->addrlo_reg);
85
+ } else {
86
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R1, TCG_REG_R2);
87
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R2, TCG_REG_R3);
88
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R3, TCG_REG_R1);
89
+ }
90
+ } else {
91
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R1, l->addrlo_reg);
92
+ }
93
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_R0, TCG_AREG0);
94
+
95
+ /*
36
+ /*
96
+ * Tail call to the helper, with the return address back inline,
37
+ * On Linux, page zero is normally marked execute only + gateway.
97
+ * just for the clarity of the debugging traceback -- the helper
38
+ * Normal read or write is supposed to fail (thus PROT_NONE above),
98
+ * cannot return. We have used BLNE to arrive here, so LR is
39
+ * but specific offsets have kernel code mapped to raise permissions
99
+ * already set.
40
+ * and implement syscalls. Here, simply mark the page executable.
41
+ * Special case the entry points during translation (see do_page_zero).
100
+ */
42
+ */
101
+ tcg_out_goto(s, COND_AL, (const void *)
43
+ page_set_flags(LO_COMMPAGE, LO_COMMPAGE + TARGET_PAGE_SIZE,
102
+ (l->is_ld ? helper_unaligned_ld : helper_unaligned_st));
44
+ PAGE_EXEC | PAGE_VALID);
103
+ return true;
45
+ return true;
104
+}
46
+}
105
+
47
+
106
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
48
#endif /* TARGET_HPPA */
107
+{
49
108
+ return tcg_out_fail_alignment(s, l);
50
#ifdef TARGET_XTENSA
109
+}
51
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
110
+
52
}
111
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
53
112
+{
54
#if defined(HI_COMMPAGE)
113
+ return tcg_out_fail_alignment(s, l);
55
-#define LO_COMMPAGE 0
114
+}
56
+#define LO_COMMPAGE -1
115
#endif /* SOFTMMU */
57
#elif defined(LO_COMMPAGE)
116
58
#define HI_COMMPAGE 0
117
static void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
59
#else
118
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
60
#define HI_COMMPAGE 0
119
int mem_index;
61
-#define LO_COMMPAGE 0
120
TCGReg addend;
62
+#define LO_COMMPAGE -1
121
tcg_insn_unit *label_ptr;
63
#define init_guest_commpage() true
122
+#else
123
+ unsigned a_bits;
124
#endif
64
#endif
125
65
126
datalo = *args++;
66
@@ -XXX,XX +XXX,XX @@ static void pgb_static(const char *image_name, abi_ulong orig_loaddr,
127
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
67
} else {
128
add_qemu_ldst_label(s, true, oi, datalo, datahi, addrlo, addrhi,
68
offset = -(HI_COMMPAGE & -align);
129
s->code_ptr, label_ptr);
69
}
130
#else /* !CONFIG_SOFTMMU */
70
- } else if (LO_COMMPAGE != 0) {
131
+ a_bits = get_alignment_bits(opc);
71
+ } else if (LO_COMMPAGE != -1) {
132
+ if (a_bits) {
72
loaddr = MIN(loaddr, LO_COMMPAGE & -align);
133
+ tcg_out_test_alignment(s, true, addrlo, addrhi, a_bits);
73
}
134
+ }
74
135
if (guest_base) {
136
tcg_out_qemu_ld_index(s, opc, datalo, datahi,
137
addrlo, TCG_REG_GUEST_BASE, false);
138
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
139
int mem_index;
140
TCGReg addend;
141
tcg_insn_unit *label_ptr;
142
+#else
143
+ unsigned a_bits;
144
#endif
145
146
datalo = *args++;
147
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
148
add_qemu_ldst_label(s, false, oi, datalo, datahi, addrlo, addrhi,
149
s->code_ptr, label_ptr);
150
#else /* !CONFIG_SOFTMMU */
151
+ a_bits = get_alignment_bits(opc);
152
+ if (a_bits) {
153
+ tcg_out_test_alignment(s, false, addrlo, addrhi, a_bits);
154
+ }
155
if (guest_base) {
156
tcg_out_qemu_st_index(s, COND_AL, opc, datalo, datahi,
157
addrlo, TCG_REG_GUEST_BASE, false);
158
--
75
--
159
2.25.1
76
2.34.1
160
161
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
We're about to start validating PAGE_EXEC, which means that we've
2
got to mark the vsyscall page executable. We had been special
3
casing this entirely within translate.
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
8
---
4
tcg/i386/tcg-target.h | 2 -
9
linux-user/elfload.c | 23 +++++++++++++++++++++++
5
tcg/i386/tcg-target.c.inc | 103 ++++++++++++++++++++++++++++++++++++--
10
1 file changed, 23 insertions(+)
6
2 files changed, 98 insertions(+), 7 deletions(-)
7
11
8
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
9
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/i386/tcg-target.h
14
--- a/linux-user/elfload.c
11
+++ b/tcg/i386/tcg-target.h
15
+++ b/linux-user/elfload.c
12
@@ -XXX,XX +XXX,XX @@ static inline void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_rx,
16
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
13
17
(*regs)[26] = tswapreg(env->segs[R_GS].selector & 0xffff);
14
#define TCG_TARGET_HAS_MEMORY_BSWAP have_movbe
15
16
-#ifdef CONFIG_SOFTMMU
17
#define TCG_TARGET_NEED_LDST_LABELS
18
-#endif
19
#define TCG_TARGET_NEED_POOL_LABELS
20
21
#endif
22
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
23
index XXXXXXX..XXXXXXX 100644
24
--- a/tcg/i386/tcg-target.c.inc
25
+++ b/tcg/i386/tcg-target.c.inc
26
@@ -XXX,XX +XXX,XX @@
27
* THE SOFTWARE.
28
*/
29
30
+#include "../tcg-ldst.c.inc"
31
#include "../tcg-pool.c.inc"
32
33
#ifdef CONFIG_DEBUG_TCG
34
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
35
#define OPC_VZEROUPPER (0x77 | P_EXT)
36
#define OPC_XCHG_ax_r32    (0x90)
37
38
-#define OPC_GRP3_Ev    (0xf7)
39
-#define OPC_GRP5    (0xff)
40
+#define OPC_GRP3_Eb (0xf6)
41
+#define OPC_GRP3_Ev (0xf7)
42
+#define OPC_GRP5 (0xff)
43
#define OPC_GRP14 (0x73 | P_EXT | P_DATA16)
44
45
/* Group 1 opcode extensions for 0x80-0x83.
46
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
47
#define SHIFT_SAR 7
48
49
/* Group 3 opcode extensions for 0xf6, 0xf7. To be used with OPC_GRP3. */
50
+#define EXT3_TESTi 0
51
#define EXT3_NOT 2
52
#define EXT3_NEG 3
53
#define EXT3_MUL 4
54
@@ -XXX,XX +XXX,XX @@ static void tcg_out_nopn(TCGContext *s, int n)
55
}
18
}
56
19
57
#if defined(CONFIG_SOFTMMU)
20
+#if ULONG_MAX >= TARGET_VSYSCALL_PAGE
58
-#include "../tcg-ldst.c.inc"
21
+#define INIT_GUEST_COMMPAGE
59
-
22
+static bool init_guest_commpage(void)
60
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
61
* int mmu_idx, uintptr_t ra)
62
*/
63
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
64
tcg_out_jmp(s, qemu_st_helpers[opc & (MO_BSWAP | MO_SIZE)]);
65
return true;
66
}
67
-#elif TCG_TARGET_REG_BITS == 32
68
+#else
69
+
70
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addrlo,
71
+ TCGReg addrhi, unsigned a_bits)
72
+{
23
+{
73
+ unsigned a_mask = (1 << a_bits) - 1;
74
+ TCGLabelQemuLdst *label;
75
+
76
+ /*
24
+ /*
77
+ * We are expecting a_bits to max out at 7, so we can usually use testb.
25
+ * The vsyscall page is at a high negative address aka kernel space,
78
+ * For i686, we have to use testl for %esi/%edi.
26
+ * which means that we cannot actually allocate it with target_mmap.
27
+ * We still should be able to use page_set_flags, unless the user
28
+ * has specified -R reserved_va, which would trigger an assert().
79
+ */
29
+ */
80
+ if (a_mask <= 0xff && (TCG_TARGET_REG_BITS == 64 || addrlo < 4)) {
30
+ if (reserved_va != 0 &&
81
+ tcg_out_modrm(s, OPC_GRP3_Eb | P_REXB_RM, EXT3_TESTi, addrlo);
31
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE >= reserved_va) {
82
+ tcg_out8(s, a_mask);
32
+ error_report("Cannot allocate vsyscall page");
83
+ } else {
33
+ exit(EXIT_FAILURE);
84
+ tcg_out_modrm(s, OPC_GRP3_Ev, EXT3_TESTi, addrlo);
85
+ tcg_out32(s, a_mask);
86
+ }
34
+ }
87
+
35
+ page_set_flags(TARGET_VSYSCALL_PAGE,
88
+ /* jne slow_path */
36
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE,
89
+ tcg_out_opc(s, OPC_JCC_long + JCC_JNE, 0, 0, 0);
37
+ PAGE_EXEC | PAGE_VALID);
90
+
91
+ label = new_ldst_label(s);
92
+ label->is_ld = is_ld;
93
+ label->addrlo_reg = addrlo;
94
+ label->addrhi_reg = addrhi;
95
+ label->raddr = tcg_splitwx_to_rx(s->code_ptr + 4);
96
+ label->label_ptr[0] = s->code_ptr;
97
+
98
+ s->code_ptr += 4;
99
+}
100
+
101
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
102
+{
103
+ /* resolve label address */
104
+ tcg_patch32(l->label_ptr[0], s->code_ptr - l->label_ptr[0] - 4);
105
+
106
+ if (TCG_TARGET_REG_BITS == 32) {
107
+ int ofs = 0;
108
+
109
+ tcg_out_st(s, TCG_TYPE_PTR, TCG_AREG0, TCG_REG_ESP, ofs);
110
+ ofs += 4;
111
+
112
+ tcg_out_st(s, TCG_TYPE_I32, l->addrlo_reg, TCG_REG_ESP, ofs);
113
+ ofs += 4;
114
+ if (TARGET_LONG_BITS == 64) {
115
+ tcg_out_st(s, TCG_TYPE_I32, l->addrhi_reg, TCG_REG_ESP, ofs);
116
+ ofs += 4;
117
+ }
118
+
119
+ tcg_out_pushi(s, (uintptr_t)l->raddr);
120
+ } else {
121
+ tcg_out_mov(s, TCG_TYPE_TL, tcg_target_call_iarg_regs[1],
122
+ l->addrlo_reg);
123
+ tcg_out_mov(s, TCG_TYPE_PTR, tcg_target_call_iarg_regs[0], TCG_AREG0);
124
+
125
+ tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RAX, (uintptr_t)l->raddr);
126
+ tcg_out_push(s, TCG_REG_RAX);
127
+ }
128
+
129
+ /* "Tail call" to the helper, with the return address back inline. */
130
+ tcg_out_jmp(s, (const void *)(l->is_ld ? helper_unaligned_ld
131
+ : helper_unaligned_st));
132
+ return true;
38
+ return true;
133
+}
39
+}
134
+
135
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
136
+{
137
+ return tcg_out_fail_alignment(s, l);
138
+}
139
+
140
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
141
+{
142
+ return tcg_out_fail_alignment(s, l);
143
+}
144
+
145
+#if TCG_TARGET_REG_BITS == 32
146
# define x86_guest_base_seg 0
147
# define x86_guest_base_index -1
148
# define x86_guest_base_offset guest_base
149
@@ -XXX,XX +XXX,XX @@ static inline int setup_guest_base_seg(void)
150
return 0;
151
}
152
# endif
153
+#endif
40
+#endif
154
#endif /* SOFTMMU */
41
#else
155
42
156
static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg datalo, TCGReg datahi,
43
#define ELF_START_MMAP 0x80000000
157
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
44
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
158
#if defined(CONFIG_SOFTMMU)
45
#else
159
int mem_index;
46
#define HI_COMMPAGE 0
160
tcg_insn_unit *label_ptr[2];
47
#define LO_COMMPAGE -1
161
+#else
48
+#ifndef INIT_GUEST_COMMPAGE
162
+ unsigned a_bits;
49
#define init_guest_commpage() true
163
#endif
50
#endif
164
51
+#endif
165
datalo = *args++;
52
166
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
53
static void pgb_fail_in_use(const char *image_name)
167
add_qemu_ldst_label(s, true, is64, oi, datalo, datahi, addrlo, addrhi,
54
{
168
s->code_ptr, label_ptr);
169
#else
170
+ a_bits = get_alignment_bits(opc);
171
+ if (a_bits) {
172
+ tcg_out_test_alignment(s, true, addrlo, addrhi, a_bits);
173
+ }
174
+
175
tcg_out_qemu_ld_direct(s, datalo, datahi, addrlo, x86_guest_base_index,
176
x86_guest_base_offset, x86_guest_base_seg,
177
is64, opc);
178
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
179
#if defined(CONFIG_SOFTMMU)
180
int mem_index;
181
tcg_insn_unit *label_ptr[2];
182
+#else
183
+ unsigned a_bits;
184
#endif
185
186
datalo = *args++;
187
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
188
add_qemu_ldst_label(s, false, is64, oi, datalo, datahi, addrlo, addrhi,
189
s->code_ptr, label_ptr);
190
#else
191
+ a_bits = get_alignment_bits(opc);
192
+ if (a_bits) {
193
+ tcg_out_test_alignment(s, false, addrlo, addrhi, a_bits);
194
+ }
195
+
196
tcg_out_qemu_st_direct(s, datalo, datahi, addrlo, x86_guest_base_index,
197
x86_guest_base_offset, x86_guest_base_seg, opc);
198
#endif
199
--
55
--
200
2.25.1
56
2.34.1
201
202
diff view generated by jsdifflib
1
We can use the routines just added for user-only to emit
1
We cannot deliver two interrupts simultaneously;
2
unaligned accesses in softmmu mode too.
2
the first interrupt handler must execute first.
3
3
4
Tested-by: Jiaxun Yang <jiaxun.yang@flygoat.com>
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
5
Reviewed-by: Jiaxun Yang <jiaxun.yang@flygoat.com>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
7
---
9
tcg/mips/tcg-target.c.inc | 91 ++++++++++++++++++++++-----------------
8
target/avr/helper.c | 9 +++------
10
1 file changed, 51 insertions(+), 40 deletions(-)
9
1 file changed, 3 insertions(+), 6 deletions(-)
11
10
12
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/mips/tcg-target.c.inc
13
--- a/target/avr/helper.c
15
+++ b/tcg/mips/tcg-target.c.inc
14
+++ b/target/avr/helper.c
16
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
15
@@ -XXX,XX +XXX,XX @@
17
tcg_insn_unit *label_ptr[2], bool is_load)
16
17
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
18
{
18
{
19
MemOp opc = get_memop(oi);
19
- bool ret = false;
20
- unsigned s_bits = opc & MO_SIZE;
20
AVRCPU *cpu = AVR_CPU(cs);
21
unsigned a_bits = get_alignment_bits(opc);
21
CPUAVRState *env = &cpu->env;
22
+ unsigned s_bits = opc & MO_SIZE;
22
23
+ unsigned a_mask = (1 << a_bits) - 1;
23
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
24
+ unsigned s_mask = (1 << s_bits) - 1;
24
avr_cpu_do_interrupt(cs);
25
int mem_index = get_mmuidx(oi);
25
26
int fast_off = TLB_MASK_TABLE_OFS(mem_index);
26
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
27
int mask_off = fast_off + offsetof(CPUTLBDescFast, mask);
28
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
29
int add_off = offsetof(CPUTLBEntry, addend);
30
int cmp_off = (is_load ? offsetof(CPUTLBEntry, addr_read)
31
: offsetof(CPUTLBEntry, addr_write));
32
- target_ulong mask;
33
+ target_ulong tlb_mask;
34
35
/* Load tlb_mask[mmu_idx] and tlb_table[mmu_idx]. */
36
tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP0, TCG_AREG0, mask_off);
37
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
38
/* Add the tlb_table pointer, creating the CPUTLBEntry address in TMP3. */
39
tcg_out_opc_reg(s, ALIAS_PADD, TCG_TMP3, TCG_TMP3, TCG_TMP1);
40
41
- /* We don't currently support unaligned accesses.
42
- We could do so with mips32r6. */
43
- if (a_bits < s_bits) {
44
- a_bits = s_bits;
45
- }
46
-
27
-
47
- /* Mask the page bits, keeping the alignment bits to compare against. */
28
- ret = true;
48
- mask = (target_ulong)TARGET_PAGE_MASK | ((1 << a_bits) - 1);
29
+ return true;
30
}
31
}
32
if (interrupt_request & CPU_INTERRUPT_HARD) {
33
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
34
if (!env->intsrc) {
35
cs->interrupt_request &= ~CPU_INTERRUPT_HARD;
36
}
49
-
37
-
50
/* Load the (low-half) tlb comparator. */
38
- ret = true;
51
if (TCG_TARGET_REG_BITS < TARGET_LONG_BITS) {
39
+ return true;
52
- tcg_out_ld(s, TCG_TYPE_I32, TCG_TMP0, TCG_TMP3, cmp_off + LO_OFF);
40
}
53
- tcg_out_movi(s, TCG_TYPE_I32, TCG_TMP1, mask);
54
+ tcg_out_ldst(s, OPC_LW, TCG_TMP0, TCG_TMP3, cmp_off + LO_OFF);
55
} else {
56
tcg_out_ldst(s, (TARGET_LONG_BITS == 64 ? OPC_LD
57
: TCG_TARGET_REG_BITS == 64 ? OPC_LWU : OPC_LW),
58
TCG_TMP0, TCG_TMP3, cmp_off);
59
- tcg_out_movi(s, TCG_TYPE_TL, TCG_TMP1, mask);
60
- /* No second compare is required here;
61
- load the tlb addend for the fast path. */
62
- tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP2, TCG_TMP3, add_off);
63
}
41
}
64
42
- return ret;
65
/* Zero extend a 32-bit guest address for a 64-bit host. */
43
+ return false;
66
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
67
tcg_out_ext32u(s, base, addrl);
68
addrl = base;
69
}
70
- tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, addrl);
71
+
72
+ /*
73
+ * Mask the page bits, keeping the alignment bits to compare against.
74
+ * For unaligned accesses, compare against the end of the access to
75
+ * verify that it does not cross a page boundary.
76
+ */
77
+ tlb_mask = (target_ulong)TARGET_PAGE_MASK | a_mask;
78
+ tcg_out_movi(s, TCG_TYPE_I32, TCG_TMP1, tlb_mask);
79
+ if (a_mask >= s_mask) {
80
+ tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, addrl);
81
+ } else {
82
+ tcg_out_opc_imm(s, ALIAS_PADDI, TCG_TMP2, addrl, s_mask - a_mask);
83
+ tcg_out_opc_reg(s, OPC_AND, TCG_TMP1, TCG_TMP1, TCG_TMP2);
84
+ }
85
+
86
+ if (TCG_TARGET_REG_BITS >= TARGET_LONG_BITS) {
87
+ /* Load the tlb addend for the fast path. */
88
+ tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP2, TCG_TMP3, add_off);
89
+ }
90
91
label_ptr[0] = s->code_ptr;
92
tcg_out_opc_br(s, OPC_BNE, TCG_TMP1, TCG_TMP0);
93
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
94
/* Load and test the high half tlb comparator. */
95
if (TCG_TARGET_REG_BITS < TARGET_LONG_BITS) {
96
/* delay slot */
97
- tcg_out_ld(s, TCG_TYPE_I32, TCG_TMP0, TCG_TMP3, cmp_off + HI_OFF);
98
+ tcg_out_ldst(s, OPC_LW, TCG_TMP0, TCG_TMP3, cmp_off + HI_OFF);
99
100
/* Load the tlb addend for the fast path. */
101
tcg_out_ld(s, TCG_TYPE_PTR, TCG_TMP2, TCG_TMP3, add_off);
102
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
103
}
104
}
44
}
105
45
106
-static void __attribute__((unused))
46
void avr_cpu_do_interrupt(CPUState *cs)
107
-tcg_out_qemu_ld_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
108
+static void tcg_out_qemu_ld_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
109
TCGReg base, MemOp opc, bool is_64)
110
{
111
const MIPSInsn lw1 = MIPS_BE ? OPC_LWL : OPC_LWR;
112
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
113
#if defined(CONFIG_SOFTMMU)
114
tcg_insn_unit *label_ptr[2];
115
#else
116
- unsigned a_bits, s_bits;
117
#endif
118
+ unsigned a_bits, s_bits;
119
TCGReg base = TCG_REG_A0;
120
121
data_regl = *args++;
122
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
123
addr_regh = (TCG_TARGET_REG_BITS < TARGET_LONG_BITS ? *args++ : 0);
124
oi = *args++;
125
opc = get_memop(oi);
126
+ a_bits = get_alignment_bits(opc);
127
+ s_bits = opc & MO_SIZE;
128
129
+ /*
130
+ * R6 removes the left/right instructions but requires the
131
+ * system to support misaligned memory accesses.
132
+ */
133
#if defined(CONFIG_SOFTMMU)
134
tcg_out_tlb_load(s, base, addr_regl, addr_regh, oi, label_ptr, 1);
135
- tcg_out_qemu_ld_direct(s, data_regl, data_regh, base, opc, is_64);
136
+ if (use_mips32r6_instructions || a_bits >= s_bits) {
137
+ tcg_out_qemu_ld_direct(s, data_regl, data_regh, base, opc, is_64);
138
+ } else {
139
+ tcg_out_qemu_ld_unalign(s, data_regl, data_regh, base, opc, is_64);
140
+ }
141
add_qemu_ldst_label(s, 1, oi,
142
(is_64 ? TCG_TYPE_I64 : TCG_TYPE_I32),
143
data_regl, data_regh, addr_regl, addr_regh,
144
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
145
} else {
146
tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_GUEST_BASE_REG, addr_regl);
147
}
148
- a_bits = get_alignment_bits(opc);
149
- s_bits = opc & MO_SIZE;
150
- /*
151
- * R6 removes the left/right instructions but requires the
152
- * system to support misaligned memory accesses.
153
- */
154
if (use_mips32r6_instructions) {
155
if (a_bits) {
156
tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
157
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
158
}
159
}
160
161
-static void __attribute__((unused))
162
-tcg_out_qemu_st_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
163
+static void tcg_out_qemu_st_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
164
TCGReg base, MemOp opc)
165
{
166
const MIPSInsn sw1 = MIPS_BE ? OPC_SWL : OPC_SWR;
167
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
168
MemOp opc;
169
#if defined(CONFIG_SOFTMMU)
170
tcg_insn_unit *label_ptr[2];
171
-#else
172
- unsigned a_bits, s_bits;
173
#endif
174
+ unsigned a_bits, s_bits;
175
TCGReg base = TCG_REG_A0;
176
177
data_regl = *args++;
178
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
179
addr_regh = (TCG_TARGET_REG_BITS < TARGET_LONG_BITS ? *args++ : 0);
180
oi = *args++;
181
opc = get_memop(oi);
182
+ a_bits = get_alignment_bits(opc);
183
+ s_bits = opc & MO_SIZE;
184
185
+ /*
186
+ * R6 removes the left/right instructions but requires the
187
+ * system to support misaligned memory accesses.
188
+ */
189
#if defined(CONFIG_SOFTMMU)
190
tcg_out_tlb_load(s, base, addr_regl, addr_regh, oi, label_ptr, 0);
191
- tcg_out_qemu_st_direct(s, data_regl, data_regh, base, opc);
192
+ if (use_mips32r6_instructions || a_bits >= s_bits) {
193
+ tcg_out_qemu_st_direct(s, data_regl, data_regh, base, opc);
194
+ } else {
195
+ tcg_out_qemu_st_unalign(s, data_regl, data_regh, base, opc);
196
+ }
197
add_qemu_ldst_label(s, 0, oi,
198
(is_64 ? TCG_TYPE_I64 : TCG_TYPE_I32),
199
data_regl, data_regh, addr_regl, addr_regh,
200
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
201
} else {
202
tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_GUEST_BASE_REG, addr_regl);
203
}
204
- a_bits = get_alignment_bits(opc);
205
- s_bits = opc & MO_SIZE;
206
- /*
207
- * R6 removes the left/right instructions but requires the
208
- * system to support misaligned memory accesses.
209
- */
210
if (use_mips32r6_instructions) {
211
if (a_bits) {
212
tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
213
--
47
--
214
2.25.1
48
2.34.1
215
49
216
50
diff view generated by jsdifflib
1
This is kinda sorta the opposite of the other tcg hosts, where
1
This bit is not saved across interrupts, so we must
2
we get (normal) alignment checks for free with host SIGBUS and
2
delay delivering the interrupt until the skip has
3
need to add code to support unaligned accesses.
3
been processed.
4
4
5
Fortunately, the ISA contains pairs of instructions that are
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1118
6
used to implement unaligned memory accesses. Use them.
6
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
7
8
Tested-by: Jiaxun Yang <jiaxun.yang@flygoat.com>
9
Reviewed-by: Jiaxun Yang <jiaxun.yang@flygoat.com>
10
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
---
9
---
13
tcg/mips/tcg-target.h | 2 -
10
target/avr/helper.c | 9 +++++++++
14
tcg/mips/tcg-target.c.inc | 334 +++++++++++++++++++++++++++++++++++++-
11
target/avr/translate.c | 26 ++++++++++++++++++++++----
15
2 files changed, 328 insertions(+), 8 deletions(-)
12
2 files changed, 31 insertions(+), 4 deletions(-)
16
13
17
diff --git a/tcg/mips/tcg-target.h b/tcg/mips/tcg-target.h
14
diff --git a/target/avr/helper.c b/target/avr/helper.c
18
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
19
--- a/tcg/mips/tcg-target.h
16
--- a/target/avr/helper.c
20
+++ b/tcg/mips/tcg-target.h
17
+++ b/target/avr/helper.c
21
@@ -XXX,XX +XXX,XX @@ extern bool use_mips32r2_instructions;
18
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
22
void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t)
19
AVRCPU *cpu = AVR_CPU(cs);
23
QEMU_ERROR("code path is reachable");
20
CPUAVRState *env = &cpu->env;
24
21
25
-#ifdef CONFIG_SOFTMMU
22
+ /*
26
#define TCG_TARGET_NEED_LDST_LABELS
23
+ * We cannot separate a skip from the next instruction,
27
-#endif
24
+ * as the skip would not be preserved across the interrupt.
28
25
+ * Separating the two insn normally only happens at page boundaries.
29
#endif
26
+ */
30
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
27
+ if (env->skip) {
31
index XXXXXXX..XXXXXXX 100644
32
--- a/tcg/mips/tcg-target.c.inc
33
+++ b/tcg/mips/tcg-target.c.inc
34
@@ -XXX,XX +XXX,XX @@
35
* THE SOFTWARE.
36
*/
37
38
+#include "../tcg-ldst.c.inc"
39
+
40
#ifdef HOST_WORDS_BIGENDIAN
41
# define MIPS_BE 1
42
#else
43
@@ -XXX,XX +XXX,XX @@ typedef enum {
44
OPC_ORI = 015 << 26,
45
OPC_XORI = 016 << 26,
46
OPC_LUI = 017 << 26,
47
+ OPC_BNEL = 025 << 26,
48
+ OPC_BNEZALC_R6 = 030 << 26,
49
OPC_DADDIU = 031 << 26,
50
+ OPC_LDL = 032 << 26,
51
+ OPC_LDR = 033 << 26,
52
OPC_LB = 040 << 26,
53
OPC_LH = 041 << 26,
54
+ OPC_LWL = 042 << 26,
55
OPC_LW = 043 << 26,
56
OPC_LBU = 044 << 26,
57
OPC_LHU = 045 << 26,
58
+ OPC_LWR = 046 << 26,
59
OPC_LWU = 047 << 26,
60
OPC_SB = 050 << 26,
61
OPC_SH = 051 << 26,
62
+ OPC_SWL = 052 << 26,
63
OPC_SW = 053 << 26,
64
+ OPC_SDL = 054 << 26,
65
+ OPC_SDR = 055 << 26,
66
+ OPC_SWR = 056 << 26,
67
OPC_LD = 067 << 26,
68
OPC_SD = 077 << 26,
69
70
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call(TCGContext *s, const tcg_insn_unit *arg)
71
}
72
73
#if defined(CONFIG_SOFTMMU)
74
-#include "../tcg-ldst.c.inc"
75
-
76
static void * const qemu_ld_helpers[(MO_SSIZE | MO_BSWAP) + 1] = {
77
[MO_UB] = helper_ret_ldub_mmu,
78
[MO_SB] = helper_ret_ldsb_mmu,
79
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
80
tcg_out_mov(s, TCG_TYPE_PTR, tcg_target_call_iarg_regs[0], TCG_AREG0);
81
return true;
82
}
83
-#endif
84
+
85
+#else
86
+
87
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addrlo,
88
+ TCGReg addrhi, unsigned a_bits)
89
+{
90
+ unsigned a_mask = (1 << a_bits) - 1;
91
+ TCGLabelQemuLdst *l = new_ldst_label(s);
92
+
93
+ l->is_ld = is_ld;
94
+ l->addrlo_reg = addrlo;
95
+ l->addrhi_reg = addrhi;
96
+
97
+ /* We are expecting a_bits to max out at 7, much lower than ANDI. */
98
+ tcg_debug_assert(a_bits < 16);
99
+ tcg_out_opc_imm(s, OPC_ANDI, TCG_TMP0, addrlo, a_mask);
100
+
101
+ l->label_ptr[0] = s->code_ptr;
102
+ if (use_mips32r6_instructions) {
103
+ tcg_out_opc_br(s, OPC_BNEZALC_R6, TCG_REG_ZERO, TCG_TMP0);
104
+ } else {
105
+ tcg_out_opc_br(s, OPC_BNEL, TCG_TMP0, TCG_REG_ZERO);
106
+ tcg_out_nop(s);
107
+ }
108
+
109
+ l->raddr = tcg_splitwx_to_rx(s->code_ptr);
110
+}
111
+
112
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
113
+{
114
+ void *target;
115
+
116
+ if (!reloc_pc16(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
117
+ return false;
28
+ return false;
118
+ }
29
+ }
119
+
30
+
120
+ if (TCG_TARGET_REG_BITS < TARGET_LONG_BITS) {
31
if (interrupt_request & CPU_INTERRUPT_RESET) {
121
+ /* A0 is env, A1 is skipped, A2:A3 is the uint64_t address. */
32
if (cpu_interrupts_enabled(env)) {
122
+ TCGReg a2 = MIPS_BE ? l->addrhi_reg : l->addrlo_reg;
33
cs->exception_index = EXCP_RESET;
123
+ TCGReg a3 = MIPS_BE ? l->addrlo_reg : l->addrhi_reg;
34
diff --git a/target/avr/translate.c b/target/avr/translate.c
35
index XXXXXXX..XXXXXXX 100644
36
--- a/target/avr/translate.c
37
+++ b/target/avr/translate.c
38
@@ -XXX,XX +XXX,XX @@ static void avr_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
39
if (skip_label) {
40
canonicalize_skip(ctx);
41
gen_set_label(skip_label);
42
- if (ctx->base.is_jmp == DISAS_NORETURN) {
124
+
43
+
125
+ if (a3 != TCG_REG_A2) {
44
+ switch (ctx->base.is_jmp) {
126
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A2, a2);
45
+ case DISAS_NORETURN:
127
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A3, a3);
46
ctx->base.is_jmp = DISAS_CHAIN;
128
+ } else if (a2 != TCG_REG_A3) {
47
+ break;
129
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A3, a3);
48
+ case DISAS_NEXT:
130
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A2, a2);
49
+ if (ctx->base.tb->flags & TB_FLAGS_SKIP) {
131
+ } else {
50
+ ctx->base.is_jmp = DISAS_TOO_MANY;
132
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_TMP0, TCG_REG_A2);
51
+ }
133
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A2, TCG_REG_A3);
52
+ break;
134
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_A3, TCG_TMP0);
53
+ default:
135
+ }
54
+ break;
136
+ } else {
55
}
137
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_A1, l->addrlo_reg);
56
}
138
+ }
57
139
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_A0, TCG_AREG0);
58
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
140
+
59
{
60
DisasContext *ctx = container_of(dcbase, DisasContext, base);
61
bool nonconst_skip = canonicalize_skip(ctx);
141
+ /*
62
+ /*
142
+ * Tail call to the helper, with the return address back inline.
63
+ * Because we disable interrupts while env->skip is set,
143
+ * We have arrived here via BNEL, so $31 is already set.
64
+ * we must return to the main loop to re-evaluate afterward.
144
+ */
65
+ */
145
+ target = (l->is_ld ? helper_unaligned_ld : helper_unaligned_st);
66
+ bool force_exit = ctx->base.tb->flags & TB_FLAGS_SKIP;
146
+ tcg_out_call_int(s, target, true);
67
147
+ return true;
68
switch (ctx->base.is_jmp) {
148
+}
69
case DISAS_NORETURN:
149
+
70
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
150
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
71
case DISAS_NEXT:
151
+{
72
case DISAS_TOO_MANY:
152
+ return tcg_out_fail_alignment(s, l);
73
case DISAS_CHAIN:
153
+}
74
- if (!nonconst_skip) {
154
+
75
+ if (!nonconst_skip && !force_exit) {
155
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
76
/* Note gen_goto_tb checks singlestep. */
156
+{
77
gen_goto_tb(ctx, 1, ctx->npc);
157
+ return tcg_out_fail_alignment(s, l);
78
break;
158
+}
79
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
159
+#endif /* SOFTMMU */
80
tcg_gen_movi_tl(cpu_pc, ctx->npc);
160
81
/* fall through */
161
static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
82
case DISAS_LOOKUP:
162
TCGReg base, MemOp opc, bool is_64)
83
- tcg_gen_lookup_and_goto_ptr();
163
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
84
- break;
164
}
85
+ if (!force_exit) {
165
}
86
+ tcg_gen_lookup_and_goto_ptr();
166
167
+static void __attribute__((unused))
168
+tcg_out_qemu_ld_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
169
+ TCGReg base, MemOp opc, bool is_64)
170
+{
171
+ const MIPSInsn lw1 = MIPS_BE ? OPC_LWL : OPC_LWR;
172
+ const MIPSInsn lw2 = MIPS_BE ? OPC_LWR : OPC_LWL;
173
+ const MIPSInsn ld1 = MIPS_BE ? OPC_LDL : OPC_LDR;
174
+ const MIPSInsn ld2 = MIPS_BE ? OPC_LDR : OPC_LDL;
175
+
176
+ bool sgn = (opc & MO_SIGN);
177
+
178
+ switch (opc & (MO_SSIZE | MO_BSWAP)) {
179
+ case MO_SW | MO_BE:
180
+ case MO_UW | MO_BE:
181
+ tcg_out_opc_imm(s, sgn ? OPC_LB : OPC_LBU, TCG_TMP0, base, 0);
182
+ tcg_out_opc_imm(s, OPC_LBU, lo, base, 1);
183
+ if (use_mips32r2_instructions) {
184
+ tcg_out_opc_bf(s, OPC_INS, lo, TCG_TMP0, 31, 8);
185
+ } else {
186
+ tcg_out_opc_sa(s, OPC_SLL, TCG_TMP0, TCG_TMP0, 8);
187
+ tcg_out_opc_reg(s, OPC_OR, lo, TCG_TMP0, TCG_TMP1);
188
+ }
189
+ break;
190
+
191
+ case MO_SW | MO_LE:
192
+ case MO_UW | MO_LE:
193
+ if (use_mips32r2_instructions && lo != base) {
194
+ tcg_out_opc_imm(s, OPC_LBU, lo, base, 0);
195
+ tcg_out_opc_imm(s, sgn ? OPC_LB : OPC_LBU, TCG_TMP0, base, 1);
196
+ tcg_out_opc_bf(s, OPC_INS, lo, TCG_TMP0, 31, 8);
197
+ } else {
198
+ tcg_out_opc_imm(s, OPC_LBU, TCG_TMP0, base, 0);
199
+ tcg_out_opc_imm(s, sgn ? OPC_LB : OPC_LBU, TCG_TMP1, base, 1);
200
+ tcg_out_opc_sa(s, OPC_SLL, TCG_TMP1, TCG_TMP1, 8);
201
+ tcg_out_opc_reg(s, OPC_OR, lo, TCG_TMP0, TCG_TMP1);
202
+ }
203
+ break;
204
+
205
+ case MO_SL:
206
+ case MO_UL:
207
+ tcg_out_opc_imm(s, lw1, lo, base, 0);
208
+ tcg_out_opc_imm(s, lw2, lo, base, 3);
209
+ if (TCG_TARGET_REG_BITS == 64 && is_64 && !sgn) {
210
+ tcg_out_ext32u(s, lo, lo);
211
+ }
212
+ break;
213
+
214
+ case MO_UL | MO_BSWAP:
215
+ case MO_SL | MO_BSWAP:
216
+ if (use_mips32r2_instructions) {
217
+ tcg_out_opc_imm(s, lw1, lo, base, 0);
218
+ tcg_out_opc_imm(s, lw2, lo, base, 3);
219
+ tcg_out_bswap32(s, lo, lo,
220
+ TCG_TARGET_REG_BITS == 64 && is_64
221
+ ? (sgn ? TCG_BSWAP_OS : TCG_BSWAP_OZ) : 0);
222
+ } else {
223
+ const tcg_insn_unit *subr =
224
+ (TCG_TARGET_REG_BITS == 64 && is_64 && !sgn
225
+ ? bswap32u_addr : bswap32_addr);
226
+
227
+ tcg_out_opc_imm(s, lw1, TCG_TMP0, base, 0);
228
+ tcg_out_bswap_subr(s, subr);
229
+ /* delay slot */
230
+ tcg_out_opc_imm(s, lw2, TCG_TMP0, base, 3);
231
+ tcg_out_mov(s, is_64 ? TCG_TYPE_I64 : TCG_TYPE_I32, lo, TCG_TMP3);
232
+ }
233
+ break;
234
+
235
+ case MO_UQ:
236
+ if (TCG_TARGET_REG_BITS == 64) {
237
+ tcg_out_opc_imm(s, ld1, lo, base, 0);
238
+ tcg_out_opc_imm(s, ld2, lo, base, 7);
239
+ } else {
240
+ tcg_out_opc_imm(s, lw1, MIPS_BE ? hi : lo, base, 0 + 0);
241
+ tcg_out_opc_imm(s, lw2, MIPS_BE ? hi : lo, base, 0 + 3);
242
+ tcg_out_opc_imm(s, lw1, MIPS_BE ? lo : hi, base, 4 + 0);
243
+ tcg_out_opc_imm(s, lw2, MIPS_BE ? lo : hi, base, 4 + 3);
244
+ }
245
+ break;
246
+
247
+ case MO_UQ | MO_BSWAP:
248
+ if (TCG_TARGET_REG_BITS == 64) {
249
+ if (use_mips32r2_instructions) {
250
+ tcg_out_opc_imm(s, ld1, lo, base, 0);
251
+ tcg_out_opc_imm(s, ld2, lo, base, 7);
252
+ tcg_out_bswap64(s, lo, lo);
253
+ } else {
254
+ tcg_out_opc_imm(s, ld1, TCG_TMP0, base, 0);
255
+ tcg_out_bswap_subr(s, bswap64_addr);
256
+ /* delay slot */
257
+ tcg_out_opc_imm(s, ld2, TCG_TMP0, base, 7);
258
+ tcg_out_mov(s, TCG_TYPE_I64, lo, TCG_TMP3);
259
+ }
260
+ } else if (use_mips32r2_instructions) {
261
+ tcg_out_opc_imm(s, lw1, TCG_TMP0, base, 0 + 0);
262
+ tcg_out_opc_imm(s, lw2, TCG_TMP0, base, 0 + 3);
263
+ tcg_out_opc_imm(s, lw1, TCG_TMP1, base, 4 + 0);
264
+ tcg_out_opc_imm(s, lw2, TCG_TMP1, base, 4 + 3);
265
+ tcg_out_opc_reg(s, OPC_WSBH, TCG_TMP0, 0, TCG_TMP0);
266
+ tcg_out_opc_reg(s, OPC_WSBH, TCG_TMP1, 0, TCG_TMP1);
267
+ tcg_out_opc_sa(s, OPC_ROTR, MIPS_BE ? lo : hi, TCG_TMP0, 16);
268
+ tcg_out_opc_sa(s, OPC_ROTR, MIPS_BE ? hi : lo, TCG_TMP1, 16);
269
+ } else {
270
+ tcg_out_opc_imm(s, lw1, TCG_TMP0, base, 0 + 0);
271
+ tcg_out_bswap_subr(s, bswap32_addr);
272
+ /* delay slot */
273
+ tcg_out_opc_imm(s, lw2, TCG_TMP0, base, 0 + 3);
274
+ tcg_out_opc_imm(s, lw1, TCG_TMP0, base, 4 + 0);
275
+ tcg_out_mov(s, TCG_TYPE_I32, MIPS_BE ? lo : hi, TCG_TMP3);
276
+ tcg_out_bswap_subr(s, bswap32_addr);
277
+ /* delay slot */
278
+ tcg_out_opc_imm(s, lw2, TCG_TMP0, base, 4 + 3);
279
+ tcg_out_mov(s, TCG_TYPE_I32, MIPS_BE ? hi : lo, TCG_TMP3);
280
+ }
281
+ break;
282
+
283
+ default:
284
+ g_assert_not_reached();
285
+ }
286
+}
287
+
288
static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
289
{
290
TCGReg addr_regl, addr_regh __attribute__((unused));
291
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
292
MemOp opc;
293
#if defined(CONFIG_SOFTMMU)
294
tcg_insn_unit *label_ptr[2];
295
+#else
296
+ unsigned a_bits, s_bits;
297
#endif
298
TCGReg base = TCG_REG_A0;
299
300
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
301
} else {
302
tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_GUEST_BASE_REG, addr_regl);
303
}
304
- tcg_out_qemu_ld_direct(s, data_regl, data_regh, base, opc, is_64);
305
+ a_bits = get_alignment_bits(opc);
306
+ s_bits = opc & MO_SIZE;
307
+ /*
308
+ * R6 removes the left/right instructions but requires the
309
+ * system to support misaligned memory accesses.
310
+ */
311
+ if (use_mips32r6_instructions) {
312
+ if (a_bits) {
313
+ tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
314
+ }
315
+ tcg_out_qemu_ld_direct(s, data_regl, data_regh, base, opc, is_64);
316
+ } else {
317
+ if (a_bits && a_bits != s_bits) {
318
+ tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
319
+ }
320
+ if (a_bits >= s_bits) {
321
+ tcg_out_qemu_ld_direct(s, data_regl, data_regh, base, opc, is_64);
322
+ } else {
323
+ tcg_out_qemu_ld_unalign(s, data_regl, data_regh, base, opc, is_64);
324
+ }
325
+ }
326
#endif
327
}
328
329
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, TCGReg lo, TCGReg hi,
330
}
331
}
332
333
+static void __attribute__((unused))
334
+tcg_out_qemu_st_unalign(TCGContext *s, TCGReg lo, TCGReg hi,
335
+ TCGReg base, MemOp opc)
336
+{
337
+ const MIPSInsn sw1 = MIPS_BE ? OPC_SWL : OPC_SWR;
338
+ const MIPSInsn sw2 = MIPS_BE ? OPC_SWR : OPC_SWL;
339
+ const MIPSInsn sd1 = MIPS_BE ? OPC_SDL : OPC_SDR;
340
+ const MIPSInsn sd2 = MIPS_BE ? OPC_SDR : OPC_SDL;
341
+
342
+ /* Don't clutter the code below with checks to avoid bswapping ZERO. */
343
+ if ((lo | hi) == 0) {
344
+ opc &= ~MO_BSWAP;
345
+ }
346
+
347
+ switch (opc & (MO_SIZE | MO_BSWAP)) {
348
+ case MO_16 | MO_BE:
349
+ tcg_out_opc_sa(s, OPC_SRL, TCG_TMP0, lo, 8);
350
+ tcg_out_opc_imm(s, OPC_SB, TCG_TMP0, base, 0);
351
+ tcg_out_opc_imm(s, OPC_SB, lo, base, 1);
352
+ break;
353
+
354
+ case MO_16 | MO_LE:
355
+ tcg_out_opc_sa(s, OPC_SRL, TCG_TMP0, lo, 8);
356
+ tcg_out_opc_imm(s, OPC_SB, lo, base, 0);
357
+ tcg_out_opc_imm(s, OPC_SB, TCG_TMP0, base, 1);
358
+ break;
359
+
360
+ case MO_32 | MO_BSWAP:
361
+ tcg_out_bswap32(s, TCG_TMP3, lo, 0);
362
+ lo = TCG_TMP3;
363
+ /* fall through */
364
+ case MO_32:
365
+ tcg_out_opc_imm(s, sw1, lo, base, 0);
366
+ tcg_out_opc_imm(s, sw2, lo, base, 3);
367
+ break;
368
+
369
+ case MO_64 | MO_BSWAP:
370
+ if (TCG_TARGET_REG_BITS == 64) {
371
+ tcg_out_bswap64(s, TCG_TMP3, lo);
372
+ lo = TCG_TMP3;
373
+ } else if (use_mips32r2_instructions) {
374
+ tcg_out_opc_reg(s, OPC_WSBH, TCG_TMP0, 0, MIPS_BE ? hi : lo);
375
+ tcg_out_opc_reg(s, OPC_WSBH, TCG_TMP1, 0, MIPS_BE ? lo : hi);
376
+ tcg_out_opc_sa(s, OPC_ROTR, TCG_TMP0, TCG_TMP0, 16);
377
+ tcg_out_opc_sa(s, OPC_ROTR, TCG_TMP1, TCG_TMP1, 16);
378
+ hi = MIPS_BE ? TCG_TMP0 : TCG_TMP1;
379
+ lo = MIPS_BE ? TCG_TMP1 : TCG_TMP0;
380
+ } else {
381
+ tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? lo : hi, 0);
382
+ tcg_out_opc_imm(s, sw1, TCG_TMP3, base, 0 + 0);
383
+ tcg_out_opc_imm(s, sw2, TCG_TMP3, base, 0 + 3);
384
+ tcg_out_bswap32(s, TCG_TMP3, MIPS_BE ? hi : lo, 0);
385
+ tcg_out_opc_imm(s, sw1, TCG_TMP3, base, 4 + 0);
386
+ tcg_out_opc_imm(s, sw2, TCG_TMP3, base, 4 + 3);
387
+ break;
87
+ break;
388
+ }
88
+ }
389
+ /* fall through */
89
+ /* fall through */
390
+ case MO_64:
90
case DISAS_EXIT:
391
+ if (TCG_TARGET_REG_BITS == 64) {
91
tcg_gen_exit_tb(NULL, 0);
392
+ tcg_out_opc_imm(s, sd1, lo, base, 0);
92
break;
393
+ tcg_out_opc_imm(s, sd2, lo, base, 7);
394
+ } else {
395
+ tcg_out_opc_imm(s, sw1, MIPS_BE ? hi : lo, base, 0 + 0);
396
+ tcg_out_opc_imm(s, sw2, MIPS_BE ? hi : lo, base, 0 + 3);
397
+ tcg_out_opc_imm(s, sw1, MIPS_BE ? lo : hi, base, 4 + 0);
398
+ tcg_out_opc_imm(s, sw2, MIPS_BE ? lo : hi, base, 4 + 3);
399
+ }
400
+ break;
401
+
402
+ default:
403
+ tcg_abort();
404
+ }
405
+}
406
static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
407
{
408
TCGReg addr_regl, addr_regh __attribute__((unused));
409
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
410
MemOp opc;
411
#if defined(CONFIG_SOFTMMU)
412
tcg_insn_unit *label_ptr[2];
413
+#else
414
+ unsigned a_bits, s_bits;
415
#endif
416
TCGReg base = TCG_REG_A0;
417
418
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
419
data_regl, data_regh, addr_regl, addr_regh,
420
s->code_ptr, label_ptr);
421
#else
422
- base = TCG_REG_A0;
423
if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
424
tcg_out_ext32u(s, base, addr_regl);
425
addr_regl = base;
426
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
427
} else {
428
tcg_out_opc_reg(s, ALIAS_PADD, base, TCG_GUEST_BASE_REG, addr_regl);
429
}
430
- tcg_out_qemu_st_direct(s, data_regl, data_regh, base, opc);
431
+ a_bits = get_alignment_bits(opc);
432
+ s_bits = opc & MO_SIZE;
433
+ /*
434
+ * R6 removes the left/right instructions but requires the
435
+ * system to support misaligned memory accesses.
436
+ */
437
+ if (use_mips32r6_instructions) {
438
+ if (a_bits) {
439
+ tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
440
+ }
441
+ tcg_out_qemu_st_direct(s, data_regl, data_regh, base, opc);
442
+ } else {
443
+ if (a_bits && a_bits != s_bits) {
444
+ tcg_out_test_alignment(s, true, addr_regl, addr_regh, a_bits);
445
+ }
446
+ if (a_bits >= s_bits) {
447
+ tcg_out_qemu_st_direct(s, data_regl, data_regh, base, opc);
448
+ } else {
449
+ tcg_out_qemu_st_unalign(s, data_regl, data_regh, base, opc);
450
+ }
451
+ }
452
#endif
453
}
454
455
--
93
--
456
2.25.1
94
2.34.1
457
95
458
96
diff view generated by jsdifflib
1
Handle 32-bit constants with a separate function, so that
1
Map the stack executable if required by default or on demand.
2
tcg_out_movi_int does not need to recurse. This slightly
3
rearranges the order of tests for small constants, but
4
produces the same output.
5
2
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
6
---
9
tcg/sparc/tcg-target.c.inc | 36 +++++++++++++++++++++---------------
7
include/elf.h | 1 +
10
1 file changed, 21 insertions(+), 15 deletions(-)
8
linux-user/qemu.h | 1 +
9
linux-user/elfload.c | 19 ++++++++++++++++++-
10
3 files changed, 20 insertions(+), 1 deletion(-)
11
11
12
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
12
diff --git a/include/elf.h b/include/elf.h
13
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/sparc/tcg-target.c.inc
14
--- a/include/elf.h
15
+++ b/tcg/sparc/tcg-target.c.inc
15
+++ b/include/elf.h
16
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_imm13(TCGContext *s, TCGReg ret, int32_t arg)
16
@@ -XXX,XX +XXX,XX @@ typedef int64_t Elf64_Sxword;
17
tcg_out_arithi(s, ret, TCG_REG_G0, arg, ARITH_OR);
17
#define PT_LOPROC 0x70000000
18
}
18
#define PT_HIPROC 0x7fffffff
19
19
20
+static void tcg_out_movi_imm32(TCGContext *s, TCGReg ret, int32_t arg)
20
+#define PT_GNU_STACK (PT_LOOS + 0x474e551)
21
+{
21
#define PT_GNU_PROPERTY (PT_LOOS + 0x474e553)
22
+ if (check_fit_i32(arg, 13)) {
22
23
+ /* A 13-bit constant sign-extended to 64-bits. */
23
#define PT_MIPS_REGINFO 0x70000000
24
+ tcg_out_movi_imm13(s, ret, arg);
24
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
25
+ } else {
25
index XXXXXXX..XXXXXXX 100644
26
+ /* A 32-bit constant zero-extended to 64 bits. */
26
--- a/linux-user/qemu.h
27
+ tcg_out_sethi(s, ret, arg);
27
+++ b/linux-user/qemu.h
28
+ if (arg & 0x3ff) {
28
@@ -XXX,XX +XXX,XX @@ struct image_info {
29
+ tcg_out_arithi(s, ret, ret, arg & 0x3ff, ARITH_OR);
29
uint32_t elf_flags;
30
+ }
30
int personality;
31
abi_ulong alignment;
32
+ bool exec_stack;
33
34
/* Generic semihosting knows about these pointers. */
35
abi_ulong arg_strings; /* strings for argv */
36
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/linux-user/elfload.c
39
+++ b/linux-user/elfload.c
40
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
41
#define ELF_ARCH EM_386
42
43
#define ELF_PLATFORM get_elf_platform()
44
+#define EXSTACK_DEFAULT true
45
46
static const char *get_elf_platform(void)
47
{
48
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
49
50
#define ELF_ARCH EM_ARM
51
#define ELF_CLASS ELFCLASS32
52
+#define EXSTACK_DEFAULT true
53
54
static inline void init_thread(struct target_pt_regs *regs,
55
struct image_info *infop)
56
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
57
#else
58
59
#define ELF_CLASS ELFCLASS32
60
+#define EXSTACK_DEFAULT true
61
62
#endif
63
64
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUPPCState *en
65
66
#define ELF_CLASS ELFCLASS64
67
#define ELF_ARCH EM_LOONGARCH
68
+#define EXSTACK_DEFAULT true
69
70
#define elf_check_arch(x) ((x) == EM_LOONGARCH)
71
72
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
73
#define ELF_CLASS ELFCLASS32
74
#endif
75
#define ELF_ARCH EM_MIPS
76
+#define EXSTACK_DEFAULT true
77
78
#ifdef TARGET_ABI_MIPSN32
79
#define elf_check_abi(x) ((x) & EF_MIPS_ABI2)
80
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
81
#define bswaptls(ptr) bswap32s(ptr)
82
#endif
83
84
+#ifndef EXSTACK_DEFAULT
85
+#define EXSTACK_DEFAULT false
86
+#endif
87
+
88
#include "elf.h"
89
90
/* We must delay the following stanzas until after "elf.h". */
91
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
92
struct image_info *info)
93
{
94
abi_ulong size, error, guard;
95
+ int prot;
96
97
size = guest_stack_size;
98
if (size < STACK_LOWER_LIMIT) {
99
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
100
guard = qemu_real_host_page_size();
101
}
102
103
- error = target_mmap(0, size + guard, PROT_READ | PROT_WRITE,
104
+ prot = PROT_READ | PROT_WRITE;
105
+ if (info->exec_stack) {
106
+ prot |= PROT_EXEC;
31
+ }
107
+ }
32
+}
108
+ error = target_mmap(0, size + guard, prot,
33
+
109
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
34
static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
110
if (error == -1) {
35
tcg_target_long arg, bool in_prologue)
111
perror("mmap stack");
36
{
112
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
37
tcg_target_long hi, lo = (int32_t)arg;
113
*/
38
tcg_target_long test, lsb;
114
loaddr = -1, hiaddr = 0;
39
115
info->alignment = 0;
40
- /* Make sure we test 32-bit constants for imm13 properly. */
116
+ info->exec_stack = EXSTACK_DEFAULT;
41
- if (type == TCG_TYPE_I32) {
117
for (i = 0; i < ehdr->e_phnum; ++i) {
42
- arg = lo;
118
struct elf_phdr *eppnt = phdr + i;
43
+ /* A 32-bit constant, or 32-bit zero-extended to 64-bits. */
119
if (eppnt->p_type == PT_LOAD) {
44
+ if (type == TCG_TYPE_I32 || arg == (uint32_t)arg) {
120
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
45
+ tcg_out_movi_imm32(s, ret, arg);
121
if (!parse_elf_properties(image_fd, info, eppnt, bprm_buf, &err)) {
46
+ return;
122
goto exit_errmsg;
47
}
123
}
48
124
+ } else if (eppnt->p_type == PT_GNU_STACK) {
49
/* A 13-bit constant sign-extended to 64-bits. */
125
+ info->exec_stack = eppnt->p_flags & PF_X;
50
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
51
}
126
}
52
}
127
}
53
128
54
- /* A 32-bit constant, or 32-bit zero-extended to 64-bits. */
55
- if (type == TCG_TYPE_I32 || arg == (uint32_t)arg) {
56
- tcg_out_sethi(s, ret, arg);
57
- if (arg & 0x3ff) {
58
- tcg_out_arithi(s, ret, ret, arg & 0x3ff, ARITH_OR);
59
- }
60
- return;
61
- }
62
-
63
/* A 32-bit constant sign-extended to 64-bits. */
64
if (arg == lo) {
65
tcg_out_sethi(s, ret, ~arg);
66
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
67
/* A 64-bit constant decomposed into 2 32-bit pieces. */
68
if (check_fit_i32(lo, 13)) {
69
hi = (arg - lo) >> 32;
70
- tcg_out_movi(s, TCG_TYPE_I32, ret, hi);
71
+ tcg_out_movi_imm32(s, ret, hi);
72
tcg_out_arithi(s, ret, ret, 32, SHIFT_SLLX);
73
tcg_out_arithi(s, ret, ret, lo, ARITH_ADD);
74
} else {
75
hi = arg >> 32;
76
- tcg_out_movi(s, TCG_TYPE_I32, ret, hi);
77
- tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_T2, lo);
78
+ tcg_out_movi_imm32(s, ret, hi);
79
+ tcg_out_movi_imm32(s, TCG_REG_T2, lo);
80
tcg_out_arithi(s, ret, ret, 32, SHIFT_SLLX);
81
tcg_out_arith(s, ret, ret, TCG_REG_T2, ARITH_OR);
82
}
83
--
129
--
84
2.25.1
130
2.34.1
85
86
diff view generated by jsdifflib
1
This is now always true, since we require armv6.
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Currently it's possible to execute pages that do not have PAGE_EXEC
4
if there is an existing translation block. Fix by invalidating TBs
5
that touch the affected pages.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Message-Id: <20220817150506.592862-2-iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
10
---
6
tcg/arm/tcg-target.h | 1 -
11
linux-user/mmap.c | 6 ++++--
7
tcg/arm/tcg-target.c.inc | 192 ++++++---------------------------------
12
1 file changed, 4 insertions(+), 2 deletions(-)
8
2 files changed, 27 insertions(+), 166 deletions(-)
9
13
10
diff --git a/tcg/arm/tcg-target.h b/tcg/arm/tcg-target.h
14
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
11
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/arm/tcg-target.h
16
--- a/linux-user/mmap.c
13
+++ b/tcg/arm/tcg-target.h
17
+++ b/linux-user/mmap.c
14
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
15
19
goto error;
16
extern int arm_arch;
17
18
-#define use_armv6_instructions (__ARM_ARCH >= 6 || arm_arch >= 6)
19
#define use_armv7_instructions (__ARM_ARCH >= 7 || arm_arch >= 7)
20
21
#undef TCG_TARGET_STACK_GROWSUP
22
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
23
index XXXXXXX..XXXXXXX 100644
24
--- a/tcg/arm/tcg-target.c.inc
25
+++ b/tcg/arm/tcg-target.c.inc
26
@@ -XXX,XX +XXX,XX @@ static void tcg_out_dat_rIN(TCGContext *s, ARMCond cond, ARMInsn opc,
27
static void tcg_out_mul32(TCGContext *s, ARMCond cond, TCGReg rd,
28
TCGReg rn, TCGReg rm)
29
{
30
- /* if ArchVersion() < 6 && d == n then UNPREDICTABLE; */
31
- if (!use_armv6_instructions && rd == rn) {
32
- if (rd == rm) {
33
- /* rd == rn == rm; copy an input to tmp first. */
34
- tcg_out_mov_reg(s, cond, TCG_REG_TMP, rn);
35
- rm = rn = TCG_REG_TMP;
36
- } else {
37
- rn = rm;
38
- rm = rd;
39
- }
40
- }
41
/* mul */
42
tcg_out32(s, (cond << 28) | 0x90 | (rd << 16) | (rm << 8) | rn);
43
}
44
@@ -XXX,XX +XXX,XX @@ static void tcg_out_mul32(TCGContext *s, ARMCond cond, TCGReg rd,
45
static void tcg_out_umull32(TCGContext *s, ARMCond cond, TCGReg rd0,
46
TCGReg rd1, TCGReg rn, TCGReg rm)
47
{
48
- /* if ArchVersion() < 6 && (dHi == n || dLo == n) then UNPREDICTABLE; */
49
- if (!use_armv6_instructions && (rd0 == rn || rd1 == rn)) {
50
- if (rd0 == rm || rd1 == rm) {
51
- tcg_out_mov_reg(s, cond, TCG_REG_TMP, rn);
52
- rn = TCG_REG_TMP;
53
- } else {
54
- TCGReg t = rn;
55
- rn = rm;
56
- rm = t;
57
- }
58
- }
59
/* umull */
60
tcg_out32(s, (cond << 28) | 0x00800090 |
61
(rd1 << 16) | (rd0 << 12) | (rm << 8) | rn);
62
@@ -XXX,XX +XXX,XX @@ static void tcg_out_umull32(TCGContext *s, ARMCond cond, TCGReg rd0,
63
static void tcg_out_smull32(TCGContext *s, ARMCond cond, TCGReg rd0,
64
TCGReg rd1, TCGReg rn, TCGReg rm)
65
{
66
- /* if ArchVersion() < 6 && (dHi == n || dLo == n) then UNPREDICTABLE; */
67
- if (!use_armv6_instructions && (rd0 == rn || rd1 == rn)) {
68
- if (rd0 == rm || rd1 == rm) {
69
- tcg_out_mov_reg(s, cond, TCG_REG_TMP, rn);
70
- rn = TCG_REG_TMP;
71
- } else {
72
- TCGReg t = rn;
73
- rn = rm;
74
- rm = t;
75
- }
76
- }
77
/* smull */
78
tcg_out32(s, (cond << 28) | 0x00c00090 |
79
(rd1 << 16) | (rd0 << 12) | (rm << 8) | rn);
80
@@ -XXX,XX +XXX,XX @@ static void tcg_out_udiv(TCGContext *s, ARMCond cond,
81
82
static void tcg_out_ext8s(TCGContext *s, ARMCond cond, TCGReg rd, TCGReg rn)
83
{
84
- if (use_armv6_instructions) {
85
- /* sxtb */
86
- tcg_out32(s, 0x06af0070 | (cond << 28) | (rd << 12) | rn);
87
- } else {
88
- tcg_out_dat_reg(s, cond, ARITH_MOV,
89
- rd, 0, rn, SHIFT_IMM_LSL(24));
90
- tcg_out_dat_reg(s, cond, ARITH_MOV,
91
- rd, 0, rd, SHIFT_IMM_ASR(24));
92
- }
93
+ /* sxtb */
94
+ tcg_out32(s, 0x06af0070 | (cond << 28) | (rd << 12) | rn);
95
}
96
97
static void __attribute__((unused))
98
@@ -XXX,XX +XXX,XX @@ tcg_out_ext8u(TCGContext *s, ARMCond cond, TCGReg rd, TCGReg rn)
99
100
static void tcg_out_ext16s(TCGContext *s, ARMCond cond, TCGReg rd, TCGReg rn)
101
{
102
- if (use_armv6_instructions) {
103
- /* sxth */
104
- tcg_out32(s, 0x06bf0070 | (cond << 28) | (rd << 12) | rn);
105
- } else {
106
- tcg_out_dat_reg(s, cond, ARITH_MOV,
107
- rd, 0, rn, SHIFT_IMM_LSL(16));
108
- tcg_out_dat_reg(s, cond, ARITH_MOV,
109
- rd, 0, rd, SHIFT_IMM_ASR(16));
110
- }
111
+ /* sxth */
112
+ tcg_out32(s, 0x06bf0070 | (cond << 28) | (rd << 12) | rn);
113
}
114
115
static void tcg_out_ext16u(TCGContext *s, ARMCond cond, TCGReg rd, TCGReg rn)
116
{
117
- if (use_armv6_instructions) {
118
- /* uxth */
119
- tcg_out32(s, 0x06ff0070 | (cond << 28) | (rd << 12) | rn);
120
- } else {
121
- tcg_out_dat_reg(s, cond, ARITH_MOV,
122
- rd, 0, rn, SHIFT_IMM_LSL(16));
123
- tcg_out_dat_reg(s, cond, ARITH_MOV,
124
- rd, 0, rd, SHIFT_IMM_LSR(16));
125
- }
126
+ /* uxth */
127
+ tcg_out32(s, 0x06ff0070 | (cond << 28) | (rd << 12) | rn);
128
}
129
130
static void tcg_out_bswap16(TCGContext *s, ARMCond cond,
131
TCGReg rd, TCGReg rn, int flags)
132
{
133
- if (use_armv6_instructions) {
134
- if (flags & TCG_BSWAP_OS) {
135
- /* revsh */
136
- tcg_out32(s, 0x06ff0fb0 | (cond << 28) | (rd << 12) | rn);
137
- return;
138
- }
139
-
140
- /* rev16 */
141
- tcg_out32(s, 0x06bf0fb0 | (cond << 28) | (rd << 12) | rn);
142
- if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
143
- /* uxth */
144
- tcg_out32(s, 0x06ff0070 | (cond << 28) | (rd << 12) | rd);
145
- }
146
+ if (flags & TCG_BSWAP_OS) {
147
+ /* revsh */
148
+ tcg_out32(s, 0x06ff0fb0 | (cond << 28) | (rd << 12) | rn);
149
return;
150
}
151
152
- if (flags == 0) {
153
- /*
154
- * For stores, no input or output extension:
155
- * rn = xxAB
156
- * lsr tmp, rn, #8 tmp = 0xxA
157
- * and tmp, tmp, #0xff tmp = 000A
158
- * orr rd, tmp, rn, lsl #8 rd = xABA
159
- */
160
- tcg_out_dat_reg(s, cond, ARITH_MOV,
161
- TCG_REG_TMP, 0, rn, SHIFT_IMM_LSR(8));
162
- tcg_out_dat_imm(s, cond, ARITH_AND, TCG_REG_TMP, TCG_REG_TMP, 0xff);
163
- tcg_out_dat_reg(s, cond, ARITH_ORR,
164
- rd, TCG_REG_TMP, rn, SHIFT_IMM_LSL(8));
165
- return;
166
+ /* rev16 */
167
+ tcg_out32(s, 0x06bf0fb0 | (cond << 28) | (rd << 12) | rn);
168
+ if ((flags & (TCG_BSWAP_IZ | TCG_BSWAP_OZ)) == TCG_BSWAP_OZ) {
169
+ /* uxth */
170
+ tcg_out32(s, 0x06ff0070 | (cond << 28) | (rd << 12) | rd);
171
}
172
-
173
- /*
174
- * Byte swap, leaving the result at the top of the register.
175
- * We will then shift down, zero or sign-extending.
176
- */
177
- if (flags & TCG_BSWAP_IZ) {
178
- /*
179
- * rn = 00AB
180
- * ror tmp, rn, #8 tmp = B00A
181
- * orr tmp, tmp, tmp, lsl #16 tmp = BA00
182
- */
183
- tcg_out_dat_reg(s, cond, ARITH_MOV,
184
- TCG_REG_TMP, 0, rn, SHIFT_IMM_ROR(8));
185
- tcg_out_dat_reg(s, cond, ARITH_ORR,
186
- TCG_REG_TMP, TCG_REG_TMP, TCG_REG_TMP,
187
- SHIFT_IMM_LSL(16));
188
- } else {
189
- /*
190
- * rn = xxAB
191
- * and tmp, rn, #0xff00 tmp = 00A0
192
- * lsl tmp, tmp, #8 tmp = 0A00
193
- * orr tmp, tmp, rn, lsl #24 tmp = BA00
194
- */
195
- tcg_out_dat_rI(s, cond, ARITH_AND, TCG_REG_TMP, rn, 0xff00, 1);
196
- tcg_out_dat_reg(s, cond, ARITH_MOV,
197
- TCG_REG_TMP, 0, TCG_REG_TMP, SHIFT_IMM_LSL(8));
198
- tcg_out_dat_reg(s, cond, ARITH_ORR,
199
- TCG_REG_TMP, TCG_REG_TMP, rn, SHIFT_IMM_LSL(24));
200
- }
201
- tcg_out_dat_reg(s, cond, ARITH_MOV, rd, 0, TCG_REG_TMP,
202
- (flags & TCG_BSWAP_OS
203
- ? SHIFT_IMM_ASR(8) : SHIFT_IMM_LSR(8)));
204
}
205
206
static void tcg_out_bswap32(TCGContext *s, ARMCond cond, TCGReg rd, TCGReg rn)
207
{
208
- if (use_armv6_instructions) {
209
- /* rev */
210
- tcg_out32(s, 0x06bf0f30 | (cond << 28) | (rd << 12) | rn);
211
- } else {
212
- tcg_out_dat_reg(s, cond, ARITH_EOR,
213
- TCG_REG_TMP, rn, rn, SHIFT_IMM_ROR(16));
214
- tcg_out_dat_imm(s, cond, ARITH_BIC,
215
- TCG_REG_TMP, TCG_REG_TMP, 0xff | 0x800);
216
- tcg_out_dat_reg(s, cond, ARITH_MOV,
217
- rd, 0, rn, SHIFT_IMM_ROR(8));
218
- tcg_out_dat_reg(s, cond, ARITH_EOR,
219
- rd, rd, TCG_REG_TMP, SHIFT_IMM_LSR(8));
220
- }
221
+ /* rev */
222
+ tcg_out32(s, 0x06bf0f30 | (cond << 28) | (rd << 12) | rn);
223
}
224
225
static void tcg_out_deposit(TCGContext *s, ARMCond cond, TCGReg rd,
226
@@ -XXX,XX +XXX,XX @@ static void tcg_out_mb(TCGContext *s, TCGArg a0)
227
{
228
if (use_armv7_instructions) {
229
tcg_out32(s, INSN_DMB_ISH);
230
- } else if (use_armv6_instructions) {
231
+ } else {
232
tcg_out32(s, INSN_DMB_MCR);
233
}
234
}
235
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_arg_reg64(TCGContext *s, TCGReg argreg,
236
if (argreg & 1) {
237
argreg++;
238
}
239
- if (use_armv6_instructions && argreg >= 4
240
- && (arglo & 1) == 0 && arghi == arglo + 1) {
241
+ if (argreg >= 4 && (arglo & 1) == 0 && arghi == arglo + 1) {
242
tcg_out_strd_8(s, COND_AL, arglo,
243
TCG_REG_CALL_STACK, (argreg - 4) * 4);
244
return argreg + 2;
245
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
246
int cmp_off = (is_load ? offsetof(CPUTLBEntry, addr_read)
247
: offsetof(CPUTLBEntry, addr_write));
248
int fast_off = TLB_MASK_TABLE_OFS(mem_index);
249
- int mask_off = fast_off + offsetof(CPUTLBDescFast, mask);
250
- int table_off = fast_off + offsetof(CPUTLBDescFast, table);
251
unsigned s_bits = opc & MO_SIZE;
252
unsigned a_bits = get_alignment_bits(opc);
253
254
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
255
}
256
257
/* Load env_tlb(env)->f[mmu_idx].{mask,table} into {r0,r1}. */
258
- if (use_armv6_instructions) {
259
- tcg_out_ldrd_8(s, COND_AL, TCG_REG_R0, TCG_AREG0, fast_off);
260
- } else {
261
- tcg_out_ld(s, TCG_TYPE_I32, TCG_REG_R0, TCG_AREG0, mask_off);
262
- tcg_out_ld(s, TCG_TYPE_I32, TCG_REG_R1, TCG_AREG0, table_off);
263
- }
264
+ tcg_out_ldrd_8(s, COND_AL, TCG_REG_R0, TCG_AREG0, fast_off);
265
266
/* Extract the tlb index from the address into R0. */
267
tcg_out_dat_reg(s, COND_AL, ARITH_AND, TCG_REG_R0, TCG_REG_R0, addrlo,
268
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
269
* Load the tlb comparator into R2/R3 and the fast path addend into R1.
270
*/
271
if (cmp_off == 0) {
272
- if (use_armv6_instructions && TARGET_LONG_BITS == 64) {
273
+ if (TARGET_LONG_BITS == 64) {
274
tcg_out_ldrd_rwb(s, COND_AL, TCG_REG_R2, TCG_REG_R1, TCG_REG_R0);
275
} else {
276
tcg_out_ld32_rwb(s, COND_AL, TCG_REG_R2, TCG_REG_R1, TCG_REG_R0);
277
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
278
} else {
279
tcg_out_dat_reg(s, COND_AL, ARITH_ADD,
280
TCG_REG_R1, TCG_REG_R1, TCG_REG_R0, 0);
281
- if (use_armv6_instructions && TARGET_LONG_BITS == 64) {
282
+ if (TARGET_LONG_BITS == 64) {
283
tcg_out_ldrd_8(s, COND_AL, TCG_REG_R2, TCG_REG_R1, cmp_off);
284
} else {
285
tcg_out_ld32_12(s, COND_AL, TCG_REG_R2, TCG_REG_R1, cmp_off);
286
}
20
}
287
}
21
}
288
- if (!use_armv6_instructions && TARGET_LONG_BITS == 64) {
22
+
289
- tcg_out_ld32_12(s, COND_AL, TCG_REG_R3, TCG_REG_R1, cmp_off + 4);
23
page_set_flags(start, start + len, page_flags);
290
- }
24
- mmap_unlock();
291
25
- return 0;
292
/* Load the tlb addend. */
26
+ tb_invalidate_phys_range(start, start + len);
293
tcg_out_ld32_12(s, COND_AL, TCG_REG_R1, TCG_REG_R1,
27
+ ret = 0;
294
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
28
+
295
TCGReg argreg, datalo, datahi;
29
error:
296
MemOpIdx oi = lb->oi;
30
mmap_unlock();
297
MemOp opc = get_memop(oi);
31
return ret;
298
- void *func;
299
300
if (!reloc_pc24(lb->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
301
return false;
302
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
303
argreg = tcg_out_arg_imm32(s, argreg, oi);
304
argreg = tcg_out_arg_reg32(s, argreg, TCG_REG_R14);
305
306
- /* For armv6 we can use the canonical unsigned helpers and minimize
307
- icache usage. For pre-armv6, use the signed helpers since we do
308
- not have a single insn sign-extend. */
309
- if (use_armv6_instructions) {
310
- func = qemu_ld_helpers[opc & MO_SIZE];
311
- } else {
312
- func = qemu_ld_helpers[opc & MO_SSIZE];
313
- if (opc & MO_SIGN) {
314
- opc = MO_UL;
315
- }
316
- }
317
- tcg_out_call(s, func);
318
+ /* Use the canonical unsigned helpers and minimize icache usage. */
319
+ tcg_out_call(s, qemu_ld_helpers[opc & MO_SIZE]);
320
321
datalo = lb->datalo_reg;
322
datahi = lb->datahi_reg;
323
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
324
break;
325
case MO_UQ:
326
/* Avoid ldrd for user-only emulation, to handle unaligned. */
327
- if (USING_SOFTMMU && use_armv6_instructions
328
+ if (USING_SOFTMMU
329
&& (datalo & 1) == 0 && datahi == datalo + 1) {
330
tcg_out_ldrd_r(s, COND_AL, datalo, addrlo, addend);
331
} else if (datalo != addend) {
332
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc, TCGReg datalo,
333
break;
334
case MO_UQ:
335
/* Avoid ldrd for user-only emulation, to handle unaligned. */
336
- if (USING_SOFTMMU && use_armv6_instructions
337
+ if (USING_SOFTMMU
338
&& (datalo & 1) == 0 && datahi == datalo + 1) {
339
tcg_out_ldrd_8(s, COND_AL, datalo, addrlo, 0);
340
} else if (datalo == addrlo) {
341
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_index(TCGContext *s, ARMCond cond, MemOp opc,
342
break;
343
case MO_64:
344
/* Avoid strd for user-only emulation, to handle unaligned. */
345
- if (USING_SOFTMMU && use_armv6_instructions
346
+ if (USING_SOFTMMU
347
&& (datalo & 1) == 0 && datahi == datalo + 1) {
348
tcg_out_strd_r(s, cond, datalo, addrlo, addend);
349
} else {
350
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc, TCGReg datalo,
351
break;
352
case MO_64:
353
/* Avoid strd for user-only emulation, to handle unaligned. */
354
- if (USING_SOFTMMU && use_armv6_instructions
355
+ if (USING_SOFTMMU
356
&& (datalo & 1) == 0 && datahi == datalo + 1) {
357
tcg_out_strd_8(s, COND_AL, datalo, addrlo, 0);
358
} else {
359
--
32
--
360
2.25.1
33
2.34.1
361
362
diff view generated by jsdifflib
1
From: WANG Xuerui <git@xen0n.name>
1
We're about to start validating PAGE_EXEC, which means
2
that we've got to put this code into a section that is
3
both writable and executable.
2
4
3
Apparently we were left behind; just renaming MO_Q to MO_UQ is enough.
5
Note that this test did not run on hardware beforehand either.
4
6
5
Fixes: fc313c64345453c7 ("exec/memop: Adding signedness to quad definitions")
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: WANG Xuerui <git@xen0n.name>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Message-Id: <20220206162106.1092364-1-i.qemu@xen0n.name>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
10
---
10
tcg/loongarch64/tcg-target.c.inc | 2 +-
11
tests/tcg/i386/test-i386.c | 2 +-
11
1 file changed, 1 insertion(+), 1 deletion(-)
12
1 file changed, 1 insertion(+), 1 deletion(-)
12
13
13
diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc
14
diff --git a/tests/tcg/i386/test-i386.c b/tests/tcg/i386/test-i386.c
14
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
15
--- a/tcg/loongarch64/tcg-target.c.inc
16
--- a/tests/tcg/i386/test-i386.c
16
+++ b/tcg/loongarch64/tcg-target.c.inc
17
+++ b/tests/tcg/i386/test-i386.c
17
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_indexed(TCGContext *s, TCGReg rd, TCGReg rj,
18
@@ -XXX,XX +XXX,XX @@ uint8_t code[] = {
18
case MO_SL:
19
0xc3, /* ret */
19
tcg_out_opc_ldx_w(s, rd, rj, rk);
20
};
20
break;
21
21
- case MO_Q:
22
-asm(".section \".data\"\n"
22
+ case MO_UQ:
23
+asm(".section \".data_x\",\"awx\"\n"
23
tcg_out_opc_ldx_d(s, rd, rj, rk);
24
"smc_code2:\n"
24
break;
25
"movl 4(%esp), %eax\n"
25
default:
26
"movl %eax, smc_patch_addr2 + 1\n"
26
--
27
--
27
2.25.1
28
2.34.1
28
29
diff view generated by jsdifflib
1
Due to mapping changes, we now rarely place the code_gen_buffer
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
near the main executable. Which means that direct calls will
3
now rarely be in range.
4
2
5
So, always use indirect calls for tail calls, which allows us to
3
Introduce a function that checks whether a given address is on the same
6
avoid clobbering %o7, and therefore we need not save and restore it.
4
page as where disassembly started. Having it improves readability of
5
the following patches.
7
6
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Message-Id: <20220811095534.241224-3-iii@linux.ibm.com>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
[rth: Make the DisasContextBase parameter const.]
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
13
---
11
tcg/sparc/tcg-target.c.inc | 37 +++++++++++++++++++++++--------------
14
include/exec/translator.h | 10 ++++++++++
12
1 file changed, 23 insertions(+), 14 deletions(-)
15
1 file changed, 10 insertions(+)
13
16
14
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
17
diff --git a/include/exec/translator.h b/include/exec/translator.h
15
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
16
--- a/tcg/sparc/tcg-target.c.inc
19
--- a/include/exec/translator.h
17
+++ b/tcg/sparc/tcg-target.c.inc
20
+++ b/include/exec/translator.h
18
@@ -XXX,XX +XXX,XX @@ static void tcg_out_addsub2_i64(TCGContext *s, TCGReg rl, TCGReg rh,
21
@@ -XXX,XX +XXX,XX @@ FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
19
tcg_out_mov(s, TCG_TYPE_I64, rl, tmp);
22
20
}
23
#undef GEN_TRANSLATOR_LD
21
24
22
+static void tcg_out_jmpl_const(TCGContext *s, const tcg_insn_unit *dest,
25
+/*
23
+ bool in_prologue, bool tail_call)
26
+ * Return whether addr is on the same page as where disassembly started.
27
+ * Translators can use this to enforce the rule that only single-insn
28
+ * translation blocks are allowed to cross page boundaries.
29
+ */
30
+static inline bool is_same_page(const DisasContextBase *db, target_ulong addr)
24
+{
31
+{
25
+ uintptr_t desti = (uintptr_t)dest;
32
+ return ((addr ^ db->pc_first) & TARGET_PAGE_MASK) == 0;
26
+
27
+ /* Be careful not to clobber %o7 for a tail call. */
28
+ tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_REG_T1,
29
+ desti & ~0xfff, in_prologue,
30
+ tail_call ? TCG_REG_G2 : TCG_REG_O7);
31
+ tcg_out_arithi(s, tail_call ? TCG_REG_G0 : TCG_REG_O7,
32
+ TCG_REG_T1, desti & 0xfff, JMPL);
33
+}
33
+}
34
+
34
+
35
static void tcg_out_call_nodelay(TCGContext *s, const tcg_insn_unit *dest,
35
#endif /* EXEC__TRANSLATOR_H */
36
bool in_prologue)
37
{
38
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call_nodelay(TCGContext *s, const tcg_insn_unit *dest,
39
if (disp == (int32_t)disp) {
40
tcg_out32(s, CALL | (uint32_t)disp >> 2);
41
} else {
42
- uintptr_t desti = (uintptr_t)dest;
43
- tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_REG_T1,
44
- desti & ~0xfff, in_prologue, TCG_REG_O7);
45
- tcg_out_arithi(s, TCG_REG_O7, TCG_REG_T1, desti & 0xfff, JMPL);
46
+ tcg_out_jmpl_const(s, dest, in_prologue, false);
47
}
48
}
49
50
@@ -XXX,XX +XXX,XX @@ static void build_trampolines(TCGContext *s)
51
52
/* Set the retaddr operand. */
53
tcg_out_mov(s, TCG_TYPE_PTR, ra, TCG_REG_O7);
54
- /* Set the env operand. */
55
- tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0);
56
/* Tail call. */
57
- tcg_out_call_nodelay(s, qemu_ld_helpers[i], true);
58
- tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra);
59
+ tcg_out_jmpl_const(s, qemu_ld_helpers[i], true, true);
60
+ /* delay slot -- set the env argument */
61
+ tcg_out_mov_delay(s, TCG_REG_O0, TCG_AREG0);
62
}
63
64
for (i = 0; i < ARRAY_SIZE(qemu_st_helpers); ++i) {
65
@@ -XXX,XX +XXX,XX @@ static void build_trampolines(TCGContext *s)
66
if (ra >= TCG_REG_O6) {
67
tcg_out_st(s, TCG_TYPE_PTR, TCG_REG_O7, TCG_REG_CALL_STACK,
68
TCG_TARGET_CALL_STACK_OFFSET);
69
- ra = TCG_REG_G1;
70
+ } else {
71
+ tcg_out_mov(s, TCG_TYPE_PTR, ra, TCG_REG_O7);
72
}
73
- tcg_out_mov(s, TCG_TYPE_PTR, ra, TCG_REG_O7);
74
- /* Set the env operand. */
75
- tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O0, TCG_AREG0);
76
+
77
/* Tail call. */
78
- tcg_out_call_nodelay(s, qemu_st_helpers[i], true);
79
- tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_O7, ra);
80
+ tcg_out_jmpl_const(s, qemu_st_helpers[i], true, true);
81
+ /* delay slot -- set the env argument */
82
+ tcg_out_mov_delay(s, TCG_REG_O0, TCG_AREG0);
83
}
84
}
85
#endif
86
--
36
--
87
2.25.1
37
2.34.1
88
89
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
The current implementation is a no-op, simply returning addr.
2
This is incorrect, because we ought to be checking the page
3
permissions for execution.
4
5
Make get_page_addr_code inline for both implementations.
6
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Acked-by: Alistair Francis <alistair.francis@wdc.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
11
---
4
tcg/s390x/tcg-target.h | 2 --
12
include/exec/exec-all.h | 85 ++++++++++++++---------------------------
5
tcg/s390x/tcg-target.c.inc | 59 ++++++++++++++++++++++++++++++++++++--
13
accel/tcg/cputlb.c | 5 ---
6
2 files changed, 57 insertions(+), 4 deletions(-)
14
accel/tcg/user-exec.c | 14 +++++++
15
3 files changed, 42 insertions(+), 62 deletions(-)
7
16
8
diff --git a/tcg/s390x/tcg-target.h b/tcg/s390x/tcg-target.h
17
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
9
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/s390x/tcg-target.h
19
--- a/include/exec/exec-all.h
11
+++ b/tcg/s390x/tcg-target.h
20
+++ b/include/exec/exec-all.h
12
@@ -XXX,XX +XXX,XX @@ static inline void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_rx,
21
@@ -XXX,XX +XXX,XX @@ struct MemoryRegionSection *iotlb_to_section(CPUState *cpu,
13
/* no need to flush icache explicitly */
22
hwaddr index, MemTxAttrs attrs);
23
#endif
24
25
-#if defined(CONFIG_USER_ONLY)
26
-void mmap_lock(void);
27
-void mmap_unlock(void);
28
-bool have_mmap_lock(void);
29
-
30
/**
31
- * get_page_addr_code() - user-mode version
32
+ * get_page_addr_code_hostp()
33
* @env: CPUArchState
34
* @addr: guest virtual address of guest code
35
*
36
- * Returns @addr.
37
+ * See get_page_addr_code() (full-system version) for documentation on the
38
+ * return value.
39
+ *
40
+ * Sets *@hostp (when @hostp is non-NULL) as follows.
41
+ * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
42
+ * to the host address where @addr's content is kept.
43
+ *
44
+ * Note: this function can trigger an exception.
45
+ */
46
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
47
+ void **hostp);
48
+
49
+/**
50
+ * get_page_addr_code()
51
+ * @env: CPUArchState
52
+ * @addr: guest virtual address of guest code
53
+ *
54
+ * If we cannot translate and execute from the entire RAM page, or if
55
+ * the region is not backed by RAM, returns -1. Otherwise, returns the
56
+ * ram_addr_t corresponding to the guest code at @addr.
57
+ *
58
+ * Note: this function can trigger an exception.
59
*/
60
static inline tb_page_addr_t get_page_addr_code(CPUArchState *env,
61
target_ulong addr)
62
{
63
- return addr;
64
+ return get_page_addr_code_hostp(env, addr, NULL);
14
}
65
}
15
66
16
-#ifdef CONFIG_SOFTMMU
67
-/**
17
#define TCG_TARGET_NEED_LDST_LABELS
68
- * get_page_addr_code_hostp() - user-mode version
18
-#endif
69
- * @env: CPUArchState
19
#define TCG_TARGET_NEED_POOL_LABELS
70
- * @addr: guest virtual address of guest code
20
71
- *
21
#endif
72
- * Returns @addr.
22
diff --git a/tcg/s390x/tcg-target.c.inc b/tcg/s390x/tcg-target.c.inc
73
- *
74
- * If @hostp is non-NULL, sets *@hostp to the host address where @addr's content
75
- * is kept.
76
- */
77
-static inline tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env,
78
- target_ulong addr,
79
- void **hostp)
80
-{
81
- if (hostp) {
82
- *hostp = g2h_untagged(addr);
83
- }
84
- return addr;
85
-}
86
+#if defined(CONFIG_USER_ONLY)
87
+void mmap_lock(void);
88
+void mmap_unlock(void);
89
+bool have_mmap_lock(void);
90
91
/**
92
* adjust_signal_pc:
93
@@ -XXX,XX +XXX,XX @@ G_NORETURN void cpu_loop_exit_sigbus(CPUState *cpu, target_ulong addr,
94
static inline void mmap_lock(void) {}
95
static inline void mmap_unlock(void) {}
96
97
-/**
98
- * get_page_addr_code() - full-system version
99
- * @env: CPUArchState
100
- * @addr: guest virtual address of guest code
101
- *
102
- * If we cannot translate and execute from the entire RAM page, or if
103
- * the region is not backed by RAM, returns -1. Otherwise, returns the
104
- * ram_addr_t corresponding to the guest code at @addr.
105
- *
106
- * Note: this function can trigger an exception.
107
- */
108
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr);
109
-
110
-/**
111
- * get_page_addr_code_hostp() - full-system version
112
- * @env: CPUArchState
113
- * @addr: guest virtual address of guest code
114
- *
115
- * See get_page_addr_code() (full-system version) for documentation on the
116
- * return value.
117
- *
118
- * Sets *@hostp (when @hostp is non-NULL) as follows.
119
- * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
120
- * to the host address where @addr's content is kept.
121
- *
122
- * Note: this function can trigger an exception.
123
- */
124
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
125
- void **hostp);
126
-
127
void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length);
128
void tlb_set_dirty(CPUState *cpu, target_ulong vaddr);
129
130
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
23
index XXXXXXX..XXXXXXX 100644
131
index XXXXXXX..XXXXXXX 100644
24
--- a/tcg/s390x/tcg-target.c.inc
132
--- a/accel/tcg/cputlb.c
25
+++ b/tcg/s390x/tcg-target.c.inc
133
+++ b/accel/tcg/cputlb.c
26
@@ -XXX,XX +XXX,XX @@
134
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
27
#error "unsupported code generation mode"
135
return qemu_ram_addr_from_host_nofail(p);
28
#endif
29
30
+#include "../tcg-ldst.c.inc"
31
#include "../tcg-pool.c.inc"
32
#include "elf.h"
33
34
@@ -XXX,XX +XXX,XX @@ typedef enum S390Opcode {
35
RI_OIHL = 0xa509,
36
RI_OILH = 0xa50a,
37
RI_OILL = 0xa50b,
38
+ RI_TMLL = 0xa701,
39
40
RIE_CGIJ = 0xec7c,
41
RIE_CGRJ = 0xec64,
42
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc, TCGReg data,
43
}
136
}
44
137
45
#if defined(CONFIG_SOFTMMU)
138
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr)
46
-#include "../tcg-ldst.c.inc"
139
-{
140
- return get_page_addr_code_hostp(env, addr, NULL);
141
-}
47
-
142
-
48
/* We're expecting to use a 20-bit negative offset on the tlb memory ops. */
143
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
49
QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) > 0);
144
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
50
QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) < -(1 << 19));
145
{
51
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
146
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
52
return true;
147
index XXXXXXX..XXXXXXX 100644
148
--- a/accel/tcg/user-exec.c
149
+++ b/accel/tcg/user-exec.c
150
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
151
return size ? g2h(env_cpu(env), addr) : NULL;
53
}
152
}
54
#else
153
55
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld,
154
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
56
+ TCGReg addrlo, unsigned a_bits)
155
+ void **hostp)
57
+{
156
+{
58
+ unsigned a_mask = (1 << a_bits) - 1;
157
+ int flags;
59
+ TCGLabelQemuLdst *l = new_ldst_label(s);
60
+
158
+
61
+ l->is_ld = is_ld;
159
+ flags = probe_access_internal(env, addr, 1, MMU_INST_FETCH, false, 0);
62
+ l->addrlo_reg = addrlo;
160
+ g_assert(flags == 0);
63
+
161
+
64
+ /* We are expecting a_bits to max out at 7, much lower than TMLL. */
162
+ if (hostp) {
65
+ tcg_debug_assert(a_bits < 16);
163
+ *hostp = g2h_untagged(addr);
66
+ tcg_out_insn(s, RI, TMLL, addrlo, a_mask);
164
+ }
67
+
165
+ return addr;
68
+ tcg_out16(s, RI_BRC | (7 << 4)); /* CC in {1,2,3} */
69
+ l->label_ptr[0] = s->code_ptr;
70
+ s->code_ptr += 1;
71
+
72
+ l->raddr = tcg_splitwx_to_rx(s->code_ptr);
73
+}
166
+}
74
+
167
+
75
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
168
/* The softmmu versions of these helpers are in cputlb.c. */
76
+{
169
77
+ if (!patch_reloc(l->label_ptr[0], R_390_PC16DBL,
170
/*
78
+ (intptr_t)tcg_splitwx_to_rx(s->code_ptr), 2)) {
79
+ return false;
80
+ }
81
+
82
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_R3, l->addrlo_reg);
83
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_R2, TCG_AREG0);
84
+
85
+ /* "Tail call" to the helper, with the return address back inline. */
86
+ tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_R14, (uintptr_t)l->raddr);
87
+ tgen_gotoi(s, S390_CC_ALWAYS, (const void *)(l->is_ld ? helper_unaligned_ld
88
+ : helper_unaligned_st));
89
+ return true;
90
+}
91
+
92
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
93
+{
94
+ return tcg_out_fail_alignment(s, l);
95
+}
96
+
97
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
98
+{
99
+ return tcg_out_fail_alignment(s, l);
100
+}
101
+
102
static void tcg_prepare_user_ldst(TCGContext *s, TCGReg *addr_reg,
103
TCGReg *index_reg, tcg_target_long *disp)
104
{
105
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext* s, TCGReg data_reg, TCGReg addr_reg,
106
#else
107
TCGReg index_reg;
108
tcg_target_long disp;
109
+ unsigned a_bits = get_alignment_bits(opc);
110
111
+ if (a_bits) {
112
+ tcg_out_test_alignment(s, true, addr_reg, a_bits);
113
+ }
114
tcg_prepare_user_ldst(s, &addr_reg, &index_reg, &disp);
115
tcg_out_qemu_ld_direct(s, opc, data_reg, addr_reg, index_reg, disp);
116
#endif
117
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext* s, TCGReg data_reg, TCGReg addr_reg,
118
#else
119
TCGReg index_reg;
120
tcg_target_long disp;
121
+ unsigned a_bits = get_alignment_bits(opc);
122
123
+ if (a_bits) {
124
+ tcg_out_test_alignment(s, false, addr_reg, a_bits);
125
+ }
126
tcg_prepare_user_ldst(s, &addr_reg, &index_reg, &disp);
127
tcg_out_qemu_st_direct(s, opc, data_reg, addr_reg, index_reg, disp);
128
#endif
129
--
171
--
130
2.25.1
172
2.34.1
131
132
diff view generated by jsdifflib
1
From armv6, the architecture supports unaligned accesses.
1
The mmap_lock is held around tb_gen_code. While the comment
2
All we need to do is perform the correct alignment check
2
is correct that the lock is dropped when tb_gen_code runs out
3
in tcg_out_tlb_read.
3
of memory, the lock is *not* dropped when an exception is
4
raised reading code for translation.
4
5
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Acked-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
10
---
8
tcg/arm/tcg-target.c.inc | 41 ++++++++++++++++++++--------------------
11
accel/tcg/cpu-exec.c | 12 ++++++------
9
1 file changed, 21 insertions(+), 20 deletions(-)
12
accel/tcg/user-exec.c | 3 ---
13
2 files changed, 6 insertions(+), 9 deletions(-)
10
14
11
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
15
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
12
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/arm/tcg-target.c.inc
17
--- a/accel/tcg/cpu-exec.c
14
+++ b/tcg/arm/tcg-target.c.inc
18
+++ b/accel/tcg/cpu-exec.c
15
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
19
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
16
int cmp_off = (is_load ? offsetof(CPUTLBEntry, addr_read)
20
cpu_tb_exec(cpu, tb, &tb_exit);
17
: offsetof(CPUTLBEntry, addr_write));
21
cpu_exec_exit(cpu);
18
int fast_off = TLB_MASK_TABLE_OFS(mem_index);
19
- unsigned s_bits = opc & MO_SIZE;
20
- unsigned a_bits = get_alignment_bits(opc);
21
-
22
- /*
23
- * We don't support inline unaligned acceses, but we can easily
24
- * support overalignment checks.
25
- */
26
- if (a_bits < s_bits) {
27
- a_bits = s_bits;
28
- }
29
+ unsigned s_mask = (1 << (opc & MO_SIZE)) - 1;
30
+ unsigned a_mask = (1 << get_alignment_bits(opc)) - 1;
31
+ TCGReg t_addr;
32
33
/* Load env_tlb(env)->f[mmu_idx].{mask,table} into {r0,r1}. */
34
tcg_out_ldrd_8(s, COND_AL, TCG_REG_R0, TCG_AREG0, fast_off);
35
@@ -XXX,XX +XXX,XX @@ static TCGReg tcg_out_tlb_read(TCGContext *s, TCGReg addrlo, TCGReg addrhi,
36
37
/*
38
* Check alignment, check comparators.
39
- * Do this in no more than 3 insns. Use MOVW for v7, if possible,
40
+ * Do this in 2-4 insns. Use MOVW for v7, if possible,
41
* to reduce the number of sequential conditional instructions.
42
* Almost all guests have at least 4k pages, which means that we need
43
* to clear at least 9 bits even for an 8-byte memory, which means it
44
* isn't worth checking for an immediate operand for BIC.
45
+ *
46
+ * For unaligned accesses, test the page of the last unit of alignment.
47
+ * This leaves the least significant alignment bits unchanged, and of
48
+ * course must be zero.
49
*/
50
+ t_addr = addrlo;
51
+ if (a_mask < s_mask) {
52
+ t_addr = TCG_REG_R0;
53
+ tcg_out_dat_imm(s, COND_AL, ARITH_ADD, t_addr,
54
+ addrlo, s_mask - a_mask);
55
+ }
56
if (use_armv7_instructions && TARGET_PAGE_BITS <= 16) {
57
- tcg_target_ulong mask = ~(TARGET_PAGE_MASK | ((1 << a_bits) - 1));
58
-
59
- tcg_out_movi32(s, COND_AL, TCG_REG_TMP, mask);
60
+ tcg_out_movi32(s, COND_AL, TCG_REG_TMP, ~(TARGET_PAGE_MASK | a_mask));
61
tcg_out_dat_reg(s, COND_AL, ARITH_BIC, TCG_REG_TMP,
62
- addrlo, TCG_REG_TMP, 0);
63
+ t_addr, TCG_REG_TMP, 0);
64
tcg_out_dat_reg(s, COND_AL, ARITH_CMP, 0, TCG_REG_R2, TCG_REG_TMP, 0);
65
} else {
22
} else {
66
- if (a_bits) {
23
- /*
67
- tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addrlo,
24
- * The mmap_lock is dropped by tb_gen_code if it runs out of
68
- (1 << a_bits) - 1);
25
- * memory.
69
+ if (a_mask) {
26
- */
70
+ tcg_debug_assert(a_mask <= 0xff);
27
#ifndef CONFIG_SOFTMMU
71
+ tcg_out_dat_imm(s, COND_AL, ARITH_TST, 0, addrlo, a_mask);
28
clear_helper_retaddr();
72
}
29
- tcg_debug_assert(!have_mmap_lock());
73
- tcg_out_dat_reg(s, COND_AL, ARITH_MOV, TCG_REG_TMP, 0, addrlo,
30
+ if (have_mmap_lock()) {
74
+ tcg_out_dat_reg(s, COND_AL, ARITH_MOV, TCG_REG_TMP, 0, t_addr,
31
+ mmap_unlock();
75
SHIFT_IMM_LSR(TARGET_PAGE_BITS));
32
+ }
76
- tcg_out_dat_reg(s, (a_bits ? COND_EQ : COND_AL), ARITH_CMP,
33
#endif
77
+ tcg_out_dat_reg(s, (a_mask ? COND_EQ : COND_AL), ARITH_CMP,
34
if (qemu_mutex_iothread_locked()) {
78
0, TCG_REG_R2, TCG_REG_TMP,
35
qemu_mutex_unlock_iothread();
79
SHIFT_IMM_LSL(TARGET_PAGE_BITS));
36
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
37
38
#ifndef CONFIG_SOFTMMU
39
clear_helper_retaddr();
40
- tcg_debug_assert(!have_mmap_lock());
41
+ if (have_mmap_lock()) {
42
+ mmap_unlock();
43
+ }
44
#endif
45
if (qemu_mutex_iothread_locked()) {
46
qemu_mutex_unlock_iothread();
47
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/accel/tcg/user-exec.c
50
+++ b/accel/tcg/user-exec.c
51
@@ -XXX,XX +XXX,XX @@ MMUAccessType adjust_signal_pc(uintptr_t *pc, bool is_write)
52
* (and if the translator doesn't handle page boundaries correctly
53
* there's little we can do about that here). Therefore, do not
54
* trigger the unwinder.
55
- *
56
- * Like tb_gen_code, release the memory lock before cpu_loop_exit.
57
*/
58
- mmap_unlock();
59
*pc = 0;
60
return MMU_INST_FETCH;
80
}
61
}
81
--
62
--
82
2.25.1
63
2.34.1
83
84
diff view generated by jsdifflib
1
We do not support sparc32 as a host, so there's no point in
1
The function is not used outside of cpu-exec.c. Move it and
2
sparc64 redirecting to sparc.
2
its subroutines up in the file, before the first use.
3
3
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
8
---
7
linux-user/include/host/sparc/host-signal.h | 71 -------------------
9
include/exec/exec-all.h | 3 -
8
linux-user/include/host/sparc64/host-signal.h | 64 ++++++++++++++++-
10
accel/tcg/cpu-exec.c | 122 ++++++++++++++++++++--------------------
9
2 files changed, 63 insertions(+), 72 deletions(-)
11
2 files changed, 61 insertions(+), 64 deletions(-)
10
delete mode 100644 linux-user/include/host/sparc/host-signal.h
11
12
12
diff --git a/linux-user/include/host/sparc/host-signal.h b/linux-user/include/host/sparc/host-signal.h
13
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
13
deleted file mode 100644
14
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX
15
--- a/include/exec/exec-all.h
15
--- a/linux-user/include/host/sparc/host-signal.h
16
+++ b/include/exec/exec-all.h
16
+++ /dev/null
17
@@ -XXX,XX +XXX,XX @@ void tb_invalidate_phys_addr(AddressSpace *as, hwaddr addr, MemTxAttrs attrs);
17
@@ -XXX,XX +XXX,XX @@
18
#endif
18
-/*
19
void tb_flush(CPUState *cpu);
19
- * host-signal.h: signal info dependent on the host architecture
20
void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr);
20
- *
21
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
21
- * Copyright (c) 2003-2005 Fabrice Bellard
22
- target_ulong cs_base, uint32_t flags,
22
- * Copyright (c) 2021 Linaro Limited
23
- uint32_t cflags);
23
- *
24
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr);
24
- * This work is licensed under the terms of the GNU LGPL, version 2.1 or later.
25
25
- * See the COPYING file in the top-level directory.
26
/* GETPC is the true target of the return instruction that we'll execute. */
26
- */
27
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/accel/tcg/cpu-exec.c
30
+++ b/accel/tcg/cpu-exec.c
31
@@ -XXX,XX +XXX,XX @@ uint32_t curr_cflags(CPUState *cpu)
32
return cflags;
33
}
34
35
+struct tb_desc {
36
+ target_ulong pc;
37
+ target_ulong cs_base;
38
+ CPUArchState *env;
39
+ tb_page_addr_t phys_page1;
40
+ uint32_t flags;
41
+ uint32_t cflags;
42
+ uint32_t trace_vcpu_dstate;
43
+};
44
+
45
+static bool tb_lookup_cmp(const void *p, const void *d)
46
+{
47
+ const TranslationBlock *tb = p;
48
+ const struct tb_desc *desc = d;
49
+
50
+ if (tb->pc == desc->pc &&
51
+ tb->page_addr[0] == desc->phys_page1 &&
52
+ tb->cs_base == desc->cs_base &&
53
+ tb->flags == desc->flags &&
54
+ tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
55
+ tb_cflags(tb) == desc->cflags) {
56
+ /* check next page if needed */
57
+ if (tb->page_addr[1] == -1) {
58
+ return true;
59
+ } else {
60
+ tb_page_addr_t phys_page2;
61
+ target_ulong virt_page2;
62
+
63
+ virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
64
+ phys_page2 = get_page_addr_code(desc->env, virt_page2);
65
+ if (tb->page_addr[1] == phys_page2) {
66
+ return true;
67
+ }
68
+ }
69
+ }
70
+ return false;
71
+}
72
+
73
+static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
74
+ target_ulong cs_base, uint32_t flags,
75
+ uint32_t cflags)
76
+{
77
+ tb_page_addr_t phys_pc;
78
+ struct tb_desc desc;
79
+ uint32_t h;
80
+
81
+ desc.env = cpu->env_ptr;
82
+ desc.cs_base = cs_base;
83
+ desc.flags = flags;
84
+ desc.cflags = cflags;
85
+ desc.trace_vcpu_dstate = *cpu->trace_dstate;
86
+ desc.pc = pc;
87
+ phys_pc = get_page_addr_code(desc.env, pc);
88
+ if (phys_pc == -1) {
89
+ return NULL;
90
+ }
91
+ desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
92
+ h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
93
+ return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
94
+}
95
+
96
/* Might cause an exception, so have a longjmp destination ready */
97
static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
98
target_ulong cs_base,
99
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
100
end_exclusive();
101
}
102
103
-struct tb_desc {
104
- target_ulong pc;
105
- target_ulong cs_base;
106
- CPUArchState *env;
107
- tb_page_addr_t phys_page1;
108
- uint32_t flags;
109
- uint32_t cflags;
110
- uint32_t trace_vcpu_dstate;
111
-};
27
-
112
-
28
-#ifndef SPARC_HOST_SIGNAL_H
113
-static bool tb_lookup_cmp(const void *p, const void *d)
29
-#define SPARC_HOST_SIGNAL_H
114
-{
115
- const TranslationBlock *tb = p;
116
- const struct tb_desc *desc = d;
30
-
117
-
31
-/* FIXME: the third argument to a SA_SIGINFO handler is *not* ucontext_t. */
118
- if (tb->pc == desc->pc &&
32
-typedef ucontext_t host_sigcontext;
119
- tb->page_addr[0] == desc->phys_page1 &&
120
- tb->cs_base == desc->cs_base &&
121
- tb->flags == desc->flags &&
122
- tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
123
- tb_cflags(tb) == desc->cflags) {
124
- /* check next page if needed */
125
- if (tb->page_addr[1] == -1) {
126
- return true;
127
- } else {
128
- tb_page_addr_t phys_page2;
129
- target_ulong virt_page2;
33
-
130
-
34
-static inline uintptr_t host_signal_pc(host_sigcontext *uc)
131
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
35
-{
132
- phys_page2 = get_page_addr_code(desc->env, virt_page2);
36
-#ifdef __arch64__
133
- if (tb->page_addr[1] == phys_page2) {
37
- return uc->uc_mcontext.mc_gregs[MC_PC];
134
- return true;
38
-#else
135
- }
39
- return uc->uc_mcontext.gregs[REG_PC];
40
-#endif
41
-}
42
-
43
-static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
44
-{
45
-#ifdef __arch64__
46
- uc->uc_mcontext.mc_gregs[MC_PC] = pc;
47
-#else
48
- uc->uc_mcontext.gregs[REG_PC] = pc;
49
-#endif
50
-}
51
-
52
-static inline void *host_signal_mask(host_sigcontext *uc)
53
-{
54
- return &uc->uc_sigmask;
55
-}
56
-
57
-static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
58
-{
59
- uint32_t insn = *(uint32_t *)host_signal_pc(uc);
60
-
61
- if ((insn >> 30) == 3) {
62
- switch ((insn >> 19) & 0x3f) {
63
- case 0x05: /* stb */
64
- case 0x15: /* stba */
65
- case 0x06: /* sth */
66
- case 0x16: /* stha */
67
- case 0x04: /* st */
68
- case 0x14: /* sta */
69
- case 0x07: /* std */
70
- case 0x17: /* stda */
71
- case 0x0e: /* stx */
72
- case 0x1e: /* stxa */
73
- case 0x24: /* stf */
74
- case 0x34: /* stfa */
75
- case 0x27: /* stdf */
76
- case 0x37: /* stdfa */
77
- case 0x26: /* stqf */
78
- case 0x36: /* stqfa */
79
- case 0x25: /* stfsr */
80
- case 0x3c: /* casa */
81
- case 0x3e: /* casxa */
82
- return true;
83
- }
136
- }
84
- }
137
- }
85
- return false;
138
- return false;
86
-}
139
-}
87
-
140
-
88
-#endif
141
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
89
diff --git a/linux-user/include/host/sparc64/host-signal.h b/linux-user/include/host/sparc64/host-signal.h
142
- target_ulong cs_base, uint32_t flags,
90
index XXXXXXX..XXXXXXX 100644
143
- uint32_t cflags)
91
--- a/linux-user/include/host/sparc64/host-signal.h
144
-{
92
+++ b/linux-user/include/host/sparc64/host-signal.h
145
- tb_page_addr_t phys_pc;
93
@@ -1 +1,63 @@
146
- struct tb_desc desc;
94
-#include "../sparc/host-signal.h"
147
- uint32_t h;
95
+/*
148
-
96
+ * host-signal.h: signal info dependent on the host architecture
149
- desc.env = cpu->env_ptr;
97
+ *
150
- desc.cs_base = cs_base;
98
+ * Copyright (c) 2003-2005 Fabrice Bellard
151
- desc.flags = flags;
99
+ * Copyright (c) 2021 Linaro Limited
152
- desc.cflags = cflags;
100
+ *
153
- desc.trace_vcpu_dstate = *cpu->trace_dstate;
101
+ * This work is licensed under the terms of the GNU LGPL, version 2.1 or later.
154
- desc.pc = pc;
102
+ * See the COPYING file in the top-level directory.
155
- phys_pc = get_page_addr_code(desc.env, pc);
103
+ */
156
- if (phys_pc == -1) {
104
+
157
- return NULL;
105
+#ifndef SPARC64_HOST_SIGNAL_H
158
- }
106
+#define SPARC64_HOST_SIGNAL_H
159
- desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
107
+
160
- h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
108
+/* FIXME: the third argument to a SA_SIGINFO handler is *not* ucontext_t. */
161
- return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
109
+typedef ucontext_t host_sigcontext;
162
-}
110
+
163
-
111
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
164
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr)
112
+{
165
{
113
+ return uc->uc_mcontext.mc_gregs[MC_PC];
166
if (TCG_TARGET_HAS_direct_jump) {
114
+}
115
+
116
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
117
+{
118
+ uc->uc_mcontext.mc_gregs[MC_PC] = pc;
119
+}
120
+
121
+static inline void *host_signal_mask(host_sigcontext *uc)
122
+{
123
+ return &uc->uc_sigmask;
124
+}
125
+
126
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
127
+{
128
+ uint32_t insn = *(uint32_t *)host_signal_pc(uc);
129
+
130
+ if ((insn >> 30) == 3) {
131
+ switch ((insn >> 19) & 0x3f) {
132
+ case 0x05: /* stb */
133
+ case 0x15: /* stba */
134
+ case 0x06: /* sth */
135
+ case 0x16: /* stha */
136
+ case 0x04: /* st */
137
+ case 0x14: /* sta */
138
+ case 0x07: /* std */
139
+ case 0x17: /* stda */
140
+ case 0x0e: /* stx */
141
+ case 0x1e: /* stxa */
142
+ case 0x24: /* stf */
143
+ case 0x34: /* stfa */
144
+ case 0x27: /* stdf */
145
+ case 0x37: /* stdfa */
146
+ case 0x26: /* stqf */
147
+ case 0x36: /* stqfa */
148
+ case 0x25: /* stfsr */
149
+ case 0x3c: /* casa */
150
+ case 0x3e: /* casxa */
151
+ return true;
152
+ }
153
+ }
154
+ return false;
155
+}
156
+
157
+#endif
158
--
167
--
159
2.25.1
168
2.34.1
160
161
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
The base qemu_ram_addr_from_host function is already in
2
softmmu/physmem.c; move the nofail version to be adjacent.
3
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
8
---
4
tcg/aarch64/tcg-target.h | 2 -
9
include/exec/cpu-common.h | 1 +
5
tcg/aarch64/tcg-target.c.inc | 91 +++++++++++++++++++++++++++++-------
10
accel/tcg/cputlb.c | 12 ------------
6
2 files changed, 74 insertions(+), 19 deletions(-)
11
softmmu/physmem.c | 12 ++++++++++++
12
3 files changed, 13 insertions(+), 12 deletions(-)
7
13
8
diff --git a/tcg/aarch64/tcg-target.h b/tcg/aarch64/tcg-target.h
14
diff --git a/include/exec/cpu-common.h b/include/exec/cpu-common.h
9
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/aarch64/tcg-target.h
16
--- a/include/exec/cpu-common.h
11
+++ b/tcg/aarch64/tcg-target.h
17
+++ b/include/exec/cpu-common.h
12
@@ -XXX,XX +XXX,XX @@ typedef enum {
18
@@ -XXX,XX +XXX,XX @@ typedef uintptr_t ram_addr_t;
13
19
void qemu_ram_remap(ram_addr_t addr, ram_addr_t length);
14
void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
20
/* This should not be used by devices. */
15
21
ram_addr_t qemu_ram_addr_from_host(void *ptr);
16
-#ifdef CONFIG_SOFTMMU
22
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr);
17
#define TCG_TARGET_NEED_LDST_LABELS
23
RAMBlock *qemu_ram_block_by_name(const char *name);
18
-#endif
24
RAMBlock *qemu_ram_block_from_host(void *ptr, bool round_offset,
19
#define TCG_TARGET_NEED_POOL_LABELS
25
ram_addr_t *offset);
20
26
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
21
#endif /* AARCH64_TCG_TARGET_H */
22
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
23
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
24
--- a/tcg/aarch64/tcg-target.c.inc
28
--- a/accel/tcg/cputlb.c
25
+++ b/tcg/aarch64/tcg-target.c.inc
29
+++ b/accel/tcg/cputlb.c
26
@@ -XXX,XX +XXX,XX @@
30
@@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr,
27
* See the COPYING file in the top-level directory for details.
31
prot, mmu_idx, size);
28
*/
29
30
+#include "../tcg-ldst.c.inc"
31
#include "../tcg-pool.c.inc"
32
#include "qemu/bitops.h"
33
34
@@ -XXX,XX +XXX,XX @@ typedef enum {
35
I3404_ANDI = 0x12000000,
36
I3404_ORRI = 0x32000000,
37
I3404_EORI = 0x52000000,
38
+ I3404_ANDSI = 0x72000000,
39
40
/* Move wide immediate instructions. */
41
I3405_MOVN = 0x12800000,
42
@@ -XXX,XX +XXX,XX @@ static void tcg_out_goto_long(TCGContext *s, const tcg_insn_unit *target)
43
if (offset == sextract64(offset, 0, 26)) {
44
tcg_out_insn(s, 3206, B, offset);
45
} else {
46
- tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP, (intptr_t)target);
47
- tcg_out_insn(s, 3207, BR, TCG_REG_TMP);
48
+ /* Choose X9 as a call-clobbered non-LR temporary. */
49
+ tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_X9, (intptr_t)target);
50
+ tcg_out_insn(s, 3207, BR, TCG_REG_X9);
51
}
52
}
32
}
53
33
54
@@ -XXX,XX +XXX,XX @@ static void tcg_out_cltz(TCGContext *s, TCGType ext, TCGReg d,
34
-static inline ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
55
}
56
}
57
58
-#ifdef CONFIG_SOFTMMU
59
-#include "../tcg-ldst.c.inc"
60
+static void tcg_out_adr(TCGContext *s, TCGReg rd, const void *target)
61
+{
62
+ ptrdiff_t offset = tcg_pcrel_diff(s, target);
63
+ tcg_debug_assert(offset == sextract64(offset, 0, 21));
64
+ tcg_out_insn(s, 3406, ADR, rd, offset);
65
+}
66
67
+#ifdef CONFIG_SOFTMMU
68
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
69
* MemOpIdx oi, uintptr_t ra)
70
*/
71
@@ -XXX,XX +XXX,XX @@ static void * const qemu_st_helpers[MO_SIZE + 1] = {
72
#endif
73
};
74
75
-static inline void tcg_out_adr(TCGContext *s, TCGReg rd, const void *target)
76
-{
35
-{
77
- ptrdiff_t offset = tcg_pcrel_diff(s, target);
36
- ram_addr_t ram_addr;
78
- tcg_debug_assert(offset == sextract64(offset, 0, 21));
37
-
79
- tcg_out_insn(s, 3406, ADR, rd, offset);
38
- ram_addr = qemu_ram_addr_from_host(ptr);
39
- if (ram_addr == RAM_ADDR_INVALID) {
40
- error_report("Bad ram pointer %p", ptr);
41
- abort();
42
- }
43
- return ram_addr;
80
-}
44
-}
81
-
45
-
82
static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
46
/*
83
{
47
* Note: tlb_fill() can trigger a resize of the TLB. This means that all of the
84
MemOpIdx oi = lb->oi;
48
* caller's prior references to the TLB table (e.g. CPUTLBEntry pointers) must
85
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_read(TCGContext *s, TCGReg addr_reg, MemOp opc,
49
diff --git a/softmmu/physmem.c b/softmmu/physmem.c
86
tcg_out_insn(s, 3202, B_C, TCG_COND_NE, 0);
50
index XXXXXXX..XXXXXXX 100644
51
--- a/softmmu/physmem.c
52
+++ b/softmmu/physmem.c
53
@@ -XXX,XX +XXX,XX @@ ram_addr_t qemu_ram_addr_from_host(void *ptr)
54
return block->offset + offset;
87
}
55
}
88
56
89
+#else
57
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
90
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addr_reg,
91
+ unsigned a_bits)
92
+{
58
+{
93
+ unsigned a_mask = (1 << a_bits) - 1;
59
+ ram_addr_t ram_addr;
94
+ TCGLabelQemuLdst *label = new_ldst_label(s);
95
+
60
+
96
+ label->is_ld = is_ld;
61
+ ram_addr = qemu_ram_addr_from_host(ptr);
97
+ label->addrlo_reg = addr_reg;
62
+ if (ram_addr == RAM_ADDR_INVALID) {
98
+
63
+ error_report("Bad ram pointer %p", ptr);
99
+ /* tst addr, #mask */
64
+ abort();
100
+ tcg_out_logicali(s, I3404_ANDSI, 0, TCG_REG_XZR, addr_reg, a_mask);
65
+ }
101
+
66
+ return ram_addr;
102
+ label->label_ptr[0] = s->code_ptr;
103
+
104
+ /* b.ne slow_path */
105
+ tcg_out_insn(s, 3202, B_C, TCG_COND_NE, 0);
106
+
107
+ label->raddr = tcg_splitwx_to_rx(s->code_ptr);
108
+}
67
+}
109
+
68
+
110
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
69
static MemTxResult flatview_read(FlatView *fv, hwaddr addr,
111
+{
70
MemTxAttrs attrs, void *buf, hwaddr len);
112
+ if (!reloc_pc19(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
71
static MemTxResult flatview_write(FlatView *fv, hwaddr addr, MemTxAttrs attrs,
113
+ return false;
114
+ }
115
+
116
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_X1, l->addrlo_reg);
117
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_X0, TCG_AREG0);
118
+
119
+ /* "Tail call" to the helper, with the return address back inline. */
120
+ tcg_out_adr(s, TCG_REG_LR, l->raddr);
121
+ tcg_out_goto_long(s, (const void *)(l->is_ld ? helper_unaligned_ld
122
+ : helper_unaligned_st));
123
+ return true;
124
+}
125
+
126
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
127
+{
128
+ return tcg_out_fail_alignment(s, l);
129
+}
130
+
131
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
132
+{
133
+ return tcg_out_fail_alignment(s, l);
134
+}
135
#endif /* CONFIG_SOFTMMU */
136
137
static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
138
TCGReg data_r, TCGReg addr_r,
139
TCGType otype, TCGReg off_r)
140
{
141
- /* Byte swapping is left to middle-end expansion. */
142
- tcg_debug_assert((memop & MO_BSWAP) == 0);
143
-
144
switch (memop & MO_SSIZE) {
145
case MO_UB:
146
tcg_out_ldst_r(s, I3312_LDRB, data_r, addr_r, otype, off_r);
147
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp memop,
148
TCGReg data_r, TCGReg addr_r,
149
TCGType otype, TCGReg off_r)
150
{
151
- /* Byte swapping is left to middle-end expansion. */
152
- tcg_debug_assert((memop & MO_BSWAP) == 0);
153
-
154
switch (memop & MO_SIZE) {
155
case MO_8:
156
tcg_out_ldst_r(s, I3312_STRB, data_r, addr_r, otype, off_r);
157
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
158
{
159
MemOp memop = get_memop(oi);
160
const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
161
+
162
+ /* Byte swapping is left to middle-end expansion. */
163
+ tcg_debug_assert((memop & MO_BSWAP) == 0);
164
+
165
#ifdef CONFIG_SOFTMMU
166
unsigned mem_index = get_mmuidx(oi);
167
tcg_insn_unit *label_ptr;
168
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
169
add_qemu_ldst_label(s, true, oi, ext, data_reg, addr_reg,
170
s->code_ptr, label_ptr);
171
#else /* !CONFIG_SOFTMMU */
172
+ unsigned a_bits = get_alignment_bits(memop);
173
+ if (a_bits) {
174
+ tcg_out_test_alignment(s, true, addr_reg, a_bits);
175
+ }
176
if (USE_GUEST_BASE) {
177
tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
178
TCG_REG_GUEST_BASE, otype, addr_reg);
179
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
180
{
181
MemOp memop = get_memop(oi);
182
const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
183
+
184
+ /* Byte swapping is left to middle-end expansion. */
185
+ tcg_debug_assert((memop & MO_BSWAP) == 0);
186
+
187
#ifdef CONFIG_SOFTMMU
188
unsigned mem_index = get_mmuidx(oi);
189
tcg_insn_unit *label_ptr;
190
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
191
add_qemu_ldst_label(s, false, oi, (memop & MO_SIZE)== MO_64,
192
data_reg, addr_reg, s->code_ptr, label_ptr);
193
#else /* !CONFIG_SOFTMMU */
194
+ unsigned a_bits = get_alignment_bits(memop);
195
+ if (a_bits) {
196
+ tcg_out_test_alignment(s, false, addr_reg, a_bits);
197
+ }
198
if (USE_GUEST_BASE) {
199
tcg_out_qemu_st_direct(s, memop, data_reg,
200
TCG_REG_GUEST_BASE, otype, addr_reg);
201
--
72
--
202
2.25.1
73
2.34.1
203
204
diff view generated by jsdifflib
1
From: Idan Horowitz <idan.horowitz@gmail.com>
1
Simplify the implementation of get_page_addr_code_hostp
2
by reusing the existing probe_access infrastructure.
2
3
3
When the length of the range is large enough, clearing the whole cache is
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
faster than iterating over the (possibly extremely large) set of pages
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
contained in the range.
6
7
This mimics the pre-existing similar optimization done on the flush of the
8
tlb itself.
9
10
Signed-off-by: Idan Horowitz <idan.horowitz@gmail.com>
11
Message-Id: <20220110164754.1066025-1-idan.horowitz@gmail.com>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
14
---
7
---
15
accel/tcg/cputlb.c | 9 +++++++++
8
accel/tcg/cputlb.c | 76 ++++++++++++++++------------------------------
16
1 file changed, 9 insertions(+)
9
1 file changed, 26 insertions(+), 50 deletions(-)
17
10
18
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
11
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
19
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
20
--- a/accel/tcg/cputlb.c
13
--- a/accel/tcg/cputlb.c
21
+++ b/accel/tcg/cputlb.c
14
+++ b/accel/tcg/cputlb.c
22
@@ -XXX,XX +XXX,XX @@ static void tlb_flush_range_by_mmuidx_async_0(CPUState *cpu,
15
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
23
}
16
victim_tlb_hit(env, mmu_idx, index, offsetof(CPUTLBEntry, TY), \
24
qemu_spin_unlock(&env_tlb(env)->c.lock);
17
(ADDR) & TARGET_PAGE_MASK)
25
18
26
+ /*
19
-/*
27
+ * If the length is larger than the jump cache size, then it will take
20
- * Return a ram_addr_t for the virtual address for execution.
28
+ * longer to clear each entry individually than it will to clear it all.
21
- *
29
+ */
22
- * Return -1 if we can't translate and execute from an entire page
30
+ if (d.len >= (TARGET_PAGE_SIZE * TB_JMP_CACHE_SIZE)) {
23
- * of RAM. This will force us to execute by loading and translating
31
+ cpu_tb_jmp_cache_clear(cpu);
24
- * one insn at a time, without caching.
32
+ return;
25
- *
26
- * NOTE: This function will trigger an exception if the page is
27
- * not executable.
28
- */
29
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
30
- void **hostp)
31
-{
32
- uintptr_t mmu_idx = cpu_mmu_index(env, true);
33
- uintptr_t index = tlb_index(env, mmu_idx, addr);
34
- CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr);
35
- void *p;
36
-
37
- if (unlikely(!tlb_hit(entry->addr_code, addr))) {
38
- if (!VICTIM_TLB_HIT(addr_code, addr)) {
39
- tlb_fill(env_cpu(env), addr, 0, MMU_INST_FETCH, mmu_idx, 0);
40
- index = tlb_index(env, mmu_idx, addr);
41
- entry = tlb_entry(env, mmu_idx, addr);
42
-
43
- if (unlikely(entry->addr_code & TLB_INVALID_MASK)) {
44
- /*
45
- * The MMU protection covers a smaller range than a target
46
- * page, so we must redo the MMU check for every insn.
47
- */
48
- return -1;
49
- }
50
- }
51
- assert(tlb_hit(entry->addr_code, addr));
52
- }
53
-
54
- if (unlikely(entry->addr_code & TLB_MMIO)) {
55
- /* The region is not backed by RAM. */
56
- if (hostp) {
57
- *hostp = NULL;
58
- }
59
- return -1;
60
- }
61
-
62
- p = (void *)((uintptr_t)addr + entry->addend);
63
- if (hostp) {
64
- *hostp = p;
65
- }
66
- return qemu_ram_addr_from_host_nofail(p);
67
-}
68
-
69
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
70
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
71
{
72
@@ -XXX,XX +XXX,XX @@ void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
73
return flags ? NULL : host;
74
}
75
76
+/*
77
+ * Return a ram_addr_t for the virtual address for execution.
78
+ *
79
+ * Return -1 if we can't translate and execute from an entire page
80
+ * of RAM. This will force us to execute by loading and translating
81
+ * one insn at a time, without caching.
82
+ *
83
+ * NOTE: This function will trigger an exception if the page is
84
+ * not executable.
85
+ */
86
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
87
+ void **hostp)
88
+{
89
+ void *p;
90
+
91
+ (void)probe_access_internal(env, addr, 1, MMU_INST_FETCH,
92
+ cpu_mmu_index(env, true), false, &p, 0);
93
+ if (p == NULL) {
94
+ return -1;
33
+ }
95
+ }
96
+ if (hostp) {
97
+ *hostp = p;
98
+ }
99
+ return qemu_ram_addr_from_host_nofail(p);
100
+}
34
+
101
+
35
for (target_ulong i = 0; i < d.len; i += TARGET_PAGE_SIZE) {
102
#ifdef CONFIG_PLUGIN
36
tb_flush_jmp_cache(cpu, d.addr + i);
103
/*
37
}
104
* Perform a TLB lookup and populate the qemu_plugin_hwaddr structure.
38
--
105
--
39
2.25.1
106
2.34.1
40
41
diff view generated by jsdifflib
1
From: Pavel Dovgalyuk <pavel.dovgalyuk@ispras.ru>
1
It was non-obvious to me why we can raise an exception in
2
the middle of a comparison function, but it works.
3
While nearby, use TARGET_PAGE_ALIGN instead of open-coding.
2
4
3
Commit aff0e204cb1f1c036a496c94c15f5dfafcd9b4b4 introduced CF_NOIRQ usage,
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
but one case was forgotten. Record/replay uses one special TB which is not
5
really executed, but used to cause a correct exception in replay mode.
6
This patch adds CF_NOIRQ flag for such block.
7
8
Signed-off-by: Pavel Dovgalyuk <Pavel.Dovgalyuk@ispras.ru>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-Id: <164362834054.1754532.7678416881159817273.stgit@pasha-ThinkPad-X280>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
---
7
---
13
accel/tcg/cpu-exec.c | 3 ++-
8
accel/tcg/cpu-exec.c | 11 ++++++++++-
14
1 file changed, 2 insertions(+), 1 deletion(-)
9
1 file changed, 10 insertions(+), 1 deletion(-)
15
10
16
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
11
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
17
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
18
--- a/accel/tcg/cpu-exec.c
13
--- a/accel/tcg/cpu-exec.c
19
+++ b/accel/tcg/cpu-exec.c
14
+++ b/accel/tcg/cpu-exec.c
20
@@ -XXX,XX +XXX,XX @@ static inline bool cpu_handle_exception(CPUState *cpu, int *ret)
15
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
21
if (replay_has_exception()
16
tb_page_addr_t phys_page2;
22
&& cpu_neg(cpu)->icount_decr.u16.low + cpu->icount_extra == 0) {
17
target_ulong virt_page2;
23
/* Execute just one insn to trigger exception pending in the log */
18
24
- cpu->cflags_next_tb = (curr_cflags(cpu) & ~CF_USE_ICOUNT) | 1;
19
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
25
+ cpu->cflags_next_tb = (curr_cflags(cpu) & ~CF_USE_ICOUNT)
20
+ /*
26
+ | CF_NOIRQ | 1;
21
+ * We know that the first page matched, and an otherwise valid TB
27
}
22
+ * encountered an incomplete instruction at the end of that page,
28
#endif
23
+ * therefore we know that generating a new TB from the current PC
29
return false;
24
+ * must also require reading from the next page -- even if the
25
+ * second pages do not match, and therefore the resulting insn
26
+ * is different for the new TB. Therefore any exception raised
27
+ * here by the faulting lookup is not premature.
28
+ */
29
+ virt_page2 = TARGET_PAGE_ALIGN(desc->pc);
30
phys_page2 = get_page_addr_code(desc->env, virt_page2);
31
if (tb->page_addr[1] == phys_page2) {
32
return true;
30
--
33
--
31
2.25.1
34
2.34.1
32
33
diff view generated by jsdifflib
1
This will allow us to control exactly what scratch register is
1
The only user can easily use translator_lduw and
2
used for loading the constant.
2
adjust the type to signed during the return.
3
3
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
8
---
7
tcg/sparc/tcg-target.c.inc | 15 +++++++++------
9
include/exec/translator.h | 1 -
8
1 file changed, 9 insertions(+), 6 deletions(-)
10
target/i386/tcg/translate.c | 2 +-
11
2 files changed, 1 insertion(+), 2 deletions(-)
9
12
10
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
13
diff --git a/include/exec/translator.h b/include/exec/translator.h
11
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/sparc/tcg-target.c.inc
15
--- a/include/exec/translator.h
13
+++ b/tcg/sparc/tcg-target.c.inc
16
+++ b/include/exec/translator.h
14
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_imm32(TCGContext *s, TCGReg ret, int32_t arg)
17
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
18
19
#define FOR_EACH_TRANSLATOR_LD(F) \
20
F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
21
- F(translator_ldsw, int16_t, cpu_ldsw_code, bswap16) \
22
F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
23
F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
24
F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
25
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/i386/tcg/translate.c
28
+++ b/target/i386/tcg/translate.c
29
@@ -XXX,XX +XXX,XX @@ static inline uint8_t x86_ldub_code(CPUX86State *env, DisasContext *s)
30
31
static inline int16_t x86_ldsw_code(CPUX86State *env, DisasContext *s)
32
{
33
- return translator_ldsw(env, &s->base, advance_pc(env, s, 2));
34
+ return translator_lduw(env, &s->base, advance_pc(env, s, 2));
15
}
35
}
16
36
17
static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
37
static inline uint16_t x86_lduw_code(CPUX86State *env, DisasContext *s)
18
- tcg_target_long arg, bool in_prologue)
19
+ tcg_target_long arg, bool in_prologue,
20
+ TCGReg scratch)
21
{
22
tcg_target_long hi, lo = (int32_t)arg;
23
tcg_target_long test, lsb;
24
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
25
} else {
26
hi = arg >> 32;
27
tcg_out_movi_imm32(s, ret, hi);
28
- tcg_out_movi_imm32(s, TCG_REG_T2, lo);
29
+ tcg_out_movi_imm32(s, scratch, lo);
30
tcg_out_arithi(s, ret, ret, 32, SHIFT_SLLX);
31
- tcg_out_arith(s, ret, ret, TCG_REG_T2, ARITH_OR);
32
+ tcg_out_arith(s, ret, ret, scratch, ARITH_OR);
33
}
34
}
35
36
static void tcg_out_movi(TCGContext *s, TCGType type,
37
TCGReg ret, tcg_target_long arg)
38
{
39
- tcg_out_movi_int(s, type, ret, arg, false);
40
+ tcg_debug_assert(ret != TCG_REG_T2);
41
+ tcg_out_movi_int(s, type, ret, arg, false, TCG_REG_T2);
42
}
43
44
static void tcg_out_ldst_rr(TCGContext *s, TCGReg data, TCGReg a1,
45
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call_nodelay(TCGContext *s, const tcg_insn_unit *dest,
46
} else {
47
uintptr_t desti = (uintptr_t)dest;
48
tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_REG_T1,
49
- desti & ~0xfff, in_prologue);
50
+ desti & ~0xfff, in_prologue, TCG_REG_O7);
51
tcg_out_arithi(s, TCG_REG_O7, TCG_REG_T1, desti & 0xfff, JMPL);
52
}
53
}
54
@@ -XXX,XX +XXX,XX @@ static void tcg_target_qemu_prologue(TCGContext *s)
55
56
#ifndef CONFIG_SOFTMMU
57
if (guest_base != 0) {
58
- tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, guest_base, true);
59
+ tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG,
60
+ guest_base, true, TCG_REG_T1);
61
tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
62
}
63
#endif
64
--
38
--
65
2.25.1
39
2.34.1
66
67
diff view generated by jsdifflib
1
We will shortly allow the use of unaligned memory accesses,
1
Pass these along to translator_loop -- pc may be used instead
2
and these require proper alignment. Use get_alignment_bits
2
of tb->pc, and host_pc is currently unused. Adjust all targets
3
to verify and remove USING_SOFTMMU.
3
at one time.
4
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Acked-by: Alistair Francis <alistair.francis@wdc.com>
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
9
---
8
tcg/arm/tcg-target.c.inc | 23 ++++++++---------------
10
include/exec/exec-all.h | 1 -
9
1 file changed, 8 insertions(+), 15 deletions(-)
11
include/exec/translator.h | 24 ++++++++++++++++++++----
12
accel/tcg/translate-all.c | 6 ++++--
13
accel/tcg/translator.c | 9 +++++----
14
target/alpha/translate.c | 5 +++--
15
target/arm/translate.c | 5 +++--
16
target/avr/translate.c | 5 +++--
17
target/cris/translate.c | 5 +++--
18
target/hexagon/translate.c | 6 ++++--
19
target/hppa/translate.c | 5 +++--
20
target/i386/tcg/translate.c | 5 +++--
21
target/loongarch/translate.c | 6 ++++--
22
target/m68k/translate.c | 5 +++--
23
target/microblaze/translate.c | 5 +++--
24
target/mips/tcg/translate.c | 5 +++--
25
target/nios2/translate.c | 5 +++--
26
target/openrisc/translate.c | 6 ++++--
27
target/ppc/translate.c | 5 +++--
28
target/riscv/translate.c | 5 +++--
29
target/rx/translate.c | 5 +++--
30
target/s390x/tcg/translate.c | 5 +++--
31
target/sh4/translate.c | 5 +++--
32
target/sparc/translate.c | 5 +++--
33
target/tricore/translate.c | 6 ++++--
34
target/xtensa/translate.c | 6 ++++--
35
25 files changed, 97 insertions(+), 53 deletions(-)
10
36
11
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
37
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
12
index XXXXXXX..XXXXXXX 100644
38
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/arm/tcg-target.c.inc
39
--- a/include/exec/exec-all.h
14
+++ b/tcg/arm/tcg-target.c.inc
40
+++ b/include/exec/exec-all.h
15
@@ -XXX,XX +XXX,XX @@ bool use_idiv_instructions;
41
@@ -XXX,XX +XXX,XX @@ typedef ram_addr_t tb_page_addr_t;
16
bool use_neon_instructions;
42
#define TB_PAGE_ADDR_FMT RAM_ADDR_FMT
17
#endif
43
#endif
18
44
19
-/* ??? Ought to think about changing CONFIG_SOFTMMU to always defined. */
45
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns);
20
-#ifdef CONFIG_SOFTMMU
46
void restore_state_to_opc(CPUArchState *env, TranslationBlock *tb,
21
-# define USING_SOFTMMU 1
47
target_ulong *data);
22
-#else
48
23
-# define USING_SOFTMMU 0
49
diff --git a/include/exec/translator.h b/include/exec/translator.h
24
-#endif
50
index XXXXXXX..XXXXXXX 100644
25
-
51
--- a/include/exec/translator.h
26
#ifdef CONFIG_DEBUG_TCG
52
+++ b/include/exec/translator.h
27
static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
53
@@ -XXX,XX +XXX,XX @@
28
"%r0", "%r1", "%r2", "%r3", "%r4", "%r5", "%r6", "%r7",
54
#include "exec/translate-all.h"
29
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
55
#include "tcg/tcg.h"
30
tcg_out_ld32_r(s, COND_AL, datalo, addrlo, addend);
56
31
break;
57
+/**
32
case MO_UQ:
58
+ * gen_intermediate_code
33
- /* Avoid ldrd for user-only emulation, to handle unaligned. */
59
+ * @cpu: cpu context
34
- if (USING_SOFTMMU
60
+ * @tb: translation block
35
+ /* LDRD requires alignment; double-check that. */
61
+ * @max_insns: max number of instructions to translate
36
+ if (get_alignment_bits(opc) >= MO_64
62
+ * @pc: guest virtual program counter address
37
&& (datalo & 1) == 0 && datahi == datalo + 1) {
63
+ * @host_pc: host physical program counter address
38
tcg_out_ldrd_r(s, COND_AL, datalo, addrlo, addend);
64
+ *
39
} else if (datalo != addend) {
65
+ * This function must be provided by the target, which should create
40
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc, TCGReg datalo,
66
+ * the target-specific DisasContext, and then invoke translator_loop.
41
tcg_out_ld32_12(s, COND_AL, datalo, addrlo, 0);
67
+ */
42
break;
68
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
43
case MO_UQ:
69
+ target_ulong pc, void *host_pc);
44
- /* Avoid ldrd for user-only emulation, to handle unaligned. */
70
45
- if (USING_SOFTMMU
71
/**
46
+ /* LDRD requires alignment; double-check that. */
72
* DisasJumpType:
47
+ if (get_alignment_bits(opc) >= MO_64
73
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
48
&& (datalo & 1) == 0 && datahi == datalo + 1) {
74
49
tcg_out_ldrd_8(s, COND_AL, datalo, addrlo, 0);
75
/**
50
} else if (datalo == addrlo) {
76
* translator_loop:
51
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_index(TCGContext *s, ARMCond cond, MemOp opc,
77
- * @ops: Target-specific operations.
52
tcg_out_st32_r(s, cond, datalo, addrlo, addend);
78
- * @db: Disassembly context.
53
break;
79
* @cpu: Target vCPU.
54
case MO_64:
80
* @tb: Translation block.
55
- /* Avoid strd for user-only emulation, to handle unaligned. */
81
* @max_insns: Maximum number of insns to translate.
56
- if (USING_SOFTMMU
82
+ * @pc: guest virtual program counter address
57
+ /* STRD requires alignment; double-check that. */
83
+ * @host_pc: host physical program counter address
58
+ if (get_alignment_bits(opc) >= MO_64
84
+ * @ops: Target-specific operations.
59
&& (datalo & 1) == 0 && datahi == datalo + 1) {
85
+ * @db: Disassembly context.
60
tcg_out_strd_r(s, cond, datalo, addrlo, addend);
86
*
61
} else {
87
* Generic translator loop.
62
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc, TCGReg datalo,
88
*
63
tcg_out_st32_12(s, COND_AL, datalo, addrlo, 0);
89
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
64
break;
90
* - When single-stepping is enabled (system-wide or on the current vCPU).
65
case MO_64:
91
* - When too many instructions have been translated.
66
- /* Avoid strd for user-only emulation, to handle unaligned. */
92
*/
67
- if (USING_SOFTMMU
93
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
68
+ /* STRD requires alignment; double-check that. */
94
- CPUState *cpu, TranslationBlock *tb, int max_insns);
69
+ if (get_alignment_bits(opc) >= MO_64
95
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
70
&& (datalo & 1) == 0 && datahi == datalo + 1) {
96
+ target_ulong pc, void *host_pc,
71
tcg_out_strd_8(s, COND_AL, datalo, addrlo, 0);
97
+ const TranslatorOps *ops, DisasContextBase *db);
72
} else {
98
99
void translator_loop_temp_check(DisasContextBase *db);
100
101
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
102
index XXXXXXX..XXXXXXX 100644
103
--- a/accel/tcg/translate-all.c
104
+++ b/accel/tcg/translate-all.c
105
@@ -XXX,XX +XXX,XX @@
106
107
#include "exec/cputlb.h"
108
#include "exec/translate-all.h"
109
+#include "exec/translator.h"
110
#include "qemu/bitmap.h"
111
#include "qemu/qemu-print.h"
112
#include "qemu/timer.h"
113
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
114
TCGProfile *prof = &tcg_ctx->prof;
115
int64_t ti;
116
#endif
117
+ void *host_pc;
118
119
assert_memory_lock();
120
qemu_thread_jit_write();
121
122
- phys_pc = get_page_addr_code(env, pc);
123
+ phys_pc = get_page_addr_code_hostp(env, pc, &host_pc);
124
125
if (phys_pc == -1) {
126
/* Generate a one-shot TB with 1 insn in it */
127
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
128
tcg_func_start(tcg_ctx);
129
130
tcg_ctx->cpu = env_cpu(env);
131
- gen_intermediate_code(cpu, tb, max_insns);
132
+ gen_intermediate_code(cpu, tb, max_insns, pc, host_pc);
133
assert(tb->size != 0);
134
tcg_ctx->cpu = NULL;
135
max_insns = tb->icount;
136
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
137
index XXXXXXX..XXXXXXX 100644
138
--- a/accel/tcg/translator.c
139
+++ b/accel/tcg/translator.c
140
@@ -XXX,XX +XXX,XX @@ static inline void translator_page_protect(DisasContextBase *dcbase,
141
#endif
142
}
143
144
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
145
- CPUState *cpu, TranslationBlock *tb, int max_insns)
146
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
147
+ target_ulong pc, void *host_pc,
148
+ const TranslatorOps *ops, DisasContextBase *db)
149
{
150
uint32_t cflags = tb_cflags(tb);
151
bool plugin_enabled;
152
153
/* Initialize DisasContext */
154
db->tb = tb;
155
- db->pc_first = tb->pc;
156
- db->pc_next = db->pc_first;
157
+ db->pc_first = pc;
158
+ db->pc_next = pc;
159
db->is_jmp = DISAS_NEXT;
160
db->num_insns = 0;
161
db->max_insns = max_insns;
162
diff --git a/target/alpha/translate.c b/target/alpha/translate.c
163
index XXXXXXX..XXXXXXX 100644
164
--- a/target/alpha/translate.c
165
+++ b/target/alpha/translate.c
166
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps alpha_tr_ops = {
167
.disas_log = alpha_tr_disas_log,
168
};
169
170
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
171
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
172
+ target_ulong pc, void *host_pc)
173
{
174
DisasContext dc;
175
- translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
176
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &alpha_tr_ops, &dc.base);
177
}
178
179
void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
180
diff --git a/target/arm/translate.c b/target/arm/translate.c
181
index XXXXXXX..XXXXXXX 100644
182
--- a/target/arm/translate.c
183
+++ b/target/arm/translate.c
184
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps thumb_translator_ops = {
185
};
186
187
/* generate intermediate code for basic block 'tb'. */
188
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
189
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
190
+ target_ulong pc, void *host_pc)
191
{
192
DisasContext dc = { };
193
const TranslatorOps *ops = &arm_translator_ops;
194
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
195
}
196
#endif
197
198
- translator_loop(ops, &dc.base, cpu, tb, max_insns);
199
+ translator_loop(cpu, tb, max_insns, pc, host_pc, ops, &dc.base);
200
}
201
202
void restore_state_to_opc(CPUARMState *env, TranslationBlock *tb,
203
diff --git a/target/avr/translate.c b/target/avr/translate.c
204
index XXXXXXX..XXXXXXX 100644
205
--- a/target/avr/translate.c
206
+++ b/target/avr/translate.c
207
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps avr_tr_ops = {
208
.disas_log = avr_tr_disas_log,
209
};
210
211
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
212
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
213
+ target_ulong pc, void *host_pc)
214
{
215
DisasContext dc = { };
216
- translator_loop(&avr_tr_ops, &dc.base, cs, tb, max_insns);
217
+ translator_loop(cs, tb, max_insns, pc, host_pc, &avr_tr_ops, &dc.base);
218
}
219
220
void restore_state_to_opc(CPUAVRState *env, TranslationBlock *tb,
221
diff --git a/target/cris/translate.c b/target/cris/translate.c
222
index XXXXXXX..XXXXXXX 100644
223
--- a/target/cris/translate.c
224
+++ b/target/cris/translate.c
225
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps cris_tr_ops = {
226
.disas_log = cris_tr_disas_log,
227
};
228
229
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
230
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
231
+ target_ulong pc, void *host_pc)
232
{
233
DisasContext dc;
234
- translator_loop(&cris_tr_ops, &dc.base, cs, tb, max_insns);
235
+ translator_loop(cs, tb, max_insns, pc, host_pc, &cris_tr_ops, &dc.base);
236
}
237
238
void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
239
diff --git a/target/hexagon/translate.c b/target/hexagon/translate.c
240
index XXXXXXX..XXXXXXX 100644
241
--- a/target/hexagon/translate.c
242
+++ b/target/hexagon/translate.c
243
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hexagon_tr_ops = {
244
.disas_log = hexagon_tr_disas_log,
245
};
246
247
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
248
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
249
+ target_ulong pc, void *host_pc)
250
{
251
DisasContext ctx;
252
253
- translator_loop(&hexagon_tr_ops, &ctx.base, cs, tb, max_insns);
254
+ translator_loop(cs, tb, max_insns, pc, host_pc,
255
+ &hexagon_tr_ops, &ctx.base);
256
}
257
258
#define NAME_LEN 64
259
diff --git a/target/hppa/translate.c b/target/hppa/translate.c
260
index XXXXXXX..XXXXXXX 100644
261
--- a/target/hppa/translate.c
262
+++ b/target/hppa/translate.c
263
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hppa_tr_ops = {
264
.disas_log = hppa_tr_disas_log,
265
};
266
267
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
268
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
269
+ target_ulong pc, void *host_pc)
270
{
271
DisasContext ctx;
272
- translator_loop(&hppa_tr_ops, &ctx.base, cs, tb, max_insns);
273
+ translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base);
274
}
275
276
void restore_state_to_opc(CPUHPPAState *env, TranslationBlock *tb,
277
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
278
index XXXXXXX..XXXXXXX 100644
279
--- a/target/i386/tcg/translate.c
280
+++ b/target/i386/tcg/translate.c
281
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps i386_tr_ops = {
282
};
283
284
/* generate intermediate code for basic block 'tb'. */
285
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
286
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
287
+ target_ulong pc, void *host_pc)
288
{
289
DisasContext dc;
290
291
- translator_loop(&i386_tr_ops, &dc.base, cpu, tb, max_insns);
292
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &i386_tr_ops, &dc.base);
293
}
294
295
void restore_state_to_opc(CPUX86State *env, TranslationBlock *tb,
296
diff --git a/target/loongarch/translate.c b/target/loongarch/translate.c
297
index XXXXXXX..XXXXXXX 100644
298
--- a/target/loongarch/translate.c
299
+++ b/target/loongarch/translate.c
300
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps loongarch_tr_ops = {
301
.disas_log = loongarch_tr_disas_log,
302
};
303
304
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
305
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
306
+ target_ulong pc, void *host_pc)
307
{
308
DisasContext ctx;
309
310
- translator_loop(&loongarch_tr_ops, &ctx.base, cs, tb, max_insns);
311
+ translator_loop(cs, tb, max_insns, pc, host_pc,
312
+ &loongarch_tr_ops, &ctx.base);
313
}
314
315
void loongarch_translate_init(void)
316
diff --git a/target/m68k/translate.c b/target/m68k/translate.c
317
index XXXXXXX..XXXXXXX 100644
318
--- a/target/m68k/translate.c
319
+++ b/target/m68k/translate.c
320
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps m68k_tr_ops = {
321
.disas_log = m68k_tr_disas_log,
322
};
323
324
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
325
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
326
+ target_ulong pc, void *host_pc)
327
{
328
DisasContext dc;
329
- translator_loop(&m68k_tr_ops, &dc.base, cpu, tb, max_insns);
330
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &m68k_tr_ops, &dc.base);
331
}
332
333
static double floatx80_to_double(CPUM68KState *env, uint16_t high, uint64_t low)
334
diff --git a/target/microblaze/translate.c b/target/microblaze/translate.c
335
index XXXXXXX..XXXXXXX 100644
336
--- a/target/microblaze/translate.c
337
+++ b/target/microblaze/translate.c
338
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mb_tr_ops = {
339
.disas_log = mb_tr_disas_log,
340
};
341
342
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
343
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
344
+ target_ulong pc, void *host_pc)
345
{
346
DisasContext dc;
347
- translator_loop(&mb_tr_ops, &dc.base, cpu, tb, max_insns);
348
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &mb_tr_ops, &dc.base);
349
}
350
351
void mb_cpu_dump_state(CPUState *cs, FILE *f, int flags)
352
diff --git a/target/mips/tcg/translate.c b/target/mips/tcg/translate.c
353
index XXXXXXX..XXXXXXX 100644
354
--- a/target/mips/tcg/translate.c
355
+++ b/target/mips/tcg/translate.c
356
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mips_tr_ops = {
357
.disas_log = mips_tr_disas_log,
358
};
359
360
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
361
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
362
+ target_ulong pc, void *host_pc)
363
{
364
DisasContext ctx;
365
366
- translator_loop(&mips_tr_ops, &ctx.base, cs, tb, max_insns);
367
+ translator_loop(cs, tb, max_insns, pc, host_pc, &mips_tr_ops, &ctx.base);
368
}
369
370
void mips_tcg_init(void)
371
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
372
index XXXXXXX..XXXXXXX 100644
373
--- a/target/nios2/translate.c
374
+++ b/target/nios2/translate.c
375
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps nios2_tr_ops = {
376
.disas_log = nios2_tr_disas_log,
377
};
378
379
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
380
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
381
+ target_ulong pc, void *host_pc)
382
{
383
DisasContext dc;
384
- translator_loop(&nios2_tr_ops, &dc.base, cs, tb, max_insns);
385
+ translator_loop(cs, tb, max_insns, pc, host_pc, &nios2_tr_ops, &dc.base);
386
}
387
388
void nios2_cpu_dump_state(CPUState *cs, FILE *f, int flags)
389
diff --git a/target/openrisc/translate.c b/target/openrisc/translate.c
390
index XXXXXXX..XXXXXXX 100644
391
--- a/target/openrisc/translate.c
392
+++ b/target/openrisc/translate.c
393
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps openrisc_tr_ops = {
394
.disas_log = openrisc_tr_disas_log,
395
};
396
397
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
398
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
399
+ target_ulong pc, void *host_pc)
400
{
401
DisasContext ctx;
402
403
- translator_loop(&openrisc_tr_ops, &ctx.base, cs, tb, max_insns);
404
+ translator_loop(cs, tb, max_insns, pc, host_pc,
405
+ &openrisc_tr_ops, &ctx.base);
406
}
407
408
void openrisc_cpu_dump_state(CPUState *cs, FILE *f, int flags)
409
diff --git a/target/ppc/translate.c b/target/ppc/translate.c
410
index XXXXXXX..XXXXXXX 100644
411
--- a/target/ppc/translate.c
412
+++ b/target/ppc/translate.c
413
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps ppc_tr_ops = {
414
.disas_log = ppc_tr_disas_log,
415
};
416
417
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
418
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
419
+ target_ulong pc, void *host_pc)
420
{
421
DisasContext ctx;
422
423
- translator_loop(&ppc_tr_ops, &ctx.base, cs, tb, max_insns);
424
+ translator_loop(cs, tb, max_insns, pc, host_pc, &ppc_tr_ops, &ctx.base);
425
}
426
427
void restore_state_to_opc(CPUPPCState *env, TranslationBlock *tb,
428
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
429
index XXXXXXX..XXXXXXX 100644
430
--- a/target/riscv/translate.c
431
+++ b/target/riscv/translate.c
432
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps riscv_tr_ops = {
433
.disas_log = riscv_tr_disas_log,
434
};
435
436
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
437
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
438
+ target_ulong pc, void *host_pc)
439
{
440
DisasContext ctx;
441
442
- translator_loop(&riscv_tr_ops, &ctx.base, cs, tb, max_insns);
443
+ translator_loop(cs, tb, max_insns, pc, host_pc, &riscv_tr_ops, &ctx.base);
444
}
445
446
void riscv_translate_init(void)
447
diff --git a/target/rx/translate.c b/target/rx/translate.c
448
index XXXXXXX..XXXXXXX 100644
449
--- a/target/rx/translate.c
450
+++ b/target/rx/translate.c
451
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps rx_tr_ops = {
452
.disas_log = rx_tr_disas_log,
453
};
454
455
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
456
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
457
+ target_ulong pc, void *host_pc)
458
{
459
DisasContext dc;
460
461
- translator_loop(&rx_tr_ops, &dc.base, cs, tb, max_insns);
462
+ translator_loop(cs, tb, max_insns, pc, host_pc, &rx_tr_ops, &dc.base);
463
}
464
465
void restore_state_to_opc(CPURXState *env, TranslationBlock *tb,
466
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
467
index XXXXXXX..XXXXXXX 100644
468
--- a/target/s390x/tcg/translate.c
469
+++ b/target/s390x/tcg/translate.c
470
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps s390x_tr_ops = {
471
.disas_log = s390x_tr_disas_log,
472
};
473
474
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
475
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
476
+ target_ulong pc, void *host_pc)
477
{
478
DisasContext dc;
479
480
- translator_loop(&s390x_tr_ops, &dc.base, cs, tb, max_insns);
481
+ translator_loop(cs, tb, max_insns, pc, host_pc, &s390x_tr_ops, &dc.base);
482
}
483
484
void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb,
485
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
486
index XXXXXXX..XXXXXXX 100644
487
--- a/target/sh4/translate.c
488
+++ b/target/sh4/translate.c
489
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sh4_tr_ops = {
490
.disas_log = sh4_tr_disas_log,
491
};
492
493
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
494
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
495
+ target_ulong pc, void *host_pc)
496
{
497
DisasContext ctx;
498
499
- translator_loop(&sh4_tr_ops, &ctx.base, cs, tb, max_insns);
500
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base);
501
}
502
503
void restore_state_to_opc(CPUSH4State *env, TranslationBlock *tb,
504
diff --git a/target/sparc/translate.c b/target/sparc/translate.c
505
index XXXXXXX..XXXXXXX 100644
506
--- a/target/sparc/translate.c
507
+++ b/target/sparc/translate.c
508
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sparc_tr_ops = {
509
.disas_log = sparc_tr_disas_log,
510
};
511
512
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
513
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
514
+ target_ulong pc, void *host_pc)
515
{
516
DisasContext dc = {};
517
518
- translator_loop(&sparc_tr_ops, &dc.base, cs, tb, max_insns);
519
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sparc_tr_ops, &dc.base);
520
}
521
522
void sparc_tcg_init(void)
523
diff --git a/target/tricore/translate.c b/target/tricore/translate.c
524
index XXXXXXX..XXXXXXX 100644
525
--- a/target/tricore/translate.c
526
+++ b/target/tricore/translate.c
527
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps tricore_tr_ops = {
528
};
529
530
531
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
532
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
533
+ target_ulong pc, void *host_pc)
534
{
535
DisasContext ctx;
536
- translator_loop(&tricore_tr_ops, &ctx.base, cs, tb, max_insns);
537
+ translator_loop(cs, tb, max_insns, pc, host_pc,
538
+ &tricore_tr_ops, &ctx.base);
539
}
540
541
void
542
diff --git a/target/xtensa/translate.c b/target/xtensa/translate.c
543
index XXXXXXX..XXXXXXX 100644
544
--- a/target/xtensa/translate.c
545
+++ b/target/xtensa/translate.c
546
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps xtensa_translator_ops = {
547
.disas_log = xtensa_tr_disas_log,
548
};
549
550
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
551
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
552
+ target_ulong pc, void *host_pc)
553
{
554
DisasContext dc = {};
555
- translator_loop(&xtensa_translator_ops, &dc.base, cpu, tb, max_insns);
556
+ translator_loop(cpu, tb, max_insns, pc, host_pc,
557
+ &xtensa_translator_ops, &dc.base);
558
}
559
560
void xtensa_cpu_dump_state(CPUState *cs, FILE *f, int flags)
73
--
561
--
74
2.25.1
562
2.34.1
75
76
diff view generated by jsdifflib
1
Do not directly access the uc_sigmask member.
1
Cache the translation from guest to host address, so we may
2
This is preparation for a sparc64 fix.
2
use direct loads when we hit on the primary translation page.
3
3
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Look up the second translation page only once, during translation.
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
This obviates another lookup of the second page within tb_gen_code
6
after translation.
7
8
Fixes a bug in that plugin_insn_append should be passed the bytes
9
in the original memory order, not bswapped by pieces.
10
11
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
12
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
14
---
8
linux-user/include/host/aarch64/host-signal.h | 5 +++++
15
include/exec/translator.h | 63 +++++++++++--------
9
linux-user/include/host/alpha/host-signal.h | 5 +++++
16
accel/tcg/translate-all.c | 23 +++----
10
linux-user/include/host/arm/host-signal.h | 5 +++++
17
accel/tcg/translator.c | 126 +++++++++++++++++++++++++++++---------
11
linux-user/include/host/i386/host-signal.h | 5 +++++
18
3 files changed, 141 insertions(+), 71 deletions(-)
12
.../include/host/loongarch64/host-signal.h | 5 +++++
13
linux-user/include/host/mips/host-signal.h | 5 +++++
14
linux-user/include/host/ppc/host-signal.h | 5 +++++
15
linux-user/include/host/riscv/host-signal.h | 5 +++++
16
linux-user/include/host/s390/host-signal.h | 5 +++++
17
linux-user/include/host/sparc/host-signal.h | 5 +++++
18
linux-user/include/host/x86_64/host-signal.h | 5 +++++
19
linux-user/signal.c | 18 ++++++++----------
20
12 files changed, 63 insertions(+), 10 deletions(-)
21
19
22
diff --git a/linux-user/include/host/aarch64/host-signal.h b/linux-user/include/host/aarch64/host-signal.h
20
diff --git a/include/exec/translator.h b/include/exec/translator.h
23
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
24
--- a/linux-user/include/host/aarch64/host-signal.h
22
--- a/include/exec/translator.h
25
+++ b/linux-user/include/host/aarch64/host-signal.h
23
+++ b/include/exec/translator.h
26
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
24
@@ -XXX,XX +XXX,XX @@ typedef enum DisasJumpType {
27
uc->uc_mcontext.pc = pc;
25
* Architecture-agnostic disassembly context.
26
*/
27
typedef struct DisasContextBase {
28
- const TranslationBlock *tb;
29
+ TranslationBlock *tb;
30
target_ulong pc_first;
31
target_ulong pc_next;
32
DisasJumpType is_jmp;
33
int num_insns;
34
int max_insns;
35
bool singlestep_enabled;
36
-#ifdef CONFIG_USER_ONLY
37
- /*
38
- * Guest address of the last byte of the last protected page.
39
- *
40
- * Pages containing the translated instructions are made non-writable in
41
- * order to achieve consistency in case another thread is modifying the
42
- * code while translate_insn() fetches the instruction bytes piecemeal.
43
- * Such writer threads are blocked on mmap_lock() in page_unprotect().
44
- */
45
- target_ulong page_protect_end;
46
-#endif
47
+ void *host_addr[2];
48
} DisasContextBase;
49
50
/**
51
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
52
* the relevant information at translation time.
53
*/
54
55
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
56
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
57
- abi_ptr pc, bool do_swap); \
58
- static inline type fullname(CPUArchState *env, \
59
- DisasContextBase *dcbase, abi_ptr pc) \
60
- { \
61
- return fullname ## _swap(env, dcbase, pc, false); \
62
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
63
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
64
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
65
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
66
+
67
+static inline uint16_t
68
+translator_lduw_swap(CPUArchState *env, DisasContextBase *db,
69
+ abi_ptr pc, bool do_swap)
70
+{
71
+ uint16_t ret = translator_lduw(env, db, pc);
72
+ if (do_swap) {
73
+ ret = bswap16(ret);
74
}
75
+ return ret;
76
+}
77
78
-#define FOR_EACH_TRANSLATOR_LD(F) \
79
- F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
80
- F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
81
- F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
82
- F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
83
+static inline uint32_t
84
+translator_ldl_swap(CPUArchState *env, DisasContextBase *db,
85
+ abi_ptr pc, bool do_swap)
86
+{
87
+ uint32_t ret = translator_ldl(env, db, pc);
88
+ if (do_swap) {
89
+ ret = bswap32(ret);
90
+ }
91
+ return ret;
92
+}
93
94
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
95
-
96
-#undef GEN_TRANSLATOR_LD
97
+static inline uint64_t
98
+translator_ldq_swap(CPUArchState *env, DisasContextBase *db,
99
+ abi_ptr pc, bool do_swap)
100
+{
101
+ uint64_t ret = translator_ldq_swap(env, db, pc, false);
102
+ if (do_swap) {
103
+ ret = bswap64(ret);
104
+ }
105
+ return ret;
106
+}
107
108
/*
109
* Return whether addr is on the same page as where disassembly started.
110
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
111
index XXXXXXX..XXXXXXX 100644
112
--- a/accel/tcg/translate-all.c
113
+++ b/accel/tcg/translate-all.c
114
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
115
{
116
CPUArchState *env = cpu->env_ptr;
117
TranslationBlock *tb, *existing_tb;
118
- tb_page_addr_t phys_pc, phys_page2;
119
- target_ulong virt_page2;
120
+ tb_page_addr_t phys_pc;
121
tcg_insn_unit *gen_code_buf;
122
int gen_code_size, search_size, max_insns;
123
#ifdef CONFIG_PROFILER
124
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
125
tb->flags = flags;
126
tb->cflags = cflags;
127
tb->trace_vcpu_dstate = *cpu->trace_dstate;
128
+ tb->page_addr[0] = phys_pc;
129
+ tb->page_addr[1] = -1;
130
tcg_ctx->tb_cflags = cflags;
131
tb_overflow:
132
133
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
134
}
135
136
/*
137
- * If the TB is not associated with a physical RAM page then
138
- * it must be a temporary one-insn TB, and we have nothing to do
139
- * except fill in the page_addr[] fields. Return early before
140
- * attempting to link to other TBs or add to the lookup table.
141
+ * If the TB is not associated with a physical RAM page then it must be
142
+ * a temporary one-insn TB, and we have nothing left to do. Return early
143
+ * before attempting to link to other TBs or add to the lookup table.
144
*/
145
- if (phys_pc == -1) {
146
- tb->page_addr[0] = tb->page_addr[1] = -1;
147
+ if (tb->page_addr[0] == -1) {
148
return tb;
149
}
150
151
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
152
*/
153
tcg_tb_insert(tb);
154
155
- /* check next page if needed */
156
- virt_page2 = (pc + tb->size - 1) & TARGET_PAGE_MASK;
157
- phys_page2 = -1;
158
- if ((pc & TARGET_PAGE_MASK) != virt_page2) {
159
- phys_page2 = get_page_addr_code(env, virt_page2);
160
- }
161
/*
162
* No explicit memory barrier is required -- tb_link_page() makes the
163
* TB visible in a consistent state.
164
*/
165
- existing_tb = tb_link_page(tb, phys_pc, phys_page2);
166
+ existing_tb = tb_link_page(tb, tb->page_addr[0], tb->page_addr[1]);
167
/* if the TB already exists, discard what we just translated */
168
if (unlikely(existing_tb != tb)) {
169
uintptr_t orig_aligned = (uintptr_t)gen_code_buf;
170
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
171
index XXXXXXX..XXXXXXX 100644
172
--- a/accel/tcg/translator.c
173
+++ b/accel/tcg/translator.c
174
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest)
175
return ((db->pc_first ^ dest) & TARGET_PAGE_MASK) == 0;
28
}
176
}
29
177
30
+static inline void *host_signal_mask(ucontext_t *uc)
178
-static inline void translator_page_protect(DisasContextBase *dcbase,
31
+{
179
- target_ulong pc)
32
+ return &uc->uc_sigmask;
180
-{
33
+}
181
-#ifdef CONFIG_USER_ONLY
34
+
182
- dcbase->page_protect_end = pc | ~TARGET_PAGE_MASK;
35
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
183
- page_protect(pc);
36
{
184
-#endif
37
struct _aarch64_ctx *hdr;
185
-}
38
diff --git a/linux-user/include/host/alpha/host-signal.h b/linux-user/include/host/alpha/host-signal.h
186
-
39
index XXXXXXX..XXXXXXX 100644
187
void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
40
--- a/linux-user/include/host/alpha/host-signal.h
188
target_ulong pc, void *host_pc,
41
+++ b/linux-user/include/host/alpha/host-signal.h
189
const TranslatorOps *ops, DisasContextBase *db)
42
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
190
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
43
uc->uc_mcontext.sc_pc = pc;
191
db->num_insns = 0;
44
}
192
db->max_insns = max_insns;
45
193
db->singlestep_enabled = cflags & CF_SINGLE_STEP;
46
+static inline void *host_signal_mask(ucontext_t *uc)
194
- translator_page_protect(db, db->pc_next);
47
+{
195
+ db->host_addr[0] = host_pc;
48
+ return &uc->uc_sigmask;
196
+ db->host_addr[1] = NULL;
49
+}
197
+
50
+
198
+#ifdef CONFIG_USER_ONLY
51
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
199
+ page_protect(pc);
52
{
200
+#endif
53
uint32_t *pc = (uint32_t *)host_signal_pc(uc);
201
54
diff --git a/linux-user/include/host/arm/host-signal.h b/linux-user/include/host/arm/host-signal.h
202
ops->init_disas_context(db, cpu);
55
index XXXXXXX..XXXXXXX 100644
203
tcg_debug_assert(db->is_jmp == DISAS_NEXT); /* no early exit */
56
--- a/linux-user/include/host/arm/host-signal.h
204
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
57
+++ b/linux-user/include/host/arm/host-signal.h
58
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
59
uc->uc_mcontext.arm_pc = pc;
60
}
61
62
+static inline void *host_signal_mask(ucontext_t *uc)
63
+{
64
+ return &uc->uc_sigmask;
65
+}
66
+
67
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
68
{
69
/*
70
diff --git a/linux-user/include/host/i386/host-signal.h b/linux-user/include/host/i386/host-signal.h
71
index XXXXXXX..XXXXXXX 100644
72
--- a/linux-user/include/host/i386/host-signal.h
73
+++ b/linux-user/include/host/i386/host-signal.h
74
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
75
uc->uc_mcontext.gregs[REG_EIP] = pc;
76
}
77
78
+static inline void *host_signal_mask(ucontext_t *uc)
79
+{
80
+ return &uc->uc_sigmask;
81
+}
82
+
83
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
84
{
85
return uc->uc_mcontext.gregs[REG_TRAPNO] == 0xe
86
diff --git a/linux-user/include/host/loongarch64/host-signal.h b/linux-user/include/host/loongarch64/host-signal.h
87
index XXXXXXX..XXXXXXX 100644
88
--- a/linux-user/include/host/loongarch64/host-signal.h
89
+++ b/linux-user/include/host/loongarch64/host-signal.h
90
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
91
uc->uc_mcontext.__pc = pc;
92
}
93
94
+static inline void *host_signal_mask(ucontext_t *uc)
95
+{
96
+ return &uc->uc_sigmask;
97
+}
98
+
99
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
100
{
101
const uint32_t *pinsn = (const uint32_t *)host_signal_pc(uc);
102
diff --git a/linux-user/include/host/mips/host-signal.h b/linux-user/include/host/mips/host-signal.h
103
index XXXXXXX..XXXXXXX 100644
104
--- a/linux-user/include/host/mips/host-signal.h
105
+++ b/linux-user/include/host/mips/host-signal.h
106
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
107
uc->uc_mcontext.pc = pc;
108
}
109
110
+static inline void *host_signal_mask(ucontext_t *uc)
111
+{
112
+ return &uc->uc_sigmask;
113
+}
114
+
115
#if defined(__misp16) || defined(__mips_micromips)
116
#error "Unsupported encoding"
117
#endif
118
diff --git a/linux-user/include/host/ppc/host-signal.h b/linux-user/include/host/ppc/host-signal.h
119
index XXXXXXX..XXXXXXX 100644
120
--- a/linux-user/include/host/ppc/host-signal.h
121
+++ b/linux-user/include/host/ppc/host-signal.h
122
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
123
uc->uc_mcontext.regs->nip = pc;
124
}
125
126
+static inline void *host_signal_mask(ucontext_t *uc)
127
+{
128
+ return &uc->uc_sigmask;
129
+}
130
+
131
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
132
{
133
return uc->uc_mcontext.regs->trap != 0x400
134
diff --git a/linux-user/include/host/riscv/host-signal.h b/linux-user/include/host/riscv/host-signal.h
135
index XXXXXXX..XXXXXXX 100644
136
--- a/linux-user/include/host/riscv/host-signal.h
137
+++ b/linux-user/include/host/riscv/host-signal.h
138
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
139
uc->uc_mcontext.__gregs[REG_PC] = pc;
140
}
141
142
+static inline void *host_signal_mask(ucontext_t *uc)
143
+{
144
+ return &uc->uc_sigmask;
145
+}
146
+
147
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
148
{
149
/*
150
diff --git a/linux-user/include/host/s390/host-signal.h b/linux-user/include/host/s390/host-signal.h
151
index XXXXXXX..XXXXXXX 100644
152
--- a/linux-user/include/host/s390/host-signal.h
153
+++ b/linux-user/include/host/s390/host-signal.h
154
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
155
uc->uc_mcontext.psw.addr = pc;
156
}
157
158
+static inline void *host_signal_mask(ucontext_t *uc)
159
+{
160
+ return &uc->uc_sigmask;
161
+}
162
+
163
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
164
{
165
uint16_t *pinsn = (uint16_t *)host_signal_pc(uc);
166
diff --git a/linux-user/include/host/sparc/host-signal.h b/linux-user/include/host/sparc/host-signal.h
167
index XXXXXXX..XXXXXXX 100644
168
--- a/linux-user/include/host/sparc/host-signal.h
169
+++ b/linux-user/include/host/sparc/host-signal.h
170
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
171
#endif
205
#endif
172
}
206
}
173
207
174
+static inline void *host_signal_mask(ucontext_t *uc)
208
-static inline void translator_maybe_page_protect(DisasContextBase *dcbase,
175
+{
209
- target_ulong pc, size_t len)
176
+ return &uc->uc_sigmask;
210
+static void *translator_access(CPUArchState *env, DisasContextBase *db,
177
+}
211
+ target_ulong pc, size_t len)
178
+
179
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
180
{
212
{
181
uint32_t insn = *(uint32_t *)host_signal_pc(uc);
213
-#ifdef CONFIG_USER_ONLY
182
diff --git a/linux-user/include/host/x86_64/host-signal.h b/linux-user/include/host/x86_64/host-signal.h
214
- target_ulong end = pc + len - 1;
183
index XXXXXXX..XXXXXXX 100644
215
+ void *host;
184
--- a/linux-user/include/host/x86_64/host-signal.h
216
+ target_ulong base, end;
185
+++ b/linux-user/include/host/x86_64/host-signal.h
217
+ TranslationBlock *tb;
186
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
218
187
uc->uc_mcontext.gregs[REG_RIP] = pc;
219
- if (end > dcbase->page_protect_end) {
220
- translator_page_protect(dcbase, end);
221
+ tb = db->tb;
222
+
223
+ /* Use slow path if first page is MMIO. */
224
+ if (unlikely(tb->page_addr[0] == -1)) {
225
+ return NULL;
226
}
227
+
228
+ end = pc + len - 1;
229
+ if (likely(is_same_page(db, end))) {
230
+ host = db->host_addr[0];
231
+ base = db->pc_first;
232
+ } else {
233
+ host = db->host_addr[1];
234
+ base = TARGET_PAGE_ALIGN(db->pc_first);
235
+ if (host == NULL) {
236
+ tb->page_addr[1] =
237
+ get_page_addr_code_hostp(env, base, &db->host_addr[1]);
238
+#ifdef CONFIG_USER_ONLY
239
+ page_protect(end);
240
#endif
241
+ /* We cannot handle MMIO as second page. */
242
+ assert(tb->page_addr[1] != -1);
243
+ host = db->host_addr[1];
244
+ }
245
+
246
+ /* Use slow path when crossing pages. */
247
+ if (is_same_page(db, pc)) {
248
+ return NULL;
249
+ }
250
+ }
251
+
252
+ tcg_debug_assert(pc >= base);
253
+ return host + (pc - base);
188
}
254
}
189
255
190
+static inline void *host_signal_mask(ucontext_t *uc)
256
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
191
+{
257
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
192
+ return &uc->uc_sigmask;
258
- abi_ptr pc, bool do_swap) \
193
+}
259
- { \
194
+
260
- translator_maybe_page_protect(dcbase, pc, sizeof(type)); \
195
static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
261
- type ret = load_fn(env, pc); \
196
{
262
- if (do_swap) { \
197
return uc->uc_mcontext.gregs[REG_TRAPNO] == 0xe
263
- ret = swap_fn(ret); \
198
diff --git a/linux-user/signal.c b/linux-user/signal.c
264
- } \
199
index XXXXXXX..XXXXXXX 100644
265
- plugin_insn_append(pc, &ret, sizeof(ret)); \
200
--- a/linux-user/signal.c
266
- return ret; \
201
+++ b/linux-user/signal.c
267
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
202
@@ -XXX,XX +XXX,XX @@ static void host_signal_handler(int host_sig, siginfo_t *info, void *puc)
268
+{
203
int guest_sig;
269
+ uint8_t ret;
204
uintptr_t pc = 0;
270
+ void *p = translator_access(env, db, pc, sizeof(ret));
205
bool sync_sig = false;
271
+
206
+ void *sigmask = host_signal_mask(uc);
272
+ if (p) {
207
273
+ plugin_insn_append(pc, p, sizeof(ret));
208
/*
274
+ return ldub_p(p);
209
* Non-spoofed SIGSEGV and SIGBUS are synchronous, and need special
275
}
210
@@ -XXX,XX +XXX,XX @@ static void host_signal_handler(int host_sig, siginfo_t *info, void *puc)
276
+ ret = cpu_ldub_code(env, pc);
211
if (info->si_code == SEGV_ACCERR && h2g_valid(host_addr)) {
277
+ plugin_insn_append(pc, &ret, sizeof(ret));
212
/* If this was a write to a TB protected page, restart. */
278
+ return ret;
213
if (is_write &&
279
+}
214
- handle_sigsegv_accerr_write(cpu, &uc->uc_sigmask,
280
215
- pc, guest_addr)) {
281
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
216
+ handle_sigsegv_accerr_write(cpu, sigmask, pc, guest_addr)) {
282
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
217
return;
283
+{
218
}
284
+ uint16_t ret, plug;
219
285
+ void *p = translator_access(env, db, pc, sizeof(ret));
220
@@ -XXX,XX +XXX,XX @@ static void host_signal_handler(int host_sig, siginfo_t *info, void *puc)
286
221
}
287
-#undef GEN_TRANSLATOR_LD
222
}
288
+ if (p) {
223
289
+ plugin_insn_append(pc, p, sizeof(ret));
224
- sigprocmask(SIG_SETMASK, &uc->uc_sigmask, NULL);
290
+ return lduw_p(p);
225
+ sigprocmask(SIG_SETMASK, sigmask, NULL);
291
+ }
226
cpu_loop_exit_sigsegv(cpu, guest_addr, access_type, maperr, pc);
292
+ ret = cpu_lduw_code(env, pc);
227
} else {
293
+ plug = tswap16(ret);
228
- sigprocmask(SIG_SETMASK, &uc->uc_sigmask, NULL);
294
+ plugin_insn_append(pc, &plug, sizeof(ret));
229
+ sigprocmask(SIG_SETMASK, sigmask, NULL);
295
+ return ret;
230
if (info->si_code == BUS_ADRALN) {
296
+}
231
cpu_loop_exit_sigbus(cpu, guest_addr, access_type, pc);
297
+
232
}
298
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
233
@@ -XXX,XX +XXX,XX @@ static void host_signal_handler(int host_sig, siginfo_t *info, void *puc)
299
+{
234
* now and it getting out to the main loop. Signals will be
300
+ uint32_t ret, plug;
235
* unblocked again in process_pending_signals().
301
+ void *p = translator_access(env, db, pc, sizeof(ret));
236
*
302
+
237
- * WARNING: we cannot use sigfillset() here because the uc_sigmask
303
+ if (p) {
238
+ * WARNING: we cannot use sigfillset() here because the sigmask
304
+ plugin_insn_append(pc, p, sizeof(ret));
239
* field is a kernel sigset_t, which is much smaller than the
305
+ return ldl_p(p);
240
* libc sigset_t which sigfillset() operates on. Using sigfillset()
306
+ }
241
* would write 0xff bytes off the end of the structure and trash
307
+ ret = cpu_ldl_code(env, pc);
242
* data on the struct.
308
+ plug = tswap32(ret);
243
- * We can't use sizeof(uc->uc_sigmask) either, because the libc
309
+ plugin_insn_append(pc, &plug, sizeof(ret));
244
- * headers define the struct field with the wrong (too large) type.
310
+ return ret;
245
*/
311
+}
246
- memset(&uc->uc_sigmask, 0xff, SIGSET_T_SIZE);
312
+
247
- sigdelset(&uc->uc_sigmask, SIGSEGV);
313
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
248
- sigdelset(&uc->uc_sigmask, SIGBUS);
314
+{
249
+ memset(sigmask, 0xff, SIGSET_T_SIZE);
315
+ uint64_t ret, plug;
250
+ sigdelset(sigmask, SIGSEGV);
316
+ void *p = translator_access(env, db, pc, sizeof(ret));
251
+ sigdelset(sigmask, SIGBUS);
317
+
252
318
+ if (p) {
253
/* interrupt the virtual CPU as soon as possible */
319
+ plugin_insn_append(pc, p, sizeof(ret));
254
cpu_exit(thread_cpu);
320
+ return ldq_p(p);
321
+ }
322
+ ret = cpu_ldq_code(env, pc);
323
+ plug = tswap64(ret);
324
+ plugin_insn_append(pc, &plug, sizeof(ret));
325
+ return ret;
326
+}
255
--
327
--
256
2.25.1
328
2.34.1
257
258
diff view generated by jsdifflib
Deleted patch
1
Do not directly access ucontext_t as the third signal parameter.
2
This is preparation for a sparc64 fix.
3
1
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
linux-user/include/host/aarch64/host-signal.h | 13 ++++++++-----
9
linux-user/include/host/alpha/host-signal.h | 11 +++++++----
10
linux-user/include/host/arm/host-signal.h | 11 +++++++----
11
linux-user/include/host/i386/host-signal.h | 11 +++++++----
12
linux-user/include/host/loongarch64/host-signal.h | 11 +++++++----
13
linux-user/include/host/mips/host-signal.h | 11 +++++++----
14
linux-user/include/host/ppc/host-signal.h | 11 +++++++----
15
linux-user/include/host/riscv/host-signal.h | 11 +++++++----
16
linux-user/include/host/s390/host-signal.h | 11 +++++++----
17
linux-user/include/host/sparc/host-signal.h | 11 +++++++----
18
linux-user/include/host/x86_64/host-signal.h | 11 +++++++----
19
linux-user/signal.c | 4 ++--
20
12 files changed, 80 insertions(+), 47 deletions(-)
21
22
diff --git a/linux-user/include/host/aarch64/host-signal.h b/linux-user/include/host/aarch64/host-signal.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/linux-user/include/host/aarch64/host-signal.h
25
+++ b/linux-user/include/host/aarch64/host-signal.h
26
@@ -XXX,XX +XXX,XX @@
27
#ifndef AARCH64_HOST_SIGNAL_H
28
#define AARCH64_HOST_SIGNAL_H
29
30
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
31
+typedef ucontext_t host_sigcontext;
32
+
33
/* Pre-3.16 kernel headers don't have these, so provide fallback definitions */
34
#ifndef ESR_MAGIC
35
#define ESR_MAGIC 0x45535201
36
@@ -XXX,XX +XXX,XX @@ struct esr_context {
37
};
38
#endif
39
40
-static inline struct _aarch64_ctx *first_ctx(ucontext_t *uc)
41
+static inline struct _aarch64_ctx *first_ctx(host_sigcontext *uc)
42
{
43
return (struct _aarch64_ctx *)&uc->uc_mcontext.__reserved;
44
}
45
@@ -XXX,XX +XXX,XX @@ static inline struct _aarch64_ctx *next_ctx(struct _aarch64_ctx *hdr)
46
return (struct _aarch64_ctx *)((char *)hdr + hdr->size);
47
}
48
49
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
50
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
51
{
52
return uc->uc_mcontext.pc;
53
}
54
55
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
56
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
57
{
58
uc->uc_mcontext.pc = pc;
59
}
60
61
-static inline void *host_signal_mask(ucontext_t *uc)
62
+static inline void *host_signal_mask(host_sigcontext *uc)
63
{
64
return &uc->uc_sigmask;
65
}
66
67
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
68
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
69
{
70
struct _aarch64_ctx *hdr;
71
uint32_t insn;
72
diff --git a/linux-user/include/host/alpha/host-signal.h b/linux-user/include/host/alpha/host-signal.h
73
index XXXXXXX..XXXXXXX 100644
74
--- a/linux-user/include/host/alpha/host-signal.h
75
+++ b/linux-user/include/host/alpha/host-signal.h
76
@@ -XXX,XX +XXX,XX @@
77
#ifndef ALPHA_HOST_SIGNAL_H
78
#define ALPHA_HOST_SIGNAL_H
79
80
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
81
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
82
+typedef ucontext_t host_sigcontext;
83
+
84
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
85
{
86
return uc->uc_mcontext.sc_pc;
87
}
88
89
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
90
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
91
{
92
uc->uc_mcontext.sc_pc = pc;
93
}
94
95
-static inline void *host_signal_mask(ucontext_t *uc)
96
+static inline void *host_signal_mask(host_sigcontext *uc)
97
{
98
return &uc->uc_sigmask;
99
}
100
101
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
102
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
103
{
104
uint32_t *pc = (uint32_t *)host_signal_pc(uc);
105
uint32_t insn = *pc;
106
diff --git a/linux-user/include/host/arm/host-signal.h b/linux-user/include/host/arm/host-signal.h
107
index XXXXXXX..XXXXXXX 100644
108
--- a/linux-user/include/host/arm/host-signal.h
109
+++ b/linux-user/include/host/arm/host-signal.h
110
@@ -XXX,XX +XXX,XX @@
111
#ifndef ARM_HOST_SIGNAL_H
112
#define ARM_HOST_SIGNAL_H
113
114
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
115
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
116
+typedef ucontext_t host_sigcontext;
117
+
118
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
119
{
120
return uc->uc_mcontext.arm_pc;
121
}
122
123
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
124
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
125
{
126
uc->uc_mcontext.arm_pc = pc;
127
}
128
129
-static inline void *host_signal_mask(ucontext_t *uc)
130
+static inline void *host_signal_mask(host_sigcontext *uc)
131
{
132
return &uc->uc_sigmask;
133
}
134
135
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
136
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
137
{
138
/*
139
* In the FSR, bit 11 is WnR, assuming a v6 or
140
diff --git a/linux-user/include/host/i386/host-signal.h b/linux-user/include/host/i386/host-signal.h
141
index XXXXXXX..XXXXXXX 100644
142
--- a/linux-user/include/host/i386/host-signal.h
143
+++ b/linux-user/include/host/i386/host-signal.h
144
@@ -XXX,XX +XXX,XX @@
145
#ifndef I386_HOST_SIGNAL_H
146
#define I386_HOST_SIGNAL_H
147
148
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
149
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
150
+typedef ucontext_t host_sigcontext;
151
+
152
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
153
{
154
return uc->uc_mcontext.gregs[REG_EIP];
155
}
156
157
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
158
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
159
{
160
uc->uc_mcontext.gregs[REG_EIP] = pc;
161
}
162
163
-static inline void *host_signal_mask(ucontext_t *uc)
164
+static inline void *host_signal_mask(host_sigcontext *uc)
165
{
166
return &uc->uc_sigmask;
167
}
168
169
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
170
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
171
{
172
return uc->uc_mcontext.gregs[REG_TRAPNO] == 0xe
173
&& (uc->uc_mcontext.gregs[REG_ERR] & 0x2);
174
diff --git a/linux-user/include/host/loongarch64/host-signal.h b/linux-user/include/host/loongarch64/host-signal.h
175
index XXXXXXX..XXXXXXX 100644
176
--- a/linux-user/include/host/loongarch64/host-signal.h
177
+++ b/linux-user/include/host/loongarch64/host-signal.h
178
@@ -XXX,XX +XXX,XX @@
179
#ifndef LOONGARCH64_HOST_SIGNAL_H
180
#define LOONGARCH64_HOST_SIGNAL_H
181
182
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
183
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
184
+typedef ucontext_t host_sigcontext;
185
+
186
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
187
{
188
return uc->uc_mcontext.__pc;
189
}
190
191
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
192
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
193
{
194
uc->uc_mcontext.__pc = pc;
195
}
196
197
-static inline void *host_signal_mask(ucontext_t *uc)
198
+static inline void *host_signal_mask(host_sigcontext *uc)
199
{
200
return &uc->uc_sigmask;
201
}
202
203
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
204
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
205
{
206
const uint32_t *pinsn = (const uint32_t *)host_signal_pc(uc);
207
uint32_t insn = pinsn[0];
208
diff --git a/linux-user/include/host/mips/host-signal.h b/linux-user/include/host/mips/host-signal.h
209
index XXXXXXX..XXXXXXX 100644
210
--- a/linux-user/include/host/mips/host-signal.h
211
+++ b/linux-user/include/host/mips/host-signal.h
212
@@ -XXX,XX +XXX,XX @@
213
#ifndef MIPS_HOST_SIGNAL_H
214
#define MIPS_HOST_SIGNAL_H
215
216
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
217
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
218
+typedef ucontext_t host_sigcontext;
219
+
220
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
221
{
222
return uc->uc_mcontext.pc;
223
}
224
225
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
226
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
227
{
228
uc->uc_mcontext.pc = pc;
229
}
230
231
-static inline void *host_signal_mask(ucontext_t *uc)
232
+static inline void *host_signal_mask(host_sigcontext *uc)
233
{
234
return &uc->uc_sigmask;
235
}
236
@@ -XXX,XX +XXX,XX @@ static inline void *host_signal_mask(ucontext_t *uc)
237
#error "Unsupported encoding"
238
#endif
239
240
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
241
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
242
{
243
uint32_t insn = *(uint32_t *)host_signal_pc(uc);
244
245
diff --git a/linux-user/include/host/ppc/host-signal.h b/linux-user/include/host/ppc/host-signal.h
246
index XXXXXXX..XXXXXXX 100644
247
--- a/linux-user/include/host/ppc/host-signal.h
248
+++ b/linux-user/include/host/ppc/host-signal.h
249
@@ -XXX,XX +XXX,XX @@
250
#ifndef PPC_HOST_SIGNAL_H
251
#define PPC_HOST_SIGNAL_H
252
253
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
254
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
255
+typedef ucontext_t host_sigcontext;
256
+
257
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
258
{
259
return uc->uc_mcontext.regs->nip;
260
}
261
262
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
263
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
264
{
265
uc->uc_mcontext.regs->nip = pc;
266
}
267
268
-static inline void *host_signal_mask(ucontext_t *uc)
269
+static inline void *host_signal_mask(host_sigcontext *uc)
270
{
271
return &uc->uc_sigmask;
272
}
273
274
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
275
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
276
{
277
return uc->uc_mcontext.regs->trap != 0x400
278
&& (uc->uc_mcontext.regs->dsisr & 0x02000000);
279
diff --git a/linux-user/include/host/riscv/host-signal.h b/linux-user/include/host/riscv/host-signal.h
280
index XXXXXXX..XXXXXXX 100644
281
--- a/linux-user/include/host/riscv/host-signal.h
282
+++ b/linux-user/include/host/riscv/host-signal.h
283
@@ -XXX,XX +XXX,XX @@
284
#ifndef RISCV_HOST_SIGNAL_H
285
#define RISCV_HOST_SIGNAL_H
286
287
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
288
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
289
+typedef ucontext_t host_sigcontext;
290
+
291
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
292
{
293
return uc->uc_mcontext.__gregs[REG_PC];
294
}
295
296
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
297
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
298
{
299
uc->uc_mcontext.__gregs[REG_PC] = pc;
300
}
301
302
-static inline void *host_signal_mask(ucontext_t *uc)
303
+static inline void *host_signal_mask(host_sigcontext *uc)
304
{
305
return &uc->uc_sigmask;
306
}
307
308
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
309
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
310
{
311
/*
312
* Detect store by reading the instruction at the program counter.
313
diff --git a/linux-user/include/host/s390/host-signal.h b/linux-user/include/host/s390/host-signal.h
314
index XXXXXXX..XXXXXXX 100644
315
--- a/linux-user/include/host/s390/host-signal.h
316
+++ b/linux-user/include/host/s390/host-signal.h
317
@@ -XXX,XX +XXX,XX @@
318
#ifndef S390_HOST_SIGNAL_H
319
#define S390_HOST_SIGNAL_H
320
321
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
322
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
323
+typedef ucontext_t host_sigcontext;
324
+
325
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
326
{
327
return uc->uc_mcontext.psw.addr;
328
}
329
330
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
331
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
332
{
333
uc->uc_mcontext.psw.addr = pc;
334
}
335
336
-static inline void *host_signal_mask(ucontext_t *uc)
337
+static inline void *host_signal_mask(host_sigcontext *uc)
338
{
339
return &uc->uc_sigmask;
340
}
341
342
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
343
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
344
{
345
uint16_t *pinsn = (uint16_t *)host_signal_pc(uc);
346
347
diff --git a/linux-user/include/host/sparc/host-signal.h b/linux-user/include/host/sparc/host-signal.h
348
index XXXXXXX..XXXXXXX 100644
349
--- a/linux-user/include/host/sparc/host-signal.h
350
+++ b/linux-user/include/host/sparc/host-signal.h
351
@@ -XXX,XX +XXX,XX @@
352
#ifndef SPARC_HOST_SIGNAL_H
353
#define SPARC_HOST_SIGNAL_H
354
355
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
356
+/* FIXME: the third argument to a SA_SIGINFO handler is *not* ucontext_t. */
357
+typedef ucontext_t host_sigcontext;
358
+
359
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
360
{
361
#ifdef __arch64__
362
return uc->uc_mcontext.mc_gregs[MC_PC];
363
@@ -XXX,XX +XXX,XX @@ static inline uintptr_t host_signal_pc(ucontext_t *uc)
364
#endif
365
}
366
367
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
368
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
369
{
370
#ifdef __arch64__
371
uc->uc_mcontext.mc_gregs[MC_PC] = pc;
372
@@ -XXX,XX +XXX,XX @@ static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
373
#endif
374
}
375
376
-static inline void *host_signal_mask(ucontext_t *uc)
377
+static inline void *host_signal_mask(host_sigcontext *uc)
378
{
379
return &uc->uc_sigmask;
380
}
381
382
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
383
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
384
{
385
uint32_t insn = *(uint32_t *)host_signal_pc(uc);
386
387
diff --git a/linux-user/include/host/x86_64/host-signal.h b/linux-user/include/host/x86_64/host-signal.h
388
index XXXXXXX..XXXXXXX 100644
389
--- a/linux-user/include/host/x86_64/host-signal.h
390
+++ b/linux-user/include/host/x86_64/host-signal.h
391
@@ -XXX,XX +XXX,XX @@
392
#ifndef X86_64_HOST_SIGNAL_H
393
#define X86_64_HOST_SIGNAL_H
394
395
-static inline uintptr_t host_signal_pc(ucontext_t *uc)
396
+/* The third argument to a SA_SIGINFO handler is ucontext_t. */
397
+typedef ucontext_t host_sigcontext;
398
+
399
+static inline uintptr_t host_signal_pc(host_sigcontext *uc)
400
{
401
return uc->uc_mcontext.gregs[REG_RIP];
402
}
403
404
-static inline void host_signal_set_pc(ucontext_t *uc, uintptr_t pc)
405
+static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
406
{
407
uc->uc_mcontext.gregs[REG_RIP] = pc;
408
}
409
410
-static inline void *host_signal_mask(ucontext_t *uc)
411
+static inline void *host_signal_mask(host_sigcontext *uc)
412
{
413
return &uc->uc_sigmask;
414
}
415
416
-static inline bool host_signal_write(siginfo_t *info, ucontext_t *uc)
417
+static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
418
{
419
return uc->uc_mcontext.gregs[REG_TRAPNO] == 0xe
420
&& (uc->uc_mcontext.gregs[REG_ERR] & 0x2);
421
diff --git a/linux-user/signal.c b/linux-user/signal.c
422
index XXXXXXX..XXXXXXX 100644
423
--- a/linux-user/signal.c
424
+++ b/linux-user/signal.c
425
@@ -XXX,XX +XXX,XX @@ void queue_signal(CPUArchState *env, int sig, int si_type,
426
/* Adjust the signal context to rewind out of safe-syscall if we're in it */
427
static inline void rewind_if_in_safe_syscall(void *puc)
428
{
429
- ucontext_t *uc = (ucontext_t *)puc;
430
+ host_sigcontext *uc = (host_sigcontext *)puc;
431
uintptr_t pcreg = host_signal_pc(uc);
432
433
if (pcreg > (uintptr_t)safe_syscall_start
434
@@ -XXX,XX +XXX,XX @@ static void host_signal_handler(int host_sig, siginfo_t *info, void *puc)
435
CPUState *cpu = env_cpu(env);
436
TaskState *ts = cpu->opaque;
437
target_siginfo_t tinfo;
438
- ucontext_t *uc = puc;
439
+ host_sigcontext *uc = puc;
440
struct emulated_sigtable *k;
441
int guest_sig;
442
uintptr_t pc = 0;
443
--
444
2.25.1
445
446
diff view generated by jsdifflib
Deleted patch
1
Sparc64 is unique on linux in *not* passing ucontext_t as
2
the third argument to a SA_SIGINFO handler. It passes the
3
old struct sigcontext instead.
4
1
5
Set both pc and npc in host_signal_set_pc.
6
7
Fixes: 8b5bd461935b ("linux-user/host/sparc: Populate host_signal.h")
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
11
linux-user/include/host/sparc64/host-signal.h | 17 +++++++++--------
12
1 file changed, 9 insertions(+), 8 deletions(-)
13
14
diff --git a/linux-user/include/host/sparc64/host-signal.h b/linux-user/include/host/sparc64/host-signal.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/include/host/sparc64/host-signal.h
17
+++ b/linux-user/include/host/sparc64/host-signal.h
18
@@ -XXX,XX +XXX,XX @@
19
#ifndef SPARC64_HOST_SIGNAL_H
20
#define SPARC64_HOST_SIGNAL_H
21
22
-/* FIXME: the third argument to a SA_SIGINFO handler is *not* ucontext_t. */
23
-typedef ucontext_t host_sigcontext;
24
+/* The third argument to a SA_SIGINFO handler is struct sigcontext. */
25
+typedef struct sigcontext host_sigcontext;
26
27
-static inline uintptr_t host_signal_pc(host_sigcontext *uc)
28
+static inline uintptr_t host_signal_pc(host_sigcontext *sc)
29
{
30
- return uc->uc_mcontext.mc_gregs[MC_PC];
31
+ return sc->sigc_regs.tpc;
32
}
33
34
-static inline void host_signal_set_pc(host_sigcontext *uc, uintptr_t pc)
35
+static inline void host_signal_set_pc(host_sigcontext *sc, uintptr_t pc)
36
{
37
- uc->uc_mcontext.mc_gregs[MC_PC] = pc;
38
+ sc->sigc_regs.tpc = pc;
39
+ sc->sigc_regs.tnpc = pc + 4;
40
}
41
42
-static inline void *host_signal_mask(host_sigcontext *uc)
43
+static inline void *host_signal_mask(host_sigcontext *sc)
44
{
45
- return &uc->uc_sigmask;
46
+ return &sc->sigc_mask;
47
}
48
49
static inline bool host_signal_write(siginfo_t *info, host_sigcontext *uc)
50
--
51
2.25.1
52
53
diff view generated by jsdifflib
Deleted patch
1
From: Idan Horowitz <idan.horowitz@gmail.com>
2
1
3
Instead of taking the lock of the cpu work list in order to check if it's
4
empty, we can just read the head pointer atomically. This decreases
5
cpu_work_list_empty's share from 5% to 1.3% in a profile of icount-enabled
6
aarch64-softmmu.
7
8
Signed-off-by: Idan Horowitz <idan.horowitz@gmail.com>
9
Message-Id: <20220114004358.299534-1-idan.horowitz@gmail.com>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
---
13
softmmu/cpus.c | 7 +------
14
1 file changed, 1 insertion(+), 6 deletions(-)
15
16
diff --git a/softmmu/cpus.c b/softmmu/cpus.c
17
index XXXXXXX..XXXXXXX 100644
18
--- a/softmmu/cpus.c
19
+++ b/softmmu/cpus.c
20
@@ -XXX,XX +XXX,XX @@ bool cpu_is_stopped(CPUState *cpu)
21
22
bool cpu_work_list_empty(CPUState *cpu)
23
{
24
- bool ret;
25
-
26
- qemu_mutex_lock(&cpu->work_mutex);
27
- ret = QSIMPLEQ_EMPTY(&cpu->work_list);
28
- qemu_mutex_unlock(&cpu->work_mutex);
29
- return ret;
30
+ return QSIMPLEQ_EMPTY_ATOMIC(&cpu->work_list);
31
}
32
33
bool cpu_thread_is_idle(CPUState *cpu)
34
--
35
2.25.1
36
37
diff view generated by jsdifflib
1
A mostly generic test for unaligned access raising SIGBUS.
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Right now translator stops right *after* the end of a page, which
4
breaks reporting of fault locations when the last instruction of a
5
multi-insn translation block crosses a page boundary.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-Id: <20220817150506.592862-3-iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
11
---
6
tests/tcg/multiarch/sigbus.c | 68 ++++++++++++++++++++++++++++++++++++
12
target/s390x/tcg/translate.c | 15 +++-
7
1 file changed, 68 insertions(+)
13
tests/tcg/s390x/noexec.c | 106 +++++++++++++++++++++++
8
create mode 100644 tests/tcg/multiarch/sigbus.c
14
tests/tcg/multiarch/noexec.c.inc | 139 +++++++++++++++++++++++++++++++
15
tests/tcg/s390x/Makefile.target | 1 +
16
4 files changed, 257 insertions(+), 4 deletions(-)
17
create mode 100644 tests/tcg/s390x/noexec.c
18
create mode 100644 tests/tcg/multiarch/noexec.c.inc
9
19
10
diff --git a/tests/tcg/multiarch/sigbus.c b/tests/tcg/multiarch/sigbus.c
20
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/target/s390x/tcg/translate.c
23
+++ b/target/s390x/tcg/translate.c
24
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
25
dc->insn_start = tcg_last_op();
26
}
27
28
+static target_ulong get_next_pc(CPUS390XState *env, DisasContext *s,
29
+ uint64_t pc)
30
+{
31
+ uint64_t insn = ld_code2(env, s, pc);
32
+
33
+ return pc + get_ilen((insn >> 8) & 0xff);
34
+}
35
+
36
static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
37
{
38
CPUS390XState *env = cs->env_ptr;
39
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
40
41
dc->base.is_jmp = translate_one(env, dc);
42
if (dc->base.is_jmp == DISAS_NEXT) {
43
- uint64_t page_start;
44
-
45
- page_start = dc->base.pc_first & TARGET_PAGE_MASK;
46
- if (dc->base.pc_next - page_start >= TARGET_PAGE_SIZE || dc->ex_value) {
47
+ if (!is_same_page(dcbase, dc->base.pc_next) ||
48
+ !is_same_page(dcbase, get_next_pc(env, dc, dc->base.pc_next)) ||
49
+ dc->ex_value) {
50
dc->base.is_jmp = DISAS_TOO_MANY;
51
}
52
}
53
diff --git a/tests/tcg/s390x/noexec.c b/tests/tcg/s390x/noexec.c
11
new file mode 100644
54
new file mode 100644
12
index XXXXXXX..XXXXXXX
55
index XXXXXXX..XXXXXXX
13
--- /dev/null
56
--- /dev/null
14
+++ b/tests/tcg/multiarch/sigbus.c
57
+++ b/tests/tcg/s390x/noexec.c
15
@@ -XXX,XX +XXX,XX @@
58
@@ -XXX,XX +XXX,XX @@
16
+#define _GNU_SOURCE 1
59
+#include "../multiarch/noexec.c.inc"
60
+
61
+static void *arch_mcontext_pc(const mcontext_t *ctx)
62
+{
63
+ return (void *)ctx->psw.addr;
64
+}
65
+
66
+static int arch_mcontext_arg(const mcontext_t *ctx)
67
+{
68
+ return ctx->gregs[2];
69
+}
70
+
71
+static void arch_flush(void *p, int len)
72
+{
73
+}
74
+
75
+extern char noexec_1[];
76
+extern char noexec_2[];
77
+extern char noexec_end[];
78
+
79
+asm("noexec_1:\n"
80
+ " lgfi %r2,1\n" /* %r2 is 0 on entry, set 1. */
81
+ "noexec_2:\n"
82
+ " lgfi %r2,2\n" /* %r2 is 0/1; set 2. */
83
+ " br %r14\n" /* return */
84
+ "noexec_end:");
85
+
86
+extern char exrl_1[];
87
+extern char exrl_2[];
88
+extern char exrl_end[];
89
+
90
+asm("exrl_1:\n"
91
+ " exrl %r0, exrl_2\n"
92
+ " br %r14\n"
93
+ "exrl_2:\n"
94
+ " lgfi %r2,2\n"
95
+ "exrl_end:");
96
+
97
+int main(void)
98
+{
99
+ struct noexec_test noexec_tests[] = {
100
+ {
101
+ .name = "fallthrough",
102
+ .test_code = noexec_1,
103
+ .test_len = noexec_end - noexec_1,
104
+ .page_ofs = noexec_1 - noexec_2,
105
+ .entry_ofs = noexec_1 - noexec_2,
106
+ .expected_si_ofs = 0,
107
+ .expected_pc_ofs = 0,
108
+ .expected_arg = 1,
109
+ },
110
+ {
111
+ .name = "jump",
112
+ .test_code = noexec_1,
113
+ .test_len = noexec_end - noexec_1,
114
+ .page_ofs = noexec_1 - noexec_2,
115
+ .entry_ofs = 0,
116
+ .expected_si_ofs = 0,
117
+ .expected_pc_ofs = 0,
118
+ .expected_arg = 0,
119
+ },
120
+ {
121
+ .name = "exrl",
122
+ .test_code = exrl_1,
123
+ .test_len = exrl_end - exrl_1,
124
+ .page_ofs = exrl_1 - exrl_2,
125
+ .entry_ofs = exrl_1 - exrl_2,
126
+ .expected_si_ofs = 0,
127
+ .expected_pc_ofs = exrl_1 - exrl_2,
128
+ .expected_arg = 0,
129
+ },
130
+ {
131
+ .name = "fallthrough [cross]",
132
+ .test_code = noexec_1,
133
+ .test_len = noexec_end - noexec_1,
134
+ .page_ofs = noexec_1 - noexec_2 - 2,
135
+ .entry_ofs = noexec_1 - noexec_2 - 2,
136
+ .expected_si_ofs = 0,
137
+ .expected_pc_ofs = -2,
138
+ .expected_arg = 1,
139
+ },
140
+ {
141
+ .name = "jump [cross]",
142
+ .test_code = noexec_1,
143
+ .test_len = noexec_end - noexec_1,
144
+ .page_ofs = noexec_1 - noexec_2 - 2,
145
+ .entry_ofs = -2,
146
+ .expected_si_ofs = 0,
147
+ .expected_pc_ofs = -2,
148
+ .expected_arg = 0,
149
+ },
150
+ {
151
+ .name = "exrl [cross]",
152
+ .test_code = exrl_1,
153
+ .test_len = exrl_end - exrl_1,
154
+ .page_ofs = exrl_1 - exrl_2 - 2,
155
+ .entry_ofs = exrl_1 - exrl_2 - 2,
156
+ .expected_si_ofs = 0,
157
+ .expected_pc_ofs = exrl_1 - exrl_2 - 2,
158
+ .expected_arg = 0,
159
+ },
160
+ };
161
+
162
+ return test_noexec(noexec_tests,
163
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
164
+}
165
diff --git a/tests/tcg/multiarch/noexec.c.inc b/tests/tcg/multiarch/noexec.c.inc
166
new file mode 100644
167
index XXXXXXX..XXXXXXX
168
--- /dev/null
169
+++ b/tests/tcg/multiarch/noexec.c.inc
170
@@ -XXX,XX +XXX,XX @@
171
+/*
172
+ * Common code for arch-specific MMU_INST_FETCH fault testing.
173
+ */
174
+
175
+#define _GNU_SOURCE
17
+
176
+
18
+#include <assert.h>
177
+#include <assert.h>
178
+#include <signal.h>
179
+#include <stdio.h>
19
+#include <stdlib.h>
180
+#include <stdlib.h>
20
+#include <signal.h>
181
+#include <string.h>
21
+#include <endian.h>
182
+#include <errno.h>
22
+
183
+#include <unistd.h>
23
+
184
+#include <sys/mman.h>
24
+unsigned long long x = 0x8877665544332211ull;
185
+#include <sys/ucontext.h>
25
+void * volatile p = (void *)&x + 1;
186
+
26
+
187
+/* Forward declarations. */
27
+void sigbus(int sig, siginfo_t *info, void *uc)
188
+
28
+{
189
+static void *arch_mcontext_pc(const mcontext_t *ctx);
29
+ assert(sig == SIGBUS);
190
+static int arch_mcontext_arg(const mcontext_t *ctx);
30
+ assert(info->si_signo == SIGBUS);
191
+static void arch_flush(void *p, int len);
31
+#ifdef BUS_ADRALN
192
+
32
+ assert(info->si_code == BUS_ADRALN);
193
+/* Testing infrastructure. */
33
+#endif
194
+
34
+ assert(info->si_addr == p);
195
+struct noexec_test {
35
+ exit(EXIT_SUCCESS);
196
+ const char *name;
36
+}
197
+ const char *test_code;
37
+
198
+ int test_len;
38
+int main()
199
+ int page_ofs;
39
+{
200
+ int entry_ofs;
40
+ struct sigaction sa = {
201
+ int expected_si_ofs;
41
+ .sa_sigaction = sigbus,
202
+ int expected_pc_ofs;
42
+ .sa_flags = SA_SIGINFO
203
+ int expected_arg;
43
+ };
204
+};
44
+ int allow_fail = 0;
205
+
45
+ int tmp;
206
+static void *page_base;
46
+
207
+static int page_size;
47
+ tmp = sigaction(SIGBUS, &sa, NULL);
208
+static const struct noexec_test *current_noexec_test;
48
+ assert(tmp == 0);
209
+
49
+
210
+static void handle_err(const char *syscall)
50
+ /*
211
+{
51
+ * Select an operation that's likely to enforce alignment.
212
+ printf("[ FAILED ] %s: %s\n", syscall, strerror(errno));
52
+ * On many guests that support unaligned accesses by default,
213
+ exit(EXIT_FAILURE);
53
+ * this is often an atomic operation.
214
+}
54
+ */
215
+
55
+#if defined(__aarch64__)
216
+static void handle_segv(int sig, siginfo_t *info, void *ucontext)
56
+ asm volatile("ldxr %w0,[%1]" : "=r"(tmp) : "r"(p) : "memory");
217
+{
57
+#elif defined(__alpha__)
218
+ const struct noexec_test *test = current_noexec_test;
58
+ asm volatile("ldl_l %0,0(%1)" : "=r"(tmp) : "r"(p) : "memory");
219
+ const mcontext_t *mc = &((ucontext_t *)ucontext)->uc_mcontext;
59
+#elif defined(__arm__)
220
+ void *expected_si;
60
+ asm volatile("ldrex %0,[%1]" : "=r"(tmp) : "r"(p) : "memory");
221
+ void *expected_pc;
61
+#elif defined(__powerpc__)
222
+ void *pc;
62
+ asm volatile("lwarx %0,0,%1" : "=r"(tmp) : "r"(p) : "memory");
223
+ int arg;
63
+#elif defined(__riscv_atomic)
224
+
64
+ asm volatile("lr.w %0,(%1)" : "=r"(tmp) : "r"(p) : "memory");
225
+ if (test == NULL) {
65
+#else
226
+ printf("[ FAILED ] unexpected SEGV\n");
66
+ /* No insn known to fault unaligned -- try for a straight load. */
227
+ exit(EXIT_FAILURE);
67
+ allow_fail = 1;
228
+ }
68
+ tmp = *(volatile int *)p;
229
+ current_noexec_test = NULL;
69
+#endif
230
+
70
+
231
+ expected_si = page_base + test->expected_si_ofs;
71
+ assert(allow_fail);
232
+ if (info->si_addr != expected_si) {
72
+
233
+ printf("[ FAILED ] wrong si_addr (%p != %p)\n",
73
+ /*
234
+ info->si_addr, expected_si);
74
+ * We didn't see a signal.
235
+ exit(EXIT_FAILURE);
75
+ * We might as well validate the unaligned load worked.
236
+ }
76
+ */
237
+
77
+ if (BYTE_ORDER == LITTLE_ENDIAN) {
238
+ pc = arch_mcontext_pc(mc);
78
+ assert(tmp == 0x55443322);
239
+ expected_pc = page_base + test->expected_pc_ofs;
79
+ } else {
240
+ if (pc != expected_pc) {
80
+ assert(tmp == 0x77665544);
241
+ printf("[ FAILED ] wrong pc (%p != %p)\n", pc, expected_pc);
81
+ }
242
+ exit(EXIT_FAILURE);
243
+ }
244
+
245
+ arg = arch_mcontext_arg(mc);
246
+ if (arg != test->expected_arg) {
247
+ printf("[ FAILED ] wrong arg (%d != %d)\n", arg, test->expected_arg);
248
+ exit(EXIT_FAILURE);
249
+ }
250
+
251
+ if (mprotect(page_base, page_size,
252
+ PROT_READ | PROT_WRITE | PROT_EXEC) < 0) {
253
+ handle_err("mprotect");
254
+ }
255
+}
256
+
257
+static void test_noexec_1(const struct noexec_test *test)
258
+{
259
+ void *start = page_base + test->page_ofs;
260
+ void (*fn)(int arg) = page_base + test->entry_ofs;
261
+
262
+ memcpy(start, test->test_code, test->test_len);
263
+ arch_flush(start, test->test_len);
264
+
265
+ /* Trigger TB creation in order to test invalidation. */
266
+ fn(0);
267
+
268
+ if (mprotect(page_base, page_size, PROT_NONE) < 0) {
269
+ handle_err("mprotect");
270
+ }
271
+
272
+ /* Trigger SEGV and check that handle_segv() ran. */
273
+ current_noexec_test = test;
274
+ fn(0);
275
+ assert(current_noexec_test == NULL);
276
+}
277
+
278
+static int test_noexec(struct noexec_test *tests, size_t n_tests)
279
+{
280
+ struct sigaction act;
281
+ size_t i;
282
+
283
+ memset(&act, 0, sizeof(act));
284
+ act.sa_sigaction = handle_segv;
285
+ act.sa_flags = SA_SIGINFO;
286
+ if (sigaction(SIGSEGV, &act, NULL) < 0) {
287
+ handle_err("sigaction");
288
+ }
289
+
290
+ page_size = getpagesize();
291
+ page_base = mmap(NULL, 2 * page_size,
292
+ PROT_READ | PROT_WRITE | PROT_EXEC,
293
+ MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
294
+ if (page_base == MAP_FAILED) {
295
+ handle_err("mmap");
296
+ }
297
+ page_base += page_size;
298
+
299
+ for (i = 0; i < n_tests; i++) {
300
+ struct noexec_test *test = &tests[i];
301
+
302
+ printf("[ RUN ] %s\n", test->name);
303
+ test_noexec_1(test);
304
+ printf("[ OK ]\n");
305
+ }
306
+
307
+ printf("[ PASSED ]\n");
82
+ return EXIT_SUCCESS;
308
+ return EXIT_SUCCESS;
83
+}
309
+}
310
diff --git a/tests/tcg/s390x/Makefile.target b/tests/tcg/s390x/Makefile.target
311
index XXXXXXX..XXXXXXX 100644
312
--- a/tests/tcg/s390x/Makefile.target
313
+++ b/tests/tcg/s390x/Makefile.target
314
@@ -XXX,XX +XXX,XX @@ TESTS+=shift
315
TESTS+=trap
316
TESTS+=signals-s390x
317
TESTS+=branch-relative-long
318
+TESTS+=noexec
319
320
Z14_TESTS=vfminmax
321
vfminmax: LDFLAGS+=-lm
84
--
322
--
85
2.25.1
323
2.34.1
86
87
diff view generated by jsdifflib
1
This is now always true, since we require armv6.
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Right now translator stops right *after* the end of a page, which
4
breaks reporting of fault locations when the last instruction of a
5
multi-insn translation block crosses a page boundary.
6
7
An implementation, like the one arm and s390x have, would require an
8
i386 length disassembler, which is burdensome to maintain. Another
9
alternative would be to single-step at the end of a guest page, but
10
this may come with a performance impact.
11
12
Fix by snapshotting disassembly state and restoring it after we figure
13
out we crossed a page boundary. This includes rolling back cc_op
14
updates and emitted ops.
15
16
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1143
19
Message-Id: <20220817150506.592862-4-iii@linux.ibm.com>
20
[rth: Simplify end-of-insn cross-page checks.]
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
21
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
22
---
6
tcg/arm/tcg-target.h | 3 +--
23
target/i386/tcg/translate.c | 64 ++++++++++++++++-----------
7
tcg/arm/tcg-target.c.inc | 35 ++++++-----------------------------
24
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++++++++++++++
8
2 files changed, 7 insertions(+), 31 deletions(-)
25
tests/tcg/x86_64/Makefile.target | 3 +-
9
26
3 files changed, 116 insertions(+), 26 deletions(-)
10
diff --git a/tcg/arm/tcg-target.h b/tcg/arm/tcg-target.h
27
create mode 100644 tests/tcg/x86_64/noexec.c
28
29
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
11
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/arm/tcg-target.h
31
--- a/target/i386/tcg/translate.c
13
+++ b/tcg/arm/tcg-target.h
32
+++ b/target/i386/tcg/translate.c
14
@@ -XXX,XX +XXX,XX @@
33
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
15
34
TCGv_i64 tmp1_i64;
16
extern int arm_arch;
35
17
36
sigjmp_buf jmpbuf;
18
-#define use_armv5t_instructions (__ARM_ARCH >= 5 || arm_arch >= 5)
37
+ TCGOp *prev_insn_end;
19
#define use_armv6_instructions (__ARM_ARCH >= 6 || arm_arch >= 6)
38
} DisasContext;
20
#define use_armv7_instructions (__ARM_ARCH >= 7 || arm_arch >= 7)
39
21
40
/* The environment in which user-only runs is constrained. */
22
@@ -XXX,XX +XXX,XX @@ extern bool use_neon_instructions;
41
@@ -XXX,XX +XXX,XX @@ static uint64_t advance_pc(CPUX86State *env, DisasContext *s, int num_bytes)
23
#define TCG_TARGET_HAS_eqv_i32 0
42
{
24
#define TCG_TARGET_HAS_nand_i32 0
43
uint64_t pc = s->pc;
25
#define TCG_TARGET_HAS_nor_i32 0
44
26
-#define TCG_TARGET_HAS_clz_i32 use_armv5t_instructions
45
+ /* This is a subsequent insn that crosses a page boundary. */
27
+#define TCG_TARGET_HAS_clz_i32 1
46
+ if (s->base.num_insns > 1 &&
28
#define TCG_TARGET_HAS_ctz_i32 use_armv7_instructions
47
+ !is_same_page(&s->base, s->pc + num_bytes - 1)) {
29
#define TCG_TARGET_HAS_ctpop_i32 0
48
+ siglongjmp(s->jmpbuf, 2);
30
#define TCG_TARGET_HAS_deposit_i32 use_armv7_instructions
49
+ }
31
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
50
+
32
index XXXXXXX..XXXXXXX 100644
51
s->pc += num_bytes;
33
--- a/tcg/arm/tcg-target.c.inc
52
if (unlikely(s->pc - s->pc_start > X86_MAX_INSN_LENGTH)) {
34
+++ b/tcg/arm/tcg-target.c.inc
53
/* If the instruction's 16th byte is on a different page than the 1st, a
35
@@ -XXX,XX +XXX,XX @@ static void tcg_out_b_reg(TCGContext *s, ARMCond cond, TCGReg rn)
54
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
36
* Unless the C portion of QEMU is compiled as thumb, we don't need
55
int modrm, reg, rm, mod, op, opreg, val;
37
* true BX semantics; merely a branch to an address held in a register.
56
target_ulong next_eip, tval;
38
*/
57
target_ulong pc_start = s->base.pc_next;
39
- if (use_armv5t_instructions) {
58
+ bool orig_cc_op_dirty = s->cc_op_dirty;
40
- tcg_out_bx_reg(s, cond, rn);
59
+ CCOp orig_cc_op = s->cc_op;
41
- } else {
60
42
- tcg_out_mov_reg(s, cond, TCG_REG_PC, rn);
61
s->pc_start = s->pc = pc_start;
43
- }
62
s->override = -1;
44
+ tcg_out_bx_reg(s, cond, rn);
63
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
64
s->rip_offset = 0; /* for relative ip address */
65
s->vex_l = 0;
66
s->vex_v = 0;
67
- if (sigsetjmp(s->jmpbuf, 0) != 0) {
68
+ switch (sigsetjmp(s->jmpbuf, 0)) {
69
+ case 0:
70
+ break;
71
+ case 1:
72
gen_exception_gpf(s);
73
return s->pc;
74
+ case 2:
75
+ /* Restore state that may affect the next instruction. */
76
+ s->cc_op_dirty = orig_cc_op_dirty;
77
+ s->cc_op = orig_cc_op;
78
+ s->base.num_insns--;
79
+ tcg_remove_ops_after(s->prev_insn_end);
80
+ s->base.is_jmp = DISAS_TOO_MANY;
81
+ return pc_start;
82
+ default:
83
+ g_assert_not_reached();
84
}
85
86
prefixes = 0;
87
@@ -XXX,XX +XXX,XX @@ static void i386_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
88
{
89
DisasContext *dc = container_of(dcbase, DisasContext, base);
90
91
+ dc->prev_insn_end = tcg_last_op();
92
tcg_gen_insn_start(dc->base.pc_next, dc->cc_op);
45
}
93
}
46
94
47
static void tcg_out_dat_imm(TCGContext *s, ARMCond cond, ARMInsn opc,
95
@@ -XXX,XX +XXX,XX @@ static void i386_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
48
@@ -XXX,XX +XXX,XX @@ static void tcg_out_goto(TCGContext *s, ARMCond cond, const tcg_insn_unit *addr)
96
#endif
49
}
97
50
98
pc_next = disas_insn(dc, cpu);
51
/* LDR is interworking from v5t. */
99
-
52
- if (arm_mode || use_armv5t_instructions) {
100
- if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
53
- tcg_out_movi_pool(s, cond, TCG_REG_PC, addri);
101
- /* if single step mode, we generate only one instruction and
54
- return;
102
- generate an exception */
103
- /* if irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
104
- the flag and abort the translation to give the irqs a
105
- chance to happen */
106
- dc->base.is_jmp = DISAS_TOO_MANY;
107
- } else if ((tb_cflags(dc->base.tb) & CF_USE_ICOUNT)
108
- && ((pc_next & TARGET_PAGE_MASK)
109
- != ((pc_next + TARGET_MAX_INSN_SIZE - 1)
110
- & TARGET_PAGE_MASK)
111
- || (pc_next & ~TARGET_PAGE_MASK) == 0)) {
112
- /* Do not cross the boundary of the pages in icount mode,
113
- it can cause an exception. Do it only when boundary is
114
- crossed by the first instruction in the block.
115
- If current instruction already crossed the bound - it's ok,
116
- because an exception hasn't stopped this code.
117
- */
118
- dc->base.is_jmp = DISAS_TOO_MANY;
119
- } else if ((pc_next - dc->base.pc_first) >= (TARGET_PAGE_SIZE - 32)) {
120
- dc->base.is_jmp = DISAS_TOO_MANY;
55
- }
121
- }
56
-
122
-
57
- /* else v4t */
123
dc->base.pc_next = pc_next;
58
- tcg_out_movi32(s, COND_AL, TCG_REG_TMP, addri);
124
+
59
- tcg_out_bx_reg(s, COND_AL, TCG_REG_TMP);
125
+ if (dc->base.is_jmp == DISAS_NEXT) {
60
+ tcg_out_movi_pool(s, cond, TCG_REG_PC, addri);
126
+ if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
127
+ /*
128
+ * If single step mode, we generate only one instruction and
129
+ * generate an exception.
130
+ * If irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
131
+ * the flag and abort the translation to give the irqs a
132
+ * chance to happen.
133
+ */
134
+ dc->base.is_jmp = DISAS_TOO_MANY;
135
+ } else if (!is_same_page(&dc->base, pc_next)) {
136
+ dc->base.is_jmp = DISAS_TOO_MANY;
137
+ }
138
+ }
61
}
139
}
62
140
63
/*
141
static void i386_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
64
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call(TCGContext *s, const tcg_insn_unit *addr)
142
diff --git a/tests/tcg/x86_64/noexec.c b/tests/tcg/x86_64/noexec.c
65
if (disp - 8 < 0x02000000 && disp - 8 >= -0x02000000) {
143
new file mode 100644
66
if (arm_mode) {
144
index XXXXXXX..XXXXXXX
67
tcg_out_bl_imm(s, COND_AL, disp);
145
--- /dev/null
68
- return;
146
+++ b/tests/tcg/x86_64/noexec.c
69
- }
147
@@ -XXX,XX +XXX,XX @@
70
- if (use_armv5t_instructions) {
148
+#include "../multiarch/noexec.c.inc"
71
+ } else {
149
+
72
tcg_out_blx_imm(s, disp);
150
+static void *arch_mcontext_pc(const mcontext_t *ctx)
73
- return;
151
+{
74
}
152
+ return (void *)ctx->gregs[REG_RIP];
75
+ return;
153
+}
76
}
154
+
77
155
+int arch_mcontext_arg(const mcontext_t *ctx)
78
- if (use_armv5t_instructions) {
156
+{
79
- tcg_out_movi32(s, COND_AL, TCG_REG_TMP, addri);
157
+ return ctx->gregs[REG_RDI];
80
- tcg_out_blx_reg(s, COND_AL, TCG_REG_TMP);
158
+}
81
- } else if (arm_mode) {
159
+
82
- /* ??? Know that movi_pool emits exactly 1 insn. */
160
+static void arch_flush(void *p, int len)
83
- tcg_out_mov_reg(s, COND_AL, TCG_REG_R14, TCG_REG_PC);
161
+{
84
- tcg_out_movi_pool(s, COND_AL, TCG_REG_PC, addri);
162
+}
85
- } else {
163
+
86
- tcg_out_movi32(s, COND_AL, TCG_REG_TMP, addri);
164
+extern char noexec_1[];
87
- tcg_out_mov_reg(s, COND_AL, TCG_REG_R14, TCG_REG_PC);
165
+extern char noexec_2[];
88
- tcg_out_bx_reg(s, COND_AL, TCG_REG_TMP);
166
+extern char noexec_end[];
89
- }
167
+
90
+ tcg_out_movi32(s, COND_AL, TCG_REG_TMP, addri);
168
+asm("noexec_1:\n"
91
+ tcg_out_blx_reg(s, COND_AL, TCG_REG_TMP);
169
+ " movq $1,%rdi\n" /* %rdi is 0 on entry, set 1. */
92
}
170
+ "noexec_2:\n"
93
171
+ " movq $2,%rdi\n" /* %rdi is 0/1; set 2. */
94
static void tcg_out_goto_label(TCGContext *s, ARMCond cond, TCGLabel *l)
172
+ " ret\n"
173
+ "noexec_end:");
174
+
175
+int main(void)
176
+{
177
+ struct noexec_test noexec_tests[] = {
178
+ {
179
+ .name = "fallthrough",
180
+ .test_code = noexec_1,
181
+ .test_len = noexec_end - noexec_1,
182
+ .page_ofs = noexec_1 - noexec_2,
183
+ .entry_ofs = noexec_1 - noexec_2,
184
+ .expected_si_ofs = 0,
185
+ .expected_pc_ofs = 0,
186
+ .expected_arg = 1,
187
+ },
188
+ {
189
+ .name = "jump",
190
+ .test_code = noexec_1,
191
+ .test_len = noexec_end - noexec_1,
192
+ .page_ofs = noexec_1 - noexec_2,
193
+ .entry_ofs = 0,
194
+ .expected_si_ofs = 0,
195
+ .expected_pc_ofs = 0,
196
+ .expected_arg = 0,
197
+ },
198
+ {
199
+ .name = "fallthrough [cross]",
200
+ .test_code = noexec_1,
201
+ .test_len = noexec_end - noexec_1,
202
+ .page_ofs = noexec_1 - noexec_2 - 2,
203
+ .entry_ofs = noexec_1 - noexec_2 - 2,
204
+ .expected_si_ofs = 0,
205
+ .expected_pc_ofs = -2,
206
+ .expected_arg = 1,
207
+ },
208
+ {
209
+ .name = "jump [cross]",
210
+ .test_code = noexec_1,
211
+ .test_len = noexec_end - noexec_1,
212
+ .page_ofs = noexec_1 - noexec_2 - 2,
213
+ .entry_ofs = -2,
214
+ .expected_si_ofs = 0,
215
+ .expected_pc_ofs = -2,
216
+ .expected_arg = 0,
217
+ },
218
+ };
219
+
220
+ return test_noexec(noexec_tests,
221
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
222
+}
223
diff --git a/tests/tcg/x86_64/Makefile.target b/tests/tcg/x86_64/Makefile.target
224
index XXXXXXX..XXXXXXX 100644
225
--- a/tests/tcg/x86_64/Makefile.target
226
+++ b/tests/tcg/x86_64/Makefile.target
227
@@ -XXX,XX +XXX,XX @@ include $(SRC_PATH)/tests/tcg/i386/Makefile.target
228
229
ifeq ($(filter %-linux-user, $(TARGET)),$(TARGET))
230
X86_64_TESTS += vsyscall
231
+X86_64_TESTS += noexec
232
TESTS=$(MULTIARCH_TESTS) $(X86_64_TESTS) test-x86_64
233
else
234
TESTS=$(MULTIARCH_TESTS)
235
@@ -XXX,XX +XXX,XX @@ test-x86_64: LDFLAGS+=-lm -lc
236
test-x86_64: test-i386.c test-i386.h test-i386-shift.h test-i386-muldiv.h
237
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
238
239
-vsyscall: $(SRC_PATH)/tests/tcg/x86_64/vsyscall.c
240
+%: $(SRC_PATH)/tests/tcg/x86_64/%.c
241
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
95
--
242
--
96
2.25.1
243
2.34.1
97
98
diff view generated by jsdifflib
1
From: WANG Xuerui <git@xen0n.name>
1
These will be useful in properly ending the TB.
2
2
3
Signed-off-by: WANG Xuerui <git@xen0n.name>
3
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Message-Id: <20220106134238.3936163-1-git@xen0n.name>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
7
---
8
tcg/loongarch64/tcg-target.h | 2 -
8
target/riscv/translate.c | 10 +++++++++-
9
tcg/loongarch64/tcg-target.c.inc | 71 +++++++++++++++++++++++++++++++-
9
1 file changed, 9 insertions(+), 1 deletion(-)
10
2 files changed, 69 insertions(+), 4 deletions(-)
11
10
12
diff --git a/tcg/loongarch64/tcg-target.h b/tcg/loongarch64/tcg-target.h
11
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/loongarch64/tcg-target.h
13
--- a/target/riscv/translate.c
15
+++ b/tcg/loongarch64/tcg-target.h
14
+++ b/target/riscv/translate.c
16
@@ -XXX,XX +XXX,XX @@ void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
15
@@ -XXX,XX +XXX,XX @@ static uint32_t opcode_at(DisasContextBase *dcbase, target_ulong pc)
17
16
/* Include decoders for factored-out extensions */
18
#define TCG_TARGET_DEFAULT_MO (0)
17
#include "decode-XVentanaCondOps.c.inc"
19
18
20
-#ifdef CONFIG_SOFTMMU
19
+/* The specification allows for longer insns, but not supported by qemu. */
21
#define TCG_TARGET_NEED_LDST_LABELS
20
+#define MAX_INSN_LEN 4
22
-#endif
23
24
#define TCG_TARGET_HAS_MEMORY_BSWAP 0
25
26
diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc
27
index XXXXXXX..XXXXXXX 100644
28
--- a/tcg/loongarch64/tcg-target.c.inc
29
+++ b/tcg/loongarch64/tcg-target.c.inc
30
@@ -XXX,XX +XXX,XX @@
31
* THE SOFTWARE.
32
*/
33
34
+#include "../tcg-ldst.c.inc"
35
+
21
+
36
#ifdef CONFIG_DEBUG_TCG
22
+static inline int insn_len(uint16_t first_word)
37
static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
38
"zero",
39
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
40
*/
41
42
#if defined(CONFIG_SOFTMMU)
43
-#include "../tcg-ldst.c.inc"
44
-
45
/*
46
* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
47
* MemOpIdx oi, uintptr_t ra)
48
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
49
50
return tcg_out_goto(s, l->raddr);
51
}
52
+#else
53
+
54
+/*
55
+ * Alignment helpers for user-mode emulation
56
+ */
57
+
58
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addr_reg,
59
+ unsigned a_bits)
60
+{
23
+{
61
+ TCGLabelQemuLdst *l = new_ldst_label(s);
24
+ return (first_word & 3) == 3 ? 4 : 2;
62
+
63
+ l->is_ld = is_ld;
64
+ l->addrlo_reg = addr_reg;
65
+
66
+ /*
67
+ * Without micro-architecture details, we don't know which of bstrpick or
68
+ * andi is faster, so use bstrpick as it's not constrained by imm field
69
+ * width. (Not to say alignments >= 2^12 are going to happen any time
70
+ * soon, though)
71
+ */
72
+ tcg_out_opc_bstrpick_d(s, TCG_REG_TMP1, addr_reg, 0, a_bits - 1);
73
+
74
+ l->label_ptr[0] = s->code_ptr;
75
+ tcg_out_opc_bne(s, TCG_REG_TMP1, TCG_REG_ZERO, 0);
76
+
77
+ l->raddr = tcg_splitwx_to_rx(s->code_ptr);
78
+}
25
+}
79
+
26
+
80
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
27
static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
81
+{
28
{
82
+ /* resolve label address */
29
/*
83
+ if (!reloc_br_sk16(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
30
@@ -XXX,XX +XXX,XX @@ static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
84
+ return false;
31
};
85
+ }
32
86
+
33
/* Check for compressed insn */
87
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_A1, l->addrlo_reg);
34
- if (extract16(opcode, 0, 2) != 3) {
88
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_A0, TCG_AREG0);
35
+ if (insn_len(opcode) == 2) {
89
+
36
if (!has_ext(ctx, RVC)) {
90
+ /* tail call, with the return address back inline. */
37
gen_exception_illegal(ctx);
91
+ tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RA, (uintptr_t)l->raddr);
38
} else {
92
+ tcg_out_call_int(s, (const void *)(l->is_ld ? helper_unaligned_ld
93
+ : helper_unaligned_st), true);
94
+ return true;
95
+}
96
+
97
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
98
+{
99
+ return tcg_out_fail_alignment(s, l);
100
+}
101
+
102
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
103
+{
104
+ return tcg_out_fail_alignment(s, l);
105
+}
106
+
107
#endif /* CONFIG_SOFTMMU */
108
109
/*
110
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, TCGType type)
111
MemOp opc;
112
#if defined(CONFIG_SOFTMMU)
113
tcg_insn_unit *label_ptr[1];
114
+#else
115
+ unsigned a_bits;
116
#endif
117
TCGReg base;
118
119
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, TCGType type)
120
data_regl, addr_regl,
121
s->code_ptr, label_ptr);
122
#else
123
+ a_bits = get_alignment_bits(opc);
124
+ if (a_bits) {
125
+ tcg_out_test_alignment(s, true, addr_regl, a_bits);
126
+ }
127
base = tcg_out_zext_addr_if_32_bit(s, addr_regl, TCG_REG_TMP0);
128
TCGReg guest_base_reg = USE_GUEST_BASE ? TCG_GUEST_BASE_REG : TCG_REG_ZERO;
129
tcg_out_qemu_ld_indexed(s, data_regl, base, guest_base_reg, opc, type);
130
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args)
131
MemOp opc;
132
#if defined(CONFIG_SOFTMMU)
133
tcg_insn_unit *label_ptr[1];
134
+#else
135
+ unsigned a_bits;
136
#endif
137
TCGReg base;
138
139
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args)
140
data_regl, addr_regl,
141
s->code_ptr, label_ptr);
142
#else
143
+ a_bits = get_alignment_bits(opc);
144
+ if (a_bits) {
145
+ tcg_out_test_alignment(s, false, addr_regl, a_bits);
146
+ }
147
base = tcg_out_zext_addr_if_32_bit(s, addr_regl, TCG_REG_TMP0);
148
TCGReg guest_base_reg = USE_GUEST_BASE ? TCG_GUEST_BASE_REG : TCG_REG_ZERO;
149
tcg_out_qemu_st_indexed(s, data_regl, base, guest_base_reg, opc);
150
--
39
--
151
2.25.1
40
2.34.1
152
153
diff view generated by jsdifflib
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
1
Right now the translator stops right *after* the end of a page, which
2
breaks reporting of fault locations when the last instruction of a
3
multi-insn translation block crosses a page boundary.
4
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1155
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
10
---
4
tcg/ppc/tcg-target.h | 2 -
11
target/riscv/translate.c | 17 +++++--
5
tcg/ppc/tcg-target.c.inc | 98 ++++++++++++++++++++++++++++++++++++----
12
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++++++++++++
6
2 files changed, 90 insertions(+), 10 deletions(-)
13
tests/tcg/riscv64/Makefile.target | 1 +
14
3 files changed, 93 insertions(+), 4 deletions(-)
15
create mode 100644 tests/tcg/riscv64/noexec.c
7
16
8
diff --git a/tcg/ppc/tcg-target.h b/tcg/ppc/tcg-target.h
17
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
9
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
10
--- a/tcg/ppc/tcg-target.h
19
--- a/target/riscv/translate.c
11
+++ b/tcg/ppc/tcg-target.h
20
+++ b/target/riscv/translate.c
12
@@ -XXX,XX +XXX,XX @@ void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
21
@@ -XXX,XX +XXX,XX @@ static void riscv_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
13
#define TCG_TARGET_DEFAULT_MO (0)
22
}
14
#define TCG_TARGET_HAS_MEMORY_BSWAP 1
23
ctx->nftemp = 0;
15
24
16
-#ifdef CONFIG_SOFTMMU
25
+ /* Only the first insn within a TB is allowed to cross a page boundary. */
17
#define TCG_TARGET_NEED_LDST_LABELS
26
if (ctx->base.is_jmp == DISAS_NEXT) {
18
-#endif
27
- target_ulong page_start;
19
#define TCG_TARGET_NEED_POOL_LABELS
28
-
20
29
- page_start = ctx->base.pc_first & TARGET_PAGE_MASK;
21
#endif
30
- if (ctx->base.pc_next - page_start >= TARGET_PAGE_SIZE) {
22
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
31
+ if (!is_same_page(&ctx->base, ctx->base.pc_next)) {
23
index XXXXXXX..XXXXXXX 100644
32
ctx->base.is_jmp = DISAS_TOO_MANY;
24
--- a/tcg/ppc/tcg-target.c.inc
33
+ } else {
25
+++ b/tcg/ppc/tcg-target.c.inc
34
+ unsigned page_ofs = ctx->base.pc_next & ~TARGET_PAGE_MASK;
26
@@ -XXX,XX +XXX,XX @@
35
+
27
36
+ if (page_ofs > TARGET_PAGE_SIZE - MAX_INSN_LEN) {
28
#include "elf.h"
37
+ uint16_t next_insn = cpu_lduw_code(env, ctx->base.pc_next);
29
#include "../tcg-pool.c.inc"
38
+ int len = insn_len(next_insn);
30
+#include "../tcg-ldst.c.inc"
39
+
31
40
+ if (!is_same_page(&ctx->base, ctx->base.pc_next + len)) {
32
/*
41
+ ctx->base.is_jmp = DISAS_TOO_MANY;
33
* Standardize on the _CALL_FOO symbols used by GCC:
42
+ }
34
@@ -XXX,XX +XXX,XX @@ void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_rx,
43
+ }
44
}
35
}
45
}
36
}
46
}
37
47
diff --git a/tests/tcg/riscv64/noexec.c b/tests/tcg/riscv64/noexec.c
38
-static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target)
48
new file mode 100644
39
+static void tcg_out_call_int(TCGContext *s, int lk,
49
index XXXXXXX..XXXXXXX
40
+ const tcg_insn_unit *target)
50
--- /dev/null
41
{
51
+++ b/tests/tcg/riscv64/noexec.c
42
#ifdef _CALL_AIX
52
@@ -XXX,XX +XXX,XX @@
43
/* Look through the descriptor. If the branch is in range, and we
53
+#include "../multiarch/noexec.c.inc"
44
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target)
54
+
45
55
+static void *arch_mcontext_pc(const mcontext_t *ctx)
46
if (in_range_b(diff) && toc == (uint32_t)toc) {
47
tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP1, toc);
48
- tcg_out_b(s, LK, tgt);
49
+ tcg_out_b(s, lk, tgt);
50
} else {
51
/* Fold the low bits of the constant into the addresses below. */
52
intptr_t arg = (intptr_t)target;
53
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target)
54
tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R0, TCG_REG_TMP1, ofs);
55
tcg_out32(s, MTSPR | RA(TCG_REG_R0) | CTR);
56
tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R2, TCG_REG_TMP1, ofs + SZP);
57
- tcg_out32(s, BCCTR | BO_ALWAYS | LK);
58
+ tcg_out32(s, BCCTR | BO_ALWAYS | lk);
59
}
60
#elif defined(_CALL_ELF) && _CALL_ELF == 2
61
intptr_t diff;
62
@@ -XXX,XX +XXX,XX @@ static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target)
63
64
diff = tcg_pcrel_diff(s, target);
65
if (in_range_b(diff)) {
66
- tcg_out_b(s, LK, target);
67
+ tcg_out_b(s, lk, target);
68
} else {
69
tcg_out32(s, MTSPR | RS(TCG_REG_R12) | CTR);
70
- tcg_out32(s, BCCTR | BO_ALWAYS | LK);
71
+ tcg_out32(s, BCCTR | BO_ALWAYS | lk);
72
}
73
#else
74
- tcg_out_b(s, LK, target);
75
+ tcg_out_b(s, lk, target);
76
#endif
77
}
78
79
+static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target)
80
+{
56
+{
81
+ tcg_out_call_int(s, LK, target);
57
+ return (void *)ctx->__gregs[REG_PC];
82
+}
58
+}
83
+
59
+
84
static const uint32_t qemu_ldx_opc[(MO_SSIZE + MO_BSWAP) + 1] = {
60
+static int arch_mcontext_arg(const mcontext_t *ctx)
85
[MO_UB] = LBZX,
86
[MO_UW] = LHZX,
87
@@ -XXX,XX +XXX,XX @@ static const uint32_t qemu_exts_opc[4] = {
88
};
89
90
#if defined (CONFIG_SOFTMMU)
91
-#include "../tcg-ldst.c.inc"
92
-
93
/* helper signature: helper_ld_mmu(CPUState *env, target_ulong addr,
94
* int mmu_idx, uintptr_t ra)
95
*/
96
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
97
tcg_out_b(s, 0, lb->raddr);
98
return true;
99
}
100
+#else
101
+
102
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addrlo,
103
+ TCGReg addrhi, unsigned a_bits)
104
+{
61
+{
105
+ unsigned a_mask = (1 << a_bits) - 1;
62
+ return ctx->__gregs[REG_A0];
106
+ TCGLabelQemuLdst *label = new_ldst_label(s);
107
+
108
+ label->is_ld = is_ld;
109
+ label->addrlo_reg = addrlo;
110
+ label->addrhi_reg = addrhi;
111
+
112
+ /* We are expecting a_bits to max out at 7, much lower than ANDI. */
113
+ tcg_debug_assert(a_bits < 16);
114
+ tcg_out32(s, ANDI | SAI(addrlo, TCG_REG_R0, a_mask));
115
+
116
+ label->label_ptr[0] = s->code_ptr;
117
+ tcg_out32(s, BC | BI(0, CR_EQ) | BO_COND_FALSE | LK);
118
+
119
+ label->raddr = tcg_splitwx_to_rx(s->code_ptr);
120
+}
63
+}
121
+
64
+
122
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
65
+static void arch_flush(void *p, int len)
123
+{
66
+{
124
+ if (!reloc_pc14(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
67
+ __builtin___clear_cache(p, p + len);
125
+ return false;
126
+ }
127
+
128
+ if (TCG_TARGET_REG_BITS < TARGET_LONG_BITS) {
129
+ TCGReg arg = TCG_REG_R4;
130
+#ifdef TCG_TARGET_CALL_ALIGN_ARGS
131
+ arg |= 1;
132
+#endif
133
+ if (l->addrlo_reg != arg) {
134
+ tcg_out_mov(s, TCG_TYPE_I32, arg, l->addrhi_reg);
135
+ tcg_out_mov(s, TCG_TYPE_I32, arg + 1, l->addrlo_reg);
136
+ } else if (l->addrhi_reg != arg + 1) {
137
+ tcg_out_mov(s, TCG_TYPE_I32, arg + 1, l->addrlo_reg);
138
+ tcg_out_mov(s, TCG_TYPE_I32, arg, l->addrhi_reg);
139
+ } else {
140
+ tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R0, arg);
141
+ tcg_out_mov(s, TCG_TYPE_I32, arg, arg + 1);
142
+ tcg_out_mov(s, TCG_TYPE_I32, arg + 1, TCG_REG_R0);
143
+ }
144
+ } else {
145
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_R4, l->addrlo_reg);
146
+ }
147
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_R3, TCG_AREG0);
148
+
149
+ /* "Tail call" to the helper, with the return address back inline. */
150
+ tcg_out_call_int(s, 0, (const void *)(l->is_ld ? helper_unaligned_ld
151
+ : helper_unaligned_st));
152
+ return true;
153
+}
68
+}
154
+
69
+
155
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
70
+extern char noexec_1[];
71
+extern char noexec_2[];
72
+extern char noexec_end[];
73
+
74
+asm(".option push\n"
75
+ ".option norvc\n"
76
+ "noexec_1:\n"
77
+ " li a0,1\n" /* a0 is 0 on entry, set 1. */
78
+ "noexec_2:\n"
79
+ " li a0,2\n" /* a0 is 0/1; set 2. */
80
+ " ret\n"
81
+ "noexec_end:\n"
82
+ ".option pop");
83
+
84
+int main(void)
156
+{
85
+{
157
+ return tcg_out_fail_alignment(s, l);
86
+ struct noexec_test noexec_tests[] = {
87
+ {
88
+ .name = "fallthrough",
89
+ .test_code = noexec_1,
90
+ .test_len = noexec_end - noexec_1,
91
+ .page_ofs = noexec_1 - noexec_2,
92
+ .entry_ofs = noexec_1 - noexec_2,
93
+ .expected_si_ofs = 0,
94
+ .expected_pc_ofs = 0,
95
+ .expected_arg = 1,
96
+ },
97
+ {
98
+ .name = "jump",
99
+ .test_code = noexec_1,
100
+ .test_len = noexec_end - noexec_1,
101
+ .page_ofs = noexec_1 - noexec_2,
102
+ .entry_ofs = 0,
103
+ .expected_si_ofs = 0,
104
+ .expected_pc_ofs = 0,
105
+ .expected_arg = 0,
106
+ },
107
+ {
108
+ .name = "fallthrough [cross]",
109
+ .test_code = noexec_1,
110
+ .test_len = noexec_end - noexec_1,
111
+ .page_ofs = noexec_1 - noexec_2 - 2,
112
+ .entry_ofs = noexec_1 - noexec_2 - 2,
113
+ .expected_si_ofs = 0,
114
+ .expected_pc_ofs = -2,
115
+ .expected_arg = 1,
116
+ },
117
+ {
118
+ .name = "jump [cross]",
119
+ .test_code = noexec_1,
120
+ .test_len = noexec_end - noexec_1,
121
+ .page_ofs = noexec_1 - noexec_2 - 2,
122
+ .entry_ofs = -2,
123
+ .expected_si_ofs = 0,
124
+ .expected_pc_ofs = -2,
125
+ .expected_arg = 0,
126
+ },
127
+ };
128
+
129
+ return test_noexec(noexec_tests,
130
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
158
+}
131
+}
159
+
132
diff --git a/tests/tcg/riscv64/Makefile.target b/tests/tcg/riscv64/Makefile.target
160
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
133
index XXXXXXX..XXXXXXX 100644
161
+{
134
--- a/tests/tcg/riscv64/Makefile.target
162
+ return tcg_out_fail_alignment(s, l);
135
+++ b/tests/tcg/riscv64/Makefile.target
163
+}
136
@@ -XXX,XX +XXX,XX @@
164
+
137
165
#endif /* SOFTMMU */
138
VPATH += $(SRC_PATH)/tests/tcg/riscv64
166
139
TESTS += test-div
167
static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
140
+TESTS += noexec
168
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
169
#ifdef CONFIG_SOFTMMU
170
int mem_index;
171
tcg_insn_unit *label_ptr;
172
+#else
173
+ unsigned a_bits;
174
#endif
175
176
datalo = *args++;
177
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
178
179
rbase = TCG_REG_R3;
180
#else /* !CONFIG_SOFTMMU */
181
+ a_bits = get_alignment_bits(opc);
182
+ if (a_bits) {
183
+ tcg_out_test_alignment(s, true, addrlo, addrhi, a_bits);
184
+ }
185
rbase = guest_base ? TCG_GUEST_BASE_REG : 0;
186
if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
187
tcg_out_ext32u(s, TCG_REG_TMP1, addrlo);
188
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
189
#ifdef CONFIG_SOFTMMU
190
int mem_index;
191
tcg_insn_unit *label_ptr;
192
+#else
193
+ unsigned a_bits;
194
#endif
195
196
datalo = *args++;
197
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
198
199
rbase = TCG_REG_R3;
200
#else /* !CONFIG_SOFTMMU */
201
+ a_bits = get_alignment_bits(opc);
202
+ if (a_bits) {
203
+ tcg_out_test_alignment(s, false, addrlo, addrhi, a_bits);
204
+ }
205
rbase = guest_base ? TCG_GUEST_BASE_REG : 0;
206
if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
207
tcg_out_ext32u(s, TCG_REG_TMP1, addrlo);
208
--
141
--
209
2.25.1
142
2.34.1
210
211
diff view generated by jsdifflib
Deleted patch
1
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
---
3
tcg/riscv/tcg-target.h | 2 --
4
tcg/riscv/tcg-target.c.inc | 63 ++++++++++++++++++++++++++++++++++++--
5
2 files changed, 61 insertions(+), 4 deletions(-)
6
1
7
diff --git a/tcg/riscv/tcg-target.h b/tcg/riscv/tcg-target.h
8
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/riscv/tcg-target.h
10
+++ b/tcg/riscv/tcg-target.h
11
@@ -XXX,XX +XXX,XX @@ void tb_target_set_jmp_target(uintptr_t, uintptr_t, uintptr_t, uintptr_t);
12
13
#define TCG_TARGET_DEFAULT_MO (0)
14
15
-#ifdef CONFIG_SOFTMMU
16
#define TCG_TARGET_NEED_LDST_LABELS
17
-#endif
18
#define TCG_TARGET_NEED_POOL_LABELS
19
20
#define TCG_TARGET_HAS_MEMORY_BSWAP 0
21
diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc
22
index XXXXXXX..XXXXXXX 100644
23
--- a/tcg/riscv/tcg-target.c.inc
24
+++ b/tcg/riscv/tcg-target.c.inc
25
@@ -XXX,XX +XXX,XX @@
26
* THE SOFTWARE.
27
*/
28
29
+#include "../tcg-ldst.c.inc"
30
#include "../tcg-pool.c.inc"
31
32
#ifdef CONFIG_DEBUG_TCG
33
@@ -XXX,XX +XXX,XX @@ static void tcg_out_mb(TCGContext *s, TCGArg a0)
34
*/
35
36
#if defined(CONFIG_SOFTMMU)
37
-#include "../tcg-ldst.c.inc"
38
-
39
/* helper signature: helper_ret_ld_mmu(CPUState *env, target_ulong addr,
40
* MemOpIdx oi, uintptr_t ra)
41
*/
42
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
43
tcg_out_goto(s, l->raddr);
44
return true;
45
}
46
+#else
47
+
48
+static void tcg_out_test_alignment(TCGContext *s, bool is_ld, TCGReg addr_reg,
49
+ unsigned a_bits)
50
+{
51
+ unsigned a_mask = (1 << a_bits) - 1;
52
+ TCGLabelQemuLdst *l = new_ldst_label(s);
53
+
54
+ l->is_ld = is_ld;
55
+ l->addrlo_reg = addr_reg;
56
+
57
+ /* We are expecting a_bits to max out at 7, so we can always use andi. */
58
+ tcg_debug_assert(a_bits < 12);
59
+ tcg_out_opc_imm(s, OPC_ANDI, TCG_REG_TMP1, addr_reg, a_mask);
60
+
61
+ l->label_ptr[0] = s->code_ptr;
62
+ tcg_out_opc_branch(s, OPC_BNE, TCG_REG_TMP1, TCG_REG_ZERO, 0);
63
+
64
+ l->raddr = tcg_splitwx_to_rx(s->code_ptr);
65
+}
66
+
67
+static bool tcg_out_fail_alignment(TCGContext *s, TCGLabelQemuLdst *l)
68
+{
69
+ /* resolve label address */
70
+ if (!reloc_sbimm12(l->label_ptr[0], tcg_splitwx_to_rx(s->code_ptr))) {
71
+ return false;
72
+ }
73
+
74
+ tcg_out_mov(s, TCG_TYPE_TL, TCG_REG_A1, l->addrlo_reg);
75
+ tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_A0, TCG_AREG0);
76
+
77
+ /* tail call, with the return address back inline. */
78
+ tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RA, (uintptr_t)l->raddr);
79
+ tcg_out_call_int(s, (const void *)(l->is_ld ? helper_unaligned_ld
80
+ : helper_unaligned_st), true);
81
+ return true;
82
+}
83
+
84
+static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
85
+{
86
+ return tcg_out_fail_alignment(s, l);
87
+}
88
+
89
+static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
90
+{
91
+ return tcg_out_fail_alignment(s, l);
92
+}
93
+
94
#endif /* CONFIG_SOFTMMU */
95
96
static void tcg_out_qemu_ld_direct(TCGContext *s, TCGReg lo, TCGReg hi,
97
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
98
MemOp opc;
99
#if defined(CONFIG_SOFTMMU)
100
tcg_insn_unit *label_ptr[1];
101
+#else
102
+ unsigned a_bits;
103
#endif
104
TCGReg base = TCG_REG_TMP0;
105
106
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
107
tcg_out_ext32u(s, base, addr_regl);
108
addr_regl = base;
109
}
110
+ a_bits = get_alignment_bits(opc);
111
+ if (a_bits) {
112
+ tcg_out_test_alignment(s, true, addr_regl, a_bits);
113
+ }
114
if (guest_base != 0) {
115
tcg_out_opc_reg(s, OPC_ADD, base, TCG_GUEST_BASE_REG, addr_regl);
116
}
117
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
118
MemOp opc;
119
#if defined(CONFIG_SOFTMMU)
120
tcg_insn_unit *label_ptr[1];
121
+#else
122
+ unsigned a_bits;
123
#endif
124
TCGReg base = TCG_REG_TMP0;
125
126
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
127
tcg_out_ext32u(s, base, addr_regl);
128
addr_regl = base;
129
}
130
+ a_bits = get_alignment_bits(opc);
131
+ if (a_bits) {
132
+ tcg_out_test_alignment(s, false, addr_regl, a_bits);
133
+ }
134
if (guest_base != 0) {
135
tcg_out_opc_reg(s, OPC_ADD, base, TCG_GUEST_BASE_REG, addr_regl);
136
}
137
--
138
2.25.1
139
140
diff view generated by jsdifflib
Deleted patch
1
Support for unaligned accesses is difficult for pre-v6 hosts.
2
While debian still builds for armv4, we cannot use a compile
3
time test, so test the architecture at runtime and error out.
4
1
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/arm/tcg-target.c.inc | 5 +++++
9
1 file changed, 5 insertions(+)
10
11
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/arm/tcg-target.c.inc
14
+++ b/tcg/arm/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static void tcg_target_init(TCGContext *s)
16
if (pl != NULL && pl[0] == 'v' && pl[1] >= '4' && pl[1] <= '9') {
17
arm_arch = pl[1] - '0';
18
}
19
+
20
+ if (arm_arch < 6) {
21
+ error_report("TCG: ARMv%d is unsupported; exiting", arm_arch);
22
+ exit(EXIT_FAILURE);
23
+ }
24
}
25
26
tcg_target_available_regs[TCG_TYPE_I32] = ALL_GENERAL_REGS;
27
--
28
2.25.1
29
30
diff view generated by jsdifflib
Deleted patch
1
Reserve a register for the guest_base using aarch64 for reference.
2
By doing so, we do not have to recompute it for every memory load.
3
1
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/arm/tcg-target.c.inc | 39 ++++++++++++++++++++++++++++-----------
8
1 file changed, 28 insertions(+), 11 deletions(-)
9
10
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/arm/tcg-target.c.inc
13
+++ b/tcg/arm/tcg-target.c.inc
14
@@ -XXX,XX +XXX,XX @@ static const int tcg_target_call_oarg_regs[2] = {
15
16
#define TCG_REG_TMP TCG_REG_R12
17
#define TCG_VEC_TMP TCG_REG_Q15
18
+#ifndef CONFIG_SOFTMMU
19
+#define TCG_REG_GUEST_BASE TCG_REG_R11
20
+#endif
21
22
typedef enum {
23
COND_EQ = 0x0,
24
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
25
26
static void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
27
TCGReg datalo, TCGReg datahi,
28
- TCGReg addrlo, TCGReg addend)
29
+ TCGReg addrlo, TCGReg addend,
30
+ bool scratch_addend)
31
{
32
/* Byte swapping is left to middle-end expansion. */
33
tcg_debug_assert((opc & MO_BSWAP) == 0);
34
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_index(TCGContext *s, MemOp opc,
35
if (get_alignment_bits(opc) >= MO_64
36
&& (datalo & 1) == 0 && datahi == datalo + 1) {
37
tcg_out_ldrd_r(s, COND_AL, datalo, addrlo, addend);
38
- } else if (datalo != addend) {
39
+ } else if (scratch_addend) {
40
tcg_out_ld32_rwb(s, COND_AL, datalo, addend, addrlo);
41
tcg_out_ld32_12(s, COND_AL, datahi, addend, 4);
42
} else {
43
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
44
label_ptr = s->code_ptr;
45
tcg_out_bl_imm(s, COND_NE, 0);
46
47
- tcg_out_qemu_ld_index(s, opc, datalo, datahi, addrlo, addend);
48
+ tcg_out_qemu_ld_index(s, opc, datalo, datahi, addrlo, addend, true);
49
50
add_qemu_ldst_label(s, true, oi, datalo, datahi, addrlo, addrhi,
51
s->code_ptr, label_ptr);
52
#else /* !CONFIG_SOFTMMU */
53
if (guest_base) {
54
- tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP, guest_base);
55
- tcg_out_qemu_ld_index(s, opc, datalo, datahi, addrlo, TCG_REG_TMP);
56
+ tcg_out_qemu_ld_index(s, opc, datalo, datahi,
57
+ addrlo, TCG_REG_GUEST_BASE, false);
58
} else {
59
tcg_out_qemu_ld_direct(s, opc, datalo, datahi, addrlo);
60
}
61
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is64)
62
63
static void tcg_out_qemu_st_index(TCGContext *s, ARMCond cond, MemOp opc,
64
TCGReg datalo, TCGReg datahi,
65
- TCGReg addrlo, TCGReg addend)
66
+ TCGReg addrlo, TCGReg addend,
67
+ bool scratch_addend)
68
{
69
/* Byte swapping is left to middle-end expansion. */
70
tcg_debug_assert((opc & MO_BSWAP) == 0);
71
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st_index(TCGContext *s, ARMCond cond, MemOp opc,
72
if (get_alignment_bits(opc) >= MO_64
73
&& (datalo & 1) == 0 && datahi == datalo + 1) {
74
tcg_out_strd_r(s, cond, datalo, addrlo, addend);
75
- } else {
76
+ } else if (scratch_addend) {
77
tcg_out_st32_rwb(s, cond, datalo, addend, addrlo);
78
tcg_out_st32_12(s, cond, datahi, addend, 4);
79
+ } else {
80
+ tcg_out_dat_reg(s, cond, ARITH_ADD, TCG_REG_TMP,
81
+ addend, addrlo, SHIFT_IMM_LSL(0));
82
+ tcg_out_st32_12(s, cond, datalo, TCG_REG_TMP, 0);
83
+ tcg_out_st32_12(s, cond, datahi, TCG_REG_TMP, 4);
84
}
85
break;
86
default:
87
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
88
mem_index = get_mmuidx(oi);
89
addend = tcg_out_tlb_read(s, addrlo, addrhi, opc, mem_index, 0);
90
91
- tcg_out_qemu_st_index(s, COND_EQ, opc, datalo, datahi, addrlo, addend);
92
+ tcg_out_qemu_st_index(s, COND_EQ, opc, datalo, datahi,
93
+ addrlo, addend, true);
94
95
/* The conditional call must come last, as we're going to return here. */
96
label_ptr = s->code_ptr;
97
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is64)
98
s->code_ptr, label_ptr);
99
#else /* !CONFIG_SOFTMMU */
100
if (guest_base) {
101
- tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_TMP, guest_base);
102
- tcg_out_qemu_st_index(s, COND_AL, opc, datalo,
103
- datahi, addrlo, TCG_REG_TMP);
104
+ tcg_out_qemu_st_index(s, COND_AL, opc, datalo, datahi,
105
+ addrlo, TCG_REG_GUEST_BASE, false);
106
} else {
107
tcg_out_qemu_st_direct(s, opc, datalo, datahi, addrlo);
108
}
109
@@ -XXX,XX +XXX,XX @@ static void tcg_target_qemu_prologue(TCGContext *s)
110
111
tcg_out_mov(s, TCG_TYPE_PTR, TCG_AREG0, tcg_target_call_iarg_regs[0]);
112
113
+#ifndef CONFIG_SOFTMMU
114
+ if (guest_base) {
115
+ tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_GUEST_BASE, guest_base);
116
+ tcg_regset_set_reg(s->reserved_regs, TCG_REG_GUEST_BASE);
117
+ }
118
+#endif
119
+
120
tcg_out_b_reg(s, COND_AL, tcg_target_call_iarg_regs[1]);
121
122
/*
123
--
124
2.25.1
125
126
diff view generated by jsdifflib
Deleted patch
1
When BH is constant, it is constrained to 11 bits for use in MOVCC.
2
For the cases in which we must load the constant BH into a register,
3
we do not need the full logic of tcg_out_movi; we can use the simpler
4
function for emitting a 13 bit constant.
5
1
6
This eliminates the only case in which TCG_REG_T2 was passed to
7
tcg_out_movi, which will shortly become invalid.
8
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
12
tcg/sparc/tcg-target.c.inc | 10 +++++++---
13
1 file changed, 7 insertions(+), 3 deletions(-)
14
15
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
16
index XXXXXXX..XXXXXXX 100644
17
--- a/tcg/sparc/tcg-target.c.inc
18
+++ b/tcg/sparc/tcg-target.c.inc
19
@@ -XXX,XX +XXX,XX @@ static void tcg_out_addsub2_i64(TCGContext *s, TCGReg rl, TCGReg rh,
20
if (use_vis3_instructions && !is_sub) {
21
/* Note that ADDXC doesn't accept immediates. */
22
if (bhconst && bh != 0) {
23
- tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_T2, bh);
24
+ tcg_out_movi_imm13(s, TCG_REG_T2, bh);
25
bh = TCG_REG_T2;
26
}
27
tcg_out_arith(s, rh, ah, bh, ARITH_ADDXC);
28
@@ -XXX,XX +XXX,XX @@ static void tcg_out_addsub2_i64(TCGContext *s, TCGReg rl, TCGReg rh,
29
     tcg_out_movcc(s, TCG_COND_GEU, MOVCC_XCC, rh, ah, 0);
30
    }
31
} else {
32
- /* Otherwise adjust BH as if there is carry into T2 ... */
33
+ /*
34
+ * Otherwise adjust BH as if there is carry into T2.
35
+ * Note that constant BH is constrained to 11 bits for the MOVCC,
36
+ * so the adjustment fits 12 bits.
37
+ */
38
if (bhconst) {
39
- tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_T2, bh + (is_sub ? -1 : 1));
40
+ tcg_out_movi_imm13(s, TCG_REG_T2, bh + (is_sub ? -1 : 1));
41
} else {
42
tcg_out_arithi(s, TCG_REG_T2, bh, 1,
43
is_sub ? ARITH_SUB : ARITH_ADD);
44
--
45
2.25.1
46
47
diff view generated by jsdifflib
Deleted patch
1
Since 7ecd02a06f8, if patch_reloc fails we restart translation
2
with a smaller TB. SPARC had its function signature changed,
3
but not the logic. Replace assert with return false.
4
1
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
tcg/sparc/tcg-target.c.inc | 8 ++++++--
10
1 file changed, 6 insertions(+), 2 deletions(-)
11
12
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
13
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/sparc/tcg-target.c.inc
15
+++ b/tcg/sparc/tcg-target.c.inc
16
@@ -XXX,XX +XXX,XX @@ static bool patch_reloc(tcg_insn_unit *src_rw, int type,
17
18
switch (type) {
19
case R_SPARC_WDISP16:
20
- assert(check_fit_ptr(pcrel >> 2, 16));
21
+ if (!check_fit_ptr(pcrel >> 2, 16)) {
22
+ return false;
23
+ }
24
insn &= ~INSN_OFF16(-1);
25
insn |= INSN_OFF16(pcrel);
26
break;
27
case R_SPARC_WDISP19:
28
- assert(check_fit_ptr(pcrel >> 2, 19));
29
+ if (!check_fit_ptr(pcrel >> 2, 19)) {
30
+ return false;
31
+ }
32
insn &= ~INSN_OFF19(-1);
33
insn |= INSN_OFF19(pcrel);
34
break;
35
--
36
2.25.1
37
38
diff view generated by jsdifflib
Deleted patch
1
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3
---
4
tcg/sparc/tcg-target.c.inc | 15 +++++++++++++++
5
1 file changed, 15 insertions(+)
6
1
7
diff --git a/tcg/sparc/tcg-target.c.inc b/tcg/sparc/tcg-target.c.inc
8
index XXXXXXX..XXXXXXX 100644
9
--- a/tcg/sparc/tcg-target.c.inc
10
+++ b/tcg/sparc/tcg-target.c.inc
11
@@ -XXX,XX +XXX,XX @@ static bool patch_reloc(tcg_insn_unit *src_rw, int type,
12
insn &= ~INSN_OFF19(-1);
13
insn |= INSN_OFF19(pcrel);
14
break;
15
+ case R_SPARC_13:
16
+ if (!check_fit_ptr(value, 13)) {
17
+ return false;
18
+ }
19
+ insn &= ~INSN_IMM13(-1);
20
+ insn |= INSN_IMM13(value);
21
+ break;
22
default:
23
g_assert_not_reached();
24
}
25
@@ -XXX,XX +XXX,XX @@ static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
26
return;
27
}
28
29
+ /* Use the constant pool, if possible. */
30
+ if (!in_prologue && USE_REG_TB) {
31
+ new_pool_label(s, arg, R_SPARC_13, s->code_ptr,
32
+ tcg_tbrel_diff(s, NULL));
33
+ tcg_out32(s, LDX | INSN_RD(ret) | INSN_RS1(TCG_REG_TB));
34
+ return;
35
+ }
36
+
37
/* A 64-bit constant decomposed into 2 32-bit pieces. */
38
if (check_fit_i32(lo, 13)) {
39
hi = (arg - lo) >> 32;
40
--
41
2.25.1
42
43
diff view generated by jsdifflib