1
The following changes since commit 36eae3a732a1f2aa81391e871ac0e9bb3233e7d7:
1
The following changes since commit e93ded1bf6c94ab95015b33e188bc8b0b0c32670:
2
2
3
Merge remote-tracking branch 'remotes/dgilbert-gitlab/tags/pull-migration-20220302b' into staging (2022-03-02 20:55:48 +0000)
3
Merge tag 'testing-pull-request-2022-08-30' of https://gitlab.com/thuth/qemu into staging (2022-08-31 18:19:03 -0400)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220303
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220901
8
8
9
for you to fetch changes up to f23e6de25c31cadd9a3b7122f9384e6b259ce37f:
9
for you to fetch changes up to 20011be2e30b8aa8ef1fc258485f00c688703deb:
10
10
11
tcg/loongarch64: Support TCG_TARGET_SIGNED_ADDR32 (2022-03-03 10:47:20 -1000)
11
target/riscv: Make translator stop before the end of a page (2022-09-01 07:43:08 +0100)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Reorder do_constant_folding_cond test to satisfy valgrind.
14
Respect PROT_EXEC in user-only mode.
15
Fix value of MAX_OPC_PARAM_IARGS.
15
Fix s390x, i386 and riscv for translations crossing a page.
16
Add opcodes for vector nand, nor, eqv.
17
Support vector nand, nor, eqv on PPC and S390X hosts.
18
Support AVX512VL, AVX512BW, AVX512DQ, and AVX512VBMI2.
19
Support 32-bit guest addresses as signed values.
20
16
21
----------------------------------------------------------------
17
----------------------------------------------------------------
22
Alex Bennée (1):
18
Ilya Leoshkevich (4):
23
tcg/optimize: only read val after const check
19
linux-user: Clear translations on mprotect()
20
accel/tcg: Introduce is_same_page()
21
target/s390x: Make translator stop before the end of a page
22
target/i386: Make translator stop before the end of a page
24
23
25
Richard Henderson (28):
24
Richard Henderson (16):
26
tcg: Add opcodes for vector nand, nor, eqv
25
linux-user/arm: Mark the commpage executable
27
tcg/ppc: Implement vector NAND, NOR, EQV
26
linux-user/hppa: Allocate page zero as a commpage
28
tcg/s390x: Implement vector NAND, NOR, EQV
27
linux-user/x86_64: Allocate vsyscall page as a commpage
29
tcg/i386: Detect AVX512
28
linux-user: Honor PT_GNU_STACK
30
tcg/i386: Add tcg_out_evex_opc
29
tests/tcg/i386: Move smc_code2 to an executable section
31
tcg/i386: Use tcg_can_emit_vec_op in expand_vec_cmp_noinv
30
accel/tcg: Properly implement get_page_addr_code for user-only
32
tcg/i386: Implement avx512 variable shifts
31
accel/tcg: Unlock mmap_lock after longjmp
33
tcg/i386: Implement avx512 scalar shift
32
accel/tcg: Make tb_htable_lookup static
34
tcg/i386: Implement avx512 immediate sari shift
33
accel/tcg: Move qemu_ram_addr_from_host_nofail to physmem.c
35
tcg/i386: Implement avx512 immediate rotate
34
accel/tcg: Use probe_access_internal for softmmu get_page_addr_code_hostp
36
tcg/i386: Implement avx512 variable rotate
35
accel/tcg: Document the faulting lookup in tb_lookup_cmp
37
tcg/i386: Support avx512vbmi2 vector shift-double instructions
36
accel/tcg: Remove translator_ldsw
38
tcg/i386: Expand vector word rotate as avx512vbmi2 shift-double
37
accel/tcg: Add pc and host_pc params to gen_intermediate_code
39
tcg/i386: Remove rotls_vec from tcg_target_op_def
38
accel/tcg: Add fast path for translator_ld*
40
tcg/i386: Expand scalar rotate with avx512 insns
39
target/riscv: Add MAX_INSN_LEN and insn_len
41
tcg/i386: Implement avx512 min/max/abs
40
target/riscv: Make translator stop before the end of a page
42
tcg/i386: Implement avx512 multiply
43
tcg/i386: Implement more logical operations for avx512
44
tcg/i386: Implement bitsel for avx512
45
tcg: Add TCG_TARGET_SIGNED_ADDR32
46
accel/tcg: Split out g2h_tlbe
47
accel/tcg: Support TCG_TARGET_SIGNED_ADDR32 for softmmu
48
accel/tcg: Add guest_base_signed_addr32 for user-only
49
linux-user: Support TCG_TARGET_SIGNED_ADDR32
50
tcg/aarch64: Support TCG_TARGET_SIGNED_ADDR32
51
tcg/mips: Support TCG_TARGET_SIGNED_ADDR32
52
tcg/riscv: Support TCG_TARGET_SIGNED_ADDR32
53
tcg/loongarch64: Support TCG_TARGET_SIGNED_ADDR32
54
41
55
Ziqiao Kong (1):
42
include/elf.h | 1 +
56
tcg: Set MAX_OPC_PARAM_IARGS to 7
43
include/exec/cpu-common.h | 1 +
57
44
include/exec/exec-all.h | 89 ++++++++----------------
58
include/exec/cpu-all.h | 20 +-
45
include/exec/translator.h | 96 ++++++++++++++++---------
59
include/exec/cpu_ldst.h | 3 +-
46
linux-user/arm/target_cpu.h | 4 +-
60
include/qemu/cpuid.h | 20 +-
47
linux-user/qemu.h | 1 +
61
include/tcg/tcg-opc.h | 3 +
48
accel/tcg/cpu-exec.c | 143 ++++++++++++++++++++------------------
62
include/tcg/tcg.h | 5 +-
49
accel/tcg/cputlb.c | 93 +++++++------------------
63
tcg/aarch64/tcg-target-sa32.h | 7 +
50
accel/tcg/translate-all.c | 29 ++++----
64
tcg/aarch64/tcg-target.h | 3 +
51
accel/tcg/translator.c | 135 ++++++++++++++++++++++++++---------
65
tcg/arm/tcg-target-sa32.h | 1 +
52
accel/tcg/user-exec.c | 17 ++++-
66
tcg/arm/tcg-target.h | 3 +
53
linux-user/elfload.c | 82 ++++++++++++++++++++--
67
tcg/i386/tcg-target-con-set.h | 1 +
54
linux-user/mmap.c | 6 +-
68
tcg/i386/tcg-target-sa32.h | 1 +
55
softmmu/physmem.c | 12 ++++
69
tcg/i386/tcg-target.h | 17 +-
56
target/alpha/translate.c | 5 +-
70
tcg/i386/tcg-target.opc.h | 3 +
57
target/arm/translate.c | 5 +-
71
tcg/loongarch64/tcg-target-sa32.h | 1 +
58
target/avr/translate.c | 5 +-
72
tcg/mips/tcg-target-sa32.h | 9 +
59
target/cris/translate.c | 5 +-
73
tcg/ppc/tcg-target-sa32.h | 1 +
60
target/hexagon/translate.c | 6 +-
74
tcg/ppc/tcg-target.h | 3 +
61
target/hppa/translate.c | 5 +-
75
tcg/riscv/tcg-target-sa32.h | 5 +
62
target/i386/tcg/translate.c | 71 +++++++++++--------
76
tcg/s390x/tcg-target-sa32.h | 1 +
63
target/loongarch/translate.c | 6 +-
77
tcg/s390x/tcg-target.h | 3 +
64
target/m68k/translate.c | 5 +-
78
tcg/sparc/tcg-target-sa32.h | 1 +
65
target/microblaze/translate.c | 5 +-
79
tcg/tci/tcg-target-sa32.h | 1 +
66
target/mips/tcg/translate.c | 5 +-
80
accel/tcg/cputlb.c | 36 ++--
67
target/nios2/translate.c | 5 +-
81
bsd-user/main.c | 4 +
68
target/openrisc/translate.c | 6 +-
82
linux-user/elfload.c | 62 ++++--
69
target/ppc/translate.c | 5 +-
83
linux-user/main.c | 3 +
70
target/riscv/translate.c | 32 +++++++--
84
tcg/optimize.c | 20 +-
71
target/rx/translate.c | 5 +-
85
tcg/tcg-op-vec.c | 27 ++-
72
target/s390x/tcg/translate.c | 20 ++++--
86
tcg/tcg.c | 10 +
73
target/sh4/translate.c | 5 +-
87
tcg/aarch64/tcg-target.c.inc | 81 +++++---
74
target/sparc/translate.c | 5 +-
88
tcg/i386/tcg-target.c.inc | 387 +++++++++++++++++++++++++++++++-------
75
target/tricore/translate.c | 6 +-
89
tcg/loongarch64/tcg-target.c.inc | 15 +-
76
target/xtensa/translate.c | 6 +-
90
tcg/mips/tcg-target.c.inc | 10 +-
77
tests/tcg/i386/test-i386.c | 2 +-
91
tcg/ppc/tcg-target.c.inc | 15 ++
78
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++
92
tcg/riscv/tcg-target.c.inc | 8 +-
79
tests/tcg/s390x/noexec.c | 106 ++++++++++++++++++++++++++++
93
tcg/s390x/tcg-target.c.inc | 17 ++
80
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++
94
tcg/tci/tcg-target.c.inc | 2 +-
81
tests/tcg/multiarch/noexec.c.inc | 139 ++++++++++++++++++++++++++++++++++++
95
37 files changed, 640 insertions(+), 169 deletions(-)
82
tests/tcg/riscv64/Makefile.target | 1 +
96
create mode 100644 tcg/aarch64/tcg-target-sa32.h
83
tests/tcg/s390x/Makefile.target | 1 +
97
create mode 100644 tcg/arm/tcg-target-sa32.h
84
tests/tcg/x86_64/Makefile.target | 3 +-
98
create mode 100644 tcg/i386/tcg-target-sa32.h
85
43 files changed, 966 insertions(+), 367 deletions(-)
99
create mode 100644 tcg/loongarch64/tcg-target-sa32.h
86
create mode 100644 tests/tcg/riscv64/noexec.c
100
create mode 100644 tcg/mips/tcg-target-sa32.h
87
create mode 100644 tests/tcg/s390x/noexec.c
101
create mode 100644 tcg/ppc/tcg-target-sa32.h
88
create mode 100644 tests/tcg/x86_64/noexec.c
102
create mode 100644 tcg/riscv/tcg-target-sa32.h
89
create mode 100644 tests/tcg/multiarch/noexec.c.inc
103
create mode 100644 tcg/s390x/tcg-target-sa32.h
104
create mode 100644 tcg/sparc/tcg-target-sa32.h
105
create mode 100644 tcg/tci/tcg-target-sa32.h
106
diff view generated by jsdifflib
Deleted patch
1
From: Alex Bennée <alex.bennee@linaro.org>
2
1
3
valgrind pointed out that arg_info()->val can be undefined which will
4
be the case if the arguments are not constant. The ordering of the
5
checks will have ensured we never relied on an undefined value but for
6
the sake of completeness re-order the code to be clear.
7
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Signed-off-by: Alex Bennée <alex.bennee@linaro.org>
10
Message-Id: <20220209112142.3367525-1-alex.bennee@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
---
13
tcg/optimize.c | 8 ++++----
14
1 file changed, 4 insertions(+), 4 deletions(-)
15
16
diff --git a/tcg/optimize.c b/tcg/optimize.c
17
index XXXXXXX..XXXXXXX 100644
18
--- a/tcg/optimize.c
19
+++ b/tcg/optimize.c
20
@@ -XXX,XX +XXX,XX @@ static bool do_constant_folding_cond_eq(TCGCond c)
21
static int do_constant_folding_cond(TCGType type, TCGArg x,
22
TCGArg y, TCGCond c)
23
{
24
- uint64_t xv = arg_info(x)->val;
25
- uint64_t yv = arg_info(y)->val;
26
-
27
if (arg_is_const(x) && arg_is_const(y)) {
28
+ uint64_t xv = arg_info(x)->val;
29
+ uint64_t yv = arg_info(y)->val;
30
+
31
switch (type) {
32
case TCG_TYPE_I32:
33
return do_constant_folding_cond_32(xv, yv, c);
34
@@ -XXX,XX +XXX,XX @@ static int do_constant_folding_cond(TCGType type, TCGArg x,
35
}
36
} else if (args_are_copies(x, y)) {
37
return do_constant_folding_cond_eq(c);
38
- } else if (arg_is_const(y) && yv == 0) {
39
+ } else if (arg_is_const(y) && arg_info(y)->val == 0) {
40
switch (c) {
41
case TCG_COND_LTU:
42
return 0;
43
--
44
2.25.1
45
46
diff view generated by jsdifflib
1
AVX512DQ has VPMULLQ.
1
We're about to start validating PAGE_EXEC, which means
2
that we've got to mark the commpage executable. We had
3
been placing the commpage outside of reserved_va, which
4
was incorrect and lead to an abort.
2
5
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
9
---
7
tcg/i386/tcg-target.c.inc | 12 ++++++------
10
linux-user/arm/target_cpu.h | 4 ++--
8
1 file changed, 6 insertions(+), 6 deletions(-)
11
linux-user/elfload.c | 6 +++++-
12
2 files changed, 7 insertions(+), 3 deletions(-)
9
13
10
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
14
diff --git a/linux-user/arm/target_cpu.h b/linux-user/arm/target_cpu.h
11
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.c.inc
16
--- a/linux-user/arm/target_cpu.h
13
+++ b/tcg/i386/tcg-target.c.inc
17
+++ b/linux-user/arm/target_cpu.h
14
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
18
@@ -XXX,XX +XXX,XX @@ static inline unsigned long arm_max_reserved_va(CPUState *cs)
15
#define OPC_PMOVZXDQ (0x35 | P_EXT38 | P_DATA16)
19
} else {
16
#define OPC_PMULLW (0xd5 | P_EXT | P_DATA16)
20
/*
17
#define OPC_PMULLD (0x40 | P_EXT38 | P_DATA16)
21
* We need to be able to map the commpage.
18
+#define OPC_VPMULLQ (0x40 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
22
- * See validate_guest_space in linux-user/elfload.c.
19
#define OPC_POR (0xeb | P_EXT | P_DATA16)
23
+ * See init_guest_commpage in linux-user/elfload.c.
20
#define OPC_PSHUFB (0x00 | P_EXT38 | P_DATA16)
24
*/
21
#define OPC_PSHUFD (0x70 | P_EXT | P_DATA16)
25
- return 0xffff0000ul;
22
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
26
+ return 0xfffffffful;
23
OPC_PSUBUB, OPC_PSUBUW, OPC_UD2, OPC_UD2
27
}
24
};
28
}
25
static int const mul_insn[4] = {
29
#define MAX_RESERVED_VA arm_max_reserved_va
26
- OPC_UD2, OPC_PMULLW, OPC_PMULLD, OPC_UD2
30
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
27
+ OPC_UD2, OPC_PMULLW, OPC_PMULLD, OPC_VPMULLQ
31
index XXXXXXX..XXXXXXX 100644
28
};
32
--- a/linux-user/elfload.c
29
static int const shift_imm_insn[4] = {
33
+++ b/linux-user/elfload.c
30
OPC_UD2, OPC_PSHIFTW_Ib, OPC_PSHIFTD_Ib, OPC_PSHIFTQ_Ib
34
@@ -XXX,XX +XXX,XX @@ enum {
31
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
35
32
return 0;
36
static bool init_guest_commpage(void)
33
37
{
34
case INDEX_op_mul_vec:
38
- void *want = g2h_untagged(HI_COMMPAGE & -qemu_host_page_size);
35
- if (vece == MO_8) {
39
+ abi_ptr commpage = HI_COMMPAGE & -qemu_host_page_size;
36
- /* We can expand the operation for MO_8. */
40
+ void *want = g2h_untagged(commpage);
37
+ switch (vece) {
41
void *addr = mmap(want, qemu_host_page_size, PROT_READ | PROT_WRITE,
38
+ case MO_8:
42
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
39
return -1;
43
40
- }
44
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
41
- if (vece == MO_64) {
45
perror("Protecting guest commpage");
42
- return 0;
46
exit(EXIT_FAILURE);
43
+ case MO_64:
47
}
44
+ return have_avx512dq;
48
+
45
}
49
+ page_set_flags(commpage, commpage + qemu_host_page_size,
46
return 1;
50
+ PAGE_READ | PAGE_EXEC | PAGE_VALID);
51
return true;
52
}
47
53
48
--
54
--
49
2.25.1
55
2.34.1
50
51
diff view generated by jsdifflib
1
All 32-bit mips operations sign-extend the output, so we are easily
1
While there are no target-specific nonfaulting probes,
2
able to keep TCG_TYPE_I32 values sign-extended in host registers.
2
generic code may grow some uses at some point.
3
4
Note that the attrs argument was incorrect -- it should have
5
been MEMTXATTRS_UNSPECIFIED. Just use the simpler interface.
3
6
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
9
---
7
tcg/mips/tcg-target-sa32.h | 8 ++++++++
10
target/avr/helper.c | 46 ++++++++++++++++++++++++++++-----------------
8
tcg/mips/tcg-target.c.inc | 10 ++--------
11
1 file changed, 29 insertions(+), 17 deletions(-)
9
2 files changed, 10 insertions(+), 8 deletions(-)
10
12
11
diff --git a/tcg/mips/tcg-target-sa32.h b/tcg/mips/tcg-target-sa32.h
13
diff --git a/target/avr/helper.c b/target/avr/helper.c
12
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/mips/tcg-target-sa32.h
15
--- a/target/avr/helper.c
14
+++ b/tcg/mips/tcg-target-sa32.h
16
+++ b/target/avr/helper.c
15
@@ -1 +1,9 @@
17
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
16
+/*
18
MMUAccessType access_type, int mmu_idx,
17
+ * Do not set TCG_TARGET_SIGNED_ADDR32 for mips32;
19
bool probe, uintptr_t retaddr)
18
+ * TCG expects this to only be set for 64-bit hosts.
20
{
19
+ */
21
- int prot = 0;
20
+#ifdef __mips64
22
- MemTxAttrs attrs = {};
21
+#define TCG_TARGET_SIGNED_ADDR32 1
23
+ int prot, page_size = TARGET_PAGE_SIZE;
22
+#else
24
uint32_t paddr;
23
#define TCG_TARGET_SIGNED_ADDR32 0
25
24
+#endif
26
address &= TARGET_PAGE_MASK;
25
diff --git a/tcg/mips/tcg-target.c.inc b/tcg/mips/tcg-target.c.inc
27
26
index XXXXXXX..XXXXXXX 100644
28
if (mmu_idx == MMU_CODE_IDX) {
27
--- a/tcg/mips/tcg-target.c.inc
29
- /* access to code in flash */
28
+++ b/tcg/mips/tcg-target.c.inc
30
+ /* Access to code in flash. */
29
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg base, TCGReg addrl,
31
paddr = OFFSET_CODE + address;
30
TCG_TMP0, TCG_TMP3, cmp_off);
32
prot = PAGE_READ | PAGE_EXEC;
33
- if (paddr + TARGET_PAGE_SIZE > OFFSET_DATA) {
34
+ if (paddr >= OFFSET_DATA) {
35
+ /*
36
+ * This should not be possible via any architectural operations.
37
+ * There is certainly not an exception that we can deliver.
38
+ * Accept probing that might come from generic code.
39
+ */
40
+ if (probe) {
41
+ return false;
42
+ }
43
error_report("execution left flash memory");
44
abort();
45
}
46
- } else if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
47
- /*
48
- * access to CPU registers, exit and rebuilt this TB to use full access
49
- * incase it touches specially handled registers like SREG or SP
50
- */
51
- AVRCPU *cpu = AVR_CPU(cs);
52
- CPUAVRState *env = &cpu->env;
53
- env->fullacc = 1;
54
- cpu_loop_exit_restore(cs, retaddr);
55
} else {
56
- /* access to memory. nothing special */
57
+ /* Access to memory. */
58
paddr = OFFSET_DATA + address;
59
prot = PAGE_READ | PAGE_WRITE;
60
+ if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
61
+ /*
62
+ * Access to CPU registers, exit and rebuilt this TB to use
63
+ * full access in case it touches specially handled registers
64
+ * like SREG or SP. For probing, set page_size = 1, in order
65
+ * to force tlb_fill to be called for the next access.
66
+ */
67
+ if (probe) {
68
+ page_size = 1;
69
+ } else {
70
+ AVRCPU *cpu = AVR_CPU(cs);
71
+ CPUAVRState *env = &cpu->env;
72
+ env->fullacc = 1;
73
+ cpu_loop_exit_restore(cs, retaddr);
74
+ }
75
+ }
31
}
76
}
32
77
33
- /* Zero extend a 32-bit guest address for a 64-bit host. */
78
- tlb_set_page_with_attrs(cs, address, paddr, attrs, prot,
34
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
79
- mmu_idx, TARGET_PAGE_SIZE);
35
- tcg_out_ext32u(s, base, addrl);
36
- addrl = base;
37
- }
38
-
80
-
39
/*
81
+ tlb_set_page(cs, address, paddr, prot, mmu_idx, page_size);
40
* Mask the page bits, keeping the alignment bits to compare against.
82
return true;
41
* For unaligned accesses, compare against the end of the access to
83
}
42
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
84
43
data_regl, data_regh, addr_regl, addr_regh,
44
s->code_ptr, label_ptr);
45
#else
46
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
47
+ if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS && !guest_base_signed_addr32) {
48
tcg_out_ext32u(s, base, addr_regl);
49
addr_regl = base;
50
}
51
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
52
data_regl, data_regh, addr_regl, addr_regh,
53
s->code_ptr, label_ptr);
54
#else
55
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
56
+ if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS && !guest_base_signed_addr32) {
57
tcg_out_ext32u(s, base, addr_regl);
58
addr_regl = base;
59
}
60
--
85
--
61
2.25.1
86
2.34.1
62
87
63
88
diff view generated by jsdifflib
1
Tested-by: Alex Bennée <alex.bennee@linaro.org>
1
There is no need to go through cc->tcg_ops when
2
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
2
we know what value that must have.
3
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
3
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
7
---
6
tcg/ppc/tcg-target.h | 6 +++---
8
target/avr/helper.c | 5 ++---
7
tcg/ppc/tcg-target.c.inc | 15 +++++++++++++++
9
1 file changed, 2 insertions(+), 3 deletions(-)
8
2 files changed, 18 insertions(+), 3 deletions(-)
9
10
10
diff --git a/tcg/ppc/tcg-target.h b/tcg/ppc/tcg-target.h
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
11
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/ppc/tcg-target.h
13
--- a/target/avr/helper.c
13
+++ b/tcg/ppc/tcg-target.h
14
+++ b/target/avr/helper.c
14
@@ -XXX,XX +XXX,XX @@ extern bool have_vsx;
15
@@ -XXX,XX +XXX,XX @@
15
16
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
16
#define TCG_TARGET_HAS_andc_vec 1
17
{
17
#define TCG_TARGET_HAS_orc_vec have_isa_2_07
18
bool ret = false;
18
-#define TCG_TARGET_HAS_nand_vec 0
19
- CPUClass *cc = CPU_GET_CLASS(cs);
19
-#define TCG_TARGET_HAS_nor_vec 0
20
AVRCPU *cpu = AVR_CPU(cs);
20
-#define TCG_TARGET_HAS_eqv_vec 0
21
CPUAVRState *env = &cpu->env;
21
+#define TCG_TARGET_HAS_nand_vec have_isa_2_07
22
22
+#define TCG_TARGET_HAS_nor_vec 1
23
if (interrupt_request & CPU_INTERRUPT_RESET) {
23
+#define TCG_TARGET_HAS_eqv_vec have_isa_2_07
24
if (cpu_interrupts_enabled(env)) {
24
#define TCG_TARGET_HAS_not_vec 1
25
cs->exception_index = EXCP_RESET;
25
#define TCG_TARGET_HAS_neg_vec have_isa_3_00
26
- cc->tcg_ops->do_interrupt(cs);
26
#define TCG_TARGET_HAS_abs_vec 0
27
+ avr_cpu_do_interrupt(cs);
27
diff --git a/tcg/ppc/tcg-target.c.inc b/tcg/ppc/tcg-target.c.inc
28
28
index XXXXXXX..XXXXXXX 100644
29
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
29
--- a/tcg/ppc/tcg-target.c.inc
30
30
+++ b/tcg/ppc/tcg-target.c.inc
31
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
31
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
32
if (cpu_interrupts_enabled(env) && env->intsrc != 0) {
32
case INDEX_op_xor_vec:
33
int index = ctz32(env->intsrc);
33
case INDEX_op_andc_vec:
34
cs->exception_index = EXCP_INT(index);
34
case INDEX_op_not_vec:
35
- cc->tcg_ops->do_interrupt(cs);
35
+ case INDEX_op_nor_vec:
36
+ avr_cpu_do_interrupt(cs);
36
+ case INDEX_op_eqv_vec:
37
37
+ case INDEX_op_nand_vec:
38
env->intsrc &= env->intsrc - 1; /* clear the interrupt */
38
return 1;
39
if (!env->intsrc) {
39
case INDEX_op_orc_vec:
40
return have_isa_2_07;
41
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
42
case INDEX_op_orc_vec:
43
insn = VORC;
44
break;
45
+ case INDEX_op_nand_vec:
46
+ insn = VNAND;
47
+ break;
48
+ case INDEX_op_nor_vec:
49
+ insn = VNOR;
50
+ break;
51
+ case INDEX_op_eqv_vec:
52
+ insn = VEQV;
53
+ break;
54
55
case INDEX_op_cmp_vec:
56
switch (args[3]) {
57
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
58
case INDEX_op_xor_vec:
59
case INDEX_op_andc_vec:
60
case INDEX_op_orc_vec:
61
+ case INDEX_op_nor_vec:
62
+ case INDEX_op_eqv_vec:
63
+ case INDEX_op_nand_vec:
64
case INDEX_op_cmp_vec:
65
case INDEX_op_ssadd_vec:
66
case INDEX_op_sssub_vec:
67
--
40
--
68
2.25.1
41
2.34.1
69
42
70
43
diff view generated by jsdifflib
1
The evex encoding is added here, for use in a subsequent patch.
1
We're about to start validating PAGE_EXEC, which means that we've
2
got to mark page zero executable. We had been special casing this
3
entirely within translate.
2
4
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
8
---
7
tcg/i386/tcg-target.c.inc | 51 ++++++++++++++++++++++++++++++++++++++-
9
linux-user/elfload.c | 34 +++++++++++++++++++++++++++++++---
8
1 file changed, 50 insertions(+), 1 deletion(-)
10
1 file changed, 31 insertions(+), 3 deletions(-)
9
11
10
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
11
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.c.inc
14
--- a/linux-user/elfload.c
13
+++ b/tcg/i386/tcg-target.c.inc
15
+++ b/linux-user/elfload.c
14
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
16
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
15
#define P_SIMDF3 0x20000 /* 0xf3 opcode prefix */
17
regs->gr[31] = infop->entry;
16
#define P_SIMDF2 0x40000 /* 0xf2 opcode prefix */
17
#define P_VEXL 0x80000 /* Set VEX.L = 1 */
18
+#define P_EVEX 0x100000 /* Requires EVEX encoding */
19
20
#define OPC_ARITH_EvIz    (0x81)
21
#define OPC_ARITH_EvIb    (0x83)
22
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vex_opc(TCGContext *s, int opc, int r, int v,
23
tcg_out8(s, opc);
24
}
18
}
25
19
26
+static void tcg_out_evex_opc(TCGContext *s, int opc, int r, int v,
20
+#define LO_COMMPAGE 0
27
+ int rm, int index)
21
+
22
+static bool init_guest_commpage(void)
28
+{
23
+{
29
+ /* The entire 4-byte evex prefix; with R' and V' set. */
24
+ void *want = g2h_untagged(LO_COMMPAGE);
30
+ uint32_t p = 0x08041062;
25
+ void *addr = mmap(want, qemu_host_page_size, PROT_NONE,
31
+ int mm, pp;
26
+ MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
32
+
27
+
33
+ tcg_debug_assert(have_avx512vl);
28
+ if (addr == MAP_FAILED) {
34
+
29
+ perror("Allocating guest commpage");
35
+ /* EVEX.mm */
30
+ exit(EXIT_FAILURE);
36
+ if (opc & P_EXT3A) {
31
+ }
37
+ mm = 3;
32
+ if (addr != want) {
38
+ } else if (opc & P_EXT38) {
33
+ return false;
39
+ mm = 2;
40
+ } else if (opc & P_EXT) {
41
+ mm = 1;
42
+ } else {
43
+ g_assert_not_reached();
44
+ }
34
+ }
45
+
35
+
46
+ /* EVEX.pp */
36
+ /*
47
+ if (opc & P_DATA16) {
37
+ * On Linux, page zero is normally marked execute only + gateway.
48
+ pp = 1; /* 0x66 */
38
+ * Normal read or write is supposed to fail (thus PROT_NONE above),
49
+ } else if (opc & P_SIMDF3) {
39
+ * but specific offsets have kernel code mapped to raise permissions
50
+ pp = 2; /* 0xf3 */
40
+ * and implement syscalls. Here, simply mark the page executable.
51
+ } else if (opc & P_SIMDF2) {
41
+ * Special case the entry points during translation (see do_page_zero).
52
+ pp = 3; /* 0xf2 */
42
+ */
53
+ } else {
43
+ page_set_flags(LO_COMMPAGE, LO_COMMPAGE + TARGET_PAGE_SIZE,
54
+ pp = 0;
44
+ PAGE_EXEC | PAGE_VALID);
55
+ }
45
+ return true;
56
+
57
+ p = deposit32(p, 8, 2, mm);
58
+ p = deposit32(p, 13, 1, (rm & 8) == 0); /* EVEX.RXB.B */
59
+ p = deposit32(p, 14, 1, (index & 8) == 0); /* EVEX.RXB.X */
60
+ p = deposit32(p, 15, 1, (r & 8) == 0); /* EVEX.RXB.R */
61
+ p = deposit32(p, 16, 2, pp);
62
+ p = deposit32(p, 19, 4, ~v);
63
+ p = deposit32(p, 23, 1, (opc & P_VEXW) != 0);
64
+ p = deposit32(p, 29, 2, (opc & P_VEXL) != 0);
65
+
66
+ tcg_out32(s, p);
67
+ tcg_out8(s, opc);
68
+}
46
+}
69
+
47
+
70
static void tcg_out_vex_modrm(TCGContext *s, int opc, int r, int v, int rm)
48
#endif /* TARGET_HPPA */
71
{
49
72
- tcg_out_vex_opc(s, opc, r, v, rm, 0);
50
#ifdef TARGET_XTENSA
73
+ if (opc & P_EVEX) {
51
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
74
+ tcg_out_evex_opc(s, opc, r, v, rm, 0);
75
+ } else {
76
+ tcg_out_vex_opc(s, opc, r, v, rm, 0);
77
+ }
78
tcg_out8(s, 0xc0 | (LOWREGMASK(r) << 3) | LOWREGMASK(rm));
79
}
52
}
80
53
54
#if defined(HI_COMMPAGE)
55
-#define LO_COMMPAGE 0
56
+#define LO_COMMPAGE -1
57
#elif defined(LO_COMMPAGE)
58
#define HI_COMMPAGE 0
59
#else
60
#define HI_COMMPAGE 0
61
-#define LO_COMMPAGE 0
62
+#define LO_COMMPAGE -1
63
#define init_guest_commpage() true
64
#endif
65
66
@@ -XXX,XX +XXX,XX @@ static void pgb_static(const char *image_name, abi_ulong orig_loaddr,
67
} else {
68
offset = -(HI_COMMPAGE & -align);
69
}
70
- } else if (LO_COMMPAGE != 0) {
71
+ } else if (LO_COMMPAGE != -1) {
72
loaddr = MIN(loaddr, LO_COMMPAGE & -align);
73
}
74
81
--
75
--
82
2.25.1
76
2.34.1
83
84
diff view generated by jsdifflib
1
The general ternary logic operation can implement BITSEL.
1
We're about to start validating PAGE_EXEC, which means that we've
2
Funnel the 4-operand operation into three variants of the
2
got to mark the vsyscall page executable. We had been special
3
3-operand instruction, depending on input operand overlap.
3
casing this entirely within translate.
4
4
5
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
8
---
9
tcg/i386/tcg-target.h | 2 +-
9
linux-user/elfload.c | 23 +++++++++++++++++++++++
10
tcg/i386/tcg-target.c.inc | 20 +++++++++++++++++++-
10
1 file changed, 23 insertions(+)
11
2 files changed, 20 insertions(+), 2 deletions(-)
12
11
13
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
14
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
15
--- a/tcg/i386/tcg-target.h
14
--- a/linux-user/elfload.c
16
+++ b/tcg/i386/tcg-target.h
15
+++ b/linux-user/elfload.c
17
@@ -XXX,XX +XXX,XX @@ extern bool have_movbe;
16
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
18
#define TCG_TARGET_HAS_mul_vec 1
17
(*regs)[26] = tswapreg(env->segs[R_GS].selector & 0xffff);
19
#define TCG_TARGET_HAS_sat_vec 1
18
}
20
#define TCG_TARGET_HAS_minmax_vec 1
19
21
-#define TCG_TARGET_HAS_bitsel_vec 0
20
+#if ULONG_MAX >= TARGET_VSYSCALL_PAGE
22
+#define TCG_TARGET_HAS_bitsel_vec have_avx512vl
21
+#define INIT_GUEST_COMMPAGE
23
#define TCG_TARGET_HAS_cmpsel_vec -1
22
+static bool init_guest_commpage(void)
24
23
+{
25
#define TCG_TARGET_deposit_i32_valid(ofs, len) \
24
+ /*
26
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
25
+ * The vsyscall page is at a high negative address aka kernel space,
27
index XXXXXXX..XXXXXXX 100644
26
+ * which means that we cannot actually allocate it with target_mmap.
28
--- a/tcg/i386/tcg-target.c.inc
27
+ * We still should be able to use page_set_flags, unless the user
29
+++ b/tcg/i386/tcg-target.c.inc
28
+ * has specified -R reserved_va, which would trigger an assert().
30
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
29
+ */
31
30
+ if (reserved_va != 0 &&
32
TCGType type = vecl + TCG_TYPE_V64;
31
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE >= reserved_va) {
33
int insn, sub;
32
+ error_report("Cannot allocate vsyscall page");
34
- TCGArg a0, a1, a2;
33
+ exit(EXIT_FAILURE);
35
+ TCGArg a0, a1, a2, a3;
34
+ }
36
35
+ page_set_flags(TARGET_VSYSCALL_PAGE,
37
a0 = args[0];
36
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE,
38
a1 = args[1];
37
+ PAGE_EXEC | PAGE_VALID);
39
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
38
+ return true;
40
sub = 0xdd; /* orB!C */
39
+}
41
goto gen_simd_imm8;
40
+#endif
42
41
#else
43
+ case INDEX_op_bitsel_vec:
42
44
+ insn = OPC_VPTERNLOGQ;
43
#define ELF_START_MMAP 0x80000000
45
+ a3 = args[3];
44
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
46
+ if (a0 == a1) {
45
#else
47
+ a1 = a2;
46
#define HI_COMMPAGE 0
48
+ a2 = a3;
47
#define LO_COMMPAGE -1
49
+ sub = 0xca; /* A?B:C */
48
+#ifndef INIT_GUEST_COMMPAGE
50
+ } else if (a0 == a2) {
49
#define init_guest_commpage() true
51
+ a2 = a3;
50
#endif
52
+ sub = 0xe2; /* B?A:C */
51
+#endif
53
+ } else {
52
54
+ tcg_out_mov(s, type, a0, a3);
53
static void pgb_fail_in_use(const char *image_name)
55
+ sub = 0xb8; /* B?C:A */
54
{
56
+ }
57
+ goto gen_simd_imm8;
58
+
59
gen_simd_imm8:
60
tcg_debug_assert(insn != OPC_UD2);
61
if (type == TCG_TYPE_V256) {
62
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
63
case INDEX_op_x86_vpshrdv_vec:
64
return C_O1_I3(x, 0, x, x);
65
66
+ case INDEX_op_bitsel_vec:
67
case INDEX_op_x86_vpblendvb_vec:
68
return C_O1_I3(x, x, x, x);
69
70
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
71
case INDEX_op_nor_vec:
72
case INDEX_op_eqv_vec:
73
case INDEX_op_not_vec:
74
+ case INDEX_op_bitsel_vec:
75
return 1;
76
case INDEX_op_cmp_vec:
77
case INDEX_op_cmpsel_vec:
78
--
55
--
79
2.25.1
56
2.34.1
80
81
diff view generated by jsdifflib
1
All 32-bit LoongArch operations sign-extend the output, so we are easily
1
We cannot deliver two interrupts simultaneously;
2
able to keep TCG_TYPE_I32 values sign-extended in host registers.
2
the first interrupt handler must execute first.
3
3
4
Cc: WANG Xuerui <git@xen0n.name>
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
7
---
8
tcg/loongarch64/tcg-target-sa32.h | 2 +-
8
target/avr/helper.c | 9 +++------
9
tcg/loongarch64/tcg-target.c.inc | 15 ++++++---------
9
1 file changed, 3 insertions(+), 6 deletions(-)
10
2 files changed, 7 insertions(+), 10 deletions(-)
11
10
12
diff --git a/tcg/loongarch64/tcg-target-sa32.h b/tcg/loongarch64/tcg-target-sa32.h
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/loongarch64/tcg-target-sa32.h
13
--- a/target/avr/helper.c
15
+++ b/tcg/loongarch64/tcg-target-sa32.h
14
+++ b/target/avr/helper.c
16
@@ -1 +1 @@
15
@@ -XXX,XX +XXX,XX @@
17
-#define TCG_TARGET_SIGNED_ADDR32 0
16
18
+#define TCG_TARGET_SIGNED_ADDR32 1
17
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
19
diff --git a/tcg/loongarch64/tcg-target.c.inc b/tcg/loongarch64/tcg-target.c.inc
18
{
20
index XXXXXXX..XXXXXXX 100644
19
- bool ret = false;
21
--- a/tcg/loongarch64/tcg-target.c.inc
20
AVRCPU *cpu = AVR_CPU(cs);
22
+++ b/tcg/loongarch64/tcg-target.c.inc
21
CPUAVRState *env = &cpu->env;
23
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
22
24
return tcg_out_fail_alignment(s, l);
23
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
24
avr_cpu_do_interrupt(cs);
25
26
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
27
-
28
- ret = true;
29
+ return true;
30
}
31
}
32
if (interrupt_request & CPU_INTERRUPT_HARD) {
33
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
34
if (!env->intsrc) {
35
cs->interrupt_request &= ~CPU_INTERRUPT_HARD;
36
}
37
-
38
- ret = true;
39
+ return true;
40
}
41
}
42
- return ret;
43
+ return false;
25
}
44
}
26
45
27
-#endif /* CONFIG_SOFTMMU */
46
void avr_cpu_do_interrupt(CPUState *cs)
28
-
29
/*
30
* `ext32u` the address register into the temp register given,
31
* if target is 32-bit, no-op otherwise.
32
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
33
static TCGReg tcg_out_zext_addr_if_32_bit(TCGContext *s,
34
TCGReg addr, TCGReg tmp)
35
{
36
- if (TARGET_LONG_BITS == 32) {
37
+ if (TARGET_LONG_BITS == 32 && !guest_base_signed_addr32) {
38
tcg_out_ext32u(s, tmp, addr);
39
return tmp;
40
}
41
return addr;
42
}
43
+#endif /* CONFIG_SOFTMMU */
44
45
static void tcg_out_qemu_ld_indexed(TCGContext *s, TCGReg rd, TCGReg rj,
46
TCGReg rk, MemOp opc, TCGType type)
47
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, TCGType type)
48
tcg_insn_unit *label_ptr[1];
49
#else
50
unsigned a_bits;
51
-#endif
52
TCGReg base;
53
+#endif
54
55
data_regl = *args++;
56
addr_regl = *args++;
57
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, TCGType type)
58
59
#if defined(CONFIG_SOFTMMU)
60
tcg_out_tlb_load(s, addr_regl, oi, label_ptr, 1);
61
- base = tcg_out_zext_addr_if_32_bit(s, addr_regl, TCG_REG_TMP0);
62
- tcg_out_qemu_ld_indexed(s, data_regl, base, TCG_REG_TMP2, opc, type);
63
+ tcg_out_qemu_ld_indexed(s, data_regl, addr_regl, TCG_REG_TMP2, opc, type);
64
add_qemu_ldst_label(s, 1, oi, type,
65
data_regl, addr_regl,
66
s->code_ptr, label_ptr);
67
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args)
68
tcg_insn_unit *label_ptr[1];
69
#else
70
unsigned a_bits;
71
-#endif
72
TCGReg base;
73
+#endif
74
75
data_regl = *args++;
76
addr_regl = *args++;
77
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args)
78
79
#if defined(CONFIG_SOFTMMU)
80
tcg_out_tlb_load(s, addr_regl, oi, label_ptr, 0);
81
- base = tcg_out_zext_addr_if_32_bit(s, addr_regl, TCG_REG_TMP0);
82
- tcg_out_qemu_st_indexed(s, data_regl, base, TCG_REG_TMP2, opc);
83
+ tcg_out_qemu_st_indexed(s, data_regl, addr_regl, TCG_REG_TMP2, opc);
84
add_qemu_ldst_label(s, 0, oi,
85
0, /* type param is unused for stores */
86
data_regl, addr_regl,
87
--
47
--
88
2.25.1
48
2.34.1
89
49
90
50
diff view generated by jsdifflib
1
While there are no specific 16-bit rotate instructions, there
1
This bit is not saved across interrupts, so we must
2
are double-word shifts, which can perform the same operation.
2
delay delivering the interrupt until the skip has
3
been processed.
3
4
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1118
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
9
---
8
tcg/i386/tcg-target.c.inc | 18 +++++++++++++++++-
10
target/avr/helper.c | 9 +++++++++
9
1 file changed, 17 insertions(+), 1 deletion(-)
11
target/avr/translate.c | 26 ++++++++++++++++++++++----
12
2 files changed, 31 insertions(+), 4 deletions(-)
10
13
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
14
diff --git a/target/avr/helper.c b/target/avr/helper.c
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
16
--- a/target/avr/helper.c
14
+++ b/tcg/i386/tcg-target.c.inc
17
+++ b/target/avr/helper.c
15
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
18
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
16
case INDEX_op_rotlv_vec:
19
AVRCPU *cpu = AVR_CPU(cs);
17
case INDEX_op_rotrv_vec:
20
CPUAVRState *env = &cpu->env;
18
switch (vece) {
21
19
+ case MO_16:
22
+ /*
20
+ return have_avx512vbmi2 ? -1 : 0;
23
+ * We cannot separate a skip from the next instruction,
21
case MO_32:
24
+ * as the skip would not be preserved across the interrupt.
22
case MO_64:
25
+ * Separating the two insn normally only happens at page boundaries.
23
return have_avx512vl ? 1 : have_avx2 ? -1 : 0;
26
+ */
24
@@ -XXX,XX +XXX,XX @@ static void expand_vec_rotli(TCGType type, unsigned vece,
27
+ if (env->skip) {
25
return;
28
+ return false;
26
}
27
28
+ if (have_avx512vbmi2) {
29
+ vec_gen_4(INDEX_op_x86_vpshldi_vec, type, vece,
30
+ tcgv_vec_arg(v0), tcgv_vec_arg(v1), tcgv_vec_arg(v1), imm);
31
+ return;
32
+ }
29
+ }
33
+
30
+
34
t = tcg_temp_new_vec(type);
31
if (interrupt_request & CPU_INTERRUPT_RESET) {
35
tcg_gen_shli_vec(vece, t, v1, imm);
32
if (cpu_interrupts_enabled(env)) {
36
tcg_gen_shri_vec(vece, v0, v1, (8 << vece) - imm);
33
cs->exception_index = EXCP_RESET;
37
@@ -XXX,XX +XXX,XX @@ static void expand_vec_rotls(TCGType type, unsigned vece,
34
diff --git a/target/avr/translate.c b/target/avr/translate.c
38
static void expand_vec_rotv(TCGType type, unsigned vece, TCGv_vec v0,
35
index XXXXXXX..XXXXXXX 100644
39
TCGv_vec v1, TCGv_vec sh, bool right)
36
--- a/target/avr/translate.c
37
+++ b/target/avr/translate.c
38
@@ -XXX,XX +XXX,XX @@ static void avr_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
39
if (skip_label) {
40
canonicalize_skip(ctx);
41
gen_set_label(skip_label);
42
- if (ctx->base.is_jmp == DISAS_NORETURN) {
43
+
44
+ switch (ctx->base.is_jmp) {
45
+ case DISAS_NORETURN:
46
ctx->base.is_jmp = DISAS_CHAIN;
47
+ break;
48
+ case DISAS_NEXT:
49
+ if (ctx->base.tb->flags & TB_FLAGS_SKIP) {
50
+ ctx->base.is_jmp = DISAS_TOO_MANY;
51
+ }
52
+ break;
53
+ default:
54
+ break;
55
}
56
}
57
58
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
40
{
59
{
41
- TCGv_vec t = tcg_temp_new_vec(type);
60
DisasContext *ctx = container_of(dcbase, DisasContext, base);
42
+ TCGv_vec t;
61
bool nonconst_skip = canonicalize_skip(ctx);
43
62
+ /*
44
+ if (have_avx512vbmi2) {
63
+ * Because we disable interrupts while env->skip is set,
45
+ vec_gen_4(right ? INDEX_op_x86_vpshrdv_vec : INDEX_op_x86_vpshldv_vec,
64
+ * we must return to the main loop to re-evaluate afterward.
46
+ type, vece, tcgv_vec_arg(v0), tcgv_vec_arg(v1),
65
+ */
47
+ tcgv_vec_arg(v1), tcgv_vec_arg(sh));
66
+ bool force_exit = ctx->base.tb->flags & TB_FLAGS_SKIP;
48
+ return;
67
49
+ }
68
switch (ctx->base.is_jmp) {
50
+
69
case DISAS_NORETURN:
51
+ t = tcg_temp_new_vec(type);
70
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
52
tcg_gen_dupi_vec(vece, t, 8 << vece);
71
case DISAS_NEXT:
53
tcg_gen_sub_vec(vece, t, t, sh);
72
case DISAS_TOO_MANY:
54
if (right) {
73
case DISAS_CHAIN:
74
- if (!nonconst_skip) {
75
+ if (!nonconst_skip && !force_exit) {
76
/* Note gen_goto_tb checks singlestep. */
77
gen_goto_tb(ctx, 1, ctx->npc);
78
break;
79
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
80
tcg_gen_movi_tl(cpu_pc, ctx->npc);
81
/* fall through */
82
case DISAS_LOOKUP:
83
- tcg_gen_lookup_and_goto_ptr();
84
- break;
85
+ if (!force_exit) {
86
+ tcg_gen_lookup_and_goto_ptr();
87
+ break;
88
+ }
89
+ /* fall through */
90
case DISAS_EXIT:
91
tcg_gen_exit_tb(NULL, 0);
92
break;
55
--
93
--
56
2.25.1
94
2.34.1
57
95
58
96
diff view generated by jsdifflib
1
When using reserved_va, which is the default for a 64-bit host
1
Map the stack executable if required by default or on demand.
2
and a 32-bit guest, set guest_base_signed_addr32 if requested
3
by TCG_TARGET_SIGNED_ADDR32, and the executable layout allows.
4
2
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
3
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
6
---
9
include/exec/cpu-all.h | 4 ---
7
include/elf.h | 1 +
10
linux-user/elfload.c | 62 ++++++++++++++++++++++++++++++++++--------
8
linux-user/qemu.h | 1 +
11
2 files changed, 50 insertions(+), 16 deletions(-)
9
linux-user/elfload.c | 19 ++++++++++++++++++-
10
3 files changed, 20 insertions(+), 1 deletion(-)
12
11
13
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
12
diff --git a/include/elf.h b/include/elf.h
14
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
15
--- a/include/exec/cpu-all.h
14
--- a/include/elf.h
16
+++ b/include/exec/cpu-all.h
15
+++ b/include/elf.h
17
@@ -XXX,XX +XXX,XX @@ extern const TargetPageBits target_page;
16
@@ -XXX,XX +XXX,XX @@ typedef int64_t Elf64_Sxword;
18
#define PAGE_RESET 0x0040
17
#define PT_LOPROC 0x70000000
19
/* For linux-user, indicates that the page is MAP_ANON. */
18
#define PT_HIPROC 0x7fffffff
20
#define PAGE_ANON 0x0080
19
21
-
20
+#define PT_GNU_STACK (PT_LOOS + 0x474e551)
22
-#if defined(CONFIG_BSD) && defined(CONFIG_USER_ONLY)
21
#define PT_GNU_PROPERTY (PT_LOOS + 0x474e553)
23
-/* FIXME: Code that sets/uses this is broken and needs to go away. */
22
24
#define PAGE_RESERVED 0x0100
23
#define PT_MIPS_REGINFO 0x70000000
25
-#endif
24
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
26
/* Target-specific bits that will be used via page_get_flags(). */
25
index XXXXXXX..XXXXXXX 100644
27
#define PAGE_TARGET_1 0x0200
26
--- a/linux-user/qemu.h
28
#define PAGE_TARGET_2 0x0400
27
+++ b/linux-user/qemu.h
28
@@ -XXX,XX +XXX,XX @@ struct image_info {
29
uint32_t elf_flags;
30
int personality;
31
abi_ulong alignment;
32
+ bool exec_stack;
33
34
/* Generic semihosting knows about these pointers. */
35
abi_ulong arg_strings; /* strings for argv */
29
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
36
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
30
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
31
--- a/linux-user/elfload.c
38
--- a/linux-user/elfload.c
32
+++ b/linux-user/elfload.c
39
+++ b/linux-user/elfload.c
33
@@ -XXX,XX +XXX,XX @@ static void pgb_dynamic(const char *image_name, long align)
40
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
34
static void pgb_reserved_va(const char *image_name, abi_ulong guest_loaddr,
41
#define ELF_ARCH EM_386
35
abi_ulong guest_hiaddr, long align)
42
43
#define ELF_PLATFORM get_elf_platform()
44
+#define EXSTACK_DEFAULT true
45
46
static const char *get_elf_platform(void)
36
{
47
{
37
- int flags = MAP_ANONYMOUS | MAP_PRIVATE | MAP_NORESERVE;
48
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
38
+ int flags = (MAP_ANONYMOUS | MAP_PRIVATE |
49
39
+ MAP_NORESERVE | MAP_FIXED_NOREPLACE);
50
#define ELF_ARCH EM_ARM
40
+ unsigned long local_rva = reserved_va;
51
#define ELF_CLASS ELFCLASS32
41
+ bool protect_wrap = false;
52
+#define EXSTACK_DEFAULT true
42
void *addr, *test;
53
43
54
static inline void init_thread(struct target_pt_regs *regs,
44
- if (guest_hiaddr > reserved_va) {
55
struct image_info *infop)
45
+ if (guest_hiaddr > local_rva) {
56
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
46
error_report("%s: requires more than reserved virtual "
57
#else
47
"address space (0x%" PRIx64 " > 0x%lx)",
58
48
- image_name, (uint64_t)guest_hiaddr, reserved_va);
59
#define ELF_CLASS ELFCLASS32
49
+ image_name, (uint64_t)guest_hiaddr, local_rva);
60
+#define EXSTACK_DEFAULT true
50
exit(EXIT_FAILURE);
61
62
#endif
63
64
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUPPCState *en
65
66
#define ELF_CLASS ELFCLASS64
67
#define ELF_ARCH EM_LOONGARCH
68
+#define EXSTACK_DEFAULT true
69
70
#define elf_check_arch(x) ((x) == EM_LOONGARCH)
71
72
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
73
#define ELF_CLASS ELFCLASS32
74
#endif
75
#define ELF_ARCH EM_MIPS
76
+#define EXSTACK_DEFAULT true
77
78
#ifdef TARGET_ABI_MIPSN32
79
#define elf_check_abi(x) ((x) & EF_MIPS_ABI2)
80
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
81
#define bswaptls(ptr) bswap32s(ptr)
82
#endif
83
84
+#ifndef EXSTACK_DEFAULT
85
+#define EXSTACK_DEFAULT false
86
+#endif
87
+
88
#include "elf.h"
89
90
/* We must delay the following stanzas until after "elf.h". */
91
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
92
struct image_info *info)
93
{
94
abi_ulong size, error, guard;
95
+ int prot;
96
97
size = guest_stack_size;
98
if (size < STACK_LOWER_LIMIT) {
99
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
100
guard = qemu_real_host_page_size();
51
}
101
}
52
102
53
- /* Widen the "image" to the entire reserved address space. */
103
- error = target_mmap(0, size + guard, PROT_READ | PROT_WRITE,
54
- pgb_static(image_name, 0, reserved_va, align);
104
+ prot = PROT_READ | PROT_WRITE;
55
+ if (TCG_TARGET_SIGNED_ADDR32 && TARGET_LONG_BITS == 32) {
105
+ if (info->exec_stack) {
56
+ if (guest_loaddr < 0x80000000u && guest_hiaddr > 0x80000000u) {
106
+ prot |= PROT_EXEC;
57
+ /*
58
+ * The executable itself wraps on signed addresses.
59
+ * Without per-page translation, we must keep the
60
+ * guest address 0x7fff_ffff adjacent to 0x8000_0000
61
+ * consecutive in host memory: unsigned addresses.
62
+ */
63
+ } else {
64
+ set_guest_base_signed_addr32();
65
+ if (local_rva <= 0x80000000u) {
66
+ /* No guest addresses are "negative": win! */
67
+ } else {
68
+ /* Begin by allocating the entire address space. */
69
+ local_rva = 0xfffffffful + 1;
70
+ protect_wrap = true;
71
+ }
72
+ }
73
+ }
107
+ }
74
108
+ error = target_mmap(0, size + guard, prot,
75
- /* osdep.h defines this as 0 if it's missing */
109
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
76
- flags |= MAP_FIXED_NOREPLACE;
110
if (error == -1) {
77
+ /* Widen the "image" to the entire reserved address space. */
111
perror("mmap stack");
78
+ pgb_static(image_name, 0, local_rva, align);
112
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
79
+ assert(guest_base != 0);
113
*/
80
114
loaddr = -1, hiaddr = 0;
81
/* Reserve the memory on the host. */
115
info->alignment = 0;
82
- assert(guest_base != 0);
116
+ info->exec_stack = EXSTACK_DEFAULT;
83
test = g2h_untagged(0);
117
for (i = 0; i < ehdr->e_phnum; ++i) {
84
- addr = mmap(test, reserved_va, PROT_NONE, flags, -1, 0);
118
struct elf_phdr *eppnt = phdr + i;
85
+ addr = mmap(test, local_rva, PROT_NONE, flags, -1, 0);
119
if (eppnt->p_type == PT_LOAD) {
86
if (addr == MAP_FAILED || addr != test) {
120
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
87
+ /*
121
if (!parse_elf_properties(image_fd, info, eppnt, bprm_buf, &err)) {
88
+ * If protect_wrap, we could try again with the original reserved_va
122
goto exit_errmsg;
89
+ * setting, but the edge case of low ulimit vm setting on a 64-bit
123
}
90
+ * host is probably useless.
124
+ } else if (eppnt->p_type == PT_GNU_STACK) {
91
+ */
125
+ info->exec_stack = eppnt->p_flags & PF_X;
92
error_report("Unable to reserve 0x%lx bytes of virtual address "
126
}
93
- "space at %p (%s) for use as guest address space (check your"
94
- "virtual memory ulimit setting, min_mmap_addr or reserve less "
95
- "using -R option)", reserved_va, test, strerror(errno));
96
+ "space at %p (%s) for use as guest address space "
97
+ "(check your virtual memory ulimit setting, "
98
+ "min_mmap_addr or reserve less using -R option)",
99
+ local_rva, test, strerror(errno));
100
exit(EXIT_FAILURE);
101
}
127
}
102
128
103
+ if (protect_wrap) {
104
+ /*
105
+ * Prevent the page just before 0x80000000 from being allocated.
106
+ * This prevents a single guest object/allocation from crossing
107
+ * the signed wrap, and thus being discontiguous in host memory.
108
+ */
109
+ page_set_flags(0x7fffffff & TARGET_PAGE_MASK, 0x80000000u,
110
+ PAGE_RESERVED);
111
+ /* Adjust guest_base so that 0 is in the middle of the reservation. */
112
+ guest_base += 0x80000000ul;
113
+ }
114
+
115
qemu_log_mask(CPU_LOG_PAGE, "%s: base @ %p for %lu bytes\n",
116
__func__, addr, reserved_va);
117
}
118
--
129
--
119
2.25.1
130
2.34.1
120
121
diff view generated by jsdifflib
1
AVX512VL has a general ternary logic operation, VPTERNLOGQ,
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
which can implement NOT, ORC, NAND, NOR, EQV.
3
2
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
3
Currently it's possible to execute pages that do not have PAGE_EXEC
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
if there is an existing translation block. Fix by invalidating TBs
5
that touch the affected pages.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Message-Id: <20220817150506.592862-2-iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
10
---
8
tcg/i386/tcg-target.h | 10 +++++-----
11
linux-user/mmap.c | 6 ++++--
9
tcg/i386/tcg-target.c.inc | 34 ++++++++++++++++++++++++++++++++++
12
1 file changed, 4 insertions(+), 2 deletions(-)
10
2 files changed, 39 insertions(+), 5 deletions(-)
11
13
12
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
14
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/i386/tcg-target.h
16
--- a/linux-user/mmap.c
15
+++ b/tcg/i386/tcg-target.h
17
+++ b/linux-user/mmap.c
16
@@ -XXX,XX +XXX,XX @@ extern bool have_movbe;
18
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
17
#define TCG_TARGET_HAS_v256 have_avx2
19
goto error;
18
20
}
19
#define TCG_TARGET_HAS_andc_vec 1
21
}
20
-#define TCG_TARGET_HAS_orc_vec 0
21
-#define TCG_TARGET_HAS_nand_vec 0
22
-#define TCG_TARGET_HAS_nor_vec 0
23
-#define TCG_TARGET_HAS_eqv_vec 0
24
-#define TCG_TARGET_HAS_not_vec 0
25
+#define TCG_TARGET_HAS_orc_vec have_avx512vl
26
+#define TCG_TARGET_HAS_nand_vec have_avx512vl
27
+#define TCG_TARGET_HAS_nor_vec have_avx512vl
28
+#define TCG_TARGET_HAS_eqv_vec have_avx512vl
29
+#define TCG_TARGET_HAS_not_vec have_avx512vl
30
#define TCG_TARGET_HAS_neg_vec 0
31
#define TCG_TARGET_HAS_abs_vec 1
32
#define TCG_TARGET_HAS_roti_vec have_avx512vl
33
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
34
index XXXXXXX..XXXXXXX 100644
35
--- a/tcg/i386/tcg-target.c.inc
36
+++ b/tcg/i386/tcg-target.c.inc
37
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
38
#define OPC_VPSRLVW (0x10 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
39
#define OPC_VPSRLVD (0x45 | P_EXT38 | P_DATA16)
40
#define OPC_VPSRLVQ (0x45 | P_EXT38 | P_DATA16 | P_VEXW)
41
+#define OPC_VPTERNLOGQ (0x25 | P_EXT3A | P_DATA16 | P_VEXW | P_EVEX)
42
#define OPC_VZEROUPPER (0x77 | P_EXT)
43
#define OPC_XCHG_ax_r32    (0x90)
44
45
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
46
insn = vpshldi_insn[vece];
47
sub = args[3];
48
goto gen_simd_imm8;
49
+
22
+
50
+ case INDEX_op_not_vec:
23
page_set_flags(start, start + len, page_flags);
51
+ insn = OPC_VPTERNLOGQ;
24
- mmap_unlock();
52
+ a2 = a1;
25
- return 0;
53
+ sub = 0x33; /* !B */
26
+ tb_invalidate_phys_range(start, start + len);
54
+ goto gen_simd_imm8;
27
+ ret = 0;
55
+ case INDEX_op_nor_vec:
56
+ insn = OPC_VPTERNLOGQ;
57
+ sub = 0x11; /* norCB */
58
+ goto gen_simd_imm8;
59
+ case INDEX_op_nand_vec:
60
+ insn = OPC_VPTERNLOGQ;
61
+ sub = 0x77; /* nandCB */
62
+ goto gen_simd_imm8;
63
+ case INDEX_op_eqv_vec:
64
+ insn = OPC_VPTERNLOGQ;
65
+ sub = 0x99; /* xnorCB */
66
+ goto gen_simd_imm8;
67
+ case INDEX_op_orc_vec:
68
+ insn = OPC_VPTERNLOGQ;
69
+ sub = 0xdd; /* orB!C */
70
+ goto gen_simd_imm8;
71
+
28
+
72
gen_simd_imm8:
29
error:
73
tcg_debug_assert(insn != OPC_UD2);
30
mmap_unlock();
74
if (type == TCG_TYPE_V256) {
31
return ret;
75
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
76
case INDEX_op_or_vec:
77
case INDEX_op_xor_vec:
78
case INDEX_op_andc_vec:
79
+ case INDEX_op_orc_vec:
80
+ case INDEX_op_nand_vec:
81
+ case INDEX_op_nor_vec:
82
+ case INDEX_op_eqv_vec:
83
case INDEX_op_ssadd_vec:
84
case INDEX_op_usadd_vec:
85
case INDEX_op_sssub_vec:
86
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
87
88
case INDEX_op_abs_vec:
89
case INDEX_op_dup_vec:
90
+ case INDEX_op_not_vec:
91
case INDEX_op_shli_vec:
92
case INDEX_op_shri_vec:
93
case INDEX_op_sari_vec:
94
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
95
case INDEX_op_or_vec:
96
case INDEX_op_xor_vec:
97
case INDEX_op_andc_vec:
98
+ case INDEX_op_orc_vec:
99
+ case INDEX_op_nand_vec:
100
+ case INDEX_op_nor_vec:
101
+ case INDEX_op_eqv_vec:
102
+ case INDEX_op_not_vec:
103
return 1;
104
case INDEX_op_cmp_vec:
105
case INDEX_op_cmpsel_vec:
106
--
32
--
107
2.25.1
33
2.34.1
108
109
diff view generated by jsdifflib
1
From: Ziqiao Kong <ziqiaokong@gmail.com>
1
We're about to start validating PAGE_EXEC, which means
2
that we've got to put this code into a section that is
3
both writable and executable.
2
4
3
The last entry of DEF_HELPERS_FLAGS_n is DEF_HELPER_FLAGS_7 and
5
Note that this test did not run on hardware beforehand either.
4
thus the MAX_OPC_PARAM_IARGS should be 7.
5
6
6
Reviewed-by: Taylor Simpson <tsimpson@quicinc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Ziqiao Kong <ziqiaokong@gmail.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Message-Id: <20220227113127.414533-2-ziqiaokong@gmail.com>
9
Fixes: e6cadf49c3d ("tcg: Add support for a helper with 7 arguments")
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
10
---
12
include/tcg/tcg.h | 2 +-
11
tests/tcg/i386/test-i386.c | 2 +-
13
tcg/tci/tcg-target.c.inc | 2 +-
12
1 file changed, 1 insertion(+), 1 deletion(-)
14
2 files changed, 2 insertions(+), 2 deletions(-)
15
13
16
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
14
diff --git a/tests/tcg/i386/test-i386.c b/tests/tcg/i386/test-i386.c
17
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
18
--- a/include/tcg/tcg.h
16
--- a/tests/tcg/i386/test-i386.c
19
+++ b/include/tcg/tcg.h
17
+++ b/tests/tcg/i386/test-i386.c
20
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ uint8_t code[] = {
21
#else
19
0xc3, /* ret */
22
#define MAX_OPC_PARAM_PER_ARG 1
23
#endif
24
-#define MAX_OPC_PARAM_IARGS 6
25
+#define MAX_OPC_PARAM_IARGS 7
26
#define MAX_OPC_PARAM_OARGS 1
27
#define MAX_OPC_PARAM_ARGS (MAX_OPC_PARAM_IARGS + MAX_OPC_PARAM_OARGS)
28
29
diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc
30
index XXXXXXX..XXXXXXX 100644
31
--- a/tcg/tci/tcg-target.c.inc
32
+++ b/tcg/tci/tcg-target.c.inc
33
@@ -XXX,XX +XXX,XX @@ static const int tcg_target_reg_alloc_order[] = {
34
TCG_REG_R0,
35
};
20
};
36
21
37
-#if MAX_OPC_PARAM_IARGS != 6
22
-asm(".section \".data\"\n"
38
+#if MAX_OPC_PARAM_IARGS != 7
23
+asm(".section \".data_x\",\"awx\"\n"
39
# error Fix needed, number of supported input arguments changed!
24
"smc_code2:\n"
40
#endif
25
"movl 4(%esp), %eax\n"
41
26
"movl %eax, smc_patch_addr2 + 1\n"
42
--
27
--
43
2.25.1
28
2.34.1
diff view generated by jsdifflib
1
AVX512VL has VPABSQ, VPMAXSQ, VPMAXUQ, VPMINSQ, VPMINUQ.
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
2
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
3
Introduce a function that checks whether a given address is on the same
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
page as where disassembly started. Having it improves readability of
5
the following patches.
6
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Message-Id: <20220811095534.241224-3-iii@linux.ibm.com>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
[rth: Make the DisasContextBase parameter const.]
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
13
---
7
tcg/i386/tcg-target.c.inc | 18 +++++++++++-------
14
include/exec/translator.h | 10 ++++++++++
8
1 file changed, 11 insertions(+), 7 deletions(-)
15
1 file changed, 10 insertions(+)
9
16
10
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
17
diff --git a/include/exec/translator.h b/include/exec/translator.h
11
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.c.inc
19
--- a/include/exec/translator.h
13
+++ b/tcg/i386/tcg-target.c.inc
20
+++ b/include/exec/translator.h
14
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
21
@@ -XXX,XX +XXX,XX @@ FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
15
#define OPC_PABSB (0x1c | P_EXT38 | P_DATA16)
22
16
#define OPC_PABSW (0x1d | P_EXT38 | P_DATA16)
23
#undef GEN_TRANSLATOR_LD
17
#define OPC_PABSD (0x1e | P_EXT38 | P_DATA16)
24
18
+#define OPC_VPABSQ (0x1f | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
25
+/*
19
#define OPC_PACKSSDW (0x6b | P_EXT | P_DATA16)
26
+ * Return whether addr is on the same page as where disassembly started.
20
#define OPC_PACKSSWB (0x63 | P_EXT | P_DATA16)
27
+ * Translators can use this to enforce the rule that only single-insn
21
#define OPC_PACKUSDW (0x2b | P_EXT38 | P_DATA16)
28
+ * translation blocks are allowed to cross page boundaries.
22
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
29
+ */
23
#define OPC_PMAXSB (0x3c | P_EXT38 | P_DATA16)
30
+static inline bool is_same_page(const DisasContextBase *db, target_ulong addr)
24
#define OPC_PMAXSW (0xee | P_EXT | P_DATA16)
31
+{
25
#define OPC_PMAXSD (0x3d | P_EXT38 | P_DATA16)
32
+ return ((addr ^ db->pc_first) & TARGET_PAGE_MASK) == 0;
26
+#define OPC_VPMAXSQ (0x3d | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
33
+}
27
#define OPC_PMAXUB (0xde | P_EXT | P_DATA16)
34
+
28
#define OPC_PMAXUW (0x3e | P_EXT38 | P_DATA16)
35
#endif /* EXEC__TRANSLATOR_H */
29
#define OPC_PMAXUD (0x3f | P_EXT38 | P_DATA16)
30
+#define OPC_VPMAXUQ (0x3f | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
31
#define OPC_PMINSB (0x38 | P_EXT38 | P_DATA16)
32
#define OPC_PMINSW (0xea | P_EXT | P_DATA16)
33
#define OPC_PMINSD (0x39 | P_EXT38 | P_DATA16)
34
+#define OPC_VPMINSQ (0x39 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
35
#define OPC_PMINUB (0xda | P_EXT | P_DATA16)
36
#define OPC_PMINUW (0x3a | P_EXT38 | P_DATA16)
37
#define OPC_PMINUD (0x3b | P_EXT38 | P_DATA16)
38
+#define OPC_VPMINUQ (0x3b | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
39
#define OPC_PMOVSXBW (0x20 | P_EXT38 | P_DATA16)
40
#define OPC_PMOVSXWD (0x23 | P_EXT38 | P_DATA16)
41
#define OPC_PMOVSXDQ (0x25 | P_EXT38 | P_DATA16)
42
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
43
OPC_PACKUSWB, OPC_PACKUSDW, OPC_UD2, OPC_UD2
44
};
45
static int const smin_insn[4] = {
46
- OPC_PMINSB, OPC_PMINSW, OPC_PMINSD, OPC_UD2
47
+ OPC_PMINSB, OPC_PMINSW, OPC_PMINSD, OPC_VPMINSQ
48
};
49
static int const smax_insn[4] = {
50
- OPC_PMAXSB, OPC_PMAXSW, OPC_PMAXSD, OPC_UD2
51
+ OPC_PMAXSB, OPC_PMAXSW, OPC_PMAXSD, OPC_VPMAXSQ
52
};
53
static int const umin_insn[4] = {
54
- OPC_PMINUB, OPC_PMINUW, OPC_PMINUD, OPC_UD2
55
+ OPC_PMINUB, OPC_PMINUW, OPC_PMINUD, OPC_VPMINUQ
56
};
57
static int const umax_insn[4] = {
58
- OPC_PMAXUB, OPC_PMAXUW, OPC_PMAXUD, OPC_UD2
59
+ OPC_PMAXUB, OPC_PMAXUW, OPC_PMAXUD, OPC_VPMAXUQ
60
};
61
static int const rotlv_insn[4] = {
62
OPC_UD2, OPC_UD2, OPC_VPROLVD, OPC_VPROLVQ
63
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
64
OPC_UD2, OPC_VPSHRDVW, OPC_VPSHRDVD, OPC_VPSHRDVQ
65
};
66
static int const abs_insn[4] = {
67
- /* TODO: AVX512 adds support for MO_64. */
68
- OPC_PABSB, OPC_PABSW, OPC_PABSD, OPC_UD2
69
+ OPC_PABSB, OPC_PABSW, OPC_PABSD, OPC_VPABSQ
70
};
71
72
TCGType type = vecl + TCG_TYPE_V64;
73
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
74
case INDEX_op_umin_vec:
75
case INDEX_op_umax_vec:
76
case INDEX_op_abs_vec:
77
- return vece <= MO_32;
78
+ return vece <= MO_32 || have_avx512vl;
79
80
default:
81
return 0;
82
--
36
--
83
2.25.1
37
2.34.1
84
85
diff view generated by jsdifflib
1
When TCG_TARGET_SIGNED_ADDR32 is set, adjust the tlb addend to
1
The current implementation is a no-op, simply returning addr.
2
allow the 32-bit guest address to be sign extended within the
2
This is incorrect, because we ought to be checking the page
3
64-bit host register instead of zero extended.
3
permissions for execution.
4
4
5
This will simplify tcg hosts like MIPS, RISC-V, and LoongArch,
5
Make get_page_addr_code inline for both implementations.
6
which naturally sign-extend 32-bit values, in contrast to x86_64
7
and AArch64 which zero-extend them.
8
6
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
10
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Acked-by: Alistair Francis <alistair.francis@wdc.com>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
---
11
---
13
accel/tcg/cputlb.c | 12 +++++++++++-
12
include/exec/exec-all.h | 85 ++++++++++++++---------------------------
14
1 file changed, 11 insertions(+), 1 deletion(-)
13
accel/tcg/cputlb.c | 5 ---
14
accel/tcg/user-exec.c | 14 +++++++
15
3 files changed, 42 insertions(+), 62 deletions(-)
15
16
17
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/exec/exec-all.h
20
+++ b/include/exec/exec-all.h
21
@@ -XXX,XX +XXX,XX @@ struct MemoryRegionSection *iotlb_to_section(CPUState *cpu,
22
hwaddr index, MemTxAttrs attrs);
23
#endif
24
25
-#if defined(CONFIG_USER_ONLY)
26
-void mmap_lock(void);
27
-void mmap_unlock(void);
28
-bool have_mmap_lock(void);
29
-
30
/**
31
- * get_page_addr_code() - user-mode version
32
+ * get_page_addr_code_hostp()
33
* @env: CPUArchState
34
* @addr: guest virtual address of guest code
35
*
36
- * Returns @addr.
37
+ * See get_page_addr_code() (full-system version) for documentation on the
38
+ * return value.
39
+ *
40
+ * Sets *@hostp (when @hostp is non-NULL) as follows.
41
+ * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
42
+ * to the host address where @addr's content is kept.
43
+ *
44
+ * Note: this function can trigger an exception.
45
+ */
46
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
47
+ void **hostp);
48
+
49
+/**
50
+ * get_page_addr_code()
51
+ * @env: CPUArchState
52
+ * @addr: guest virtual address of guest code
53
+ *
54
+ * If we cannot translate and execute from the entire RAM page, or if
55
+ * the region is not backed by RAM, returns -1. Otherwise, returns the
56
+ * ram_addr_t corresponding to the guest code at @addr.
57
+ *
58
+ * Note: this function can trigger an exception.
59
*/
60
static inline tb_page_addr_t get_page_addr_code(CPUArchState *env,
61
target_ulong addr)
62
{
63
- return addr;
64
+ return get_page_addr_code_hostp(env, addr, NULL);
65
}
66
67
-/**
68
- * get_page_addr_code_hostp() - user-mode version
69
- * @env: CPUArchState
70
- * @addr: guest virtual address of guest code
71
- *
72
- * Returns @addr.
73
- *
74
- * If @hostp is non-NULL, sets *@hostp to the host address where @addr's content
75
- * is kept.
76
- */
77
-static inline tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env,
78
- target_ulong addr,
79
- void **hostp)
80
-{
81
- if (hostp) {
82
- *hostp = g2h_untagged(addr);
83
- }
84
- return addr;
85
-}
86
+#if defined(CONFIG_USER_ONLY)
87
+void mmap_lock(void);
88
+void mmap_unlock(void);
89
+bool have_mmap_lock(void);
90
91
/**
92
* adjust_signal_pc:
93
@@ -XXX,XX +XXX,XX @@ G_NORETURN void cpu_loop_exit_sigbus(CPUState *cpu, target_ulong addr,
94
static inline void mmap_lock(void) {}
95
static inline void mmap_unlock(void) {}
96
97
-/**
98
- * get_page_addr_code() - full-system version
99
- * @env: CPUArchState
100
- * @addr: guest virtual address of guest code
101
- *
102
- * If we cannot translate and execute from the entire RAM page, or if
103
- * the region is not backed by RAM, returns -1. Otherwise, returns the
104
- * ram_addr_t corresponding to the guest code at @addr.
105
- *
106
- * Note: this function can trigger an exception.
107
- */
108
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr);
109
-
110
-/**
111
- * get_page_addr_code_hostp() - full-system version
112
- * @env: CPUArchState
113
- * @addr: guest virtual address of guest code
114
- *
115
- * See get_page_addr_code() (full-system version) for documentation on the
116
- * return value.
117
- *
118
- * Sets *@hostp (when @hostp is non-NULL) as follows.
119
- * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
120
- * to the host address where @addr's content is kept.
121
- *
122
- * Note: this function can trigger an exception.
123
- */
124
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
125
- void **hostp);
126
-
127
void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length);
128
void tlb_set_dirty(CPUState *cpu, target_ulong vaddr);
129
16
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
130
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
17
index XXXXXXX..XXXXXXX 100644
131
index XXXXXXX..XXXXXXX 100644
18
--- a/accel/tcg/cputlb.c
132
--- a/accel/tcg/cputlb.c
19
+++ b/accel/tcg/cputlb.c
133
+++ b/accel/tcg/cputlb.c
20
@@ -XXX,XX +XXX,XX @@
134
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
21
#include "qemu/plugin-memory.h"
135
return qemu_ram_addr_from_host_nofail(p);
22
#endif
136
}
23
#include "tcg/tcg-ldst.h"
137
24
+#include "tcg-target-sa32.h"
138
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr)
25
139
-{
26
/* DEBUG defines, enable DEBUG_TLB_LOG to log to the CPU_LOG_MMU target */
140
- return get_page_addr_code_hostp(env, addr, NULL);
27
/* #define DEBUG_TLB */
141
-}
28
@@ -XXX,XX +XXX,XX @@ static inline size_t sizeof_tlb(CPUTLBDescFast *fast)
142
-
29
143
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
30
static inline uintptr_t g2h_tlbe(const CPUTLBEntry *tlb, target_ulong gaddr)
144
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
31
{
145
{
32
+ if (TCG_TARGET_SIGNED_ADDR32 && TARGET_LONG_BITS == 32) {
146
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
33
+ return tlb->addend + (int32_t)gaddr;
147
index XXXXXXX..XXXXXXX 100644
148
--- a/accel/tcg/user-exec.c
149
+++ b/accel/tcg/user-exec.c
150
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
151
return size ? g2h(env_cpu(env), addr) : NULL;
152
}
153
154
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
155
+ void **hostp)
156
+{
157
+ int flags;
158
+
159
+ flags = probe_access_internal(env, addr, 1, MMU_INST_FETCH, false, 0);
160
+ g_assert(flags == 0);
161
+
162
+ if (hostp) {
163
+ *hostp = g2h_untagged(addr);
34
+ }
164
+ }
35
return tlb->addend + (uintptr_t)gaddr;
165
+ return addr;
36
}
166
+}
37
38
@@ -XXX,XX +XXX,XX @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
39
desc->iotlb[index].attrs = attrs;
40
41
/* Now calculate the new entry */
42
- tn.addend = addend - vaddr_page;
43
+
167
+
44
+ if (TCG_TARGET_SIGNED_ADDR32 && TARGET_LONG_BITS == 32) {
168
/* The softmmu versions of these helpers are in cputlb.c. */
45
+ tn.addend = addend - (int32_t)vaddr_page;
169
46
+ } else {
170
/*
47
+ tn.addend = addend - vaddr_page;
48
+ }
49
+
50
if (prot & PAGE_READ) {
51
tn.addr_read = address;
52
if (wp_flags & BP_MEM_READ) {
53
--
171
--
54
2.25.1
172
2.34.1
55
56
diff view generated by jsdifflib
1
All RV64 32-bit operations sign-extend the output, so we are easily
1
The mmap_lock is held around tb_gen_code. While the comment
2
able to keep TCG_TYPE_I32 values sign-extended in host registers.
2
is correct that the lock is dropped when tb_gen_code runs out
3
of memory, the lock is *not* dropped when an exception is
4
raised reading code for translation.
3
5
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Acked-by: Alistair Francis <alistair.francis@wdc.com>
5
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
10
---
8
tcg/riscv/tcg-target-sa32.h | 6 +++++-
11
accel/tcg/cpu-exec.c | 12 ++++++------
9
tcg/riscv/tcg-target.c.inc | 8 ++------
12
accel/tcg/user-exec.c | 3 ---
10
2 files changed, 7 insertions(+), 7 deletions(-)
13
2 files changed, 6 insertions(+), 9 deletions(-)
11
14
12
diff --git a/tcg/riscv/tcg-target-sa32.h b/tcg/riscv/tcg-target-sa32.h
15
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
13
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/riscv/tcg-target-sa32.h
17
--- a/accel/tcg/cpu-exec.c
15
+++ b/tcg/riscv/tcg-target-sa32.h
18
+++ b/accel/tcg/cpu-exec.c
16
@@ -1 +1,5 @@
19
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
17
-#define TCG_TARGET_SIGNED_ADDR32 0
20
cpu_tb_exec(cpu, tb, &tb_exit);
18
+/*
21
cpu_exec_exit(cpu);
19
+ * Do not set TCG_TARGET_SIGNED_ADDR32 for RV32;
22
} else {
20
+ * TCG expects this to only be set for 64-bit hosts.
23
- /*
21
+ */
24
- * The mmap_lock is dropped by tb_gen_code if it runs out of
22
+#define TCG_TARGET_SIGNED_ADDR32 (__riscv_xlen == 64)
25
- * memory.
23
diff --git a/tcg/riscv/tcg-target.c.inc b/tcg/riscv/tcg-target.c.inc
26
- */
27
#ifndef CONFIG_SOFTMMU
28
clear_helper_retaddr();
29
- tcg_debug_assert(!have_mmap_lock());
30
+ if (have_mmap_lock()) {
31
+ mmap_unlock();
32
+ }
33
#endif
34
if (qemu_mutex_iothread_locked()) {
35
qemu_mutex_unlock_iothread();
36
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
37
38
#ifndef CONFIG_SOFTMMU
39
clear_helper_retaddr();
40
- tcg_debug_assert(!have_mmap_lock());
41
+ if (have_mmap_lock()) {
42
+ mmap_unlock();
43
+ }
44
#endif
45
if (qemu_mutex_iothread_locked()) {
46
qemu_mutex_unlock_iothread();
47
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
24
index XXXXXXX..XXXXXXX 100644
48
index XXXXXXX..XXXXXXX 100644
25
--- a/tcg/riscv/tcg-target.c.inc
49
--- a/accel/tcg/user-exec.c
26
+++ b/tcg/riscv/tcg-target.c.inc
50
+++ b/accel/tcg/user-exec.c
27
@@ -XXX,XX +XXX,XX @@ static void tcg_out_tlb_load(TCGContext *s, TCGReg addrl,
51
@@ -XXX,XX +XXX,XX @@ MMUAccessType adjust_signal_pc(uintptr_t *pc, bool is_write)
28
tcg_out_opc_branch(s, OPC_BNE, TCG_REG_TMP0, TCG_REG_TMP1, 0);
52
* (and if the translator doesn't handle page boundaries correctly
29
53
* there's little we can do about that here). Therefore, do not
30
/* TLB Hit - translate address using addend. */
54
* trigger the unwinder.
31
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
55
- *
32
- tcg_out_ext32u(s, TCG_REG_TMP0, addrl);
56
- * Like tb_gen_code, release the memory lock before cpu_loop_exit.
33
- addrl = TCG_REG_TMP0;
57
*/
34
- }
58
- mmap_unlock();
35
tcg_out_opc_reg(s, OPC_ADD, TCG_REG_TMP0, TCG_REG_TMP2, addrl);
59
*pc = 0;
36
}
60
return MMU_INST_FETCH;
37
38
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, bool is_64)
39
data_regl, data_regh, addr_regl, addr_regh,
40
s->code_ptr, label_ptr);
41
#else
42
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
43
+ if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS && !guest_base_signed_addr32) {
44
tcg_out_ext32u(s, base, addr_regl);
45
addr_regl = base;
46
}
47
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, bool is_64)
48
data_regl, data_regh, addr_regl, addr_regh,
49
s->code_ptr, label_ptr);
50
#else
51
- if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS) {
52
+ if (TCG_TARGET_REG_BITS > TARGET_LONG_BITS && !guest_base_signed_addr32) {
53
tcg_out_ext32u(s, base, addr_regl);
54
addr_regl = base;
55
}
61
}
56
--
62
--
57
2.25.1
63
2.34.1
58
59
diff view generated by jsdifflib
1
There is no such instruction on x86, so we should
1
The function is not used outside of cpu-exec.c. Move it and
2
not be pretending it has arguments.
2
its subroutines up in the file, before the first use.
3
3
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
---
8
tcg/i386/tcg-target.c.inc | 1 -
9
include/exec/exec-all.h | 3 -
9
1 file changed, 1 deletion(-)
10
accel/tcg/cpu-exec.c | 122 ++++++++++++++++++++--------------------
11
2 files changed, 61 insertions(+), 64 deletions(-)
10
12
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
13
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
12
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
15
--- a/include/exec/exec-all.h
14
+++ b/tcg/i386/tcg-target.c.inc
16
+++ b/include/exec/exec-all.h
15
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
17
@@ -XXX,XX +XXX,XX @@ void tb_invalidate_phys_addr(AddressSpace *as, hwaddr addr, MemTxAttrs attrs);
16
case INDEX_op_shls_vec:
18
#endif
17
case INDEX_op_shrs_vec:
19
void tb_flush(CPUState *cpu);
18
case INDEX_op_sars_vec:
20
void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr);
19
- case INDEX_op_rotls_vec:
21
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
20
case INDEX_op_cmp_vec:
22
- target_ulong cs_base, uint32_t flags,
21
case INDEX_op_x86_shufps_vec:
23
- uint32_t cflags);
22
case INDEX_op_x86_blend_vec:
24
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr);
25
26
/* GETPC is the true target of the return instruction that we'll execute. */
27
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/accel/tcg/cpu-exec.c
30
+++ b/accel/tcg/cpu-exec.c
31
@@ -XXX,XX +XXX,XX @@ uint32_t curr_cflags(CPUState *cpu)
32
return cflags;
33
}
34
35
+struct tb_desc {
36
+ target_ulong pc;
37
+ target_ulong cs_base;
38
+ CPUArchState *env;
39
+ tb_page_addr_t phys_page1;
40
+ uint32_t flags;
41
+ uint32_t cflags;
42
+ uint32_t trace_vcpu_dstate;
43
+};
44
+
45
+static bool tb_lookup_cmp(const void *p, const void *d)
46
+{
47
+ const TranslationBlock *tb = p;
48
+ const struct tb_desc *desc = d;
49
+
50
+ if (tb->pc == desc->pc &&
51
+ tb->page_addr[0] == desc->phys_page1 &&
52
+ tb->cs_base == desc->cs_base &&
53
+ tb->flags == desc->flags &&
54
+ tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
55
+ tb_cflags(tb) == desc->cflags) {
56
+ /* check next page if needed */
57
+ if (tb->page_addr[1] == -1) {
58
+ return true;
59
+ } else {
60
+ tb_page_addr_t phys_page2;
61
+ target_ulong virt_page2;
62
+
63
+ virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
64
+ phys_page2 = get_page_addr_code(desc->env, virt_page2);
65
+ if (tb->page_addr[1] == phys_page2) {
66
+ return true;
67
+ }
68
+ }
69
+ }
70
+ return false;
71
+}
72
+
73
+static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
74
+ target_ulong cs_base, uint32_t flags,
75
+ uint32_t cflags)
76
+{
77
+ tb_page_addr_t phys_pc;
78
+ struct tb_desc desc;
79
+ uint32_t h;
80
+
81
+ desc.env = cpu->env_ptr;
82
+ desc.cs_base = cs_base;
83
+ desc.flags = flags;
84
+ desc.cflags = cflags;
85
+ desc.trace_vcpu_dstate = *cpu->trace_dstate;
86
+ desc.pc = pc;
87
+ phys_pc = get_page_addr_code(desc.env, pc);
88
+ if (phys_pc == -1) {
89
+ return NULL;
90
+ }
91
+ desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
92
+ h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
93
+ return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
94
+}
95
+
96
/* Might cause an exception, so have a longjmp destination ready */
97
static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
98
target_ulong cs_base,
99
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
100
end_exclusive();
101
}
102
103
-struct tb_desc {
104
- target_ulong pc;
105
- target_ulong cs_base;
106
- CPUArchState *env;
107
- tb_page_addr_t phys_page1;
108
- uint32_t flags;
109
- uint32_t cflags;
110
- uint32_t trace_vcpu_dstate;
111
-};
112
-
113
-static bool tb_lookup_cmp(const void *p, const void *d)
114
-{
115
- const TranslationBlock *tb = p;
116
- const struct tb_desc *desc = d;
117
-
118
- if (tb->pc == desc->pc &&
119
- tb->page_addr[0] == desc->phys_page1 &&
120
- tb->cs_base == desc->cs_base &&
121
- tb->flags == desc->flags &&
122
- tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
123
- tb_cflags(tb) == desc->cflags) {
124
- /* check next page if needed */
125
- if (tb->page_addr[1] == -1) {
126
- return true;
127
- } else {
128
- tb_page_addr_t phys_page2;
129
- target_ulong virt_page2;
130
-
131
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
132
- phys_page2 = get_page_addr_code(desc->env, virt_page2);
133
- if (tb->page_addr[1] == phys_page2) {
134
- return true;
135
- }
136
- }
137
- }
138
- return false;
139
-}
140
-
141
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
142
- target_ulong cs_base, uint32_t flags,
143
- uint32_t cflags)
144
-{
145
- tb_page_addr_t phys_pc;
146
- struct tb_desc desc;
147
- uint32_t h;
148
-
149
- desc.env = cpu->env_ptr;
150
- desc.cs_base = cs_base;
151
- desc.flags = flags;
152
- desc.cflags = cflags;
153
- desc.trace_vcpu_dstate = *cpu->trace_dstate;
154
- desc.pc = pc;
155
- phys_pc = get_page_addr_code(desc.env, pc);
156
- if (phys_pc == -1) {
157
- return NULL;
158
- }
159
- desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
160
- h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
161
- return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
162
-}
163
-
164
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr)
165
{
166
if (TCG_TARGET_HAS_direct_jump) {
23
--
167
--
24
2.25.1
168
2.34.1
25
26
diff view generated by jsdifflib
1
Expand 32-bit and 64-bit scalar rotate with VPRO[LR]V;
1
The base qemu_ram_addr_from_host function is already in
2
expand 16-bit scalar rotate with VPSHLDV.
2
softmmu/physmem.c; move the nofail version to be adjacent.
3
3
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
---
8
tcg/i386/tcg-target.c.inc | 49 +++++++++++++++++++++++----------------
9
include/exec/cpu-common.h | 1 +
9
1 file changed, 29 insertions(+), 20 deletions(-)
10
accel/tcg/cputlb.c | 12 ------------
11
softmmu/physmem.c | 12 ++++++++++++
12
3 files changed, 13 insertions(+), 12 deletions(-)
10
13
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
14
diff --git a/include/exec/cpu-common.h b/include/exec/cpu-common.h
12
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
16
--- a/include/exec/cpu-common.h
14
+++ b/tcg/i386/tcg-target.c.inc
17
+++ b/include/exec/cpu-common.h
15
@@ -XXX,XX +XXX,XX @@ static void expand_vec_rotli(TCGType type, unsigned vece,
18
@@ -XXX,XX +XXX,XX @@ typedef uintptr_t ram_addr_t;
16
tcg_temp_free_vec(t);
19
void qemu_ram_remap(ram_addr_t addr, ram_addr_t length);
20
/* This should not be used by devices. */
21
ram_addr_t qemu_ram_addr_from_host(void *ptr);
22
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr);
23
RAMBlock *qemu_ram_block_by_name(const char *name);
24
RAMBlock *qemu_ram_block_from_host(void *ptr, bool round_offset,
25
ram_addr_t *offset);
26
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
27
index XXXXXXX..XXXXXXX 100644
28
--- a/accel/tcg/cputlb.c
29
+++ b/accel/tcg/cputlb.c
30
@@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr,
31
prot, mmu_idx, size);
17
}
32
}
18
33
19
-static void expand_vec_rotls(TCGType type, unsigned vece,
34
-static inline ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
20
- TCGv_vec v0, TCGv_vec v1, TCGv_i32 lsh)
21
-{
35
-{
22
- TCGv_i32 rsh;
36
- ram_addr_t ram_addr;
23
- TCGv_vec t;
24
-
37
-
25
- tcg_debug_assert(vece != MO_8);
38
- ram_addr = qemu_ram_addr_from_host(ptr);
26
-
39
- if (ram_addr == RAM_ADDR_INVALID) {
27
- t = tcg_temp_new_vec(type);
40
- error_report("Bad ram pointer %p", ptr);
28
- rsh = tcg_temp_new_i32();
41
- abort();
29
-
42
- }
30
- tcg_gen_neg_i32(rsh, lsh);
43
- return ram_addr;
31
- tcg_gen_andi_i32(rsh, rsh, (8 << vece) - 1);
32
- tcg_gen_shls_vec(vece, t, v1, lsh);
33
- tcg_gen_shrs_vec(vece, v0, v1, rsh);
34
- tcg_gen_or_vec(vece, v0, v0, t);
35
- tcg_temp_free_vec(t);
36
- tcg_temp_free_i32(rsh);
37
-}
44
-}
38
-
45
-
39
static void expand_vec_rotv(TCGType type, unsigned vece, TCGv_vec v0,
46
/*
40
TCGv_vec v1, TCGv_vec sh, bool right)
47
* Note: tlb_fill() can trigger a resize of the TLB. This means that all of the
41
{
48
* caller's prior references to the TLB table (e.g. CPUTLBEntry pointers) must
42
@@ -XXX,XX +XXX,XX @@ static void expand_vec_rotv(TCGType type, unsigned vece, TCGv_vec v0,
49
diff --git a/softmmu/physmem.c b/softmmu/physmem.c
43
tcg_temp_free_vec(t);
50
index XXXXXXX..XXXXXXX 100644
51
--- a/softmmu/physmem.c
52
+++ b/softmmu/physmem.c
53
@@ -XXX,XX +XXX,XX @@ ram_addr_t qemu_ram_addr_from_host(void *ptr)
54
return block->offset + offset;
44
}
55
}
45
56
46
+static void expand_vec_rotls(TCGType type, unsigned vece,
57
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
47
+ TCGv_vec v0, TCGv_vec v1, TCGv_i32 lsh)
48
+{
58
+{
49
+ TCGv_vec t = tcg_temp_new_vec(type);
59
+ ram_addr_t ram_addr;
50
+
60
+
51
+ tcg_debug_assert(vece != MO_8);
61
+ ram_addr = qemu_ram_addr_from_host(ptr);
52
+
62
+ if (ram_addr == RAM_ADDR_INVALID) {
53
+ if (vece >= MO_32 ? have_avx512vl : have_avx512vbmi2) {
63
+ error_report("Bad ram pointer %p", ptr);
54
+ tcg_gen_dup_i32_vec(vece, t, lsh);
64
+ abort();
55
+ if (vece >= MO_32) {
56
+ tcg_gen_rotlv_vec(vece, v0, v1, t);
57
+ } else {
58
+ expand_vec_rotv(type, vece, v0, v1, t, false);
59
+ }
60
+ } else {
61
+ TCGv_i32 rsh = tcg_temp_new_i32();
62
+
63
+ tcg_gen_neg_i32(rsh, lsh);
64
+ tcg_gen_andi_i32(rsh, rsh, (8 << vece) - 1);
65
+ tcg_gen_shls_vec(vece, t, v1, lsh);
66
+ tcg_gen_shrs_vec(vece, v0, v1, rsh);
67
+ tcg_gen_or_vec(vece, v0, v0, t);
68
+
69
+ tcg_temp_free_i32(rsh);
70
+ }
65
+ }
71
+
66
+ return ram_addr;
72
+ tcg_temp_free_vec(t);
73
+}
67
+}
74
+
68
+
75
static void expand_vec_mul(TCGType type, unsigned vece,
69
static MemTxResult flatview_read(FlatView *fv, hwaddr addr,
76
TCGv_vec v0, TCGv_vec v1, TCGv_vec v2)
70
MemTxAttrs attrs, void *buf, hwaddr len);
77
{
71
static MemTxResult flatview_write(FlatView *fv, hwaddr addr, MemTxAttrs attrs,
78
--
72
--
79
2.25.1
73
2.34.1
80
81
diff view generated by jsdifflib
1
Create a new function to combine a CPUTLBEntry addend
1
Simplify the implementation of get_page_addr_code_hostp
2
with the guest address to form a host address.
2
by reusing the existing probe_access infrastructure.
3
3
4
Reviewed-by: WANG Xuerui <git@xen0n.name>
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
7
---
10
accel/tcg/cputlb.c | 24 ++++++++++++++----------
8
accel/tcg/cputlb.c | 76 ++++++++++++++++------------------------------
11
1 file changed, 14 insertions(+), 10 deletions(-)
9
1 file changed, 26 insertions(+), 50 deletions(-)
12
10
13
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
11
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
14
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
15
--- a/accel/tcg/cputlb.c
13
--- a/accel/tcg/cputlb.c
16
+++ b/accel/tcg/cputlb.c
14
+++ b/accel/tcg/cputlb.c
17
@@ -XXX,XX +XXX,XX @@ static inline size_t sizeof_tlb(CPUTLBDescFast *fast)
15
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
18
return fast->mask + (1 << CPU_TLB_ENTRY_BITS);
16
victim_tlb_hit(env, mmu_idx, index, offsetof(CPUTLBEntry, TY), \
17
(ADDR) & TARGET_PAGE_MASK)
18
19
-/*
20
- * Return a ram_addr_t for the virtual address for execution.
21
- *
22
- * Return -1 if we can't translate and execute from an entire page
23
- * of RAM. This will force us to execute by loading and translating
24
- * one insn at a time, without caching.
25
- *
26
- * NOTE: This function will trigger an exception if the page is
27
- * not executable.
28
- */
29
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
30
- void **hostp)
31
-{
32
- uintptr_t mmu_idx = cpu_mmu_index(env, true);
33
- uintptr_t index = tlb_index(env, mmu_idx, addr);
34
- CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr);
35
- void *p;
36
-
37
- if (unlikely(!tlb_hit(entry->addr_code, addr))) {
38
- if (!VICTIM_TLB_HIT(addr_code, addr)) {
39
- tlb_fill(env_cpu(env), addr, 0, MMU_INST_FETCH, mmu_idx, 0);
40
- index = tlb_index(env, mmu_idx, addr);
41
- entry = tlb_entry(env, mmu_idx, addr);
42
-
43
- if (unlikely(entry->addr_code & TLB_INVALID_MASK)) {
44
- /*
45
- * The MMU protection covers a smaller range than a target
46
- * page, so we must redo the MMU check for every insn.
47
- */
48
- return -1;
49
- }
50
- }
51
- assert(tlb_hit(entry->addr_code, addr));
52
- }
53
-
54
- if (unlikely(entry->addr_code & TLB_MMIO)) {
55
- /* The region is not backed by RAM. */
56
- if (hostp) {
57
- *hostp = NULL;
58
- }
59
- return -1;
60
- }
61
-
62
- p = (void *)((uintptr_t)addr + entry->addend);
63
- if (hostp) {
64
- *hostp = p;
65
- }
66
- return qemu_ram_addr_from_host_nofail(p);
67
-}
68
-
69
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
70
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
71
{
72
@@ -XXX,XX +XXX,XX @@ void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
73
return flags ? NULL : host;
19
}
74
}
20
75
21
+static inline uintptr_t g2h_tlbe(const CPUTLBEntry *tlb, target_ulong gaddr)
76
+/*
77
+ * Return a ram_addr_t for the virtual address for execution.
78
+ *
79
+ * Return -1 if we can't translate and execute from an entire page
80
+ * of RAM. This will force us to execute by loading and translating
81
+ * one insn at a time, without caching.
82
+ *
83
+ * NOTE: This function will trigger an exception if the page is
84
+ * not executable.
85
+ */
86
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
87
+ void **hostp)
22
+{
88
+{
23
+ return tlb->addend + (uintptr_t)gaddr;
89
+ void *p;
90
+
91
+ (void)probe_access_internal(env, addr, 1, MMU_INST_FETCH,
92
+ cpu_mmu_index(env, true), false, &p, 0);
93
+ if (p == NULL) {
94
+ return -1;
95
+ }
96
+ if (hostp) {
97
+ *hostp = p;
98
+ }
99
+ return qemu_ram_addr_from_host_nofail(p);
24
+}
100
+}
25
+
101
+
26
static void tlb_window_reset(CPUTLBDesc *desc, int64_t ns,
102
#ifdef CONFIG_PLUGIN
27
size_t max_entries)
103
/*
28
{
104
* Perform a TLB lookup and populate the qemu_plugin_hwaddr structure.
29
@@ -XXX,XX +XXX,XX @@ static void tlb_reset_dirty_range_locked(CPUTLBEntry *tlb_entry,
30
31
if ((addr & (TLB_INVALID_MASK | TLB_MMIO |
32
TLB_DISCARD_WRITE | TLB_NOTDIRTY)) == 0) {
33
- addr &= TARGET_PAGE_MASK;
34
- addr += tlb_entry->addend;
35
+ addr = g2h_tlbe(tlb_entry, addr & TARGET_PAGE_MASK);
36
if ((addr - start) < length) {
37
#if TCG_OVERSIZED_GUEST
38
tlb_entry->addr_write |= TLB_NOTDIRTY;
39
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
40
return -1;
41
}
42
43
- p = (void *)((uintptr_t)addr + entry->addend);
44
+ p = (void *)g2h_tlbe(entry, addr);
45
if (hostp) {
46
*hostp = p;
47
}
48
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
49
}
50
51
/* Everything else is RAM. */
52
- *phost = (void *)((uintptr_t)addr + entry->addend);
53
+ *phost = (void *)g2h_tlbe(entry, addr);
54
return flags;
55
}
56
57
@@ -XXX,XX +XXX,XX @@ bool tlb_plugin_lookup(CPUState *cpu, target_ulong addr, int mmu_idx,
58
data->v.io.offset = (iotlbentry->addr & TARGET_PAGE_MASK) + addr;
59
} else {
60
data->is_io = false;
61
- data->v.ram.hostaddr = (void *)((uintptr_t)addr + tlbe->addend);
62
+ data->v.ram.hostaddr = (void *)g2h_tlbe(tlbe, addr);
63
}
64
return true;
65
} else {
66
@@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr,
67
goto stop_the_world;
68
}
69
70
- hostaddr = (void *)((uintptr_t)addr + tlbe->addend);
71
+ hostaddr = (void *)g2h_tlbe(tlbe, addr);
72
73
if (unlikely(tlb_addr & TLB_NOTDIRTY)) {
74
notdirty_write(env_cpu(env), addr, size,
75
@@ -XXX,XX +XXX,XX @@ load_helper(CPUArchState *env, target_ulong addr, MemOpIdx oi,
76
access_type, op ^ (need_swap * MO_BSWAP));
77
}
78
79
- haddr = (void *)((uintptr_t)addr + entry->addend);
80
+ haddr = (void *)g2h_tlbe(entry, addr);
81
82
/*
83
* Keep these two load_memop separate to ensure that the compiler
84
@@ -XXX,XX +XXX,XX @@ load_helper(CPUArchState *env, target_ulong addr, MemOpIdx oi,
85
return res & MAKE_64BIT_MASK(0, size * 8);
86
}
87
88
- haddr = (void *)((uintptr_t)addr + entry->addend);
89
+ haddr = (void *)g2h_tlbe(entry, addr);
90
return load_memop(haddr, op);
91
}
92
93
@@ -XXX,XX +XXX,XX @@ store_helper(CPUArchState *env, target_ulong addr, uint64_t val,
94
notdirty_write(env_cpu(env), addr, size, iotlbentry, retaddr);
95
}
96
97
- haddr = (void *)((uintptr_t)addr + entry->addend);
98
+ haddr = (void *)g2h_tlbe(entry, addr);
99
100
/*
101
* Keep these two store_memop separate to ensure that the compiler
102
@@ -XXX,XX +XXX,XX @@ store_helper(CPUArchState *env, target_ulong addr, uint64_t val,
103
return;
104
}
105
106
- haddr = (void *)((uintptr_t)addr + entry->addend);
107
+ haddr = (void *)g2h_tlbe(entry, addr);
108
store_memop(haddr, val, op);
109
}
110
111
--
105
--
112
2.25.1
106
2.34.1
113
114
diff view generated by jsdifflib
1
We will use VPSHLD, VPSHLDV and VPSHRDV for 16-bit rotates.
1
It was non-obvious to me why we can raise an exception in
2
the middle of a comparison function, but it works.
3
While nearby, use TARGET_PAGE_ALIGN instead of open-coding.
2
4
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
---
7
tcg/i386/tcg-target-con-set.h | 1 +
8
accel/tcg/cpu-exec.c | 11 ++++++++++-
8
tcg/i386/tcg-target.opc.h | 3 +++
9
1 file changed, 10 insertions(+), 1 deletion(-)
9
tcg/i386/tcg-target.c.inc | 38 +++++++++++++++++++++++++++++++++++
10
3 files changed, 42 insertions(+)
11
10
12
diff --git a/tcg/i386/tcg-target-con-set.h b/tcg/i386/tcg-target-con-set.h
11
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
13
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/i386/tcg-target-con-set.h
13
--- a/accel/tcg/cpu-exec.c
15
+++ b/tcg/i386/tcg-target-con-set.h
14
+++ b/accel/tcg/cpu-exec.c
16
@@ -XXX,XX +XXX,XX @@ C_O1_I2(r, r, rI)
15
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
17
C_O1_I2(x, x, x)
16
tb_page_addr_t phys_page2;
18
C_N1_I2(r, r, r)
17
target_ulong virt_page2;
19
C_N1_I2(r, r, rW)
18
20
+C_O1_I3(x, 0, x, x)
19
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
21
C_O1_I3(x, x, x, x)
20
+ /*
22
C_O1_I4(r, r, re, r, 0)
21
+ * We know that the first page matched, and an otherwise valid TB
23
C_O1_I4(r, r, r, ri, ri)
22
+ * encountered an incomplete instruction at the end of that page,
24
diff --git a/tcg/i386/tcg-target.opc.h b/tcg/i386/tcg-target.opc.h
23
+ * therefore we know that generating a new TB from the current PC
25
index XXXXXXX..XXXXXXX 100644
24
+ * must also require reading from the next page -- even if the
26
--- a/tcg/i386/tcg-target.opc.h
25
+ * second pages do not match, and therefore the resulting insn
27
+++ b/tcg/i386/tcg-target.opc.h
26
+ * is different for the new TB. Therefore any exception raised
28
@@ -XXX,XX +XXX,XX @@ DEF(x86_psrldq_vec, 1, 1, 1, IMPLVEC)
27
+ * here by the faulting lookup is not premature.
29
DEF(x86_vperm2i128_vec, 1, 2, 1, IMPLVEC)
28
+ */
30
DEF(x86_punpckl_vec, 1, 2, 0, IMPLVEC)
29
+ virt_page2 = TARGET_PAGE_ALIGN(desc->pc);
31
DEF(x86_punpckh_vec, 1, 2, 0, IMPLVEC)
30
phys_page2 = get_page_addr_code(desc->env, virt_page2);
32
+DEF(x86_vpshldi_vec, 1, 2, 1, IMPLVEC)
31
if (tb->page_addr[1] == phys_page2) {
33
+DEF(x86_vpshldv_vec, 1, 3, 0, IMPLVEC)
32
return true;
34
+DEF(x86_vpshrdv_vec, 1, 3, 0, IMPLVEC)
35
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
36
index XXXXXXX..XXXXXXX 100644
37
--- a/tcg/i386/tcg-target.c.inc
38
+++ b/tcg/i386/tcg-target.c.inc
39
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
40
#define OPC_VPROLVQ (0x15 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
41
#define OPC_VPRORVD (0x14 | P_EXT38 | P_DATA16 | P_EVEX)
42
#define OPC_VPRORVQ (0x14 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
43
+#define OPC_VPSHLDW (0x70 | P_EXT3A | P_DATA16 | P_VEXW | P_EVEX)
44
+#define OPC_VPSHLDD (0x71 | P_EXT3A | P_DATA16 | P_EVEX)
45
+#define OPC_VPSHLDQ (0x71 | P_EXT3A | P_DATA16 | P_VEXW | P_EVEX)
46
+#define OPC_VPSHLDVW (0x70 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
47
+#define OPC_VPSHLDVD (0x71 | P_EXT38 | P_DATA16 | P_EVEX)
48
+#define OPC_VPSHLDVQ (0x71 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
49
+#define OPC_VPSHRDVW (0x72 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
50
+#define OPC_VPSHRDVD (0x73 | P_EXT38 | P_DATA16 | P_EVEX)
51
+#define OPC_VPSHRDVQ (0x73 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
52
#define OPC_VPSLLVW (0x12 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
53
#define OPC_VPSLLVD (0x47 | P_EXT38 | P_DATA16)
54
#define OPC_VPSLLVQ (0x47 | P_EXT38 | P_DATA16 | P_VEXW)
55
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
56
static int const sars_insn[4] = {
57
OPC_UD2, OPC_PSRAW, OPC_PSRAD, OPC_VPSRAQ
58
};
59
+ static int const vpshldi_insn[4] = {
60
+ OPC_UD2, OPC_VPSHLDW, OPC_VPSHLDD, OPC_VPSHLDQ
61
+ };
62
+ static int const vpshldv_insn[4] = {
63
+ OPC_UD2, OPC_VPSHLDVW, OPC_VPSHLDVD, OPC_VPSHLDVQ
64
+ };
65
+ static int const vpshrdv_insn[4] = {
66
+ OPC_UD2, OPC_VPSHRDVW, OPC_VPSHRDVD, OPC_VPSHRDVQ
67
+ };
68
static int const abs_insn[4] = {
69
/* TODO: AVX512 adds support for MO_64. */
70
OPC_PABSB, OPC_PABSW, OPC_PABSD, OPC_UD2
71
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
72
case INDEX_op_x86_packus_vec:
73
insn = packus_insn[vece];
74
goto gen_simd;
75
+ case INDEX_op_x86_vpshldv_vec:
76
+ insn = vpshldv_insn[vece];
77
+ a1 = a2;
78
+ a2 = args[3];
79
+ goto gen_simd;
80
+ case INDEX_op_x86_vpshrdv_vec:
81
+ insn = vpshrdv_insn[vece];
82
+ a1 = a2;
83
+ a2 = args[3];
84
+ goto gen_simd;
85
#if TCG_TARGET_REG_BITS == 32
86
case INDEX_op_dup2_vec:
87
/* First merge the two 32-bit inputs to a single 64-bit element. */
88
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
89
insn = OPC_VPERM2I128;
90
sub = args[3];
91
goto gen_simd_imm8;
92
+ case INDEX_op_x86_vpshldi_vec:
93
+ insn = vpshldi_insn[vece];
94
+ sub = args[3];
95
+ goto gen_simd_imm8;
96
gen_simd_imm8:
97
+ tcg_debug_assert(insn != OPC_UD2);
98
if (type == TCG_TYPE_V256) {
99
insn |= P_VEXL;
100
}
101
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
102
case INDEX_op_x86_vperm2i128_vec:
103
case INDEX_op_x86_punpckl_vec:
104
case INDEX_op_x86_punpckh_vec:
105
+ case INDEX_op_x86_vpshldi_vec:
106
#if TCG_TARGET_REG_BITS == 32
107
case INDEX_op_dup2_vec:
108
#endif
109
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
110
case INDEX_op_x86_psrldq_vec:
111
return C_O1_I1(x, x);
112
113
+ case INDEX_op_x86_vpshldv_vec:
114
+ case INDEX_op_x86_vpshrdv_vec:
115
+ return C_O1_I3(x, 0, x, x);
116
+
117
case INDEX_op_x86_vpblendvb_vec:
118
return C_O1_I3(x, x, x, x);
119
120
--
33
--
121
2.25.1
34
2.34.1
122
123
diff view generated by jsdifflib
1
AVX512VL has VPROLVD and VPRORVQ.
1
The only user can easily use translator_lduw and
2
adjust the type to signed during the return.
2
3
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
8
---
7
tcg/i386/tcg-target.h | 2 +-
9
include/exec/translator.h | 1 -
8
tcg/i386/tcg-target.c.inc | 25 ++++++++++++++++++++++++-
10
target/i386/tcg/translate.c | 2 +-
9
2 files changed, 25 insertions(+), 2 deletions(-)
11
2 files changed, 1 insertion(+), 2 deletions(-)
10
12
11
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
13
diff --git a/include/exec/translator.h b/include/exec/translator.h
12
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.h
15
--- a/include/exec/translator.h
14
+++ b/tcg/i386/tcg-target.h
16
+++ b/include/exec/translator.h
15
@@ -XXX,XX +XXX,XX @@ extern bool have_movbe;
17
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
16
#define TCG_TARGET_HAS_abs_vec 1
18
17
#define TCG_TARGET_HAS_roti_vec have_avx512vl
19
#define FOR_EACH_TRANSLATOR_LD(F) \
18
#define TCG_TARGET_HAS_rots_vec 0
20
F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
19
-#define TCG_TARGET_HAS_rotv_vec 0
21
- F(translator_ldsw, int16_t, cpu_ldsw_code, bswap16) \
20
+#define TCG_TARGET_HAS_rotv_vec have_avx512vl
22
F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
21
#define TCG_TARGET_HAS_shi_vec 1
23
F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
22
#define TCG_TARGET_HAS_shs_vec 1
24
F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
23
#define TCG_TARGET_HAS_shv_vec have_avx2
25
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
24
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
25
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
26
--- a/tcg/i386/tcg-target.c.inc
27
--- a/target/i386/tcg/translate.c
27
+++ b/tcg/i386/tcg-target.c.inc
28
+++ b/target/i386/tcg/translate.c
28
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
29
@@ -XXX,XX +XXX,XX @@ static inline uint8_t x86_ldub_code(CPUX86State *env, DisasContext *s)
29
#define OPC_VPBROADCASTQ (0x59 | P_EXT38 | P_DATA16)
30
30
#define OPC_VPERMQ (0x00 | P_EXT3A | P_DATA16 | P_VEXW)
31
static inline int16_t x86_ldsw_code(CPUX86State *env, DisasContext *s)
31
#define OPC_VPERM2I128 (0x46 | P_EXT3A | P_DATA16 | P_VEXL)
32
{
32
+#define OPC_VPROLVD (0x15 | P_EXT38 | P_DATA16 | P_EVEX)
33
- return translator_ldsw(env, &s->base, advance_pc(env, s, 2));
33
+#define OPC_VPROLVQ (0x15 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
34
+ return translator_lduw(env, &s->base, advance_pc(env, s, 2));
34
+#define OPC_VPRORVD (0x14 | P_EXT38 | P_DATA16 | P_EVEX)
35
}
35
+#define OPC_VPRORVQ (0x14 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
36
36
#define OPC_VPSLLVW (0x12 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
37
static inline uint16_t x86_lduw_code(CPUX86State *env, DisasContext *s)
37
#define OPC_VPSLLVD (0x47 | P_EXT38 | P_DATA16)
38
#define OPC_VPSLLVQ (0x47 | P_EXT38 | P_DATA16 | P_VEXW)
39
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
40
static int const umax_insn[4] = {
41
OPC_PMAXUB, OPC_PMAXUW, OPC_PMAXUD, OPC_UD2
42
};
43
+ static int const rotlv_insn[4] = {
44
+ OPC_UD2, OPC_UD2, OPC_VPROLVD, OPC_VPROLVQ
45
+ };
46
+ static int const rotrv_insn[4] = {
47
+ OPC_UD2, OPC_UD2, OPC_VPRORVD, OPC_VPRORVQ
48
+ };
49
static int const shlv_insn[4] = {
50
OPC_UD2, OPC_VPSLLVW, OPC_VPSLLVD, OPC_VPSLLVQ
51
};
52
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
53
case INDEX_op_sarv_vec:
54
insn = sarv_insn[vece];
55
goto gen_simd;
56
+ case INDEX_op_rotlv_vec:
57
+ insn = rotlv_insn[vece];
58
+ goto gen_simd;
59
+ case INDEX_op_rotrv_vec:
60
+ insn = rotrv_insn[vece];
61
+ goto gen_simd;
62
case INDEX_op_shls_vec:
63
insn = shls_insn[vece];
64
goto gen_simd;
65
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
66
case INDEX_op_shlv_vec:
67
case INDEX_op_shrv_vec:
68
case INDEX_op_sarv_vec:
69
+ case INDEX_op_rotlv_vec:
70
+ case INDEX_op_rotrv_vec:
71
case INDEX_op_shls_vec:
72
case INDEX_op_shrs_vec:
73
case INDEX_op_sars_vec:
74
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
75
return 0;
76
case INDEX_op_rotlv_vec:
77
case INDEX_op_rotrv_vec:
78
- return have_avx2 && vece >= MO_32 ? -1 : 0;
79
+ switch (vece) {
80
+ case MO_32:
81
+ case MO_64:
82
+ return have_avx512vl ? 1 : have_avx2 ? -1 : 0;
83
+ }
84
+ return 0;
85
86
case INDEX_op_mul_vec:
87
if (vece == MO_8) {
88
--
38
--
89
2.25.1
39
2.34.1
90
91
diff view generated by jsdifflib
1
AVX512VL has VPROLD and VPROLQ, layered onto the same
1
Pass these along to translator_loop -- pc may be used instead
2
opcode as PSHIFTD, but requires EVEX encoding and W1.
2
of tb->pc, and host_pc is currently unused. Adjust all targets
3
at one time.
3
4
5
Acked-by: Alistair Francis <alistair.francis@wdc.com>
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
9
---
6
tcg/i386/tcg-target.h | 2 +-
10
include/exec/exec-all.h | 1 -
7
tcg/i386/tcg-target.c.inc | 15 +++++++++++++--
11
include/exec/translator.h | 24 ++++++++++++++++++++----
8
2 files changed, 14 insertions(+), 3 deletions(-)
12
accel/tcg/translate-all.c | 6 ++++--
13
accel/tcg/translator.c | 9 +++++----
14
target/alpha/translate.c | 5 +++--
15
target/arm/translate.c | 5 +++--
16
target/avr/translate.c | 5 +++--
17
target/cris/translate.c | 5 +++--
18
target/hexagon/translate.c | 6 ++++--
19
target/hppa/translate.c | 5 +++--
20
target/i386/tcg/translate.c | 5 +++--
21
target/loongarch/translate.c | 6 ++++--
22
target/m68k/translate.c | 5 +++--
23
target/microblaze/translate.c | 5 +++--
24
target/mips/tcg/translate.c | 5 +++--
25
target/nios2/translate.c | 5 +++--
26
target/openrisc/translate.c | 6 ++++--
27
target/ppc/translate.c | 5 +++--
28
target/riscv/translate.c | 5 +++--
29
target/rx/translate.c | 5 +++--
30
target/s390x/tcg/translate.c | 5 +++--
31
target/sh4/translate.c | 5 +++--
32
target/sparc/translate.c | 5 +++--
33
target/tricore/translate.c | 6 ++++--
34
target/xtensa/translate.c | 6 ++++--
35
25 files changed, 97 insertions(+), 53 deletions(-)
9
36
10
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
37
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
11
index XXXXXXX..XXXXXXX 100644
38
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.h
39
--- a/include/exec/exec-all.h
13
+++ b/tcg/i386/tcg-target.h
40
+++ b/include/exec/exec-all.h
14
@@ -XXX,XX +XXX,XX @@ extern bool have_movbe;
41
@@ -XXX,XX +XXX,XX @@ typedef ram_addr_t tb_page_addr_t;
15
#define TCG_TARGET_HAS_not_vec 0
42
#define TB_PAGE_ADDR_FMT RAM_ADDR_FMT
16
#define TCG_TARGET_HAS_neg_vec 0
43
#endif
17
#define TCG_TARGET_HAS_abs_vec 1
44
18
-#define TCG_TARGET_HAS_roti_vec 0
45
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns);
19
+#define TCG_TARGET_HAS_roti_vec have_avx512vl
46
void restore_state_to_opc(CPUArchState *env, TranslationBlock *tb,
20
#define TCG_TARGET_HAS_rots_vec 0
47
target_ulong *data);
21
#define TCG_TARGET_HAS_rotv_vec 0
48
22
#define TCG_TARGET_HAS_shi_vec 1
49
diff --git a/include/exec/translator.h b/include/exec/translator.h
23
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
50
index XXXXXXX..XXXXXXX 100644
24
index XXXXXXX..XXXXXXX 100644
51
--- a/include/exec/translator.h
25
--- a/tcg/i386/tcg-target.c.inc
52
+++ b/include/exec/translator.h
26
+++ b/tcg/i386/tcg-target.c.inc
53
@@ -XXX,XX +XXX,XX @@
27
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
54
#include "exec/translate-all.h"
28
#define OPC_PSHUFLW (0x70 | P_EXT | P_SIMDF2)
55
#include "tcg/tcg.h"
29
#define OPC_PSHUFHW (0x70 | P_EXT | P_SIMDF3)
56
30
#define OPC_PSHIFTW_Ib (0x71 | P_EXT | P_DATA16) /* /2 /6 /4 */
57
+/**
31
-#define OPC_PSHIFTD_Ib (0x72 | P_EXT | P_DATA16) /* /2 /6 /4 */
58
+ * gen_intermediate_code
32
+#define OPC_PSHIFTD_Ib (0x72 | P_EXT | P_DATA16) /* /1 /2 /6 /4 */
59
+ * @cpu: cpu context
33
#define OPC_PSHIFTQ_Ib (0x73 | P_EXT | P_DATA16) /* /2 /6 /4 */
60
+ * @tb: translation block
34
#define OPC_PSLLW (0xf1 | P_EXT | P_DATA16)
61
+ * @max_insns: max number of instructions to translate
35
#define OPC_PSLLD (0xf2 | P_EXT | P_DATA16)
62
+ * @pc: guest virtual program counter address
36
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
63
+ * @host_pc: host physical program counter address
37
insn = shift_imm_insn[vece];
64
+ *
38
}
65
+ * This function must be provided by the target, which should create
39
sub = 4;
66
+ * the target-specific DisasContext, and then invoke translator_loop.
40
+ goto gen_shift;
67
+ */
41
+ case INDEX_op_rotli_vec:
68
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
42
+ insn = OPC_PSHIFTD_Ib | P_EVEX; /* VPROL[DQ] */
69
+ target_ulong pc, void *host_pc);
43
+ if (vece == MO_64) {
70
44
+ insn |= P_VEXW;
71
/**
45
+ }
72
* DisasJumpType:
46
+ sub = 1;
73
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
47
+ goto gen_shift;
74
48
gen_shift:
75
/**
49
tcg_debug_assert(vece != MO_8);
76
* translator_loop:
50
if (type == TCG_TYPE_V256) {
77
- * @ops: Target-specific operations.
51
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
78
- * @db: Disassembly context.
52
case INDEX_op_shli_vec:
79
* @cpu: Target vCPU.
53
case INDEX_op_shri_vec:
80
* @tb: Translation block.
54
case INDEX_op_sari_vec:
81
* @max_insns: Maximum number of insns to translate.
55
+ case INDEX_op_rotli_vec:
82
+ * @pc: guest virtual program counter address
56
case INDEX_op_x86_psrldq_vec:
83
+ * @host_pc: host physical program counter address
57
return C_O1_I1(x, x);
84
+ * @ops: Target-specific operations.
58
85
+ * @db: Disassembly context.
59
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
86
*
60
case INDEX_op_xor_vec:
87
* Generic translator loop.
61
case INDEX_op_andc_vec:
88
*
62
return 1;
89
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
63
- case INDEX_op_rotli_vec:
90
* - When single-stepping is enabled (system-wide or on the current vCPU).
64
case INDEX_op_cmp_vec:
91
* - When too many instructions have been translated.
65
case INDEX_op_cmpsel_vec:
92
*/
66
return -1;
93
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
67
94
- CPUState *cpu, TranslationBlock *tb, int max_insns);
68
+ case INDEX_op_rotli_vec:
95
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
69
+ return have_avx512vl && vece >= MO_32 ? 1 : -1;
96
+ target_ulong pc, void *host_pc,
70
+
97
+ const TranslatorOps *ops, DisasContextBase *db);
71
case INDEX_op_shli_vec:
98
72
case INDEX_op_shri_vec:
99
void translator_loop_temp_check(DisasContextBase *db);
73
/* We must expand the operation for MO_8. */
100
101
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
102
index XXXXXXX..XXXXXXX 100644
103
--- a/accel/tcg/translate-all.c
104
+++ b/accel/tcg/translate-all.c
105
@@ -XXX,XX +XXX,XX @@
106
107
#include "exec/cputlb.h"
108
#include "exec/translate-all.h"
109
+#include "exec/translator.h"
110
#include "qemu/bitmap.h"
111
#include "qemu/qemu-print.h"
112
#include "qemu/timer.h"
113
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
114
TCGProfile *prof = &tcg_ctx->prof;
115
int64_t ti;
116
#endif
117
+ void *host_pc;
118
119
assert_memory_lock();
120
qemu_thread_jit_write();
121
122
- phys_pc = get_page_addr_code(env, pc);
123
+ phys_pc = get_page_addr_code_hostp(env, pc, &host_pc);
124
125
if (phys_pc == -1) {
126
/* Generate a one-shot TB with 1 insn in it */
127
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
128
tcg_func_start(tcg_ctx);
129
130
tcg_ctx->cpu = env_cpu(env);
131
- gen_intermediate_code(cpu, tb, max_insns);
132
+ gen_intermediate_code(cpu, tb, max_insns, pc, host_pc);
133
assert(tb->size != 0);
134
tcg_ctx->cpu = NULL;
135
max_insns = tb->icount;
136
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
137
index XXXXXXX..XXXXXXX 100644
138
--- a/accel/tcg/translator.c
139
+++ b/accel/tcg/translator.c
140
@@ -XXX,XX +XXX,XX @@ static inline void translator_page_protect(DisasContextBase *dcbase,
141
#endif
142
}
143
144
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
145
- CPUState *cpu, TranslationBlock *tb, int max_insns)
146
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
147
+ target_ulong pc, void *host_pc,
148
+ const TranslatorOps *ops, DisasContextBase *db)
149
{
150
uint32_t cflags = tb_cflags(tb);
151
bool plugin_enabled;
152
153
/* Initialize DisasContext */
154
db->tb = tb;
155
- db->pc_first = tb->pc;
156
- db->pc_next = db->pc_first;
157
+ db->pc_first = pc;
158
+ db->pc_next = pc;
159
db->is_jmp = DISAS_NEXT;
160
db->num_insns = 0;
161
db->max_insns = max_insns;
162
diff --git a/target/alpha/translate.c b/target/alpha/translate.c
163
index XXXXXXX..XXXXXXX 100644
164
--- a/target/alpha/translate.c
165
+++ b/target/alpha/translate.c
166
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps alpha_tr_ops = {
167
.disas_log = alpha_tr_disas_log,
168
};
169
170
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
171
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
172
+ target_ulong pc, void *host_pc)
173
{
174
DisasContext dc;
175
- translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
176
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &alpha_tr_ops, &dc.base);
177
}
178
179
void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
180
diff --git a/target/arm/translate.c b/target/arm/translate.c
181
index XXXXXXX..XXXXXXX 100644
182
--- a/target/arm/translate.c
183
+++ b/target/arm/translate.c
184
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps thumb_translator_ops = {
185
};
186
187
/* generate intermediate code for basic block 'tb'. */
188
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
189
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
190
+ target_ulong pc, void *host_pc)
191
{
192
DisasContext dc = { };
193
const TranslatorOps *ops = &arm_translator_ops;
194
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
195
}
196
#endif
197
198
- translator_loop(ops, &dc.base, cpu, tb, max_insns);
199
+ translator_loop(cpu, tb, max_insns, pc, host_pc, ops, &dc.base);
200
}
201
202
void restore_state_to_opc(CPUARMState *env, TranslationBlock *tb,
203
diff --git a/target/avr/translate.c b/target/avr/translate.c
204
index XXXXXXX..XXXXXXX 100644
205
--- a/target/avr/translate.c
206
+++ b/target/avr/translate.c
207
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps avr_tr_ops = {
208
.disas_log = avr_tr_disas_log,
209
};
210
211
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
212
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
213
+ target_ulong pc, void *host_pc)
214
{
215
DisasContext dc = { };
216
- translator_loop(&avr_tr_ops, &dc.base, cs, tb, max_insns);
217
+ translator_loop(cs, tb, max_insns, pc, host_pc, &avr_tr_ops, &dc.base);
218
}
219
220
void restore_state_to_opc(CPUAVRState *env, TranslationBlock *tb,
221
diff --git a/target/cris/translate.c b/target/cris/translate.c
222
index XXXXXXX..XXXXXXX 100644
223
--- a/target/cris/translate.c
224
+++ b/target/cris/translate.c
225
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps cris_tr_ops = {
226
.disas_log = cris_tr_disas_log,
227
};
228
229
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
230
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
231
+ target_ulong pc, void *host_pc)
232
{
233
DisasContext dc;
234
- translator_loop(&cris_tr_ops, &dc.base, cs, tb, max_insns);
235
+ translator_loop(cs, tb, max_insns, pc, host_pc, &cris_tr_ops, &dc.base);
236
}
237
238
void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
239
diff --git a/target/hexagon/translate.c b/target/hexagon/translate.c
240
index XXXXXXX..XXXXXXX 100644
241
--- a/target/hexagon/translate.c
242
+++ b/target/hexagon/translate.c
243
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hexagon_tr_ops = {
244
.disas_log = hexagon_tr_disas_log,
245
};
246
247
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
248
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
249
+ target_ulong pc, void *host_pc)
250
{
251
DisasContext ctx;
252
253
- translator_loop(&hexagon_tr_ops, &ctx.base, cs, tb, max_insns);
254
+ translator_loop(cs, tb, max_insns, pc, host_pc,
255
+ &hexagon_tr_ops, &ctx.base);
256
}
257
258
#define NAME_LEN 64
259
diff --git a/target/hppa/translate.c b/target/hppa/translate.c
260
index XXXXXXX..XXXXXXX 100644
261
--- a/target/hppa/translate.c
262
+++ b/target/hppa/translate.c
263
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hppa_tr_ops = {
264
.disas_log = hppa_tr_disas_log,
265
};
266
267
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
268
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
269
+ target_ulong pc, void *host_pc)
270
{
271
DisasContext ctx;
272
- translator_loop(&hppa_tr_ops, &ctx.base, cs, tb, max_insns);
273
+ translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base);
274
}
275
276
void restore_state_to_opc(CPUHPPAState *env, TranslationBlock *tb,
277
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
278
index XXXXXXX..XXXXXXX 100644
279
--- a/target/i386/tcg/translate.c
280
+++ b/target/i386/tcg/translate.c
281
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps i386_tr_ops = {
282
};
283
284
/* generate intermediate code for basic block 'tb'. */
285
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
286
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
287
+ target_ulong pc, void *host_pc)
288
{
289
DisasContext dc;
290
291
- translator_loop(&i386_tr_ops, &dc.base, cpu, tb, max_insns);
292
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &i386_tr_ops, &dc.base);
293
}
294
295
void restore_state_to_opc(CPUX86State *env, TranslationBlock *tb,
296
diff --git a/target/loongarch/translate.c b/target/loongarch/translate.c
297
index XXXXXXX..XXXXXXX 100644
298
--- a/target/loongarch/translate.c
299
+++ b/target/loongarch/translate.c
300
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps loongarch_tr_ops = {
301
.disas_log = loongarch_tr_disas_log,
302
};
303
304
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
305
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
306
+ target_ulong pc, void *host_pc)
307
{
308
DisasContext ctx;
309
310
- translator_loop(&loongarch_tr_ops, &ctx.base, cs, tb, max_insns);
311
+ translator_loop(cs, tb, max_insns, pc, host_pc,
312
+ &loongarch_tr_ops, &ctx.base);
313
}
314
315
void loongarch_translate_init(void)
316
diff --git a/target/m68k/translate.c b/target/m68k/translate.c
317
index XXXXXXX..XXXXXXX 100644
318
--- a/target/m68k/translate.c
319
+++ b/target/m68k/translate.c
320
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps m68k_tr_ops = {
321
.disas_log = m68k_tr_disas_log,
322
};
323
324
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
325
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
326
+ target_ulong pc, void *host_pc)
327
{
328
DisasContext dc;
329
- translator_loop(&m68k_tr_ops, &dc.base, cpu, tb, max_insns);
330
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &m68k_tr_ops, &dc.base);
331
}
332
333
static double floatx80_to_double(CPUM68KState *env, uint16_t high, uint64_t low)
334
diff --git a/target/microblaze/translate.c b/target/microblaze/translate.c
335
index XXXXXXX..XXXXXXX 100644
336
--- a/target/microblaze/translate.c
337
+++ b/target/microblaze/translate.c
338
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mb_tr_ops = {
339
.disas_log = mb_tr_disas_log,
340
};
341
342
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
343
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
344
+ target_ulong pc, void *host_pc)
345
{
346
DisasContext dc;
347
- translator_loop(&mb_tr_ops, &dc.base, cpu, tb, max_insns);
348
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &mb_tr_ops, &dc.base);
349
}
350
351
void mb_cpu_dump_state(CPUState *cs, FILE *f, int flags)
352
diff --git a/target/mips/tcg/translate.c b/target/mips/tcg/translate.c
353
index XXXXXXX..XXXXXXX 100644
354
--- a/target/mips/tcg/translate.c
355
+++ b/target/mips/tcg/translate.c
356
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mips_tr_ops = {
357
.disas_log = mips_tr_disas_log,
358
};
359
360
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
361
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
362
+ target_ulong pc, void *host_pc)
363
{
364
DisasContext ctx;
365
366
- translator_loop(&mips_tr_ops, &ctx.base, cs, tb, max_insns);
367
+ translator_loop(cs, tb, max_insns, pc, host_pc, &mips_tr_ops, &ctx.base);
368
}
369
370
void mips_tcg_init(void)
371
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
372
index XXXXXXX..XXXXXXX 100644
373
--- a/target/nios2/translate.c
374
+++ b/target/nios2/translate.c
375
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps nios2_tr_ops = {
376
.disas_log = nios2_tr_disas_log,
377
};
378
379
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
380
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
381
+ target_ulong pc, void *host_pc)
382
{
383
DisasContext dc;
384
- translator_loop(&nios2_tr_ops, &dc.base, cs, tb, max_insns);
385
+ translator_loop(cs, tb, max_insns, pc, host_pc, &nios2_tr_ops, &dc.base);
386
}
387
388
void nios2_cpu_dump_state(CPUState *cs, FILE *f, int flags)
389
diff --git a/target/openrisc/translate.c b/target/openrisc/translate.c
390
index XXXXXXX..XXXXXXX 100644
391
--- a/target/openrisc/translate.c
392
+++ b/target/openrisc/translate.c
393
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps openrisc_tr_ops = {
394
.disas_log = openrisc_tr_disas_log,
395
};
396
397
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
398
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
399
+ target_ulong pc, void *host_pc)
400
{
401
DisasContext ctx;
402
403
- translator_loop(&openrisc_tr_ops, &ctx.base, cs, tb, max_insns);
404
+ translator_loop(cs, tb, max_insns, pc, host_pc,
405
+ &openrisc_tr_ops, &ctx.base);
406
}
407
408
void openrisc_cpu_dump_state(CPUState *cs, FILE *f, int flags)
409
diff --git a/target/ppc/translate.c b/target/ppc/translate.c
410
index XXXXXXX..XXXXXXX 100644
411
--- a/target/ppc/translate.c
412
+++ b/target/ppc/translate.c
413
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps ppc_tr_ops = {
414
.disas_log = ppc_tr_disas_log,
415
};
416
417
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
418
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
419
+ target_ulong pc, void *host_pc)
420
{
421
DisasContext ctx;
422
423
- translator_loop(&ppc_tr_ops, &ctx.base, cs, tb, max_insns);
424
+ translator_loop(cs, tb, max_insns, pc, host_pc, &ppc_tr_ops, &ctx.base);
425
}
426
427
void restore_state_to_opc(CPUPPCState *env, TranslationBlock *tb,
428
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
429
index XXXXXXX..XXXXXXX 100644
430
--- a/target/riscv/translate.c
431
+++ b/target/riscv/translate.c
432
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps riscv_tr_ops = {
433
.disas_log = riscv_tr_disas_log,
434
};
435
436
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
437
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
438
+ target_ulong pc, void *host_pc)
439
{
440
DisasContext ctx;
441
442
- translator_loop(&riscv_tr_ops, &ctx.base, cs, tb, max_insns);
443
+ translator_loop(cs, tb, max_insns, pc, host_pc, &riscv_tr_ops, &ctx.base);
444
}
445
446
void riscv_translate_init(void)
447
diff --git a/target/rx/translate.c b/target/rx/translate.c
448
index XXXXXXX..XXXXXXX 100644
449
--- a/target/rx/translate.c
450
+++ b/target/rx/translate.c
451
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps rx_tr_ops = {
452
.disas_log = rx_tr_disas_log,
453
};
454
455
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
456
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
457
+ target_ulong pc, void *host_pc)
458
{
459
DisasContext dc;
460
461
- translator_loop(&rx_tr_ops, &dc.base, cs, tb, max_insns);
462
+ translator_loop(cs, tb, max_insns, pc, host_pc, &rx_tr_ops, &dc.base);
463
}
464
465
void restore_state_to_opc(CPURXState *env, TranslationBlock *tb,
466
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
467
index XXXXXXX..XXXXXXX 100644
468
--- a/target/s390x/tcg/translate.c
469
+++ b/target/s390x/tcg/translate.c
470
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps s390x_tr_ops = {
471
.disas_log = s390x_tr_disas_log,
472
};
473
474
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
475
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
476
+ target_ulong pc, void *host_pc)
477
{
478
DisasContext dc;
479
480
- translator_loop(&s390x_tr_ops, &dc.base, cs, tb, max_insns);
481
+ translator_loop(cs, tb, max_insns, pc, host_pc, &s390x_tr_ops, &dc.base);
482
}
483
484
void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb,
485
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
486
index XXXXXXX..XXXXXXX 100644
487
--- a/target/sh4/translate.c
488
+++ b/target/sh4/translate.c
489
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sh4_tr_ops = {
490
.disas_log = sh4_tr_disas_log,
491
};
492
493
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
494
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
495
+ target_ulong pc, void *host_pc)
496
{
497
DisasContext ctx;
498
499
- translator_loop(&sh4_tr_ops, &ctx.base, cs, tb, max_insns);
500
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base);
501
}
502
503
void restore_state_to_opc(CPUSH4State *env, TranslationBlock *tb,
504
diff --git a/target/sparc/translate.c b/target/sparc/translate.c
505
index XXXXXXX..XXXXXXX 100644
506
--- a/target/sparc/translate.c
507
+++ b/target/sparc/translate.c
508
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sparc_tr_ops = {
509
.disas_log = sparc_tr_disas_log,
510
};
511
512
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
513
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
514
+ target_ulong pc, void *host_pc)
515
{
516
DisasContext dc = {};
517
518
- translator_loop(&sparc_tr_ops, &dc.base, cs, tb, max_insns);
519
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sparc_tr_ops, &dc.base);
520
}
521
522
void sparc_tcg_init(void)
523
diff --git a/target/tricore/translate.c b/target/tricore/translate.c
524
index XXXXXXX..XXXXXXX 100644
525
--- a/target/tricore/translate.c
526
+++ b/target/tricore/translate.c
527
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps tricore_tr_ops = {
528
};
529
530
531
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
532
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
533
+ target_ulong pc, void *host_pc)
534
{
535
DisasContext ctx;
536
- translator_loop(&tricore_tr_ops, &ctx.base, cs, tb, max_insns);
537
+ translator_loop(cs, tb, max_insns, pc, host_pc,
538
+ &tricore_tr_ops, &ctx.base);
539
}
540
541
void
542
diff --git a/target/xtensa/translate.c b/target/xtensa/translate.c
543
index XXXXXXX..XXXXXXX 100644
544
--- a/target/xtensa/translate.c
545
+++ b/target/xtensa/translate.c
546
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps xtensa_translator_ops = {
547
.disas_log = xtensa_tr_disas_log,
548
};
549
550
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
551
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
552
+ target_ulong pc, void *host_pc)
553
{
554
DisasContext dc = {};
555
- translator_loop(&xtensa_translator_ops, &dc.base, cpu, tb, max_insns);
556
+ translator_loop(cpu, tb, max_insns, pc, host_pc,
557
+ &xtensa_translator_ops, &dc.base);
558
}
559
560
void xtensa_cpu_dump_state(CPUState *cs, FILE *f, int flags)
74
--
561
--
75
2.25.1
562
2.34.1
diff view generated by jsdifflib
1
AArch64 has both sign and zero-extending addressing modes, which
1
Cache the translation from guest to host address, so we may
2
means that either treatment of guest addresses is equally efficient.
2
use direct loads when we hit on the primary translation page.
3
Enabling this for AArch64 gives us testing of the feature in CI.
4
3
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Look up the second translation page only once, during translation.
5
This obviates another lookup of the second page within tb_gen_code
6
after translation.
7
8
Fixes a bug in that plugin_insn_append should be passed the bytes
9
in the original memory order, not bswapped by pieces.
10
11
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
12
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
14
---
8
tcg/aarch64/tcg-target-sa32.h | 8 +++-
15
include/exec/translator.h | 63 +++++++++++--------
9
tcg/aarch64/tcg-target.c.inc | 81 ++++++++++++++++++++++++-----------
16
accel/tcg/translate-all.c | 23 +++----
10
2 files changed, 64 insertions(+), 25 deletions(-)
17
accel/tcg/translator.c | 126 +++++++++++++++++++++++++++++---------
18
3 files changed, 141 insertions(+), 71 deletions(-)
11
19
12
diff --git a/tcg/aarch64/tcg-target-sa32.h b/tcg/aarch64/tcg-target-sa32.h
20
diff --git a/include/exec/translator.h b/include/exec/translator.h
13
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
14
--- a/tcg/aarch64/tcg-target-sa32.h
22
--- a/include/exec/translator.h
15
+++ b/tcg/aarch64/tcg-target-sa32.h
23
+++ b/include/exec/translator.h
16
@@ -1 +1,7 @@
24
@@ -XXX,XX +XXX,XX @@ typedef enum DisasJumpType {
17
-#define TCG_TARGET_SIGNED_ADDR32 0
25
* Architecture-agnostic disassembly context.
18
+/*
26
*/
19
+ * AArch64 has both SXTW and UXTW addressing modes, which means that
27
typedef struct DisasContextBase {
20
+ * it is agnostic to how guest addresses should be represented.
28
- const TranslationBlock *tb;
21
+ * Because aarch64 is more common than the other hosts that will
29
+ TranslationBlock *tb;
22
+ * want to use this feature, enable it for continuous testing.
30
target_ulong pc_first;
23
+ */
31
target_ulong pc_next;
24
+#define TCG_TARGET_SIGNED_ADDR32 1
32
DisasJumpType is_jmp;
25
diff --git a/tcg/aarch64/tcg-target.c.inc b/tcg/aarch64/tcg-target.c.inc
33
int num_insns;
34
int max_insns;
35
bool singlestep_enabled;
36
-#ifdef CONFIG_USER_ONLY
37
- /*
38
- * Guest address of the last byte of the last protected page.
39
- *
40
- * Pages containing the translated instructions are made non-writable in
41
- * order to achieve consistency in case another thread is modifying the
42
- * code while translate_insn() fetches the instruction bytes piecemeal.
43
- * Such writer threads are blocked on mmap_lock() in page_unprotect().
44
- */
45
- target_ulong page_protect_end;
46
-#endif
47
+ void *host_addr[2];
48
} DisasContextBase;
49
50
/**
51
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
52
* the relevant information at translation time.
53
*/
54
55
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
56
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
57
- abi_ptr pc, bool do_swap); \
58
- static inline type fullname(CPUArchState *env, \
59
- DisasContextBase *dcbase, abi_ptr pc) \
60
- { \
61
- return fullname ## _swap(env, dcbase, pc, false); \
62
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
63
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
64
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
65
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
66
+
67
+static inline uint16_t
68
+translator_lduw_swap(CPUArchState *env, DisasContextBase *db,
69
+ abi_ptr pc, bool do_swap)
70
+{
71
+ uint16_t ret = translator_lduw(env, db, pc);
72
+ if (do_swap) {
73
+ ret = bswap16(ret);
74
}
75
+ return ret;
76
+}
77
78
-#define FOR_EACH_TRANSLATOR_LD(F) \
79
- F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
80
- F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
81
- F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
82
- F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
83
+static inline uint32_t
84
+translator_ldl_swap(CPUArchState *env, DisasContextBase *db,
85
+ abi_ptr pc, bool do_swap)
86
+{
87
+ uint32_t ret = translator_ldl(env, db, pc);
88
+ if (do_swap) {
89
+ ret = bswap32(ret);
90
+ }
91
+ return ret;
92
+}
93
94
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
95
-
96
-#undef GEN_TRANSLATOR_LD
97
+static inline uint64_t
98
+translator_ldq_swap(CPUArchState *env, DisasContextBase *db,
99
+ abi_ptr pc, bool do_swap)
100
+{
101
+ uint64_t ret = translator_ldq_swap(env, db, pc, false);
102
+ if (do_swap) {
103
+ ret = bswap64(ret);
104
+ }
105
+ return ret;
106
+}
107
108
/*
109
* Return whether addr is on the same page as where disassembly started.
110
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
26
index XXXXXXX..XXXXXXX 100644
111
index XXXXXXX..XXXXXXX 100644
27
--- a/tcg/aarch64/tcg-target.c.inc
112
--- a/accel/tcg/translate-all.c
28
+++ b/tcg/aarch64/tcg-target.c.inc
113
+++ b/accel/tcg/translate-all.c
29
@@ -XXX,XX +XXX,XX @@ typedef enum {
114
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
30
LDST_LD_S_W = 3, /* load and sign-extend into Wt */
115
{
31
} AArch64LdstType;
116
CPUArchState *env = cpu->env_ptr;
32
117
TranslationBlock *tb, *existing_tb;
33
+/*
118
- tb_page_addr_t phys_pc, phys_page2;
34
+ * See aarch64/instrs/extendreg/DecodeRegExtend
119
- target_ulong virt_page2;
35
+ * But note that option<1> == 0 is UNDEFINED for LDR/STR.
120
+ tb_page_addr_t phys_pc;
36
+ */
121
tcg_insn_unit *gen_code_buf;
37
+typedef enum {
122
int gen_code_size, search_size, max_insns;
38
+ LDST_EXT_UXTW = 2, /* zero-extend from uint32_t */
123
#ifdef CONFIG_PROFILER
39
+ LDST_EXT_UXTX = 3, /* zero-extend from uint64_t (i.e. no extension) */
124
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
40
+ LDST_EXT_SXTW = 6, /* sign-extend from int32_t */
125
tb->flags = flags;
41
+} AArch64LdstExt;
126
tb->cflags = cflags;
42
+
127
tb->trace_vcpu_dstate = *cpu->trace_dstate;
43
/* We encode the format of the insn into the beginning of the name, so that
128
+ tb->page_addr[0] = phys_pc;
44
we can have the preprocessor help "typecheck" the insn vs the output
129
+ tb->page_addr[1] = -1;
45
function. Arm didn't provide us with nice names for the formats, so we
130
tcg_ctx->tb_cflags = cflags;
46
@@ -XXX,XX +XXX,XX @@ static void tcg_out_insn_3617(TCGContext *s, AArch64Insn insn, bool q,
131
tb_overflow:
132
133
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
134
}
135
136
/*
137
- * If the TB is not associated with a physical RAM page then
138
- * it must be a temporary one-insn TB, and we have nothing to do
139
- * except fill in the page_addr[] fields. Return early before
140
- * attempting to link to other TBs or add to the lookup table.
141
+ * If the TB is not associated with a physical RAM page then it must be
142
+ * a temporary one-insn TB, and we have nothing left to do. Return early
143
+ * before attempting to link to other TBs or add to the lookup table.
144
*/
145
- if (phys_pc == -1) {
146
- tb->page_addr[0] = tb->page_addr[1] = -1;
147
+ if (tb->page_addr[0] == -1) {
148
return tb;
149
}
150
151
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
152
*/
153
tcg_tb_insert(tb);
154
155
- /* check next page if needed */
156
- virt_page2 = (pc + tb->size - 1) & TARGET_PAGE_MASK;
157
- phys_page2 = -1;
158
- if ((pc & TARGET_PAGE_MASK) != virt_page2) {
159
- phys_page2 = get_page_addr_code(env, virt_page2);
160
- }
161
/*
162
* No explicit memory barrier is required -- tb_link_page() makes the
163
* TB visible in a consistent state.
164
*/
165
- existing_tb = tb_link_page(tb, phys_pc, phys_page2);
166
+ existing_tb = tb_link_page(tb, tb->page_addr[0], tb->page_addr[1]);
167
/* if the TB already exists, discard what we just translated */
168
if (unlikely(existing_tb != tb)) {
169
uintptr_t orig_aligned = (uintptr_t)gen_code_buf;
170
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
171
index XXXXXXX..XXXXXXX 100644
172
--- a/accel/tcg/translator.c
173
+++ b/accel/tcg/translator.c
174
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest)
175
return ((db->pc_first ^ dest) & TARGET_PAGE_MASK) == 0;
47
}
176
}
48
177
49
static void tcg_out_insn_3310(TCGContext *s, AArch64Insn insn,
178
-static inline void translator_page_protect(DisasContextBase *dcbase,
50
- TCGReg rd, TCGReg base, TCGType ext,
179
- target_ulong pc)
51
+ TCGReg rd, TCGReg base, AArch64LdstExt option,
180
-{
52
TCGReg regoff)
181
-#ifdef CONFIG_USER_ONLY
182
- dcbase->page_protect_end = pc | ~TARGET_PAGE_MASK;
183
- page_protect(pc);
184
-#endif
185
-}
186
-
187
void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
188
target_ulong pc, void *host_pc,
189
const TranslatorOps *ops, DisasContextBase *db)
190
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
191
db->num_insns = 0;
192
db->max_insns = max_insns;
193
db->singlestep_enabled = cflags & CF_SINGLE_STEP;
194
- translator_page_protect(db, db->pc_next);
195
+ db->host_addr[0] = host_pc;
196
+ db->host_addr[1] = NULL;
197
+
198
+#ifdef CONFIG_USER_ONLY
199
+ page_protect(pc);
200
+#endif
201
202
ops->init_disas_context(db, cpu);
203
tcg_debug_assert(db->is_jmp == DISAS_NEXT); /* no early exit */
204
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
205
#endif
206
}
207
208
-static inline void translator_maybe_page_protect(DisasContextBase *dcbase,
209
- target_ulong pc, size_t len)
210
+static void *translator_access(CPUArchState *env, DisasContextBase *db,
211
+ target_ulong pc, size_t len)
53
{
212
{
54
/* Note the AArch64Insn constants above are for C3.3.12. Adjust. */
213
-#ifdef CONFIG_USER_ONLY
55
tcg_out32(s, insn | I3312_TO_I3310 | regoff << 16 |
214
- target_ulong end = pc + len - 1;
56
- 0x4000 | ext << 13 | base << 5 | (rd & 0x1f));
215
+ void *host;
57
+ option << 13 | base << 5 | (rd & 0x1f));
216
+ target_ulong base, end;
217
+ TranslationBlock *tb;
218
219
- if (end > dcbase->page_protect_end) {
220
- translator_page_protect(dcbase, end);
221
+ tb = db->tb;
222
+
223
+ /* Use slow path if first page is MMIO. */
224
+ if (unlikely(tb->page_addr[0] == -1)) {
225
+ return NULL;
226
}
227
+
228
+ end = pc + len - 1;
229
+ if (likely(is_same_page(db, end))) {
230
+ host = db->host_addr[0];
231
+ base = db->pc_first;
232
+ } else {
233
+ host = db->host_addr[1];
234
+ base = TARGET_PAGE_ALIGN(db->pc_first);
235
+ if (host == NULL) {
236
+ tb->page_addr[1] =
237
+ get_page_addr_code_hostp(env, base, &db->host_addr[1]);
238
+#ifdef CONFIG_USER_ONLY
239
+ page_protect(end);
240
#endif
241
+ /* We cannot handle MMIO as second page. */
242
+ assert(tb->page_addr[1] != -1);
243
+ host = db->host_addr[1];
244
+ }
245
+
246
+ /* Use slow path when crossing pages. */
247
+ if (is_same_page(db, pc)) {
248
+ return NULL;
249
+ }
250
+ }
251
+
252
+ tcg_debug_assert(pc >= base);
253
+ return host + (pc - base);
58
}
254
}
59
255
60
static void tcg_out_insn_3312(TCGContext *s, AArch64Insn insn,
256
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
61
@@ -XXX,XX +XXX,XX @@ static void tcg_out_ldst(TCGContext *s, AArch64Insn insn, TCGReg rd,
257
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
62
258
- abi_ptr pc, bool do_swap) \
63
/* Worst-case scenario, move offset to temp register, use reg offset. */
259
- { \
64
tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP, offset);
260
- translator_maybe_page_protect(dcbase, pc, sizeof(type)); \
65
- tcg_out_ldst_r(s, insn, rd, rn, TCG_TYPE_I64, TCG_REG_TMP);
261
- type ret = load_fn(env, pc); \
66
+ tcg_out_ldst_r(s, insn, rd, rn, LDST_EXT_UXTX, TCG_REG_TMP);
262
- if (do_swap) { \
67
}
263
- ret = swap_fn(ret); \
68
264
- } \
69
static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg)
265
- plugin_insn_append(pc, &ret, sizeof(ret)); \
70
@@ -XXX,XX +XXX,XX @@ static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *l)
266
- return ret; \
71
267
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
72
static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
268
+{
73
TCGReg data_r, TCGReg addr_r,
269
+ uint8_t ret;
74
- TCGType otype, TCGReg off_r)
270
+ void *p = translator_access(env, db, pc, sizeof(ret));
75
+ AArch64LdstExt option, TCGReg off_r)
271
+
76
{
272
+ if (p) {
77
switch (memop & MO_SSIZE) {
273
+ plugin_insn_append(pc, p, sizeof(ret));
78
case MO_UB:
274
+ return ldub_p(p);
79
- tcg_out_ldst_r(s, I3312_LDRB, data_r, addr_r, otype, off_r);
275
}
80
+ tcg_out_ldst_r(s, I3312_LDRB, data_r, addr_r, option, off_r);
276
+ ret = cpu_ldub_code(env, pc);
81
break;
277
+ plugin_insn_append(pc, &ret, sizeof(ret));
82
case MO_SB:
278
+ return ret;
83
tcg_out_ldst_r(s, ext ? I3312_LDRSBX : I3312_LDRSBW,
279
+}
84
- data_r, addr_r, otype, off_r);
280
85
+ data_r, addr_r, option, off_r);
281
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
86
break;
282
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
87
case MO_UW:
283
+{
88
- tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
284
+ uint16_t ret, plug;
89
+ tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, option, off_r);
285
+ void *p = translator_access(env, db, pc, sizeof(ret));
90
break;
286
91
case MO_SW:
287
-#undef GEN_TRANSLATOR_LD
92
tcg_out_ldst_r(s, (ext ? I3312_LDRSHX : I3312_LDRSHW),
288
+ if (p) {
93
- data_r, addr_r, otype, off_r);
289
+ plugin_insn_append(pc, p, sizeof(ret));
94
+ data_r, addr_r, option, off_r);
290
+ return lduw_p(p);
95
break;
291
+ }
96
case MO_UL:
292
+ ret = cpu_lduw_code(env, pc);
97
- tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
293
+ plug = tswap16(ret);
98
+ tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, option, off_r);
294
+ plugin_insn_append(pc, &plug, sizeof(ret));
99
break;
295
+ return ret;
100
case MO_SL:
296
+}
101
- tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, otype, off_r);
297
+
102
+ tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, option, off_r);
298
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
103
break;
299
+{
104
case MO_UQ:
300
+ uint32_t ret, plug;
105
- tcg_out_ldst_r(s, I3312_LDRX, data_r, addr_r, otype, off_r);
301
+ void *p = translator_access(env, db, pc, sizeof(ret));
106
+ tcg_out_ldst_r(s, I3312_LDRX, data_r, addr_r, option, off_r);
302
+
107
break;
303
+ if (p) {
108
default:
304
+ plugin_insn_append(pc, p, sizeof(ret));
109
tcg_abort();
305
+ return ldl_p(p);
110
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
306
+ }
111
307
+ ret = cpu_ldl_code(env, pc);
112
static void tcg_out_qemu_st_direct(TCGContext *s, MemOp memop,
308
+ plug = tswap32(ret);
113
TCGReg data_r, TCGReg addr_r,
309
+ plugin_insn_append(pc, &plug, sizeof(ret));
114
- TCGType otype, TCGReg off_r)
310
+ return ret;
115
+ AArch64LdstExt option, TCGReg off_r)
311
+}
116
{
312
+
117
switch (memop & MO_SIZE) {
313
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
118
case MO_8:
314
+{
119
- tcg_out_ldst_r(s, I3312_STRB, data_r, addr_r, otype, off_r);
315
+ uint64_t ret, plug;
120
+ tcg_out_ldst_r(s, I3312_STRB, data_r, addr_r, option, off_r);
316
+ void *p = translator_access(env, db, pc, sizeof(ret));
121
break;
317
+
122
case MO_16:
318
+ if (p) {
123
- tcg_out_ldst_r(s, I3312_STRH, data_r, addr_r, otype, off_r);
319
+ plugin_insn_append(pc, p, sizeof(ret));
124
+ tcg_out_ldst_r(s, I3312_STRH, data_r, addr_r, option, off_r);
320
+ return ldq_p(p);
125
break;
321
+ }
126
case MO_32:
322
+ ret = cpu_ldq_code(env, pc);
127
- tcg_out_ldst_r(s, I3312_STRW, data_r, addr_r, otype, off_r);
323
+ plug = tswap64(ret);
128
+ tcg_out_ldst_r(s, I3312_STRW, data_r, addr_r, option, off_r);
324
+ plugin_insn_append(pc, &plug, sizeof(ret));
129
break;
325
+ return ret;
130
case MO_64:
326
+}
131
- tcg_out_ldst_r(s, I3312_STRX, data_r, addr_r, otype, off_r);
132
+ tcg_out_ldst_r(s, I3312_STRX, data_r, addr_r, option, off_r);
133
break;
134
default:
135
tcg_abort();
136
}
137
}
138
139
+/*
140
+ * Bits for the option field of LDR/STR (register),
141
+ * for application to a guest address.
142
+ */
143
+static AArch64LdstExt ldst_ext_option(void)
144
+{
145
+#ifdef CONFIG_USER_ONLY
146
+ bool signed_addr32 = guest_base_signed_addr32;
147
+#else
148
+ bool signed_addr32 = TCG_TARGET_SIGNED_ADDR32;
149
+#endif
150
+
151
+ if (TARGET_LONG_BITS == 64) {
152
+ return LDST_EXT_UXTX;
153
+ } else if (signed_addr32) {
154
+ return LDST_EXT_SXTW;
155
+ } else {
156
+ return LDST_EXT_UXTW;
157
+ }
158
+}
159
+
160
static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
161
MemOpIdx oi, TCGType ext)
162
{
163
MemOp memop = get_memop(oi);
164
- const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
165
+ AArch64LdstExt option = ldst_ext_option();
166
167
/* Byte swapping is left to middle-end expansion. */
168
tcg_debug_assert((memop & MO_BSWAP) == 0);
169
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
170
171
tcg_out_tlb_read(s, addr_reg, memop, &label_ptr, mem_index, 1);
172
tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
173
- TCG_REG_X1, otype, addr_reg);
174
+ TCG_REG_X1, option, addr_reg);
175
add_qemu_ldst_label(s, true, oi, ext, data_reg, addr_reg,
176
s->code_ptr, label_ptr);
177
#else /* !CONFIG_SOFTMMU */
178
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
179
}
180
if (USE_GUEST_BASE) {
181
tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
182
- TCG_REG_GUEST_BASE, otype, addr_reg);
183
+ TCG_REG_GUEST_BASE, option, addr_reg);
184
} else {
185
+ /* This case is always a 64-bit guest with no extension. */
186
tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
187
- addr_reg, TCG_TYPE_I64, TCG_REG_XZR);
188
+ addr_reg, LDST_EXT_UXTX, TCG_REG_XZR);
189
}
190
#endif /* CONFIG_SOFTMMU */
191
}
192
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
193
MemOpIdx oi)
194
{
195
MemOp memop = get_memop(oi);
196
- const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
197
+ AArch64LdstExt option = ldst_ext_option();
198
199
/* Byte swapping is left to middle-end expansion. */
200
tcg_debug_assert((memop & MO_BSWAP) == 0);
201
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
202
203
tcg_out_tlb_read(s, addr_reg, memop, &label_ptr, mem_index, 0);
204
tcg_out_qemu_st_direct(s, memop, data_reg,
205
- TCG_REG_X1, otype, addr_reg);
206
+ TCG_REG_X1, option, addr_reg);
207
add_qemu_ldst_label(s, false, oi, (memop & MO_SIZE)== MO_64,
208
data_reg, addr_reg, s->code_ptr, label_ptr);
209
#else /* !CONFIG_SOFTMMU */
210
@@ -XXX,XX +XXX,XX @@ static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
211
}
212
if (USE_GUEST_BASE) {
213
tcg_out_qemu_st_direct(s, memop, data_reg,
214
- TCG_REG_GUEST_BASE, otype, addr_reg);
215
+ TCG_REG_GUEST_BASE, option, addr_reg);
216
} else {
217
+ /* This case is always a 64-bit guest with no extension. */
218
tcg_out_qemu_st_direct(s, memop, data_reg,
219
- addr_reg, TCG_TYPE_I64, TCG_REG_XZR);
220
+ addr_reg, LDST_EXT_UXTX, TCG_REG_XZR);
221
}
222
#endif /* CONFIG_SOFTMMU */
223
}
224
--
327
--
225
2.25.1
328
2.34.1
diff view generated by jsdifflib
1
Define as 0 for all tcg hosts. Put this in a separate header,
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
because we'll want this in places that do not ordinarily have
3
access to all of tcg/tcg.h.
4
2
5
Reviewed-by: WANG Xuerui <git@xen0n.name>
3
Right now translator stops right *after* the end of a page, which
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
4
breaks reporting of fault locations when the last instruction of a
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
multi-insn translation block crosses a page boundary.
8
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-Id: <20220817150506.592862-3-iii@linux.ibm.com>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
11
---
11
tcg/aarch64/tcg-target-sa32.h | 1 +
12
target/s390x/tcg/translate.c | 15 +++-
12
tcg/arm/tcg-target-sa32.h | 1 +
13
tests/tcg/s390x/noexec.c | 106 +++++++++++++++++++++++
13
tcg/i386/tcg-target-sa32.h | 1 +
14
tests/tcg/multiarch/noexec.c.inc | 139 +++++++++++++++++++++++++++++++
14
tcg/loongarch64/tcg-target-sa32.h | 1 +
15
tests/tcg/s390x/Makefile.target | 1 +
15
tcg/mips/tcg-target-sa32.h | 1 +
16
4 files changed, 257 insertions(+), 4 deletions(-)
16
tcg/ppc/tcg-target-sa32.h | 1 +
17
create mode 100644 tests/tcg/s390x/noexec.c
17
tcg/riscv/tcg-target-sa32.h | 1 +
18
create mode 100644 tests/tcg/multiarch/noexec.c.inc
18
tcg/s390x/tcg-target-sa32.h | 1 +
19
tcg/sparc/tcg-target-sa32.h | 1 +
20
tcg/tci/tcg-target-sa32.h | 1 +
21
tcg/tcg.c | 4 ++++
22
11 files changed, 14 insertions(+)
23
create mode 100644 tcg/aarch64/tcg-target-sa32.h
24
create mode 100644 tcg/arm/tcg-target-sa32.h
25
create mode 100644 tcg/i386/tcg-target-sa32.h
26
create mode 100644 tcg/loongarch64/tcg-target-sa32.h
27
create mode 100644 tcg/mips/tcg-target-sa32.h
28
create mode 100644 tcg/ppc/tcg-target-sa32.h
29
create mode 100644 tcg/riscv/tcg-target-sa32.h
30
create mode 100644 tcg/s390x/tcg-target-sa32.h
31
create mode 100644 tcg/sparc/tcg-target-sa32.h
32
create mode 100644 tcg/tci/tcg-target-sa32.h
33
19
34
diff --git a/tcg/aarch64/tcg-target-sa32.h b/tcg/aarch64/tcg-target-sa32.h
20
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/target/s390x/tcg/translate.c
23
+++ b/target/s390x/tcg/translate.c
24
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
25
dc->insn_start = tcg_last_op();
26
}
27
28
+static target_ulong get_next_pc(CPUS390XState *env, DisasContext *s,
29
+ uint64_t pc)
30
+{
31
+ uint64_t insn = ld_code2(env, s, pc);
32
+
33
+ return pc + get_ilen((insn >> 8) & 0xff);
34
+}
35
+
36
static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
37
{
38
CPUS390XState *env = cs->env_ptr;
39
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
40
41
dc->base.is_jmp = translate_one(env, dc);
42
if (dc->base.is_jmp == DISAS_NEXT) {
43
- uint64_t page_start;
44
-
45
- page_start = dc->base.pc_first & TARGET_PAGE_MASK;
46
- if (dc->base.pc_next - page_start >= TARGET_PAGE_SIZE || dc->ex_value) {
47
+ if (!is_same_page(dcbase, dc->base.pc_next) ||
48
+ !is_same_page(dcbase, get_next_pc(env, dc, dc->base.pc_next)) ||
49
+ dc->ex_value) {
50
dc->base.is_jmp = DISAS_TOO_MANY;
51
}
52
}
53
diff --git a/tests/tcg/s390x/noexec.c b/tests/tcg/s390x/noexec.c
35
new file mode 100644
54
new file mode 100644
36
index XXXXXXX..XXXXXXX
55
index XXXXXXX..XXXXXXX
37
--- /dev/null
56
--- /dev/null
38
+++ b/tcg/aarch64/tcg-target-sa32.h
57
+++ b/tests/tcg/s390x/noexec.c
39
@@ -0,0 +1 @@
58
@@ -XXX,XX +XXX,XX @@
40
+#define TCG_TARGET_SIGNED_ADDR32 0
59
+#include "../multiarch/noexec.c.inc"
41
diff --git a/tcg/arm/tcg-target-sa32.h b/tcg/arm/tcg-target-sa32.h
60
+
61
+static void *arch_mcontext_pc(const mcontext_t *ctx)
62
+{
63
+ return (void *)ctx->psw.addr;
64
+}
65
+
66
+static int arch_mcontext_arg(const mcontext_t *ctx)
67
+{
68
+ return ctx->gregs[2];
69
+}
70
+
71
+static void arch_flush(void *p, int len)
72
+{
73
+}
74
+
75
+extern char noexec_1[];
76
+extern char noexec_2[];
77
+extern char noexec_end[];
78
+
79
+asm("noexec_1:\n"
80
+ " lgfi %r2,1\n" /* %r2 is 0 on entry, set 1. */
81
+ "noexec_2:\n"
82
+ " lgfi %r2,2\n" /* %r2 is 0/1; set 2. */
83
+ " br %r14\n" /* return */
84
+ "noexec_end:");
85
+
86
+extern char exrl_1[];
87
+extern char exrl_2[];
88
+extern char exrl_end[];
89
+
90
+asm("exrl_1:\n"
91
+ " exrl %r0, exrl_2\n"
92
+ " br %r14\n"
93
+ "exrl_2:\n"
94
+ " lgfi %r2,2\n"
95
+ "exrl_end:");
96
+
97
+int main(void)
98
+{
99
+ struct noexec_test noexec_tests[] = {
100
+ {
101
+ .name = "fallthrough",
102
+ .test_code = noexec_1,
103
+ .test_len = noexec_end - noexec_1,
104
+ .page_ofs = noexec_1 - noexec_2,
105
+ .entry_ofs = noexec_1 - noexec_2,
106
+ .expected_si_ofs = 0,
107
+ .expected_pc_ofs = 0,
108
+ .expected_arg = 1,
109
+ },
110
+ {
111
+ .name = "jump",
112
+ .test_code = noexec_1,
113
+ .test_len = noexec_end - noexec_1,
114
+ .page_ofs = noexec_1 - noexec_2,
115
+ .entry_ofs = 0,
116
+ .expected_si_ofs = 0,
117
+ .expected_pc_ofs = 0,
118
+ .expected_arg = 0,
119
+ },
120
+ {
121
+ .name = "exrl",
122
+ .test_code = exrl_1,
123
+ .test_len = exrl_end - exrl_1,
124
+ .page_ofs = exrl_1 - exrl_2,
125
+ .entry_ofs = exrl_1 - exrl_2,
126
+ .expected_si_ofs = 0,
127
+ .expected_pc_ofs = exrl_1 - exrl_2,
128
+ .expected_arg = 0,
129
+ },
130
+ {
131
+ .name = "fallthrough [cross]",
132
+ .test_code = noexec_1,
133
+ .test_len = noexec_end - noexec_1,
134
+ .page_ofs = noexec_1 - noexec_2 - 2,
135
+ .entry_ofs = noexec_1 - noexec_2 - 2,
136
+ .expected_si_ofs = 0,
137
+ .expected_pc_ofs = -2,
138
+ .expected_arg = 1,
139
+ },
140
+ {
141
+ .name = "jump [cross]",
142
+ .test_code = noexec_1,
143
+ .test_len = noexec_end - noexec_1,
144
+ .page_ofs = noexec_1 - noexec_2 - 2,
145
+ .entry_ofs = -2,
146
+ .expected_si_ofs = 0,
147
+ .expected_pc_ofs = -2,
148
+ .expected_arg = 0,
149
+ },
150
+ {
151
+ .name = "exrl [cross]",
152
+ .test_code = exrl_1,
153
+ .test_len = exrl_end - exrl_1,
154
+ .page_ofs = exrl_1 - exrl_2 - 2,
155
+ .entry_ofs = exrl_1 - exrl_2 - 2,
156
+ .expected_si_ofs = 0,
157
+ .expected_pc_ofs = exrl_1 - exrl_2 - 2,
158
+ .expected_arg = 0,
159
+ },
160
+ };
161
+
162
+ return test_noexec(noexec_tests,
163
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
164
+}
165
diff --git a/tests/tcg/multiarch/noexec.c.inc b/tests/tcg/multiarch/noexec.c.inc
42
new file mode 100644
166
new file mode 100644
43
index XXXXXXX..XXXXXXX
167
index XXXXXXX..XXXXXXX
44
--- /dev/null
168
--- /dev/null
45
+++ b/tcg/arm/tcg-target-sa32.h
169
+++ b/tests/tcg/multiarch/noexec.c.inc
46
@@ -0,0 +1 @@
170
@@ -XXX,XX +XXX,XX @@
47
+#define TCG_TARGET_SIGNED_ADDR32 0
171
+/*
48
diff --git a/tcg/i386/tcg-target-sa32.h b/tcg/i386/tcg-target-sa32.h
172
+ * Common code for arch-specific MMU_INST_FETCH fault testing.
49
new file mode 100644
173
+ */
50
index XXXXXXX..XXXXXXX
174
+
51
--- /dev/null
175
+#define _GNU_SOURCE
52
+++ b/tcg/i386/tcg-target-sa32.h
176
+
53
@@ -0,0 +1 @@
177
+#include <assert.h>
54
+#define TCG_TARGET_SIGNED_ADDR32 0
178
+#include <signal.h>
55
diff --git a/tcg/loongarch64/tcg-target-sa32.h b/tcg/loongarch64/tcg-target-sa32.h
179
+#include <stdio.h>
56
new file mode 100644
180
+#include <stdlib.h>
57
index XXXXXXX..XXXXXXX
181
+#include <string.h>
58
--- /dev/null
182
+#include <errno.h>
59
+++ b/tcg/loongarch64/tcg-target-sa32.h
183
+#include <unistd.h>
60
@@ -0,0 +1 @@
184
+#include <sys/mman.h>
61
+#define TCG_TARGET_SIGNED_ADDR32 0
185
+#include <sys/ucontext.h>
62
diff --git a/tcg/mips/tcg-target-sa32.h b/tcg/mips/tcg-target-sa32.h
186
+
63
new file mode 100644
187
+/* Forward declarations. */
64
index XXXXXXX..XXXXXXX
188
+
65
--- /dev/null
189
+static void *arch_mcontext_pc(const mcontext_t *ctx);
66
+++ b/tcg/mips/tcg-target-sa32.h
190
+static int arch_mcontext_arg(const mcontext_t *ctx);
67
@@ -0,0 +1 @@
191
+static void arch_flush(void *p, int len);
68
+#define TCG_TARGET_SIGNED_ADDR32 0
192
+
69
diff --git a/tcg/ppc/tcg-target-sa32.h b/tcg/ppc/tcg-target-sa32.h
193
+/* Testing infrastructure. */
70
new file mode 100644
194
+
71
index XXXXXXX..XXXXXXX
195
+struct noexec_test {
72
--- /dev/null
196
+ const char *name;
73
+++ b/tcg/ppc/tcg-target-sa32.h
197
+ const char *test_code;
74
@@ -0,0 +1 @@
198
+ int test_len;
75
+#define TCG_TARGET_SIGNED_ADDR32 0
199
+ int page_ofs;
76
diff --git a/tcg/riscv/tcg-target-sa32.h b/tcg/riscv/tcg-target-sa32.h
200
+ int entry_ofs;
77
new file mode 100644
201
+ int expected_si_ofs;
78
index XXXXXXX..XXXXXXX
202
+ int expected_pc_ofs;
79
--- /dev/null
203
+ int expected_arg;
80
+++ b/tcg/riscv/tcg-target-sa32.h
204
+};
81
@@ -0,0 +1 @@
205
+
82
+#define TCG_TARGET_SIGNED_ADDR32 0
206
+static void *page_base;
83
diff --git a/tcg/s390x/tcg-target-sa32.h b/tcg/s390x/tcg-target-sa32.h
207
+static int page_size;
84
new file mode 100644
208
+static const struct noexec_test *current_noexec_test;
85
index XXXXXXX..XXXXXXX
209
+
86
--- /dev/null
210
+static void handle_err(const char *syscall)
87
+++ b/tcg/s390x/tcg-target-sa32.h
211
+{
88
@@ -0,0 +1 @@
212
+ printf("[ FAILED ] %s: %s\n", syscall, strerror(errno));
89
+#define TCG_TARGET_SIGNED_ADDR32 0
213
+ exit(EXIT_FAILURE);
90
diff --git a/tcg/sparc/tcg-target-sa32.h b/tcg/sparc/tcg-target-sa32.h
214
+}
91
new file mode 100644
215
+
92
index XXXXXXX..XXXXXXX
216
+static void handle_segv(int sig, siginfo_t *info, void *ucontext)
93
--- /dev/null
217
+{
94
+++ b/tcg/sparc/tcg-target-sa32.h
218
+ const struct noexec_test *test = current_noexec_test;
95
@@ -0,0 +1 @@
219
+ const mcontext_t *mc = &((ucontext_t *)ucontext)->uc_mcontext;
96
+#define TCG_TARGET_SIGNED_ADDR32 0
220
+ void *expected_si;
97
diff --git a/tcg/tci/tcg-target-sa32.h b/tcg/tci/tcg-target-sa32.h
221
+ void *expected_pc;
98
new file mode 100644
222
+ void *pc;
99
index XXXXXXX..XXXXXXX
223
+ int arg;
100
--- /dev/null
224
+
101
+++ b/tcg/tci/tcg-target-sa32.h
225
+ if (test == NULL) {
102
@@ -0,0 +1 @@
226
+ printf("[ FAILED ] unexpected SEGV\n");
103
+#define TCG_TARGET_SIGNED_ADDR32 0
227
+ exit(EXIT_FAILURE);
104
diff --git a/tcg/tcg.c b/tcg/tcg.c
228
+ }
229
+ current_noexec_test = NULL;
230
+
231
+ expected_si = page_base + test->expected_si_ofs;
232
+ if (info->si_addr != expected_si) {
233
+ printf("[ FAILED ] wrong si_addr (%p != %p)\n",
234
+ info->si_addr, expected_si);
235
+ exit(EXIT_FAILURE);
236
+ }
237
+
238
+ pc = arch_mcontext_pc(mc);
239
+ expected_pc = page_base + test->expected_pc_ofs;
240
+ if (pc != expected_pc) {
241
+ printf("[ FAILED ] wrong pc (%p != %p)\n", pc, expected_pc);
242
+ exit(EXIT_FAILURE);
243
+ }
244
+
245
+ arg = arch_mcontext_arg(mc);
246
+ if (arg != test->expected_arg) {
247
+ printf("[ FAILED ] wrong arg (%d != %d)\n", arg, test->expected_arg);
248
+ exit(EXIT_FAILURE);
249
+ }
250
+
251
+ if (mprotect(page_base, page_size,
252
+ PROT_READ | PROT_WRITE | PROT_EXEC) < 0) {
253
+ handle_err("mprotect");
254
+ }
255
+}
256
+
257
+static void test_noexec_1(const struct noexec_test *test)
258
+{
259
+ void *start = page_base + test->page_ofs;
260
+ void (*fn)(int arg) = page_base + test->entry_ofs;
261
+
262
+ memcpy(start, test->test_code, test->test_len);
263
+ arch_flush(start, test->test_len);
264
+
265
+ /* Trigger TB creation in order to test invalidation. */
266
+ fn(0);
267
+
268
+ if (mprotect(page_base, page_size, PROT_NONE) < 0) {
269
+ handle_err("mprotect");
270
+ }
271
+
272
+ /* Trigger SEGV and check that handle_segv() ran. */
273
+ current_noexec_test = test;
274
+ fn(0);
275
+ assert(current_noexec_test == NULL);
276
+}
277
+
278
+static int test_noexec(struct noexec_test *tests, size_t n_tests)
279
+{
280
+ struct sigaction act;
281
+ size_t i;
282
+
283
+ memset(&act, 0, sizeof(act));
284
+ act.sa_sigaction = handle_segv;
285
+ act.sa_flags = SA_SIGINFO;
286
+ if (sigaction(SIGSEGV, &act, NULL) < 0) {
287
+ handle_err("sigaction");
288
+ }
289
+
290
+ page_size = getpagesize();
291
+ page_base = mmap(NULL, 2 * page_size,
292
+ PROT_READ | PROT_WRITE | PROT_EXEC,
293
+ MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
294
+ if (page_base == MAP_FAILED) {
295
+ handle_err("mmap");
296
+ }
297
+ page_base += page_size;
298
+
299
+ for (i = 0; i < n_tests; i++) {
300
+ struct noexec_test *test = &tests[i];
301
+
302
+ printf("[ RUN ] %s\n", test->name);
303
+ test_noexec_1(test);
304
+ printf("[ OK ]\n");
305
+ }
306
+
307
+ printf("[ PASSED ]\n");
308
+ return EXIT_SUCCESS;
309
+}
310
diff --git a/tests/tcg/s390x/Makefile.target b/tests/tcg/s390x/Makefile.target
105
index XXXXXXX..XXXXXXX 100644
311
index XXXXXXX..XXXXXXX 100644
106
--- a/tcg/tcg.c
312
--- a/tests/tcg/s390x/Makefile.target
107
+++ b/tcg/tcg.c
313
+++ b/tests/tcg/s390x/Makefile.target
108
@@ -XXX,XX +XXX,XX @@
314
@@ -XXX,XX +XXX,XX @@ TESTS+=shift
109
#include "exec/log.h"
315
TESTS+=trap
110
#include "tcg/tcg-ldst.h"
316
TESTS+=signals-s390x
111
#include "tcg-internal.h"
317
TESTS+=branch-relative-long
112
+#include "tcg-target-sa32.h"
318
+TESTS+=noexec
113
+
319
114
+/* Sanity check for TCG_TARGET_SIGNED_ADDR32. */
320
Z14_TESTS=vfminmax
115
+QEMU_BUILD_BUG_ON(TCG_TARGET_REG_BITS == 32 && TCG_TARGET_SIGNED_ADDR32);
321
vfminmax: LDFLAGS+=-lm
116
117
#ifdef CONFIG_TCG_INTERPRETER
118
#include <ffi.h>
119
--
322
--
120
2.25.1
323
2.34.1
121
122
diff view generated by jsdifflib
1
We've had placeholders for these opcodes for a while,
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
and should have support on ppc, s390x and avx512 hosts.
2
3
3
Right now translator stops right *after* the end of a page, which
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
4
breaks reporting of fault locations when the last instruction of a
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
multi-insn translation block crosses a page boundary.
6
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
7
An implementation, like the one arm and s390x have, would require an
8
i386 length disassembler, which is burdensome to maintain. Another
9
alternative would be to single-step at the end of a guest page, but
10
this may come with a performance impact.
11
12
Fix by snapshotting disassembly state and restoring it after we figure
13
out we crossed a page boundary. This includes rolling back cc_op
14
updates and emitted ops.
15
16
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1143
19
Message-Id: <20220817150506.592862-4-iii@linux.ibm.com>
20
[rth: Simplify end-of-insn cross-page checks.]
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
21
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
22
---
9
include/tcg/tcg-opc.h | 3 +++
23
target/i386/tcg/translate.c | 64 ++++++++++++++++-----------
10
include/tcg/tcg.h | 3 +++
24
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++++++++++++++
11
tcg/aarch64/tcg-target.h | 3 +++
25
tests/tcg/x86_64/Makefile.target | 3 +-
12
tcg/arm/tcg-target.h | 3 +++
26
3 files changed, 116 insertions(+), 26 deletions(-)
13
tcg/i386/tcg-target.h | 3 +++
27
create mode 100644 tests/tcg/x86_64/noexec.c
14
tcg/ppc/tcg-target.h | 3 +++
28
15
tcg/s390x/tcg-target.h | 3 +++
29
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
16
tcg/optimize.c | 12 ++++++------
17
tcg/tcg-op-vec.c | 27 ++++++++++++++++++---------
18
tcg/tcg.c | 6 ++++++
19
10 files changed, 51 insertions(+), 15 deletions(-)
20
21
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
22
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
23
--- a/include/tcg/tcg-opc.h
31
--- a/target/i386/tcg/translate.c
24
+++ b/include/tcg/tcg-opc.h
32
+++ b/target/i386/tcg/translate.c
25
@@ -XXX,XX +XXX,XX @@ DEF(or_vec, 1, 2, 0, IMPLVEC)
33
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
26
DEF(xor_vec, 1, 2, 0, IMPLVEC)
34
TCGv_i64 tmp1_i64;
27
DEF(andc_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_andc_vec))
35
28
DEF(orc_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_orc_vec))
36
sigjmp_buf jmpbuf;
29
+DEF(nand_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_nand_vec))
37
+ TCGOp *prev_insn_end;
30
+DEF(nor_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_nor_vec))
38
} DisasContext;
31
+DEF(eqv_vec, 1, 2, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_eqv_vec))
39
32
DEF(not_vec, 1, 1, 0, IMPLVEC | IMPL(TCG_TARGET_HAS_not_vec))
40
/* The environment in which user-only runs is constrained. */
33
41
@@ -XXX,XX +XXX,XX @@ static uint64_t advance_pc(CPUX86State *env, DisasContext *s, int num_bytes)
34
DEF(shli_vec, 1, 1, 1, IMPLVEC | IMPL(TCG_TARGET_HAS_shi_vec))
35
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
36
index XXXXXXX..XXXXXXX 100644
37
--- a/include/tcg/tcg.h
38
+++ b/include/tcg/tcg.h
39
@@ -XXX,XX +XXX,XX @@ typedef uint64_t TCGRegSet;
40
#define TCG_TARGET_HAS_not_vec 0
41
#define TCG_TARGET_HAS_andc_vec 0
42
#define TCG_TARGET_HAS_orc_vec 0
43
+#define TCG_TARGET_HAS_nand_vec 0
44
+#define TCG_TARGET_HAS_nor_vec 0
45
+#define TCG_TARGET_HAS_eqv_vec 0
46
#define TCG_TARGET_HAS_roti_vec 0
47
#define TCG_TARGET_HAS_rots_vec 0
48
#define TCG_TARGET_HAS_rotv_vec 0
49
diff --git a/tcg/aarch64/tcg-target.h b/tcg/aarch64/tcg-target.h
50
index XXXXXXX..XXXXXXX 100644
51
--- a/tcg/aarch64/tcg-target.h
52
+++ b/tcg/aarch64/tcg-target.h
53
@@ -XXX,XX +XXX,XX @@ typedef enum {
54
55
#define TCG_TARGET_HAS_andc_vec 1
56
#define TCG_TARGET_HAS_orc_vec 1
57
+#define TCG_TARGET_HAS_nand_vec 0
58
+#define TCG_TARGET_HAS_nor_vec 0
59
+#define TCG_TARGET_HAS_eqv_vec 0
60
#define TCG_TARGET_HAS_not_vec 1
61
#define TCG_TARGET_HAS_neg_vec 1
62
#define TCG_TARGET_HAS_abs_vec 1
63
diff --git a/tcg/arm/tcg-target.h b/tcg/arm/tcg-target.h
64
index XXXXXXX..XXXXXXX 100644
65
--- a/tcg/arm/tcg-target.h
66
+++ b/tcg/arm/tcg-target.h
67
@@ -XXX,XX +XXX,XX @@ extern bool use_neon_instructions;
68
69
#define TCG_TARGET_HAS_andc_vec 1
70
#define TCG_TARGET_HAS_orc_vec 1
71
+#define TCG_TARGET_HAS_nand_vec 0
72
+#define TCG_TARGET_HAS_nor_vec 0
73
+#define TCG_TARGET_HAS_eqv_vec 0
74
#define TCG_TARGET_HAS_not_vec 1
75
#define TCG_TARGET_HAS_neg_vec 1
76
#define TCG_TARGET_HAS_abs_vec 1
77
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
78
index XXXXXXX..XXXXXXX 100644
79
--- a/tcg/i386/tcg-target.h
80
+++ b/tcg/i386/tcg-target.h
81
@@ -XXX,XX +XXX,XX @@ extern bool have_movbe;
82
83
#define TCG_TARGET_HAS_andc_vec 1
84
#define TCG_TARGET_HAS_orc_vec 0
85
+#define TCG_TARGET_HAS_nand_vec 0
86
+#define TCG_TARGET_HAS_nor_vec 0
87
+#define TCG_TARGET_HAS_eqv_vec 0
88
#define TCG_TARGET_HAS_not_vec 0
89
#define TCG_TARGET_HAS_neg_vec 0
90
#define TCG_TARGET_HAS_abs_vec 1
91
diff --git a/tcg/ppc/tcg-target.h b/tcg/ppc/tcg-target.h
92
index XXXXXXX..XXXXXXX 100644
93
--- a/tcg/ppc/tcg-target.h
94
+++ b/tcg/ppc/tcg-target.h
95
@@ -XXX,XX +XXX,XX @@ extern bool have_vsx;
96
97
#define TCG_TARGET_HAS_andc_vec 1
98
#define TCG_TARGET_HAS_orc_vec have_isa_2_07
99
+#define TCG_TARGET_HAS_nand_vec 0
100
+#define TCG_TARGET_HAS_nor_vec 0
101
+#define TCG_TARGET_HAS_eqv_vec 0
102
#define TCG_TARGET_HAS_not_vec 1
103
#define TCG_TARGET_HAS_neg_vec have_isa_3_00
104
#define TCG_TARGET_HAS_abs_vec 0
105
diff --git a/tcg/s390x/tcg-target.h b/tcg/s390x/tcg-target.h
106
index XXXXXXX..XXXXXXX 100644
107
--- a/tcg/s390x/tcg-target.h
108
+++ b/tcg/s390x/tcg-target.h
109
@@ -XXX,XX +XXX,XX @@ extern uint64_t s390_facilities[3];
110
111
#define TCG_TARGET_HAS_andc_vec 1
112
#define TCG_TARGET_HAS_orc_vec HAVE_FACILITY(VECTOR_ENH1)
113
+#define TCG_TARGET_HAS_nand_vec 0
114
+#define TCG_TARGET_HAS_nor_vec 0
115
+#define TCG_TARGET_HAS_eqv_vec 0
116
#define TCG_TARGET_HAS_not_vec 1
117
#define TCG_TARGET_HAS_neg_vec 1
118
#define TCG_TARGET_HAS_abs_vec 1
119
diff --git a/tcg/optimize.c b/tcg/optimize.c
120
index XXXXXXX..XXXXXXX 100644
121
--- a/tcg/optimize.c
122
+++ b/tcg/optimize.c
123
@@ -XXX,XX +XXX,XX @@ static uint64_t do_constant_folding_2(TCGOpcode op, uint64_t x, uint64_t y)
124
CASE_OP_32_64_VEC(orc):
125
return x | ~y;
126
127
- CASE_OP_32_64(eqv):
128
+ CASE_OP_32_64_VEC(eqv):
129
return ~(x ^ y);
130
131
- CASE_OP_32_64(nand):
132
+ CASE_OP_32_64_VEC(nand):
133
return ~(x & y);
134
135
- CASE_OP_32_64(nor):
136
+ CASE_OP_32_64_VEC(nor):
137
return ~(x | y);
138
139
case INDEX_op_clz_i32:
140
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
141
case INDEX_op_dup2_vec:
142
done = fold_dup2(&ctx, op);
143
break;
144
- CASE_OP_32_64(eqv):
145
+ CASE_OP_32_64_VEC(eqv):
146
done = fold_eqv(&ctx, op);
147
break;
148
CASE_OP_32_64(extract):
149
@@ -XXX,XX +XXX,XX @@ void tcg_optimize(TCGContext *s)
150
CASE_OP_32_64(mulu2):
151
done = fold_multiply2(&ctx, op);
152
break;
153
- CASE_OP_32_64(nand):
154
+ CASE_OP_32_64_VEC(nand):
155
done = fold_nand(&ctx, op);
156
break;
157
CASE_OP_32_64(neg):
158
done = fold_neg(&ctx, op);
159
break;
160
- CASE_OP_32_64(nor):
161
+ CASE_OP_32_64_VEC(nor):
162
done = fold_nor(&ctx, op);
163
break;
164
CASE_OP_32_64_VEC(not):
165
diff --git a/tcg/tcg-op-vec.c b/tcg/tcg-op-vec.c
166
index XXXXXXX..XXXXXXX 100644
167
--- a/tcg/tcg-op-vec.c
168
+++ b/tcg/tcg-op-vec.c
169
@@ -XXX,XX +XXX,XX @@ void tcg_gen_orc_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b)
170
171
void tcg_gen_nand_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b)
172
{
42
{
173
- /* TODO: Add TCG_TARGET_HAS_nand_vec when adding a backend supports it. */
43
uint64_t pc = s->pc;
174
- tcg_gen_and_vec(0, r, a, b);
44
175
- tcg_gen_not_vec(0, r, r);
45
+ /* This is a subsequent insn that crosses a page boundary. */
176
+ if (TCG_TARGET_HAS_nand_vec) {
46
+ if (s->base.num_insns > 1 &&
177
+ vec_gen_op3(INDEX_op_nand_vec, 0, r, a, b);
47
+ !is_same_page(&s->base, s->pc + num_bytes - 1)) {
178
+ } else {
48
+ siglongjmp(s->jmpbuf, 2);
179
+ tcg_gen_and_vec(0, r, a, b);
49
+ }
180
+ tcg_gen_not_vec(0, r, r);
50
+
51
s->pc += num_bytes;
52
if (unlikely(s->pc - s->pc_start > X86_MAX_INSN_LENGTH)) {
53
/* If the instruction's 16th byte is on a different page than the 1st, a
54
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
55
int modrm, reg, rm, mod, op, opreg, val;
56
target_ulong next_eip, tval;
57
target_ulong pc_start = s->base.pc_next;
58
+ bool orig_cc_op_dirty = s->cc_op_dirty;
59
+ CCOp orig_cc_op = s->cc_op;
60
61
s->pc_start = s->pc = pc_start;
62
s->override = -1;
63
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
64
s->rip_offset = 0; /* for relative ip address */
65
s->vex_l = 0;
66
s->vex_v = 0;
67
- if (sigsetjmp(s->jmpbuf, 0) != 0) {
68
+ switch (sigsetjmp(s->jmpbuf, 0)) {
69
+ case 0:
70
+ break;
71
+ case 1:
72
gen_exception_gpf(s);
73
return s->pc;
74
+ case 2:
75
+ /* Restore state that may affect the next instruction. */
76
+ s->cc_op_dirty = orig_cc_op_dirty;
77
+ s->cc_op = orig_cc_op;
78
+ s->base.num_insns--;
79
+ tcg_remove_ops_after(s->prev_insn_end);
80
+ s->base.is_jmp = DISAS_TOO_MANY;
81
+ return pc_start;
82
+ default:
83
+ g_assert_not_reached();
84
}
85
86
prefixes = 0;
87
@@ -XXX,XX +XXX,XX @@ static void i386_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
88
{
89
DisasContext *dc = container_of(dcbase, DisasContext, base);
90
91
+ dc->prev_insn_end = tcg_last_op();
92
tcg_gen_insn_start(dc->base.pc_next, dc->cc_op);
93
}
94
95
@@ -XXX,XX +XXX,XX @@ static void i386_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
96
#endif
97
98
pc_next = disas_insn(dc, cpu);
99
-
100
- if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
101
- /* if single step mode, we generate only one instruction and
102
- generate an exception */
103
- /* if irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
104
- the flag and abort the translation to give the irqs a
105
- chance to happen */
106
- dc->base.is_jmp = DISAS_TOO_MANY;
107
- } else if ((tb_cflags(dc->base.tb) & CF_USE_ICOUNT)
108
- && ((pc_next & TARGET_PAGE_MASK)
109
- != ((pc_next + TARGET_MAX_INSN_SIZE - 1)
110
- & TARGET_PAGE_MASK)
111
- || (pc_next & ~TARGET_PAGE_MASK) == 0)) {
112
- /* Do not cross the boundary of the pages in icount mode,
113
- it can cause an exception. Do it only when boundary is
114
- crossed by the first instruction in the block.
115
- If current instruction already crossed the bound - it's ok,
116
- because an exception hasn't stopped this code.
117
- */
118
- dc->base.is_jmp = DISAS_TOO_MANY;
119
- } else if ((pc_next - dc->base.pc_first) >= (TARGET_PAGE_SIZE - 32)) {
120
- dc->base.is_jmp = DISAS_TOO_MANY;
121
- }
122
-
123
dc->base.pc_next = pc_next;
124
+
125
+ if (dc->base.is_jmp == DISAS_NEXT) {
126
+ if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
127
+ /*
128
+ * If single step mode, we generate only one instruction and
129
+ * generate an exception.
130
+ * If irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
131
+ * the flag and abort the translation to give the irqs a
132
+ * chance to happen.
133
+ */
134
+ dc->base.is_jmp = DISAS_TOO_MANY;
135
+ } else if (!is_same_page(&dc->base, pc_next)) {
136
+ dc->base.is_jmp = DISAS_TOO_MANY;
137
+ }
181
+ }
138
+ }
182
}
139
}
183
140
184
void tcg_gen_nor_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b)
141
static void i386_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
185
{
142
diff --git a/tests/tcg/x86_64/noexec.c b/tests/tcg/x86_64/noexec.c
186
- /* TODO: Add TCG_TARGET_HAS_nor_vec when adding a backend supports it. */
143
new file mode 100644
187
- tcg_gen_or_vec(0, r, a, b);
144
index XXXXXXX..XXXXXXX
188
- tcg_gen_not_vec(0, r, r);
145
--- /dev/null
189
+ if (TCG_TARGET_HAS_nor_vec) {
146
+++ b/tests/tcg/x86_64/noexec.c
190
+ vec_gen_op3(INDEX_op_nor_vec, 0, r, a, b);
147
@@ -XXX,XX +XXX,XX @@
191
+ } else {
148
+#include "../multiarch/noexec.c.inc"
192
+ tcg_gen_or_vec(0, r, a, b);
149
+
193
+ tcg_gen_not_vec(0, r, r);
150
+static void *arch_mcontext_pc(const mcontext_t *ctx)
194
+ }
151
+{
195
}
152
+ return (void *)ctx->gregs[REG_RIP];
196
153
+}
197
void tcg_gen_eqv_vec(unsigned vece, TCGv_vec r, TCGv_vec a, TCGv_vec b)
154
+
198
{
155
+int arch_mcontext_arg(const mcontext_t *ctx)
199
- /* TODO: Add TCG_TARGET_HAS_eqv_vec when adding a backend supports it. */
156
+{
200
- tcg_gen_xor_vec(0, r, a, b);
157
+ return ctx->gregs[REG_RDI];
201
- tcg_gen_not_vec(0, r, r);
158
+}
202
+ if (TCG_TARGET_HAS_eqv_vec) {
159
+
203
+ vec_gen_op3(INDEX_op_eqv_vec, 0, r, a, b);
160
+static void arch_flush(void *p, int len)
204
+ } else {
161
+{
205
+ tcg_gen_xor_vec(0, r, a, b);
162
+}
206
+ tcg_gen_not_vec(0, r, r);
163
+
207
+ }
164
+extern char noexec_1[];
208
}
165
+extern char noexec_2[];
209
166
+extern char noexec_end[];
210
static bool do_op2(unsigned vece, TCGv_vec r, TCGv_vec a, TCGOpcode opc)
167
+
211
diff --git a/tcg/tcg.c b/tcg/tcg.c
168
+asm("noexec_1:\n"
169
+ " movq $1,%rdi\n" /* %rdi is 0 on entry, set 1. */
170
+ "noexec_2:\n"
171
+ " movq $2,%rdi\n" /* %rdi is 0/1; set 2. */
172
+ " ret\n"
173
+ "noexec_end:");
174
+
175
+int main(void)
176
+{
177
+ struct noexec_test noexec_tests[] = {
178
+ {
179
+ .name = "fallthrough",
180
+ .test_code = noexec_1,
181
+ .test_len = noexec_end - noexec_1,
182
+ .page_ofs = noexec_1 - noexec_2,
183
+ .entry_ofs = noexec_1 - noexec_2,
184
+ .expected_si_ofs = 0,
185
+ .expected_pc_ofs = 0,
186
+ .expected_arg = 1,
187
+ },
188
+ {
189
+ .name = "jump",
190
+ .test_code = noexec_1,
191
+ .test_len = noexec_end - noexec_1,
192
+ .page_ofs = noexec_1 - noexec_2,
193
+ .entry_ofs = 0,
194
+ .expected_si_ofs = 0,
195
+ .expected_pc_ofs = 0,
196
+ .expected_arg = 0,
197
+ },
198
+ {
199
+ .name = "fallthrough [cross]",
200
+ .test_code = noexec_1,
201
+ .test_len = noexec_end - noexec_1,
202
+ .page_ofs = noexec_1 - noexec_2 - 2,
203
+ .entry_ofs = noexec_1 - noexec_2 - 2,
204
+ .expected_si_ofs = 0,
205
+ .expected_pc_ofs = -2,
206
+ .expected_arg = 1,
207
+ },
208
+ {
209
+ .name = "jump [cross]",
210
+ .test_code = noexec_1,
211
+ .test_len = noexec_end - noexec_1,
212
+ .page_ofs = noexec_1 - noexec_2 - 2,
213
+ .entry_ofs = -2,
214
+ .expected_si_ofs = 0,
215
+ .expected_pc_ofs = -2,
216
+ .expected_arg = 0,
217
+ },
218
+ };
219
+
220
+ return test_noexec(noexec_tests,
221
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
222
+}
223
diff --git a/tests/tcg/x86_64/Makefile.target b/tests/tcg/x86_64/Makefile.target
212
index XXXXXXX..XXXXXXX 100644
224
index XXXXXXX..XXXXXXX 100644
213
--- a/tcg/tcg.c
225
--- a/tests/tcg/x86_64/Makefile.target
214
+++ b/tcg/tcg.c
226
+++ b/tests/tcg/x86_64/Makefile.target
215
@@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op)
227
@@ -XXX,XX +XXX,XX @@ include $(SRC_PATH)/tests/tcg/i386/Makefile.target
216
return have_vec && TCG_TARGET_HAS_andc_vec;
228
217
case INDEX_op_orc_vec:
229
ifeq ($(filter %-linux-user, $(TARGET)),$(TARGET))
218
return have_vec && TCG_TARGET_HAS_orc_vec;
230
X86_64_TESTS += vsyscall
219
+ case INDEX_op_nand_vec:
231
+X86_64_TESTS += noexec
220
+ return have_vec && TCG_TARGET_HAS_nand_vec;
232
TESTS=$(MULTIARCH_TESTS) $(X86_64_TESTS) test-x86_64
221
+ case INDEX_op_nor_vec:
233
else
222
+ return have_vec && TCG_TARGET_HAS_nor_vec;
234
TESTS=$(MULTIARCH_TESTS)
223
+ case INDEX_op_eqv_vec:
235
@@ -XXX,XX +XXX,XX @@ test-x86_64: LDFLAGS+=-lm -lc
224
+ return have_vec && TCG_TARGET_HAS_eqv_vec;
236
test-x86_64: test-i386.c test-i386.h test-i386-shift.h test-i386-muldiv.h
225
case INDEX_op_mul_vec:
237
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
226
return have_vec && TCG_TARGET_HAS_mul_vec;
238
227
case INDEX_op_shli_vec:
239
-vsyscall: $(SRC_PATH)/tests/tcg/x86_64/vsyscall.c
240
+%: $(SRC_PATH)/tests/tcg/x86_64/%.c
241
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
228
--
242
--
229
2.25.1
243
2.34.1
230
231
diff view generated by jsdifflib
Deleted patch
1
Tested-by: Alex Bennée <alex.bennee@linaro.org>
2
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
3
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
---
6
tcg/s390x/tcg-target.h | 6 +++---
7
tcg/s390x/tcg-target.c.inc | 17 +++++++++++++++++
8
2 files changed, 20 insertions(+), 3 deletions(-)
9
1
10
diff --git a/tcg/s390x/tcg-target.h b/tcg/s390x/tcg-target.h
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/s390x/tcg-target.h
13
+++ b/tcg/s390x/tcg-target.h
14
@@ -XXX,XX +XXX,XX @@ extern uint64_t s390_facilities[3];
15
16
#define TCG_TARGET_HAS_andc_vec 1
17
#define TCG_TARGET_HAS_orc_vec HAVE_FACILITY(VECTOR_ENH1)
18
-#define TCG_TARGET_HAS_nand_vec 0
19
-#define TCG_TARGET_HAS_nor_vec 0
20
-#define TCG_TARGET_HAS_eqv_vec 0
21
+#define TCG_TARGET_HAS_nand_vec HAVE_FACILITY(VECTOR_ENH1)
22
+#define TCG_TARGET_HAS_nor_vec 1
23
+#define TCG_TARGET_HAS_eqv_vec HAVE_FACILITY(VECTOR_ENH1)
24
#define TCG_TARGET_HAS_not_vec 1
25
#define TCG_TARGET_HAS_neg_vec 1
26
#define TCG_TARGET_HAS_abs_vec 1
27
diff --git a/tcg/s390x/tcg-target.c.inc b/tcg/s390x/tcg-target.c.inc
28
index XXXXXXX..XXXXXXX 100644
29
--- a/tcg/s390x/tcg-target.c.inc
30
+++ b/tcg/s390x/tcg-target.c.inc
31
@@ -XXX,XX +XXX,XX @@ typedef enum S390Opcode {
32
VRRc_VMXL = 0xe7fd,
33
VRRc_VN = 0xe768,
34
VRRc_VNC = 0xe769,
35
+ VRRc_VNN = 0xe76e,
36
VRRc_VNO = 0xe76b,
37
+ VRRc_VNX = 0xe76c,
38
VRRc_VO = 0xe76a,
39
VRRc_VOC = 0xe76f,
40
VRRc_VPKS = 0xe797, /* we leave the m5 cs field 0 */
41
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
42
case INDEX_op_xor_vec:
43
tcg_out_insn(s, VRRc, VX, a0, a1, a2, 0);
44
break;
45
+ case INDEX_op_nand_vec:
46
+ tcg_out_insn(s, VRRc, VNN, a0, a1, a2, 0);
47
+ break;
48
+ case INDEX_op_nor_vec:
49
+ tcg_out_insn(s, VRRc, VNO, a0, a1, a2, 0);
50
+ break;
51
+ case INDEX_op_eqv_vec:
52
+ tcg_out_insn(s, VRRc, VNX, a0, a1, a2, 0);
53
+ break;
54
55
case INDEX_op_shli_vec:
56
tcg_out_insn(s, VRSa, VESL, a0, a2, TCG_REG_NONE, a1, vece);
57
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
58
case INDEX_op_and_vec:
59
case INDEX_op_andc_vec:
60
case INDEX_op_bitsel_vec:
61
+ case INDEX_op_eqv_vec:
62
+ case INDEX_op_nand_vec:
63
case INDEX_op_neg_vec:
64
+ case INDEX_op_nor_vec:
65
case INDEX_op_not_vec:
66
case INDEX_op_or_vec:
67
case INDEX_op_orc_vec:
68
@@ -XXX,XX +XXX,XX @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
69
case INDEX_op_or_vec:
70
case INDEX_op_orc_vec:
71
case INDEX_op_xor_vec:
72
+ case INDEX_op_nand_vec:
73
+ case INDEX_op_nor_vec:
74
+ case INDEX_op_eqv_vec:
75
case INDEX_op_cmp_vec:
76
case INDEX_op_mul_vec:
77
case INDEX_op_rotlv_vec:
78
--
79
2.25.1
80
81
diff view generated by jsdifflib
1
While the host may prefer to treat 32-bit addresses as signed,
1
These will be useful in properly ending the TB.
2
there are edge cases of guests that cannot be implemented with
3
addresses 0x7fff_ffff and 0x8000_0000 being non-consecutive.
4
2
5
Therefore, default to guest_base_signed_addr32 false, and allow
3
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
6
probe_guest_base to determine whether it is possible to set it
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
to true. A tcg backend which sets TCG_TARGET_SIGNED_ADDR32 will
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
have to cope with either setting for user-only.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
---
7
---
14
include/exec/cpu-all.h | 16 ++++++++++++++++
8
target/riscv/translate.c | 10 +++++++++-
15
include/exec/cpu_ldst.h | 3 ++-
9
1 file changed, 9 insertions(+), 1 deletion(-)
16
bsd-user/main.c | 4 ++++
17
linux-user/main.c | 3 +++
18
4 files changed, 25 insertions(+), 1 deletion(-)
19
10
20
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
11
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
21
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
22
--- a/include/exec/cpu-all.h
13
--- a/target/riscv/translate.c
23
+++ b/include/exec/cpu-all.h
14
+++ b/target/riscv/translate.c
24
@@ -XXX,XX +XXX,XX @@ static inline void tswap64s(uint64_t *s)
15
@@ -XXX,XX +XXX,XX @@ static uint32_t opcode_at(DisasContextBase *dcbase, target_ulong pc)
25
16
/* Include decoders for factored-out extensions */
26
#if defined(CONFIG_USER_ONLY)
17
#include "decode-XVentanaCondOps.c.inc"
27
#include "exec/user/abitypes.h"
18
28
+#include "tcg-target-sa32.h"
19
+/* The specification allows for longer insns, but not supported by qemu. */
29
20
+#define MAX_INSN_LEN 4
30
/* On some host systems the guest address space is reserved on the host.
31
* This allows the guest address space to be offset to a convenient location.
32
@@ -XXX,XX +XXX,XX @@ extern uintptr_t guest_base;
33
extern bool have_guest_base;
34
extern unsigned long reserved_va;
35
36
+#if TCG_TARGET_SIGNED_ADDR32 && TARGET_LONG_BITS == 32
37
+extern bool guest_base_signed_addr32;
38
+#else
39
+#define guest_base_signed_addr32 false
40
+#endif
41
+
21
+
42
+static inline void set_guest_base_signed_addr32(void)
22
+static inline int insn_len(uint16_t first_word)
43
+{
23
+{
44
+#ifdef guest_base_signed_addr32
24
+ return (first_word & 3) == 3 ? 4 : 2;
45
+ qemu_build_not_reached();
46
+#else
47
+ guest_base_signed_addr32 = true;
48
+#endif
49
+}
25
+}
50
+
26
+
51
/*
27
static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
52
* Limit the guest addresses as best we can.
53
*
54
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
55
index XXXXXXX..XXXXXXX 100644
56
--- a/include/exec/cpu_ldst.h
57
+++ b/include/exec/cpu_ldst.h
58
@@ -XXX,XX +XXX,XX @@ static inline abi_ptr cpu_untagged_addr(CPUState *cs, abi_ptr x)
59
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
60
static inline void *g2h_untagged(abi_ptr x)
61
{
28
{
62
- return (void *)((uintptr_t)(x) + guest_base);
29
/*
63
+ uintptr_t hx = guest_base_signed_addr32 ? (int32_t)x : (uintptr_t)x;
30
@@ -XXX,XX +XXX,XX @@ static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
64
+ return (void *)(guest_base + hx);
31
};
65
}
32
66
33
/* Check for compressed insn */
67
static inline void *g2h(CPUState *cs, abi_ptr x)
34
- if (extract16(opcode, 0, 2) != 3) {
68
diff --git a/bsd-user/main.c b/bsd-user/main.c
35
+ if (insn_len(opcode) == 2) {
69
index XXXXXXX..XXXXXXX 100644
36
if (!has_ext(ctx, RVC)) {
70
--- a/bsd-user/main.c
37
gen_exception_illegal(ctx);
71
+++ b/bsd-user/main.c
38
} else {
72
@@ -XXX,XX +XXX,XX @@
73
int singlestep;
74
uintptr_t guest_base;
75
bool have_guest_base;
76
+#ifndef guest_base_signed_addr32
77
+bool guest_base_signed_addr32;
78
+#endif
79
+
80
/*
81
* When running 32-on-64 we should make sure we can fit all of the possible
82
* guest address space into a contiguous chunk of virtual host memory.
83
diff --git a/linux-user/main.c b/linux-user/main.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/linux-user/main.c
86
+++ b/linux-user/main.c
87
@@ -XXX,XX +XXX,XX @@ static const char *seed_optarg;
88
unsigned long mmap_min_addr;
89
uintptr_t guest_base;
90
bool have_guest_base;
91
+#ifndef guest_base_signed_addr32
92
+bool guest_base_signed_addr32;
93
+#endif
94
95
/*
96
* Used to implement backwards-compatibility for the `-strace`, and
97
--
39
--
98
2.25.1
40
2.34.1
99
100
diff view generated by jsdifflib
1
There are some operation sizes in some subsets of AVX512 that
1
Right now the translator stops right *after* the end of a page, which
2
are missing from previous iterations of AVX. Detect them.
2
breaks reporting of fault locations when the last instruction of a
3
multi-insn translation block crosses a page boundary.
3
4
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1155
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
10
---
8
include/qemu/cpuid.h | 20 +++++++++++++++++---
11
target/riscv/translate.c | 17 +++++--
9
tcg/i386/tcg-target.h | 4 ++++
12
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++++++++++++
10
tcg/i386/tcg-target.c.inc | 24 ++++++++++++++++++++++--
13
tests/tcg/riscv64/Makefile.target | 1 +
11
3 files changed, 43 insertions(+), 5 deletions(-)
14
3 files changed, 93 insertions(+), 4 deletions(-)
15
create mode 100644 tests/tcg/riscv64/noexec.c
12
16
13
diff --git a/include/qemu/cpuid.h b/include/qemu/cpuid.h
17
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
14
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
15
--- a/include/qemu/cpuid.h
19
--- a/target/riscv/translate.c
16
+++ b/include/qemu/cpuid.h
20
+++ b/target/riscv/translate.c
17
@@ -XXX,XX +XXX,XX @@
21
@@ -XXX,XX +XXX,XX @@ static void riscv_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
18
#ifndef bit_AVX2
22
}
19
#define bit_AVX2 (1 << 5)
23
ctx->nftemp = 0;
20
#endif
24
21
-#ifndef bit_AVX512F
25
+ /* Only the first insn within a TB is allowed to cross a page boundary. */
22
-#define bit_AVX512F (1 << 16)
26
if (ctx->base.is_jmp == DISAS_NEXT) {
23
-#endif
27
- target_ulong page_start;
24
#ifndef bit_BMI2
28
-
25
#define bit_BMI2 (1 << 8)
29
- page_start = ctx->base.pc_first & TARGET_PAGE_MASK;
26
#endif
30
- if (ctx->base.pc_next - page_start >= TARGET_PAGE_SIZE) {
27
+#ifndef bit_AVX512F
31
+ if (!is_same_page(&ctx->base, ctx->base.pc_next)) {
28
+#define bit_AVX512F (1 << 16)
32
ctx->base.is_jmp = DISAS_TOO_MANY;
29
+#endif
33
+ } else {
30
+#ifndef bit_AVX512DQ
34
+ unsigned page_ofs = ctx->base.pc_next & ~TARGET_PAGE_MASK;
31
+#define bit_AVX512DQ (1 << 17)
32
+#endif
33
+#ifndef bit_AVX512BW
34
+#define bit_AVX512BW (1 << 30)
35
+#endif
36
+#ifndef bit_AVX512VL
37
+#define bit_AVX512VL (1u << 31)
38
+#endif
39
+
35
+
40
+/* Leaf 7, %ecx */
36
+ if (page_ofs > TARGET_PAGE_SIZE - MAX_INSN_LEN) {
41
+#ifndef bit_AVX512VBMI2
37
+ uint16_t next_insn = cpu_lduw_code(env, ctx->base.pc_next);
42
+#define bit_AVX512VBMI2 (1 << 6)
38
+ int len = insn_len(next_insn);
43
+#endif
44
45
/* Leaf 0x80000001, %ecx */
46
#ifndef bit_LZCNT
47
diff --git a/tcg/i386/tcg-target.h b/tcg/i386/tcg-target.h
48
index XXXXXXX..XXXXXXX 100644
49
--- a/tcg/i386/tcg-target.h
50
+++ b/tcg/i386/tcg-target.h
51
@@ -XXX,XX +XXX,XX @@ extern bool have_bmi1;
52
extern bool have_popcnt;
53
extern bool have_avx1;
54
extern bool have_avx2;
55
+extern bool have_avx512bw;
56
+extern bool have_avx512dq;
57
+extern bool have_avx512vbmi2;
58
+extern bool have_avx512vl;
59
extern bool have_movbe;
60
61
/* optional instructions */
62
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
63
index XXXXXXX..XXXXXXX 100644
64
--- a/tcg/i386/tcg-target.c.inc
65
+++ b/tcg/i386/tcg-target.c.inc
66
@@ -XXX,XX +XXX,XX @@ bool have_bmi1;
67
bool have_popcnt;
68
bool have_avx1;
69
bool have_avx2;
70
+bool have_avx512bw;
71
+bool have_avx512dq;
72
+bool have_avx512vbmi2;
73
+bool have_avx512vl;
74
bool have_movbe;
75
76
#ifdef CONFIG_CPUID_H
77
@@ -XXX,XX +XXX,XX @@ static void tcg_out_nop_fill(tcg_insn_unit *p, int count)
78
static void tcg_target_init(TCGContext *s)
79
{
80
#ifdef CONFIG_CPUID_H
81
- unsigned a, b, c, d, b7 = 0;
82
+ unsigned a, b, c, d, b7 = 0, c7 = 0;
83
unsigned max = __get_cpuid_max(0, 0);
84
85
if (max >= 7) {
86
/* BMI1 is available on AMD Piledriver and Intel Haswell CPUs. */
87
- __cpuid_count(7, 0, a, b7, c, d);
88
+ __cpuid_count(7, 0, a, b7, c7, d);
89
have_bmi1 = (b7 & bit_BMI) != 0;
90
have_bmi2 = (b7 & bit_BMI2) != 0;
91
}
92
@@ -XXX,XX +XXX,XX @@ static void tcg_target_init(TCGContext *s)
93
if ((xcrl & 6) == 6) {
94
have_avx1 = (c & bit_AVX) != 0;
95
have_avx2 = (b7 & bit_AVX2) != 0;
96
+
39
+
97
+ /*
40
+ if (!is_same_page(&ctx->base, ctx->base.pc_next + len)) {
98
+ * There are interesting instructions in AVX512, so long
41
+ ctx->base.is_jmp = DISAS_TOO_MANY;
99
+ * as we have AVX512VL, which indicates support for EVEX
100
+ * on sizes smaller than 512 bits. We are required to
101
+ * check that OPMASK and all extended ZMM state are enabled
102
+ * even if we're not using them -- the insns will fault.
103
+ */
104
+ if ((xcrl & 0xe0) == 0xe0
105
+ && (b7 & bit_AVX512F)
106
+ && (b7 & bit_AVX512VL)) {
107
+ have_avx512vl = true;
108
+ have_avx512bw = (b7 & bit_AVX512BW) != 0;
109
+ have_avx512dq = (b7 & bit_AVX512DQ) != 0;
110
+ have_avx512vbmi2 = (c7 & bit_AVX512VBMI2) != 0;
111
+ }
42
+ }
112
}
43
+ }
113
}
44
}
114
}
45
}
46
}
47
diff --git a/tests/tcg/riscv64/noexec.c b/tests/tcg/riscv64/noexec.c
48
new file mode 100644
49
index XXXXXXX..XXXXXXX
50
--- /dev/null
51
+++ b/tests/tcg/riscv64/noexec.c
52
@@ -XXX,XX +XXX,XX @@
53
+#include "../multiarch/noexec.c.inc"
54
+
55
+static void *arch_mcontext_pc(const mcontext_t *ctx)
56
+{
57
+ return (void *)ctx->__gregs[REG_PC];
58
+}
59
+
60
+static int arch_mcontext_arg(const mcontext_t *ctx)
61
+{
62
+ return ctx->__gregs[REG_A0];
63
+}
64
+
65
+static void arch_flush(void *p, int len)
66
+{
67
+ __builtin___clear_cache(p, p + len);
68
+}
69
+
70
+extern char noexec_1[];
71
+extern char noexec_2[];
72
+extern char noexec_end[];
73
+
74
+asm(".option push\n"
75
+ ".option norvc\n"
76
+ "noexec_1:\n"
77
+ " li a0,1\n" /* a0 is 0 on entry, set 1. */
78
+ "noexec_2:\n"
79
+ " li a0,2\n" /* a0 is 0/1; set 2. */
80
+ " ret\n"
81
+ "noexec_end:\n"
82
+ ".option pop");
83
+
84
+int main(void)
85
+{
86
+ struct noexec_test noexec_tests[] = {
87
+ {
88
+ .name = "fallthrough",
89
+ .test_code = noexec_1,
90
+ .test_len = noexec_end - noexec_1,
91
+ .page_ofs = noexec_1 - noexec_2,
92
+ .entry_ofs = noexec_1 - noexec_2,
93
+ .expected_si_ofs = 0,
94
+ .expected_pc_ofs = 0,
95
+ .expected_arg = 1,
96
+ },
97
+ {
98
+ .name = "jump",
99
+ .test_code = noexec_1,
100
+ .test_len = noexec_end - noexec_1,
101
+ .page_ofs = noexec_1 - noexec_2,
102
+ .entry_ofs = 0,
103
+ .expected_si_ofs = 0,
104
+ .expected_pc_ofs = 0,
105
+ .expected_arg = 0,
106
+ },
107
+ {
108
+ .name = "fallthrough [cross]",
109
+ .test_code = noexec_1,
110
+ .test_len = noexec_end - noexec_1,
111
+ .page_ofs = noexec_1 - noexec_2 - 2,
112
+ .entry_ofs = noexec_1 - noexec_2 - 2,
113
+ .expected_si_ofs = 0,
114
+ .expected_pc_ofs = -2,
115
+ .expected_arg = 1,
116
+ },
117
+ {
118
+ .name = "jump [cross]",
119
+ .test_code = noexec_1,
120
+ .test_len = noexec_end - noexec_1,
121
+ .page_ofs = noexec_1 - noexec_2 - 2,
122
+ .entry_ofs = -2,
123
+ .expected_si_ofs = 0,
124
+ .expected_pc_ofs = -2,
125
+ .expected_arg = 0,
126
+ },
127
+ };
128
+
129
+ return test_noexec(noexec_tests,
130
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
131
+}
132
diff --git a/tests/tcg/riscv64/Makefile.target b/tests/tcg/riscv64/Makefile.target
133
index XXXXXXX..XXXXXXX 100644
134
--- a/tests/tcg/riscv64/Makefile.target
135
+++ b/tests/tcg/riscv64/Makefile.target
136
@@ -XXX,XX +XXX,XX @@
137
138
VPATH += $(SRC_PATH)/tests/tcg/riscv64
139
TESTS += test-div
140
+TESTS += noexec
115
--
141
--
116
2.25.1
142
2.34.1
117
118
diff view generated by jsdifflib
Deleted patch
1
The condition for UMIN/UMAX availability is about to change;
2
use the canonical version.
3
1
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/i386/tcg-target.c.inc | 8 ++++----
9
1 file changed, 4 insertions(+), 4 deletions(-)
10
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
14
+++ b/tcg/i386/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static bool expand_vec_cmp_noinv(TCGType type, unsigned vece, TCGv_vec v0,
16
fixup = NEED_SWAP | NEED_INV;
17
break;
18
case TCG_COND_LEU:
19
- if (vece <= MO_32) {
20
+ if (tcg_can_emit_vec_op(INDEX_op_umin_vec, type, vece)) {
21
fixup = NEED_UMIN;
22
} else {
23
fixup = NEED_BIAS | NEED_INV;
24
}
25
break;
26
case TCG_COND_GTU:
27
- if (vece <= MO_32) {
28
+ if (tcg_can_emit_vec_op(INDEX_op_umin_vec, type, vece)) {
29
fixup = NEED_UMIN | NEED_INV;
30
} else {
31
fixup = NEED_BIAS;
32
}
33
break;
34
case TCG_COND_GEU:
35
- if (vece <= MO_32) {
36
+ if (tcg_can_emit_vec_op(INDEX_op_umax_vec, type, vece)) {
37
fixup = NEED_UMAX;
38
} else {
39
fixup = NEED_BIAS | NEED_SWAP | NEED_INV;
40
}
41
break;
42
case TCG_COND_LTU:
43
- if (vece <= MO_32) {
44
+ if (tcg_can_emit_vec_op(INDEX_op_umax_vec, type, vece)) {
45
fixup = NEED_UMAX | NEED_INV;
46
} else {
47
fixup = NEED_BIAS | NEED_SWAP;
48
--
49
2.25.1
50
51
diff view generated by jsdifflib
Deleted patch
1
AVX512VL has VPSRAVQ, and
2
AVX512BW has VPSLLVW, VPSRAVW, VPSRLVW.
3
1
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/i386/tcg-target.c.inc | 32 ++++++++++++++++++++++++--------
9
1 file changed, 24 insertions(+), 8 deletions(-)
10
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
14
+++ b/tcg/i386/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
16
#define OPC_VPBROADCASTQ (0x59 | P_EXT38 | P_DATA16)
17
#define OPC_VPERMQ (0x00 | P_EXT3A | P_DATA16 | P_VEXW)
18
#define OPC_VPERM2I128 (0x46 | P_EXT3A | P_DATA16 | P_VEXL)
19
+#define OPC_VPSLLVW (0x12 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
20
#define OPC_VPSLLVD (0x47 | P_EXT38 | P_DATA16)
21
#define OPC_VPSLLVQ (0x47 | P_EXT38 | P_DATA16 | P_VEXW)
22
+#define OPC_VPSRAVW (0x11 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
23
#define OPC_VPSRAVD (0x46 | P_EXT38 | P_DATA16)
24
+#define OPC_VPSRAVQ (0x46 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
25
+#define OPC_VPSRLVW (0x10 | P_EXT38 | P_DATA16 | P_VEXW | P_EVEX)
26
#define OPC_VPSRLVD (0x45 | P_EXT38 | P_DATA16)
27
#define OPC_VPSRLVQ (0x45 | P_EXT38 | P_DATA16 | P_VEXW)
28
#define OPC_VZEROUPPER (0x77 | P_EXT)
29
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
30
OPC_PMAXUB, OPC_PMAXUW, OPC_PMAXUD, OPC_UD2
31
};
32
static int const shlv_insn[4] = {
33
- /* TODO: AVX512 adds support for MO_16. */
34
- OPC_UD2, OPC_UD2, OPC_VPSLLVD, OPC_VPSLLVQ
35
+ OPC_UD2, OPC_VPSLLVW, OPC_VPSLLVD, OPC_VPSLLVQ
36
};
37
static int const shrv_insn[4] = {
38
- /* TODO: AVX512 adds support for MO_16. */
39
- OPC_UD2, OPC_UD2, OPC_VPSRLVD, OPC_VPSRLVQ
40
+ OPC_UD2, OPC_VPSRLVW, OPC_VPSRLVD, OPC_VPSRLVQ
41
};
42
static int const sarv_insn[4] = {
43
- /* TODO: AVX512 adds support for MO_16, MO_64. */
44
- OPC_UD2, OPC_UD2, OPC_VPSRAVD, OPC_UD2
45
+ OPC_UD2, OPC_VPSRAVW, OPC_VPSRAVD, OPC_VPSRAVQ
46
};
47
static int const shls_insn[4] = {
48
OPC_UD2, OPC_PSLLW, OPC_PSLLD, OPC_PSLLQ
49
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
50
51
case INDEX_op_shlv_vec:
52
case INDEX_op_shrv_vec:
53
- return have_avx2 && vece >= MO_32;
54
+ switch (vece) {
55
+ case MO_16:
56
+ return have_avx512bw;
57
+ case MO_32:
58
+ case MO_64:
59
+ return have_avx2;
60
+ }
61
+ return 0;
62
case INDEX_op_sarv_vec:
63
- return have_avx2 && vece == MO_32;
64
+ switch (vece) {
65
+ case MO_16:
66
+ return have_avx512bw;
67
+ case MO_32:
68
+ return have_avx2;
69
+ case MO_64:
70
+ return have_avx512vl;
71
+ }
72
+ return 0;
73
case INDEX_op_rotlv_vec:
74
case INDEX_op_rotrv_vec:
75
return have_avx2 && vece >= MO_32 ? -1 : 0;
76
--
77
2.25.1
78
79
diff view generated by jsdifflib
Deleted patch
1
AVX512VL has VPSRAQ.
2
1
3
Tested-by: Alex Bennée <alex.bennee@linaro.org>
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
tcg/i386/tcg-target.c.inc | 12 ++++++++++--
8
1 file changed, 10 insertions(+), 2 deletions(-)
9
10
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
11
index XXXXXXX..XXXXXXX 100644
12
--- a/tcg/i386/tcg-target.c.inc
13
+++ b/tcg/i386/tcg-target.c.inc
14
@@ -XXX,XX +XXX,XX @@ static bool tcg_target_const_match(int64_t val, TCGType type, int ct)
15
#define OPC_PSLLQ (0xf3 | P_EXT | P_DATA16)
16
#define OPC_PSRAW (0xe1 | P_EXT | P_DATA16)
17
#define OPC_PSRAD (0xe2 | P_EXT | P_DATA16)
18
+#define OPC_VPSRAQ (0x72 | P_EXT | P_DATA16 | P_VEXW | P_EVEX)
19
#define OPC_PSRLW (0xd1 | P_EXT | P_DATA16)
20
#define OPC_PSRLD (0xd2 | P_EXT | P_DATA16)
21
#define OPC_PSRLQ (0xd3 | P_EXT | P_DATA16)
22
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
23
OPC_UD2, OPC_PSRLW, OPC_PSRLD, OPC_PSRLQ
24
};
25
static int const sars_insn[4] = {
26
- OPC_UD2, OPC_PSRAW, OPC_PSRAD, OPC_UD2
27
+ OPC_UD2, OPC_PSRAW, OPC_PSRAD, OPC_VPSRAQ
28
};
29
static int const abs_insn[4] = {
30
/* TODO: AVX512 adds support for MO_64. */
31
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
32
case INDEX_op_shrs_vec:
33
return vece >= MO_16;
34
case INDEX_op_sars_vec:
35
- return vece >= MO_16 && vece <= MO_32;
36
+ switch (vece) {
37
+ case MO_16:
38
+ case MO_32:
39
+ return 1;
40
+ case MO_64:
41
+ return have_avx512vl;
42
+ }
43
+ return 0;
44
case INDEX_op_rotls_vec:
45
return vece >= MO_16 ? -1 : 0;
46
47
--
48
2.25.1
49
50
diff view generated by jsdifflib
Deleted patch
1
AVX512 has VPSRAQ with immediate operand, in the same form as
2
with AVX, but requires EVEX encoding and W1.
3
1
4
Tested-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
8
tcg/i386/tcg-target.c.inc | 30 +++++++++++++++++++++---------
9
1 file changed, 21 insertions(+), 9 deletions(-)
10
11
diff --git a/tcg/i386/tcg-target.c.inc b/tcg/i386/tcg-target.c.inc
12
index XXXXXXX..XXXXXXX 100644
13
--- a/tcg/i386/tcg-target.c.inc
14
+++ b/tcg/i386/tcg-target.c.inc
15
@@ -XXX,XX +XXX,XX @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
16
break;
17
18
case INDEX_op_shli_vec:
19
+ insn = shift_imm_insn[vece];
20
sub = 6;
21
goto gen_shift;
22
case INDEX_op_shri_vec:
23
+ insn = shift_imm_insn[vece];
24
sub = 2;
25
goto gen_shift;
26
case INDEX_op_sari_vec:
27
- tcg_debug_assert(vece != MO_64);
28
+ if (vece == MO_64) {
29
+ insn = OPC_PSHIFTD_Ib | P_VEXW | P_EVEX;
30
+ } else {
31
+ insn = shift_imm_insn[vece];
32
+ }
33
sub = 4;
34
gen_shift:
35
tcg_debug_assert(vece != MO_8);
36
- insn = shift_imm_insn[vece];
37
if (type == TCG_TYPE_V256) {
38
insn |= P_VEXL;
39
}
40
@@ -XXX,XX +XXX,XX @@ int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
41
return vece == MO_8 ? -1 : 1;
42
43
case INDEX_op_sari_vec:
44
- /* We must expand the operation for MO_8. */
45
- if (vece == MO_8) {
46
+ switch (vece) {
47
+ case MO_8:
48
return -1;
49
- }
50
- /* We can emulate this for MO_64, but it does not pay off
51
- unless we're producing at least 4 values. */
52
- if (vece == MO_64) {
53
+ case MO_16:
54
+ case MO_32:
55
+ return 1;
56
+ case MO_64:
57
+ if (have_avx512vl) {
58
+ return 1;
59
+ }
60
+ /*
61
+ * We can emulate this for MO_64, but it does not pay off
62
+ * unless we're producing at least 4 values.
63
+ */
64
return type >= TCG_TYPE_V256 ? -1 : 0;
65
}
66
- return 1;
67
+ return 0;
68
69
case INDEX_op_shls_vec:
70
case INDEX_op_shrs_vec:
71
--
72
2.25.1
73
74
diff view generated by jsdifflib