1
The following changes since commit e93ded1bf6c94ab95015b33e188bc8b0b0c32670:
1
The following changes since commit 9c6c079bc6723da8061ccfb44361d67b1dd785dd:
2
2
3
Merge tag 'testing-pull-request-2022-08-30' of https://gitlab.com/thuth/qemu into staging (2022-08-31 18:19:03 -0400)
3
Merge tag 'pull-target-arm-20240430' of https://git.linaro.org/people/pmaydell/qemu-arm into staging (2024-04-30 09:58:54 -0700)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20220901
7
https://gitlab.com/rth7680/qemu.git tags/pull-tcg-20240501
8
8
9
for you to fetch changes up to 20011be2e30b8aa8ef1fc258485f00c688703deb:
9
for you to fetch changes up to 917d7f8d948d706e275c9f33169b9dd0149ded1e:
10
10
11
target/riscv: Make translator stop before the end of a page (2022-09-01 07:43:08 +0100)
11
plugins: Update the documentation block for plugin-gen.c (2024-04-30 16:12:05 -0700)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Respect PROT_EXEC in user-only mode.
14
plugins: Rewrite plugin tcg expansion
15
Fix s390x, i386 and riscv for translations crossing a page.
16
15
17
----------------------------------------------------------------
16
----------------------------------------------------------------
18
Ilya Leoshkevich (4):
17
Richard Henderson (20):
19
linux-user: Clear translations on mprotect()
18
tcg: Make tcg/helper-info.h self-contained
20
accel/tcg: Introduce is_same_page()
19
tcg: Pass function pointer to tcg_gen_call*
21
target/s390x: Make translator stop before the end of a page
20
plugins: Zero new qemu_plugin_dyn_cb entries
22
target/i386: Make translator stop before the end of a page
21
plugins: Move function pointer in qemu_plugin_dyn_cb
22
plugins: Create TCGHelperInfo for all out-of-line callbacks
23
plugins: Use emit_before_op for PLUGIN_GEN_AFTER_INSN
24
plugins: Use emit_before_op for PLUGIN_GEN_FROM_TB
25
plugins: Add PLUGIN_GEN_AFTER_TB
26
plugins: Use emit_before_op for PLUGIN_GEN_FROM_INSN
27
plugins: Use emit_before_op for PLUGIN_GEN_FROM_MEM
28
plugins: Remove plugin helpers
29
tcg: Remove TCG_CALL_PLUGIN
30
tcg: Remove INDEX_op_plugin_cb_{start,end}
31
plugins: Simplify callback queues
32
plugins: Introduce PLUGIN_CB_MEM_REGULAR
33
plugins: Replace pr_ops with a proper debug dump flag
34
plugins: Split out common cb expanders
35
plugins: Merge qemu_plugin_tb_insn_get to plugin-gen.c
36
plugins: Inline plugin_gen_empty_callback
37
plugins: Update the documentation block for plugin-gen.c
23
38
24
Richard Henderson (16):
39
accel/tcg/plugin-helpers.h | 5 -
25
linux-user/arm: Mark the commpage executable
40
include/exec/helper-gen-common.h | 4 -
26
linux-user/hppa: Allocate page zero as a commpage
41
include/exec/helper-proto-common.h | 4 -
27
linux-user/x86_64: Allocate vsyscall page as a commpage
42
include/exec/plugin-gen.h | 4 -
28
linux-user: Honor PT_GNU_STACK
43
include/qemu/log.h | 1 +
29
tests/tcg/i386: Move smc_code2 to an executable section
44
include/qemu/plugin.h | 67 +--
30
accel/tcg: Properly implement get_page_addr_code for user-only
45
include/tcg/helper-info.h | 3 +
31
accel/tcg: Unlock mmap_lock after longjmp
46
include/tcg/tcg-op-common.h | 4 +-
32
accel/tcg: Make tb_htable_lookup static
47
include/tcg/tcg-opc.h | 4 +-
33
accel/tcg: Move qemu_ram_addr_from_host_nofail to physmem.c
48
include/tcg/tcg.h | 26 +-
34
accel/tcg: Use probe_access_internal for softmmu get_page_addr_code_hostp
49
include/exec/helper-gen.h.inc | 24 +-
35
accel/tcg: Document the faulting lookup in tb_lookup_cmp
50
accel/tcg/plugin-gen.c | 1007 +++++++++---------------------------
36
accel/tcg: Remove translator_ldsw
51
plugins/api.c | 26 +-
37
accel/tcg: Add pc and host_pc params to gen_intermediate_code
52
plugins/core.c | 61 ++-
38
accel/tcg: Add fast path for translator_ld*
53
tcg/tcg-op-ldst.c | 6 +-
39
target/riscv: Add MAX_INSN_LEN and insn_len
54
tcg/tcg-op.c | 8 +-
40
target/riscv: Make translator stop before the end of a page
55
tcg/tcg.c | 78 ++-
41
56
tcg/tci.c | 1 +
42
include/elf.h | 1 +
57
util/log.c | 4 +
43
include/exec/cpu-common.h | 1 +
58
19 files changed, 399 insertions(+), 938 deletions(-)
44
include/exec/exec-all.h | 89 ++++++++----------------
59
delete mode 100644 accel/tcg/plugin-helpers.h
45
include/exec/translator.h | 96 ++++++++++++++++---------
46
linux-user/arm/target_cpu.h | 4 +-
47
linux-user/qemu.h | 1 +
48
accel/tcg/cpu-exec.c | 143 ++++++++++++++++++++------------------
49
accel/tcg/cputlb.c | 93 +++++++------------------
50
accel/tcg/translate-all.c | 29 ++++----
51
accel/tcg/translator.c | 135 ++++++++++++++++++++++++++---------
52
accel/tcg/user-exec.c | 17 ++++-
53
linux-user/elfload.c | 82 ++++++++++++++++++++--
54
linux-user/mmap.c | 6 +-
55
softmmu/physmem.c | 12 ++++
56
target/alpha/translate.c | 5 +-
57
target/arm/translate.c | 5 +-
58
target/avr/translate.c | 5 +-
59
target/cris/translate.c | 5 +-
60
target/hexagon/translate.c | 6 +-
61
target/hppa/translate.c | 5 +-
62
target/i386/tcg/translate.c | 71 +++++++++++--------
63
target/loongarch/translate.c | 6 +-
64
target/m68k/translate.c | 5 +-
65
target/microblaze/translate.c | 5 +-
66
target/mips/tcg/translate.c | 5 +-
67
target/nios2/translate.c | 5 +-
68
target/openrisc/translate.c | 6 +-
69
target/ppc/translate.c | 5 +-
70
target/riscv/translate.c | 32 +++++++--
71
target/rx/translate.c | 5 +-
72
target/s390x/tcg/translate.c | 20 ++++--
73
target/sh4/translate.c | 5 +-
74
target/sparc/translate.c | 5 +-
75
target/tricore/translate.c | 6 +-
76
target/xtensa/translate.c | 6 +-
77
tests/tcg/i386/test-i386.c | 2 +-
78
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++
79
tests/tcg/s390x/noexec.c | 106 ++++++++++++++++++++++++++++
80
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++
81
tests/tcg/multiarch/noexec.c.inc | 139 ++++++++++++++++++++++++++++++++++++
82
tests/tcg/riscv64/Makefile.target | 1 +
83
tests/tcg/s390x/Makefile.target | 1 +
84
tests/tcg/x86_64/Makefile.target | 3 +-
85
43 files changed, 966 insertions(+), 367 deletions(-)
86
create mode 100644 tests/tcg/riscv64/noexec.c
87
create mode 100644 tests/tcg/s390x/noexec.c
88
create mode 100644 tests/tcg/x86_64/noexec.c
89
create mode 100644 tests/tcg/multiarch/noexec.c.inc
diff view generated by jsdifflib
1
While there are no target-specific nonfaulting probes,
1
Move MAX_CALL_IARGS from tcg.h and include for
2
generic code may grow some uses at some point.
2
the define of TCG_TARGET_REG_BITS.
3
3
4
Note that the attrs argument was incorrect -- it should have
4
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
been MEMTXATTRS_UNSPECIFIED. Just use the simpler interface.
5
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
6
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
7
---
10
target/avr/helper.c | 46 ++++++++++++++++++++++++++++-----------------
8
include/tcg/helper-info.h | 3 +++
11
1 file changed, 29 insertions(+), 17 deletions(-)
9
include/tcg/tcg.h | 2 --
10
tcg/tci.c | 1 +
11
3 files changed, 4 insertions(+), 2 deletions(-)
12
12
13
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
diff --git a/include/tcg/helper-info.h b/include/tcg/helper-info.h
14
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/avr/helper.c
15
--- a/include/tcg/helper-info.h
16
+++ b/target/avr/helper.c
16
+++ b/include/tcg/helper-info.h
17
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
17
@@ -XXX,XX +XXX,XX @@
18
MMUAccessType access_type, int mmu_idx,
18
#ifdef CONFIG_TCG_INTERPRETER
19
bool probe, uintptr_t retaddr)
19
#include <ffi.h>
20
{
20
#endif
21
- int prot = 0;
21
+#include "tcg-target-reg-bits.h"
22
- MemTxAttrs attrs = {};
22
+
23
+ int prot, page_size = TARGET_PAGE_SIZE;
23
+#define MAX_CALL_IARGS 7
24
uint32_t paddr;
24
25
25
/*
26
address &= TARGET_PAGE_MASK;
26
* Describe the calling convention of a given argument type.
27
27
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
28
if (mmu_idx == MMU_CODE_IDX) {
28
index XXXXXXX..XXXXXXX 100644
29
- /* access to code in flash */
29
--- a/include/tcg/tcg.h
30
+ /* Access to code in flash. */
30
+++ b/include/tcg/tcg.h
31
paddr = OFFSET_CODE + address;
31
@@ -XXX,XX +XXX,XX @@
32
prot = PAGE_READ | PAGE_EXEC;
32
/* XXX: make safe guess about sizes */
33
- if (paddr + TARGET_PAGE_SIZE > OFFSET_DATA) {
33
#define MAX_OP_PER_INSTR 266
34
+ if (paddr >= OFFSET_DATA) {
34
35
+ /*
35
-#define MAX_CALL_IARGS 7
36
+ * This should not be possible via any architectural operations.
37
+ * There is certainly not an exception that we can deliver.
38
+ * Accept probing that might come from generic code.
39
+ */
40
+ if (probe) {
41
+ return false;
42
+ }
43
error_report("execution left flash memory");
44
abort();
45
}
46
- } else if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
47
- /*
48
- * access to CPU registers, exit and rebuilt this TB to use full access
49
- * incase it touches specially handled registers like SREG or SP
50
- */
51
- AVRCPU *cpu = AVR_CPU(cs);
52
- CPUAVRState *env = &cpu->env;
53
- env->fullacc = 1;
54
- cpu_loop_exit_restore(cs, retaddr);
55
} else {
56
- /* access to memory. nothing special */
57
+ /* Access to memory. */
58
paddr = OFFSET_DATA + address;
59
prot = PAGE_READ | PAGE_WRITE;
60
+ if (address < NUMBER_OF_CPU_REGISTERS + NUMBER_OF_IO_REGISTERS) {
61
+ /*
62
+ * Access to CPU registers, exit and rebuilt this TB to use
63
+ * full access in case it touches specially handled registers
64
+ * like SREG or SP. For probing, set page_size = 1, in order
65
+ * to force tlb_fill to be called for the next access.
66
+ */
67
+ if (probe) {
68
+ page_size = 1;
69
+ } else {
70
+ AVRCPU *cpu = AVR_CPU(cs);
71
+ CPUAVRState *env = &cpu->env;
72
+ env->fullacc = 1;
73
+ cpu_loop_exit_restore(cs, retaddr);
74
+ }
75
+ }
76
}
77
78
- tlb_set_page_with_attrs(cs, address, paddr, attrs, prot,
79
- mmu_idx, TARGET_PAGE_SIZE);
80
-
36
-
81
+ tlb_set_page(cs, address, paddr, prot, mmu_idx, page_size);
37
#define CPU_TEMP_BUF_NLONGS 128
82
return true;
38
#define TCG_STATIC_FRAME_SIZE (CPU_TEMP_BUF_NLONGS * sizeof(long))
83
}
39
40
diff --git a/tcg/tci.c b/tcg/tci.c
41
index XXXXXXX..XXXXXXX 100644
42
--- a/tcg/tci.c
43
+++ b/tcg/tci.c
44
@@ -XXX,XX +XXX,XX @@
45
46
#include "qemu/osdep.h"
47
#include "tcg/tcg.h"
48
+#include "tcg/helper-info.h"
49
#include "tcg/tcg-ldst.h"
50
#include <ffi.h>
84
51
85
--
52
--
86
2.34.1
53
2.34.1
87
54
88
55
diff view generated by jsdifflib
1
This bit is not saved across interrupts, so we must
1
For normal helpers, read the function pointer from the
2
delay delivering the interrupt until the skip has
2
structure earlier. For plugins, this will allow the
3
been processed.
3
function pointer to come from elsewhere.
4
4
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1118
5
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
6
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
8
---
10
target/avr/helper.c | 9 +++++++++
9
include/tcg/tcg.h | 21 +++++++++-------
11
target/avr/translate.c | 26 ++++++++++++++++++++++----
10
include/exec/helper-gen.h.inc | 24 ++++++++++++-------
12
2 files changed, 31 insertions(+), 4 deletions(-)
11
tcg/tcg.c | 45 +++++++++++++++++++----------------
13
12
3 files changed, 52 insertions(+), 38 deletions(-)
14
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
14
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/avr/helper.c
16
--- a/include/tcg/tcg.h
17
+++ b/target/avr/helper.c
17
+++ b/include/tcg/tcg.h
18
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
18
@@ -XXX,XX +XXX,XX @@ typedef struct TCGTargetOpDef {
19
AVRCPU *cpu = AVR_CPU(cs);
19
20
CPUAVRState *env = &cpu->env;
20
bool tcg_op_supported(TCGOpcode op);
21
21
22
+ /*
22
-void tcg_gen_call0(TCGHelperInfo *, TCGTemp *ret);
23
+ * We cannot separate a skip from the next instruction,
23
-void tcg_gen_call1(TCGHelperInfo *, TCGTemp *ret, TCGTemp *);
24
+ * as the skip would not be preserved across the interrupt.
24
-void tcg_gen_call2(TCGHelperInfo *, TCGTemp *ret, TCGTemp *, TCGTemp *);
25
+ * Separating the two insn normally only happens at page boundaries.
25
-void tcg_gen_call3(TCGHelperInfo *, TCGTemp *ret, TCGTemp *,
26
+ */
26
+void tcg_gen_call0(void *func, TCGHelperInfo *, TCGTemp *ret);
27
+ if (env->skip) {
27
+void tcg_gen_call1(void *func, TCGHelperInfo *, TCGTemp *ret, TCGTemp *);
28
+ return false;
28
+void tcg_gen_call2(void *func, TCGHelperInfo *, TCGTemp *ret,
29
+ }
29
TCGTemp *, TCGTemp *);
30
+
30
-void tcg_gen_call4(TCGHelperInfo *, TCGTemp *ret, TCGTemp *, TCGTemp *,
31
if (interrupt_request & CPU_INTERRUPT_RESET) {
31
- TCGTemp *, TCGTemp *);
32
if (cpu_interrupts_enabled(env)) {
32
-void tcg_gen_call5(TCGHelperInfo *, TCGTemp *ret, TCGTemp *, TCGTemp *,
33
cs->exception_index = EXCP_RESET;
33
+void tcg_gen_call3(void *func, TCGHelperInfo *, TCGTemp *ret,
34
diff --git a/target/avr/translate.c b/target/avr/translate.c
34
TCGTemp *, TCGTemp *, TCGTemp *);
35
-void tcg_gen_call6(TCGHelperInfo *, TCGTemp *ret, TCGTemp *, TCGTemp *,
36
+void tcg_gen_call4(void *func, TCGHelperInfo *, TCGTemp *ret,
37
TCGTemp *, TCGTemp *, TCGTemp *, TCGTemp *);
38
-void tcg_gen_call7(TCGHelperInfo *, TCGTemp *ret, TCGTemp *, TCGTemp *,
39
+void tcg_gen_call5(void *func, TCGHelperInfo *, TCGTemp *ret,
40
TCGTemp *, TCGTemp *, TCGTemp *, TCGTemp *, TCGTemp *);
41
+void tcg_gen_call6(void *func, TCGHelperInfo *, TCGTemp *ret,
42
+ TCGTemp *, TCGTemp *, TCGTemp *, TCGTemp *,
43
+ TCGTemp *, TCGTemp *);
44
+void tcg_gen_call7(void *func, TCGHelperInfo *, TCGTemp *ret,
45
+ TCGTemp *, TCGTemp *, TCGTemp *, TCGTemp *,
46
+ TCGTemp *, TCGTemp *, TCGTemp *);
47
48
TCGOp *tcg_emit_op(TCGOpcode opc, unsigned nargs);
49
void tcg_op_remove(TCGContext *s, TCGOp *op);
50
diff --git a/include/exec/helper-gen.h.inc b/include/exec/helper-gen.h.inc
35
index XXXXXXX..XXXXXXX 100644
51
index XXXXXXX..XXXXXXX 100644
36
--- a/target/avr/translate.c
52
--- a/include/exec/helper-gen.h.inc
37
+++ b/target/avr/translate.c
53
+++ b/include/exec/helper-gen.h.inc
38
@@ -XXX,XX +XXX,XX @@ static void avr_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
54
@@ -XXX,XX +XXX,XX @@
39
if (skip_label) {
55
extern TCGHelperInfo glue(helper_info_, name); \
40
canonicalize_skip(ctx);
56
static inline void glue(gen_helper_, name)(dh_retvar_decl0(ret)) \
41
gen_set_label(skip_label);
57
{ \
42
- if (ctx->base.is_jmp == DISAS_NORETURN) {
58
- tcg_gen_call0(&glue(helper_info_, name), dh_retvar(ret)); \
43
+
59
+ tcg_gen_call0(glue(helper_info_,name).func, \
44
+ switch (ctx->base.is_jmp) {
60
+ &glue(helper_info_,name), dh_retvar(ret)); \
45
+ case DISAS_NORETURN:
61
}
46
ctx->base.is_jmp = DISAS_CHAIN;
62
47
+ break;
63
#define DEF_HELPER_FLAGS_1(name, flags, ret, t1) \
48
+ case DISAS_NEXT:
64
@@ -XXX,XX +XXX,XX @@ extern TCGHelperInfo glue(helper_info_, name); \
49
+ if (ctx->base.tb->flags & TB_FLAGS_SKIP) {
65
static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
50
+ ctx->base.is_jmp = DISAS_TOO_MANY;
66
dh_arg_decl(t1, 1)) \
51
+ }
67
{ \
52
+ break;
68
- tcg_gen_call1(&glue(helper_info_, name), dh_retvar(ret), \
53
+ default:
69
+ tcg_gen_call1(glue(helper_info_,name).func, \
54
+ break;
70
+ &glue(helper_info_,name), dh_retvar(ret), \
71
dh_arg(t1, 1)); \
72
}
73
74
@@ -XXX,XX +XXX,XX @@ extern TCGHelperInfo glue(helper_info_, name); \
75
static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
76
dh_arg_decl(t1, 1), dh_arg_decl(t2, 2)) \
77
{ \
78
- tcg_gen_call2(&glue(helper_info_, name), dh_retvar(ret), \
79
+ tcg_gen_call2(glue(helper_info_,name).func, \
80
+ &glue(helper_info_,name), dh_retvar(ret), \
81
dh_arg(t1, 1), dh_arg(t2, 2)); \
82
}
83
84
@@ -XXX,XX +XXX,XX @@ extern TCGHelperInfo glue(helper_info_, name); \
85
static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
86
dh_arg_decl(t1, 1), dh_arg_decl(t2, 2), dh_arg_decl(t3, 3)) \
87
{ \
88
- tcg_gen_call3(&glue(helper_info_, name), dh_retvar(ret), \
89
+ tcg_gen_call3(glue(helper_info_,name).func, \
90
+ &glue(helper_info_,name), dh_retvar(ret), \
91
dh_arg(t1, 1), dh_arg(t2, 2), dh_arg(t3, 3)); \
92
}
93
94
@@ -XXX,XX +XXX,XX @@ static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
95
dh_arg_decl(t1, 1), dh_arg_decl(t2, 2), \
96
dh_arg_decl(t3, 3), dh_arg_decl(t4, 4)) \
97
{ \
98
- tcg_gen_call4(&glue(helper_info_, name), dh_retvar(ret), \
99
+ tcg_gen_call4(glue(helper_info_,name).func, \
100
+ &glue(helper_info_,name), dh_retvar(ret), \
101
dh_arg(t1, 1), dh_arg(t2, 2), \
102
dh_arg(t3, 3), dh_arg(t4, 4)); \
103
}
104
@@ -XXX,XX +XXX,XX @@ static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
105
dh_arg_decl(t1, 1), dh_arg_decl(t2, 2), dh_arg_decl(t3, 3), \
106
dh_arg_decl(t4, 4), dh_arg_decl(t5, 5)) \
107
{ \
108
- tcg_gen_call5(&glue(helper_info_, name), dh_retvar(ret), \
109
+ tcg_gen_call5(glue(helper_info_,name).func, \
110
+ &glue(helper_info_,name), dh_retvar(ret), \
111
dh_arg(t1, 1), dh_arg(t2, 2), dh_arg(t3, 3), \
112
dh_arg(t4, 4), dh_arg(t5, 5)); \
113
}
114
@@ -XXX,XX +XXX,XX @@ static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
115
dh_arg_decl(t1, 1), dh_arg_decl(t2, 2), dh_arg_decl(t3, 3), \
116
dh_arg_decl(t4, 4), dh_arg_decl(t5, 5), dh_arg_decl(t6, 6)) \
117
{ \
118
- tcg_gen_call6(&glue(helper_info_, name), dh_retvar(ret), \
119
+ tcg_gen_call6(glue(helper_info_,name).func, \
120
+ &glue(helper_info_,name), dh_retvar(ret), \
121
dh_arg(t1, 1), dh_arg(t2, 2), dh_arg(t3, 3), \
122
dh_arg(t4, 4), dh_arg(t5, 5), dh_arg(t6, 6)); \
123
}
124
@@ -XXX,XX +XXX,XX @@ static inline void glue(gen_helper_, name)(dh_retvar_decl(ret) \
125
dh_arg_decl(t4, 4), dh_arg_decl(t5, 5), dh_arg_decl(t6, 6), \
126
dh_arg_decl(t7, 7)) \
127
{ \
128
- tcg_gen_call7(&glue(helper_info_, name), dh_retvar(ret), \
129
+ tcg_gen_call7(glue(helper_info_,name).func, \
130
+ &glue(helper_info_,name), dh_retvar(ret), \
131
dh_arg(t1, 1), dh_arg(t2, 2), dh_arg(t3, 3), \
132
dh_arg(t4, 4), dh_arg(t5, 5), dh_arg(t6, 6), \
133
dh_arg(t7, 7)); \
134
diff --git a/tcg/tcg.c b/tcg/tcg.c
135
index XXXXXXX..XXXXXXX 100644
136
--- a/tcg/tcg.c
137
+++ b/tcg/tcg.c
138
@@ -XXX,XX +XXX,XX @@ bool tcg_op_supported(TCGOpcode op)
139
140
static TCGOp *tcg_op_alloc(TCGOpcode opc, unsigned nargs);
141
142
-static void tcg_gen_callN(TCGHelperInfo *info, TCGTemp *ret, TCGTemp **args)
143
+static void tcg_gen_callN(void *func, TCGHelperInfo *info,
144
+ TCGTemp *ret, TCGTemp **args)
145
{
146
TCGv_i64 extend_free[MAX_CALL_IARGS];
147
int n_extend = 0;
148
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_callN(TCGHelperInfo *info, TCGTemp *ret, TCGTemp **args)
149
g_assert_not_reached();
55
}
150
}
56
}
151
}
57
152
- op->args[pi++] = (uintptr_t)info->func;
58
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
153
+ op->args[pi++] = (uintptr_t)func;
59
{
154
op->args[pi++] = (uintptr_t)info;
60
DisasContext *ctx = container_of(dcbase, DisasContext, base);
155
tcg_debug_assert(pi == total_args);
61
bool nonconst_skip = canonicalize_skip(ctx);
156
62
+ /*
157
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_callN(TCGHelperInfo *info, TCGTemp *ret, TCGTemp **args)
63
+ * Because we disable interrupts while env->skip is set,
158
}
64
+ * we must return to the main loop to re-evaluate afterward.
159
}
65
+ */
160
66
+ bool force_exit = ctx->base.tb->flags & TB_FLAGS_SKIP;
161
-void tcg_gen_call0(TCGHelperInfo *info, TCGTemp *ret)
67
162
+void tcg_gen_call0(void *func, TCGHelperInfo *info, TCGTemp *ret)
68
switch (ctx->base.is_jmp) {
163
{
69
case DISAS_NORETURN:
164
- tcg_gen_callN(info, ret, NULL);
70
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
165
+ tcg_gen_callN(func, info, ret, NULL);
71
case DISAS_NEXT:
166
}
72
case DISAS_TOO_MANY:
167
73
case DISAS_CHAIN:
168
-void tcg_gen_call1(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1)
74
- if (!nonconst_skip) {
169
+void tcg_gen_call1(void *func, TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1)
75
+ if (!nonconst_skip && !force_exit) {
170
{
76
/* Note gen_goto_tb checks singlestep. */
171
- tcg_gen_callN(info, ret, &t1);
77
gen_goto_tb(ctx, 1, ctx->npc);
172
+ tcg_gen_callN(func, info, ret, &t1);
78
break;
173
}
79
@@ -XXX,XX +XXX,XX @@ static void avr_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
174
80
tcg_gen_movi_tl(cpu_pc, ctx->npc);
175
-void tcg_gen_call2(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1, TCGTemp *t2)
81
/* fall through */
176
+void tcg_gen_call2(void *func, TCGHelperInfo *info, TCGTemp *ret,
82
case DISAS_LOOKUP:
177
+ TCGTemp *t1, TCGTemp *t2)
83
- tcg_gen_lookup_and_goto_ptr();
178
{
84
- break;
179
TCGTemp *args[2] = { t1, t2 };
85
+ if (!force_exit) {
180
- tcg_gen_callN(info, ret, args);
86
+ tcg_gen_lookup_and_goto_ptr();
181
+ tcg_gen_callN(func, info, ret, args);
87
+ break;
182
}
88
+ }
183
89
+ /* fall through */
184
-void tcg_gen_call3(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
90
case DISAS_EXIT:
185
- TCGTemp *t2, TCGTemp *t3)
91
tcg_gen_exit_tb(NULL, 0);
186
+void tcg_gen_call3(void *func, TCGHelperInfo *info, TCGTemp *ret,
92
break;
187
+ TCGTemp *t1, TCGTemp *t2, TCGTemp *t3)
188
{
189
TCGTemp *args[3] = { t1, t2, t3 };
190
- tcg_gen_callN(info, ret, args);
191
+ tcg_gen_callN(func, info, ret, args);
192
}
193
194
-void tcg_gen_call4(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
195
- TCGTemp *t2, TCGTemp *t3, TCGTemp *t4)
196
+void tcg_gen_call4(void *func, TCGHelperInfo *info, TCGTemp *ret,
197
+ TCGTemp *t1, TCGTemp *t2, TCGTemp *t3, TCGTemp *t4)
198
{
199
TCGTemp *args[4] = { t1, t2, t3, t4 };
200
- tcg_gen_callN(info, ret, args);
201
+ tcg_gen_callN(func, info, ret, args);
202
}
203
204
-void tcg_gen_call5(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
205
+void tcg_gen_call5(void *func, TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
206
TCGTemp *t2, TCGTemp *t3, TCGTemp *t4, TCGTemp *t5)
207
{
208
TCGTemp *args[5] = { t1, t2, t3, t4, t5 };
209
- tcg_gen_callN(info, ret, args);
210
+ tcg_gen_callN(func, info, ret, args);
211
}
212
213
-void tcg_gen_call6(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1, TCGTemp *t2,
214
- TCGTemp *t3, TCGTemp *t4, TCGTemp *t5, TCGTemp *t6)
215
+void tcg_gen_call6(void *func, TCGHelperInfo *info, TCGTemp *ret,
216
+ TCGTemp *t1, TCGTemp *t2, TCGTemp *t3,
217
+ TCGTemp *t4, TCGTemp *t5, TCGTemp *t6)
218
{
219
TCGTemp *args[6] = { t1, t2, t3, t4, t5, t6 };
220
- tcg_gen_callN(info, ret, args);
221
+ tcg_gen_callN(func, info, ret, args);
222
}
223
224
-void tcg_gen_call7(TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
225
+void tcg_gen_call7(void *func, TCGHelperInfo *info, TCGTemp *ret, TCGTemp *t1,
226
TCGTemp *t2, TCGTemp *t3, TCGTemp *t4,
227
TCGTemp *t5, TCGTemp *t6, TCGTemp *t7)
228
{
229
TCGTemp *args[7] = { t1, t2, t3, t4, t5, t6, t7 };
230
- tcg_gen_callN(info, ret, args);
231
+ tcg_gen_callN(func, info, ret, args);
232
}
233
234
static void tcg_reg_alloc_start(TCGContext *s)
93
--
235
--
94
2.34.1
236
2.34.1
95
237
96
238
diff view generated by jsdifflib
1
The mmap_lock is held around tb_gen_code. While the comment
1
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
2
is correct that the lock is dropped when tb_gen_code runs out
3
of memory, the lock is *not* dropped when an exception is
4
raised reading code for translation.
5
6
Acked-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
3
---
11
accel/tcg/cpu-exec.c | 12 ++++++------
4
plugins/core.c | 2 +-
12
accel/tcg/user-exec.c | 3 ---
5
1 file changed, 1 insertion(+), 1 deletion(-)
13
2 files changed, 6 insertions(+), 9 deletions(-)
14
6
15
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
7
diff --git a/plugins/core.c b/plugins/core.c
16
index XXXXXXX..XXXXXXX 100644
8
index XXXXXXX..XXXXXXX 100644
17
--- a/accel/tcg/cpu-exec.c
9
--- a/plugins/core.c
18
+++ b/accel/tcg/cpu-exec.c
10
+++ b/plugins/core.c
19
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
11
@@ -XXX,XX +XXX,XX @@ static struct qemu_plugin_dyn_cb *plugin_get_dyn_cb(GArray **arr)
20
cpu_tb_exec(cpu, tb, &tb_exit);
12
GArray *cbs = *arr;
21
cpu_exec_exit(cpu);
13
22
} else {
14
if (!cbs) {
23
- /*
15
- cbs = g_array_sized_new(false, false,
24
- * The mmap_lock is dropped by tb_gen_code if it runs out of
16
+ cbs = g_array_sized_new(false, true,
25
- * memory.
17
sizeof(struct qemu_plugin_dyn_cb), 1);
26
- */
18
*arr = cbs;
27
#ifndef CONFIG_SOFTMMU
28
clear_helper_retaddr();
29
- tcg_debug_assert(!have_mmap_lock());
30
+ if (have_mmap_lock()) {
31
+ mmap_unlock();
32
+ }
33
#endif
34
if (qemu_mutex_iothread_locked()) {
35
qemu_mutex_unlock_iothread();
36
@@ -XXX,XX +XXX,XX @@ int cpu_exec(CPUState *cpu)
37
38
#ifndef CONFIG_SOFTMMU
39
clear_helper_retaddr();
40
- tcg_debug_assert(!have_mmap_lock());
41
+ if (have_mmap_lock()) {
42
+ mmap_unlock();
43
+ }
44
#endif
45
if (qemu_mutex_iothread_locked()) {
46
qemu_mutex_unlock_iothread();
47
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/accel/tcg/user-exec.c
50
+++ b/accel/tcg/user-exec.c
51
@@ -XXX,XX +XXX,XX @@ MMUAccessType adjust_signal_pc(uintptr_t *pc, bool is_write)
52
* (and if the translator doesn't handle page boundaries correctly
53
* there's little we can do about that here). Therefore, do not
54
* trigger the unwinder.
55
- *
56
- * Like tb_gen_code, release the memory lock before cpu_loop_exit.
57
*/
58
- mmap_unlock();
59
*pc = 0;
60
return MMU_INST_FETCH;
61
}
19
}
62
--
20
--
63
2.34.1
21
2.34.1
22
23
diff view generated by jsdifflib
1
We cannot deliver two interrupts simultaneously;
1
The out-of-line function pointer is mutually exclusive
2
the first interrupt handler must execute first.
2
with inline expansion, so move it into the union.
3
Wrap the pointer in a structure named 'regular' to match
4
PLUGIN_CB_REGULAR.
3
5
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
6
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
9
---
8
target/avr/helper.c | 9 +++------
10
include/qemu/plugin.h | 4 +++-
9
1 file changed, 3 insertions(+), 6 deletions(-)
11
accel/tcg/plugin-gen.c | 4 ++--
12
plugins/core.c | 8 ++++----
13
3 files changed, 9 insertions(+), 7 deletions(-)
10
14
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
15
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
12
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
13
--- a/target/avr/helper.c
17
--- a/include/qemu/plugin.h
14
+++ b/target/avr/helper.c
18
+++ b/include/qemu/plugin.h
15
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ enum plugin_dyn_cb_subtype {
16
20
* instance of a callback to be called upon the execution of a particular TB.
17
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
21
*/
18
{
22
struct qemu_plugin_dyn_cb {
19
- bool ret = false;
23
- union qemu_plugin_cb_sig f;
20
AVRCPU *cpu = AVR_CPU(cs);
24
void *userp;
21
CPUAVRState *env = &cpu->env;
25
enum plugin_dyn_cb_subtype type;
22
26
/* @rw applies to mem callbacks only (both regular and inline) */
23
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
27
enum qemu_plugin_mem_rw rw;
24
avr_cpu_do_interrupt(cs);
28
/* fields specific to each dyn_cb type go here */
25
29
union {
26
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
30
+ struct {
27
-
31
+ union qemu_plugin_cb_sig f;
28
- ret = true;
32
+ } regular;
29
+ return true;
33
struct {
34
qemu_plugin_u64 entry;
35
enum qemu_plugin_op op;
36
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/accel/tcg/plugin-gen.c
39
+++ b/accel/tcg/plugin-gen.c
40
@@ -XXX,XX +XXX,XX @@ static TCGOp *append_udata_cb(const struct qemu_plugin_dyn_cb *cb,
41
}
42
43
/* call */
44
- op = copy_call(&begin_op, op, cb->f.vcpu_udata, cb_idx);
45
+ op = copy_call(&begin_op, op, cb->regular.f.vcpu_udata, cb_idx);
46
47
return op;
48
}
49
@@ -XXX,XX +XXX,XX @@ static TCGOp *append_mem_cb(const struct qemu_plugin_dyn_cb *cb,
50
51
if (type == PLUGIN_GEN_CB_MEM) {
52
/* call */
53
- op = copy_call(&begin_op, op, cb->f.vcpu_udata, cb_idx);
54
+ op = copy_call(&begin_op, op, cb->regular.f.vcpu_udata, cb_idx);
55
}
56
57
return op;
58
diff --git a/plugins/core.c b/plugins/core.c
59
index XXXXXXX..XXXXXXX 100644
60
--- a/plugins/core.c
61
+++ b/plugins/core.c
62
@@ -XXX,XX +XXX,XX @@ void plugin_register_dyn_cb__udata(GArray **arr,
63
64
dyn_cb->userp = udata;
65
/* Note flags are discarded as unused. */
66
- dyn_cb->f.vcpu_udata = cb;
67
+ dyn_cb->regular.f.vcpu_udata = cb;
68
dyn_cb->type = PLUGIN_CB_REGULAR;
69
}
70
71
@@ -XXX,XX +XXX,XX @@ void plugin_register_vcpu_mem_cb(GArray **arr,
72
/* Note flags are discarded as unused. */
73
dyn_cb->type = PLUGIN_CB_REGULAR;
74
dyn_cb->rw = rw;
75
- dyn_cb->f.generic = cb;
76
+ dyn_cb->regular.f.vcpu_mem = cb;
77
}
78
79
/*
80
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_vcpu_mem_cb(CPUState *cpu, uint64_t vaddr,
30
}
81
}
31
}
82
switch (cb->type) {
32
if (interrupt_request & CPU_INTERRUPT_HARD) {
83
case PLUGIN_CB_REGULAR:
33
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
84
- cb->f.vcpu_mem(cpu->cpu_index, make_plugin_meminfo(oi, rw),
34
if (!env->intsrc) {
85
- vaddr, cb->userp);
35
cs->interrupt_request &= ~CPU_INTERRUPT_HARD;
86
+ cb->regular.f.vcpu_mem(cpu->cpu_index, make_plugin_meminfo(oi, rw),
36
}
87
+ vaddr, cb->userp);
37
-
88
break;
38
- ret = true;
89
case PLUGIN_CB_INLINE:
39
+ return true;
90
exec_inline_op(cb, cpu->cpu_index);
40
}
41
}
42
- return ret;
43
+ return false;
44
}
45
46
void avr_cpu_do_interrupt(CPUState *cs)
47
--
91
--
48
2.34.1
92
2.34.1
49
93
50
94
diff view generated by jsdifflib
1
Right now the translator stops right *after* the end of a page, which
1
TCGHelperInfo includes the ABI for every function call.
2
breaks reporting of fault locations when the last instruction of a
3
multi-insn translation block crosses a page boundary.
4
2
5
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1155
3
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
6
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
5
---
11
target/riscv/translate.c | 17 +++++--
6
include/qemu/plugin.h | 1 +
12
tests/tcg/riscv64/noexec.c | 79 +++++++++++++++++++++++++++++++
7
plugins/core.c | 51 ++++++++++++++++++++++++++++++++++++++-----
13
tests/tcg/riscv64/Makefile.target | 1 +
8
2 files changed, 46 insertions(+), 6 deletions(-)
14
3 files changed, 93 insertions(+), 4 deletions(-)
15
create mode 100644 tests/tcg/riscv64/noexec.c
16
9
17
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
10
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
18
index XXXXXXX..XXXXXXX 100644
11
index XXXXXXX..XXXXXXX 100644
19
--- a/target/riscv/translate.c
12
--- a/include/qemu/plugin.h
20
+++ b/target/riscv/translate.c
13
+++ b/include/qemu/plugin.h
21
@@ -XXX,XX +XXX,XX @@ static void riscv_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
14
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_dyn_cb {
22
}
15
union {
23
ctx->nftemp = 0;
16
struct {
24
17
union qemu_plugin_cb_sig f;
25
+ /* Only the first insn within a TB is allowed to cross a page boundary. */
18
+ TCGHelperInfo *info;
26
if (ctx->base.is_jmp == DISAS_NEXT) {
19
} regular;
27
- target_ulong page_start;
20
struct {
28
-
21
qemu_plugin_u64 entry;
29
- page_start = ctx->base.pc_first & TARGET_PAGE_MASK;
22
diff --git a/plugins/core.c b/plugins/core.c
30
- if (ctx->base.pc_next - page_start >= TARGET_PAGE_SIZE) {
23
index XXXXXXX..XXXXXXX 100644
31
+ if (!is_same_page(&ctx->base, ctx->base.pc_next)) {
24
--- a/plugins/core.c
32
ctx->base.is_jmp = DISAS_TOO_MANY;
25
+++ b/plugins/core.c
33
+ } else {
26
@@ -XXX,XX +XXX,XX @@ void plugin_register_dyn_cb__udata(GArray **arr,
34
+ unsigned page_ofs = ctx->base.pc_next & ~TARGET_PAGE_MASK;
27
enum qemu_plugin_cb_flags flags,
28
void *udata)
29
{
30
- struct qemu_plugin_dyn_cb *dyn_cb = plugin_get_dyn_cb(arr);
31
+ static TCGHelperInfo info[3] = {
32
+ [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG | TCG_CALL_PLUGIN,
33
+ [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG | TCG_CALL_PLUGIN,
34
+ [QEMU_PLUGIN_CB_RW_REGS].flags = TCG_CALL_PLUGIN,
35
+ /*
36
+ * Match qemu_plugin_vcpu_udata_cb_t:
37
+ * void (*)(uint32_t, void *)
38
+ */
39
+ [0 ... 2].typemask = (dh_typemask(void, 0) |
40
+ dh_typemask(i32, 1) |
41
+ dh_typemask(ptr, 2))
42
+ };
43
44
+ struct qemu_plugin_dyn_cb *dyn_cb = plugin_get_dyn_cb(arr);
45
dyn_cb->userp = udata;
46
- /* Note flags are discarded as unused. */
47
- dyn_cb->regular.f.vcpu_udata = cb;
48
dyn_cb->type = PLUGIN_CB_REGULAR;
49
+ dyn_cb->regular.f.vcpu_udata = cb;
35
+
50
+
36
+ if (page_ofs > TARGET_PAGE_SIZE - MAX_INSN_LEN) {
51
+ assert((unsigned)flags < ARRAY_SIZE(info));
37
+ uint16_t next_insn = cpu_lduw_code(env, ctx->base.pc_next);
52
+ dyn_cb->regular.info = &info[flags];
38
+ int len = insn_len(next_insn);
39
+
40
+ if (!is_same_page(&ctx->base, ctx->base.pc_next + len)) {
41
+ ctx->base.is_jmp = DISAS_TOO_MANY;
42
+ }
43
+ }
44
}
45
}
46
}
53
}
47
diff --git a/tests/tcg/riscv64/noexec.c b/tests/tcg/riscv64/noexec.c
54
48
new file mode 100644
55
void plugin_register_vcpu_mem_cb(GArray **arr,
49
index XXXXXXX..XXXXXXX
56
@@ -XXX,XX +XXX,XX @@ void plugin_register_vcpu_mem_cb(GArray **arr,
50
--- /dev/null
57
enum qemu_plugin_mem_rw rw,
51
+++ b/tests/tcg/riscv64/noexec.c
58
void *udata)
52
@@ -XXX,XX +XXX,XX @@
59
{
53
+#include "../multiarch/noexec.c.inc"
60
- struct qemu_plugin_dyn_cb *dyn_cb;
54
+
61
+ /*
55
+static void *arch_mcontext_pc(const mcontext_t *ctx)
62
+ * Expect that the underlying type for enum qemu_plugin_meminfo_t
56
+{
63
+ * is either int32_t or uint32_t, aka int or unsigned int.
57
+ return (void *)ctx->__gregs[REG_PC];
64
+ */
58
+}
65
+ QEMU_BUILD_BUG_ON(
59
+
66
+ !__builtin_types_compatible_p(qemu_plugin_meminfo_t, uint32_t) &&
60
+static int arch_mcontext_arg(const mcontext_t *ctx)
67
+ !__builtin_types_compatible_p(qemu_plugin_meminfo_t, int32_t));
61
+{
68
62
+ return ctx->__gregs[REG_A0];
69
- dyn_cb = plugin_get_dyn_cb(arr);
63
+}
70
+ static TCGHelperInfo info[3] = {
64
+
71
+ [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG | TCG_CALL_PLUGIN,
65
+static void arch_flush(void *p, int len)
72
+ [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG | TCG_CALL_PLUGIN,
66
+{
73
+ [QEMU_PLUGIN_CB_RW_REGS].flags = TCG_CALL_PLUGIN,
67
+ __builtin___clear_cache(p, p + len);
74
+ /*
68
+}
75
+ * Match qemu_plugin_vcpu_mem_cb_t:
69
+
76
+ * void (*)(uint32_t, qemu_plugin_meminfo_t, uint64_t, void *)
70
+extern char noexec_1[];
77
+ */
71
+extern char noexec_2[];
78
+ [0 ... 2].typemask =
72
+extern char noexec_end[];
79
+ (dh_typemask(void, 0) |
73
+
80
+ dh_typemask(i32, 1) |
74
+asm(".option push\n"
81
+ (__builtin_types_compatible_p(qemu_plugin_meminfo_t, uint32_t)
75
+ ".option norvc\n"
82
+ ? dh_typemask(i32, 2) : dh_typemask(s32, 2)) |
76
+ "noexec_1:\n"
83
+ dh_typemask(i64, 3) |
77
+ " li a0,1\n" /* a0 is 0 on entry, set 1. */
84
+ dh_typemask(ptr, 4))
78
+ "noexec_2:\n"
79
+ " li a0,2\n" /* a0 is 0/1; set 2. */
80
+ " ret\n"
81
+ "noexec_end:\n"
82
+ ".option pop");
83
+
84
+int main(void)
85
+{
86
+ struct noexec_test noexec_tests[] = {
87
+ {
88
+ .name = "fallthrough",
89
+ .test_code = noexec_1,
90
+ .test_len = noexec_end - noexec_1,
91
+ .page_ofs = noexec_1 - noexec_2,
92
+ .entry_ofs = noexec_1 - noexec_2,
93
+ .expected_si_ofs = 0,
94
+ .expected_pc_ofs = 0,
95
+ .expected_arg = 1,
96
+ },
97
+ {
98
+ .name = "jump",
99
+ .test_code = noexec_1,
100
+ .test_len = noexec_end - noexec_1,
101
+ .page_ofs = noexec_1 - noexec_2,
102
+ .entry_ofs = 0,
103
+ .expected_si_ofs = 0,
104
+ .expected_pc_ofs = 0,
105
+ .expected_arg = 0,
106
+ },
107
+ {
108
+ .name = "fallthrough [cross]",
109
+ .test_code = noexec_1,
110
+ .test_len = noexec_end - noexec_1,
111
+ .page_ofs = noexec_1 - noexec_2 - 2,
112
+ .entry_ofs = noexec_1 - noexec_2 - 2,
113
+ .expected_si_ofs = 0,
114
+ .expected_pc_ofs = -2,
115
+ .expected_arg = 1,
116
+ },
117
+ {
118
+ .name = "jump [cross]",
119
+ .test_code = noexec_1,
120
+ .test_len = noexec_end - noexec_1,
121
+ .page_ofs = noexec_1 - noexec_2 - 2,
122
+ .entry_ofs = -2,
123
+ .expected_si_ofs = 0,
124
+ .expected_pc_ofs = -2,
125
+ .expected_arg = 0,
126
+ },
127
+ };
85
+ };
128
+
86
+
129
+ return test_noexec(noexec_tests,
87
+ struct qemu_plugin_dyn_cb *dyn_cb = plugin_get_dyn_cb(arr);
130
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
88
dyn_cb->userp = udata;
131
+}
89
- /* Note flags are discarded as unused. */
132
diff --git a/tests/tcg/riscv64/Makefile.target b/tests/tcg/riscv64/Makefile.target
90
dyn_cb->type = PLUGIN_CB_REGULAR;
133
index XXXXXXX..XXXXXXX 100644
91
dyn_cb->rw = rw;
134
--- a/tests/tcg/riscv64/Makefile.target
92
dyn_cb->regular.f.vcpu_mem = cb;
135
+++ b/tests/tcg/riscv64/Makefile.target
93
+
136
@@ -XXX,XX +XXX,XX @@
94
+ assert((unsigned)flags < ARRAY_SIZE(info));
137
95
+ dyn_cb->regular.info = &info[flags];
138
VPATH += $(SRC_PATH)/tests/tcg/riscv64
96
}
139
TESTS += test-div
97
140
+TESTS += noexec
98
/*
141
--
99
--
142
2.34.1
100
2.34.1
diff view generated by jsdifflib
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
1
Introduce a new plugin_cb op and migrate one operation.
2
By using emit_before_op, we do not need to emit opcodes
3
early and modify them later -- we can simply emit the
4
final set of opcodes once.
2
5
3
Right now translator stops right *after* the end of a page, which
6
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
4
breaks reporting of fault locations when the last instruction of a
5
multi-insn translation block crosses a page boundary.
6
7
An implementation, like the one arm and s390x have, would require an
8
i386 length disassembler, which is burdensome to maintain. Another
9
alternative would be to single-step at the end of a guest page, but
10
this may come with a performance impact.
11
12
Fix by snapshotting disassembly state and restoring it after we figure
13
out we crossed a page boundary. This includes rolling back cc_op
14
updates and emitted ops.
15
16
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1143
19
Message-Id: <20220817150506.592862-4-iii@linux.ibm.com>
20
[rth: Simplify end-of-insn cross-page checks.]
21
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
22
---
8
---
23
target/i386/tcg/translate.c | 64 ++++++++++++++++-----------
9
include/tcg/tcg-op-common.h | 1 +
24
tests/tcg/x86_64/noexec.c | 75 ++++++++++++++++++++++++++++++++
10
include/tcg/tcg-opc.h | 1 +
25
tests/tcg/x86_64/Makefile.target | 3 +-
11
accel/tcg/plugin-gen.c | 74 +++++++++++++++++++++----------------
26
3 files changed, 116 insertions(+), 26 deletions(-)
12
tcg/tcg-op.c | 5 +++
27
create mode 100644 tests/tcg/x86_64/noexec.c
13
4 files changed, 50 insertions(+), 31 deletions(-)
28
14
29
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
15
diff --git a/include/tcg/tcg-op-common.h b/include/tcg/tcg-op-common.h
30
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
31
--- a/target/i386/tcg/translate.c
17
--- a/include/tcg/tcg-op-common.h
32
+++ b/target/i386/tcg/translate.c
18
+++ b/include/tcg/tcg-op-common.h
33
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
19
@@ -XXX,XX +XXX,XX @@ void tcg_gen_goto_tb(unsigned idx);
34
TCGv_i64 tmp1_i64;
20
*/
35
21
void tcg_gen_lookup_and_goto_ptr(void);
36
sigjmp_buf jmpbuf;
22
37
+ TCGOp *prev_insn_end;
23
+void tcg_gen_plugin_cb(unsigned from);
38
} DisasContext;
24
void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr);
39
25
void tcg_gen_plugin_cb_end(void);
40
/* The environment in which user-only runs is constrained. */
26
41
@@ -XXX,XX +XXX,XX @@ static uint64_t advance_pc(CPUX86State *env, DisasContext *s, int num_bytes)
27
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
28
index XXXXXXX..XXXXXXX 100644
29
--- a/include/tcg/tcg-opc.h
30
+++ b/include/tcg/tcg-opc.h
31
@@ -XXX,XX +XXX,XX @@ DEF(exit_tb, 0, 0, 1, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
32
DEF(goto_tb, 0, 0, 1, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
33
DEF(goto_ptr, 0, 1, 0, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
34
35
+DEF(plugin_cb, 0, 0, 1, TCG_OPF_NOT_PRESENT)
36
DEF(plugin_cb_start, 0, 0, 3, TCG_OPF_NOT_PRESENT)
37
DEF(plugin_cb_end, 0, 0, 0, TCG_OPF_NOT_PRESENT)
38
39
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
40
index XXXXXXX..XXXXXXX 100644
41
--- a/accel/tcg/plugin-gen.c
42
+++ b/accel/tcg/plugin-gen.c
43
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_empty_callback(enum plugin_gen_from from)
42
{
44
{
43
uint64_t pc = s->pc;
45
switch (from) {
44
46
case PLUGIN_GEN_AFTER_INSN:
45
+ /* This is a subsequent insn that crosses a page boundary. */
47
- gen_wrapped(from, PLUGIN_GEN_DISABLE_MEM_HELPER,
46
+ if (s->base.num_insns > 1 &&
48
- gen_empty_mem_helper);
47
+ !is_same_page(&s->base, s->pc + num_bytes - 1)) {
49
+ tcg_gen_plugin_cb(from);
48
+ siglongjmp(s->jmpbuf, 2);
50
break;
49
+ }
51
case PLUGIN_GEN_FROM_INSN:
50
+
52
/*
51
s->pc += num_bytes;
53
@@ -XXX,XX +XXX,XX @@ static void inject_mem_enable_helper(struct qemu_plugin_tb *ptb,
52
if (unlikely(s->pc - s->pc_start > X86_MAX_INSN_LENGTH)) {
54
inject_mem_helper(begin_op, arr);
53
/* If the instruction's 16th byte is on a different page than the 1st, a
55
}
54
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
56
55
int modrm, reg, rm, mod, op, opreg, val;
57
-static void inject_mem_disable_helper(struct qemu_plugin_insn *plugin_insn,
56
target_ulong next_eip, tval;
58
- TCGOp *begin_op)
57
target_ulong pc_start = s->base.pc_next;
59
-{
58
+ bool orig_cc_op_dirty = s->cc_op_dirty;
60
- if (likely(!plugin_insn->mem_helper)) {
59
+ CCOp orig_cc_op = s->cc_op;
61
- rm_ops(begin_op);
60
62
- return;
61
s->pc_start = s->pc = pc_start;
63
- }
62
s->override = -1;
64
- inject_mem_helper(begin_op, NULL);
63
@@ -XXX,XX +XXX,XX @@ static target_ulong disas_insn(DisasContext *s, CPUState *cpu)
65
-}
64
s->rip_offset = 0; /* for relative ip address */
66
-
65
s->vex_l = 0;
67
/* called before finishing a TB with exit_tb, goto_tb or goto_ptr */
66
s->vex_v = 0;
68
void plugin_gen_disable_mem_helpers(void)
67
- if (sigsetjmp(s->jmpbuf, 0) != 0) {
68
+ switch (sigsetjmp(s->jmpbuf, 0)) {
69
+ case 0:
70
+ break;
71
+ case 1:
72
gen_exception_gpf(s);
73
return s->pc;
74
+ case 2:
75
+ /* Restore state that may affect the next instruction. */
76
+ s->cc_op_dirty = orig_cc_op_dirty;
77
+ s->cc_op = orig_cc_op;
78
+ s->base.num_insns--;
79
+ tcg_remove_ops_after(s->prev_insn_end);
80
+ s->base.is_jmp = DISAS_TOO_MANY;
81
+ return pc_start;
82
+ default:
83
+ g_assert_not_reached();
84
}
85
86
prefixes = 0;
87
@@ -XXX,XX +XXX,XX @@ static void i386_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
88
{
69
{
89
DisasContext *dc = container_of(dcbase, DisasContext, base);
70
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
90
71
inject_mem_enable_helper(ptb, insn, begin_op);
91
+ dc->prev_insn_end = tcg_last_op();
92
tcg_gen_insn_start(dc->base.pc_next, dc->cc_op);
93
}
72
}
94
73
95
@@ -XXX,XX +XXX,XX @@ static void i386_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
74
-static void plugin_gen_disable_mem_helper(struct qemu_plugin_tb *ptb,
96
#endif
75
- TCGOp *begin_op, int insn_idx)
97
76
+static void gen_disable_mem_helper(struct qemu_plugin_tb *ptb,
98
pc_next = disas_insn(dc, cpu);
77
+ struct qemu_plugin_insn *insn)
99
-
78
{
100
- if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
79
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
101
- /* if single step mode, we generate only one instruction and
80
- inject_mem_disable_helper(insn, begin_op);
102
- generate an exception */
81
+ if (insn->mem_helper) {
103
- /* if irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
82
+ tcg_gen_st_ptr(tcg_constant_ptr(0), tcg_env,
104
- the flag and abort the translation to give the irqs a
83
+ offsetof(CPUState, plugin_mem_cbs) -
105
- chance to happen */
84
+ offsetof(ArchCPU, env));
106
- dc->base.is_jmp = DISAS_TOO_MANY;
107
- } else if ((tb_cflags(dc->base.tb) & CF_USE_ICOUNT)
108
- && ((pc_next & TARGET_PAGE_MASK)
109
- != ((pc_next + TARGET_MAX_INSN_SIZE - 1)
110
- & TARGET_PAGE_MASK)
111
- || (pc_next & ~TARGET_PAGE_MASK) == 0)) {
112
- /* Do not cross the boundary of the pages in icount mode,
113
- it can cause an exception. Do it only when boundary is
114
- crossed by the first instruction in the block.
115
- If current instruction already crossed the bound - it's ok,
116
- because an exception hasn't stopped this code.
117
- */
118
- dc->base.is_jmp = DISAS_TOO_MANY;
119
- } else if ((pc_next - dc->base.pc_first) >= (TARGET_PAGE_SIZE - 32)) {
120
- dc->base.is_jmp = DISAS_TOO_MANY;
121
- }
122
-
123
dc->base.pc_next = pc_next;
124
+
125
+ if (dc->base.is_jmp == DISAS_NEXT) {
126
+ if (dc->flags & (HF_TF_MASK | HF_INHIBIT_IRQ_MASK)) {
127
+ /*
128
+ * If single step mode, we generate only one instruction and
129
+ * generate an exception.
130
+ * If irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
131
+ * the flag and abort the translation to give the irqs a
132
+ * chance to happen.
133
+ */
134
+ dc->base.is_jmp = DISAS_TOO_MANY;
135
+ } else if (!is_same_page(&dc->base, pc_next)) {
136
+ dc->base.is_jmp = DISAS_TOO_MANY;
137
+ }
138
+ }
85
+ }
139
}
86
}
140
87
141
static void i386_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
88
/* #define DEBUG_PLUGIN_GEN_OPS */
142
diff --git a/tests/tcg/x86_64/noexec.c b/tests/tcg/x86_64/noexec.c
89
@@ -XXX,XX +XXX,XX @@ static void pr_ops(void)
143
new file mode 100644
90
144
index XXXXXXX..XXXXXXX
91
static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
145
--- /dev/null
92
{
146
+++ b/tests/tcg/x86_64/noexec.c
93
- TCGOp *op;
147
@@ -XXX,XX +XXX,XX @@
94
+ TCGOp *op, *next;
148
+#include "../multiarch/noexec.c.inc"
95
int insn_idx = -1;
96
97
pr_ops();
98
99
- QTAILQ_FOREACH(op, &tcg_ctx->ops, link) {
100
+ /*
101
+ * While injecting code, we cannot afford to reuse any ebb temps
102
+ * that might be live within the existing opcode stream.
103
+ * The simplest solution is to release them all and create new.
104
+ */
105
+ memset(tcg_ctx->free_temps, 0, sizeof(tcg_ctx->free_temps));
149
+
106
+
150
+static void *arch_mcontext_pc(const mcontext_t *ctx)
107
+ QTAILQ_FOREACH_SAFE(op, &tcg_ctx->ops, link, next) {
108
switch (op->opc) {
109
case INDEX_op_insn_start:
110
insn_idx++;
111
break;
112
+
113
+ case INDEX_op_plugin_cb:
114
+ {
115
+ enum plugin_gen_from from = op->args[0];
116
+ struct qemu_plugin_insn *insn = NULL;
117
+
118
+ if (insn_idx >= 0) {
119
+ insn = g_ptr_array_index(plugin_tb->insns, insn_idx);
120
+ }
121
+
122
+ tcg_ctx->emit_before_op = op;
123
+
124
+ switch (from) {
125
+ case PLUGIN_GEN_AFTER_INSN:
126
+ assert(insn != NULL);
127
+ gen_disable_mem_helper(plugin_tb, insn);
128
+ break;
129
+ default:
130
+ g_assert_not_reached();
131
+ }
132
+
133
+ tcg_ctx->emit_before_op = NULL;
134
+ tcg_op_remove(tcg_ctx, op);
135
+ break;
136
+ }
137
+
138
case INDEX_op_plugin_cb_start:
139
{
140
enum plugin_gen_from from = op->args[0];
141
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
142
143
break;
144
}
145
- case PLUGIN_GEN_AFTER_INSN:
146
- {
147
- g_assert(insn_idx >= 0);
148
-
149
- switch (type) {
150
- case PLUGIN_GEN_DISABLE_MEM_HELPER:
151
- plugin_gen_disable_mem_helper(plugin_tb, op, insn_idx);
152
- break;
153
- default:
154
- g_assert_not_reached();
155
- }
156
- break;
157
- }
158
default:
159
g_assert_not_reached();
160
}
161
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
162
index XXXXXXX..XXXXXXX 100644
163
--- a/tcg/tcg-op.c
164
+++ b/tcg/tcg-op.c
165
@@ -XXX,XX +XXX,XX @@ void tcg_gen_mb(TCGBar mb_type)
166
}
167
}
168
169
+void tcg_gen_plugin_cb(unsigned from)
151
+{
170
+{
152
+ return (void *)ctx->gregs[REG_RIP];
171
+ tcg_gen_op1(INDEX_op_plugin_cb, from);
153
+}
172
+}
154
+
173
+
155
+int arch_mcontext_arg(const mcontext_t *ctx)
174
void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr)
156
+{
175
{
157
+ return ctx->gregs[REG_RDI];
176
tcg_gen_op3(INDEX_op_plugin_cb_start, from, type, wr);
158
+}
159
+
160
+static void arch_flush(void *p, int len)
161
+{
162
+}
163
+
164
+extern char noexec_1[];
165
+extern char noexec_2[];
166
+extern char noexec_end[];
167
+
168
+asm("noexec_1:\n"
169
+ " movq $1,%rdi\n" /* %rdi is 0 on entry, set 1. */
170
+ "noexec_2:\n"
171
+ " movq $2,%rdi\n" /* %rdi is 0/1; set 2. */
172
+ " ret\n"
173
+ "noexec_end:");
174
+
175
+int main(void)
176
+{
177
+ struct noexec_test noexec_tests[] = {
178
+ {
179
+ .name = "fallthrough",
180
+ .test_code = noexec_1,
181
+ .test_len = noexec_end - noexec_1,
182
+ .page_ofs = noexec_1 - noexec_2,
183
+ .entry_ofs = noexec_1 - noexec_2,
184
+ .expected_si_ofs = 0,
185
+ .expected_pc_ofs = 0,
186
+ .expected_arg = 1,
187
+ },
188
+ {
189
+ .name = "jump",
190
+ .test_code = noexec_1,
191
+ .test_len = noexec_end - noexec_1,
192
+ .page_ofs = noexec_1 - noexec_2,
193
+ .entry_ofs = 0,
194
+ .expected_si_ofs = 0,
195
+ .expected_pc_ofs = 0,
196
+ .expected_arg = 0,
197
+ },
198
+ {
199
+ .name = "fallthrough [cross]",
200
+ .test_code = noexec_1,
201
+ .test_len = noexec_end - noexec_1,
202
+ .page_ofs = noexec_1 - noexec_2 - 2,
203
+ .entry_ofs = noexec_1 - noexec_2 - 2,
204
+ .expected_si_ofs = 0,
205
+ .expected_pc_ofs = -2,
206
+ .expected_arg = 1,
207
+ },
208
+ {
209
+ .name = "jump [cross]",
210
+ .test_code = noexec_1,
211
+ .test_len = noexec_end - noexec_1,
212
+ .page_ofs = noexec_1 - noexec_2 - 2,
213
+ .entry_ofs = -2,
214
+ .expected_si_ofs = 0,
215
+ .expected_pc_ofs = -2,
216
+ .expected_arg = 0,
217
+ },
218
+ };
219
+
220
+ return test_noexec(noexec_tests,
221
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
222
+}
223
diff --git a/tests/tcg/x86_64/Makefile.target b/tests/tcg/x86_64/Makefile.target
224
index XXXXXXX..XXXXXXX 100644
225
--- a/tests/tcg/x86_64/Makefile.target
226
+++ b/tests/tcg/x86_64/Makefile.target
227
@@ -XXX,XX +XXX,XX @@ include $(SRC_PATH)/tests/tcg/i386/Makefile.target
228
229
ifeq ($(filter %-linux-user, $(TARGET)),$(TARGET))
230
X86_64_TESTS += vsyscall
231
+X86_64_TESTS += noexec
232
TESTS=$(MULTIARCH_TESTS) $(X86_64_TESTS) test-x86_64
233
else
234
TESTS=$(MULTIARCH_TESTS)
235
@@ -XXX,XX +XXX,XX @@ test-x86_64: LDFLAGS+=-lm -lc
236
test-x86_64: test-i386.c test-i386.h test-i386-shift.h test-i386-muldiv.h
237
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
238
239
-vsyscall: $(SRC_PATH)/tests/tcg/x86_64/vsyscall.c
240
+%: $(SRC_PATH)/tests/tcg/x86_64/%.c
241
    $(CC) $(CFLAGS) $< -o $@ $(LDFLAGS)
242
--
177
--
243
2.34.1
178
2.34.1
diff view generated by jsdifflib
1
Cache the translation from guest to host address, so we may
1
By having the qemu_plugin_cb_flags be recorded in the TCGHelperInfo,
2
use direct loads when we hit on the primary translation page.
2
we no longer need to distinguish PLUGIN_CB_REGULAR from
3
PLUGIN_CB_REGULAR_R, so place all TB callbacks in the same queue.
3
4
4
Look up the second translation page only once, during translation.
5
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
5
This obviates another lookup of the second page within tb_gen_code
6
after translation.
7
8
Fixes a bug in that plugin_insn_append should be passed the bytes
9
in the original memory order, not bswapped by pieces.
10
11
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
12
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
14
---
7
---
15
include/exec/translator.h | 63 +++++++++++--------
8
accel/tcg/plugin-gen.c | 96 +++++++++++++++++++++++++-----------------
16
accel/tcg/translate-all.c | 23 +++----
9
plugins/api.c | 6 +--
17
accel/tcg/translator.c | 126 +++++++++++++++++++++++++++++---------
10
2 files changed, 58 insertions(+), 44 deletions(-)
18
3 files changed, 141 insertions(+), 71 deletions(-)
19
11
20
diff --git a/include/exec/translator.h b/include/exec/translator.h
12
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
21
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
22
--- a/include/exec/translator.h
14
--- a/accel/tcg/plugin-gen.c
23
+++ b/include/exec/translator.h
15
+++ b/accel/tcg/plugin-gen.c
24
@@ -XXX,XX +XXX,XX @@ typedef enum DisasJumpType {
16
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_empty_callback(enum plugin_gen_from from)
25
* Architecture-agnostic disassembly context.
26
*/
27
typedef struct DisasContextBase {
28
- const TranslationBlock *tb;
29
+ TranslationBlock *tb;
30
target_ulong pc_first;
31
target_ulong pc_next;
32
DisasJumpType is_jmp;
33
int num_insns;
34
int max_insns;
35
bool singlestep_enabled;
36
-#ifdef CONFIG_USER_ONLY
37
- /*
38
- * Guest address of the last byte of the last protected page.
39
- *
40
- * Pages containing the translated instructions are made non-writable in
41
- * order to achieve consistency in case another thread is modifying the
42
- * code while translate_insn() fetches the instruction bytes piecemeal.
43
- * Such writer threads are blocked on mmap_lock() in page_unprotect().
44
- */
45
- target_ulong page_protect_end;
46
-#endif
47
+ void *host_addr[2];
48
} DisasContextBase;
49
50
/**
51
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
52
* the relevant information at translation time.
53
*/
54
55
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
56
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
57
- abi_ptr pc, bool do_swap); \
58
- static inline type fullname(CPUArchState *env, \
59
- DisasContextBase *dcbase, abi_ptr pc) \
60
- { \
61
- return fullname ## _swap(env, dcbase, pc, false); \
62
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
63
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
64
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
65
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc);
66
+
67
+static inline uint16_t
68
+translator_lduw_swap(CPUArchState *env, DisasContextBase *db,
69
+ abi_ptr pc, bool do_swap)
70
+{
71
+ uint16_t ret = translator_lduw(env, db, pc);
72
+ if (do_swap) {
73
+ ret = bswap16(ret);
74
}
75
+ return ret;
76
+}
77
78
-#define FOR_EACH_TRANSLATOR_LD(F) \
79
- F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
80
- F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
81
- F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
82
- F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
83
+static inline uint32_t
84
+translator_ldl_swap(CPUArchState *env, DisasContextBase *db,
85
+ abi_ptr pc, bool do_swap)
86
+{
87
+ uint32_t ret = translator_ldl(env, db, pc);
88
+ if (do_swap) {
89
+ ret = bswap32(ret);
90
+ }
91
+ return ret;
92
+}
93
94
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
95
-
96
-#undef GEN_TRANSLATOR_LD
97
+static inline uint64_t
98
+translator_ldq_swap(CPUArchState *env, DisasContextBase *db,
99
+ abi_ptr pc, bool do_swap)
100
+{
101
+ uint64_t ret = translator_ldq_swap(env, db, pc, false);
102
+ if (do_swap) {
103
+ ret = bswap64(ret);
104
+ }
105
+ return ret;
106
+}
107
108
/*
109
* Return whether addr is on the same page as where disassembly started.
110
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
111
index XXXXXXX..XXXXXXX 100644
112
--- a/accel/tcg/translate-all.c
113
+++ b/accel/tcg/translate-all.c
114
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
115
{
17
{
116
CPUArchState *env = cpu->env_ptr;
18
switch (from) {
117
TranslationBlock *tb, *existing_tb;
19
case PLUGIN_GEN_AFTER_INSN:
118
- tb_page_addr_t phys_pc, phys_page2;
20
+ case PLUGIN_GEN_FROM_TB:
119
- target_ulong virt_page2;
21
tcg_gen_plugin_cb(from);
120
+ tb_page_addr_t phys_pc;
22
break;
121
tcg_insn_unit *gen_code_buf;
23
case PLUGIN_GEN_FROM_INSN:
122
int gen_code_size, search_size, max_insns;
24
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_empty_callback(enum plugin_gen_from from)
123
#ifdef CONFIG_PROFILER
25
*/
124
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
26
gen_wrapped(from, PLUGIN_GEN_ENABLE_MEM_HELPER,
125
tb->flags = flags;
27
gen_empty_mem_helper);
126
tb->cflags = cflags;
28
- /* fall through */
127
tb->trace_vcpu_dstate = *cpu->trace_dstate;
29
- case PLUGIN_GEN_FROM_TB:
128
+ tb->page_addr[0] = phys_pc;
30
gen_wrapped(from, PLUGIN_GEN_CB_UDATA, gen_empty_udata_cb_no_rwg);
129
+ tb->page_addr[1] = -1;
31
gen_wrapped(from, PLUGIN_GEN_CB_UDATA_R, gen_empty_udata_cb_no_wg);
130
tcg_ctx->tb_cflags = cflags;
32
gen_wrapped(from, PLUGIN_GEN_CB_INLINE, gen_empty_inline_cb);
131
tb_overflow:
33
@@ -XXX,XX +XXX,XX @@ void plugin_gen_disable_mem_helpers(void)
132
34
offsetof(CPUState, plugin_mem_cbs) - offsetof(ArchCPU, env));
133
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
134
}
135
136
/*
137
- * If the TB is not associated with a physical RAM page then
138
- * it must be a temporary one-insn TB, and we have nothing to do
139
- * except fill in the page_addr[] fields. Return early before
140
- * attempting to link to other TBs or add to the lookup table.
141
+ * If the TB is not associated with a physical RAM page then it must be
142
+ * a temporary one-insn TB, and we have nothing left to do. Return early
143
+ * before attempting to link to other TBs or add to the lookup table.
144
*/
145
- if (phys_pc == -1) {
146
- tb->page_addr[0] = tb->page_addr[1] = -1;
147
+ if (tb->page_addr[0] == -1) {
148
return tb;
149
}
150
151
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
152
*/
153
tcg_tb_insert(tb);
154
155
- /* check next page if needed */
156
- virt_page2 = (pc + tb->size - 1) & TARGET_PAGE_MASK;
157
- phys_page2 = -1;
158
- if ((pc & TARGET_PAGE_MASK) != virt_page2) {
159
- phys_page2 = get_page_addr_code(env, virt_page2);
160
- }
161
/*
162
* No explicit memory barrier is required -- tb_link_page() makes the
163
* TB visible in a consistent state.
164
*/
165
- existing_tb = tb_link_page(tb, phys_pc, phys_page2);
166
+ existing_tb = tb_link_page(tb, tb->page_addr[0], tb->page_addr[1]);
167
/* if the TB already exists, discard what we just translated */
168
if (unlikely(existing_tb != tb)) {
169
uintptr_t orig_aligned = (uintptr_t)gen_code_buf;
170
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
171
index XXXXXXX..XXXXXXX 100644
172
--- a/accel/tcg/translator.c
173
+++ b/accel/tcg/translator.c
174
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest)
175
return ((db->pc_first ^ dest) & TARGET_PAGE_MASK) == 0;
176
}
35
}
177
36
178
-static inline void translator_page_protect(DisasContextBase *dcbase,
37
-static void plugin_gen_tb_udata(const struct qemu_plugin_tb *ptb,
179
- target_ulong pc)
38
- TCGOp *begin_op)
180
-{
39
-{
181
-#ifdef CONFIG_USER_ONLY
40
- inject_udata_cb(ptb->cbs[PLUGIN_CB_REGULAR], begin_op);
182
- dcbase->page_protect_end = pc | ~TARGET_PAGE_MASK;
183
- page_protect(pc);
184
-#endif
185
-}
41
-}
186
-
42
-
187
void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
43
-static void plugin_gen_tb_udata_r(const struct qemu_plugin_tb *ptb,
188
target_ulong pc, void *host_pc,
44
- TCGOp *begin_op)
189
const TranslatorOps *ops, DisasContextBase *db)
45
-{
190
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
46
- inject_udata_cb(ptb->cbs[PLUGIN_CB_REGULAR_R], begin_op);
191
db->num_insns = 0;
47
-}
192
db->max_insns = max_insns;
48
-
193
db->singlestep_enabled = cflags & CF_SINGLE_STEP;
49
-static void plugin_gen_tb_inline(const struct qemu_plugin_tb *ptb,
194
- translator_page_protect(db, db->pc_next);
50
- TCGOp *begin_op)
195
+ db->host_addr[0] = host_pc;
51
-{
196
+ db->host_addr[1] = NULL;
52
- inject_inline_cb(ptb->cbs[PLUGIN_CB_INLINE], begin_op, op_ok);
53
-}
54
-
55
static void plugin_gen_insn_udata(const struct qemu_plugin_tb *ptb,
56
TCGOp *begin_op, int insn_idx)
57
{
58
@@ -XXX,XX +XXX,XX @@ static void gen_disable_mem_helper(struct qemu_plugin_tb *ptb,
59
}
60
}
61
62
+static void gen_udata_cb(struct qemu_plugin_dyn_cb *cb)
63
+{
64
+ TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
197
+
65
+
198
+#ifdef CONFIG_USER_ONLY
66
+ tcg_gen_ld_i32(cpu_index, tcg_env,
199
+ page_protect(pc);
67
+ -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
200
+#endif
68
+ tcg_gen_call2(cb->regular.f.vcpu_udata, cb->regular.info, NULL,
201
69
+ tcgv_i32_temp(cpu_index),
202
ops->init_disas_context(db, cpu);
70
+ tcgv_ptr_temp(tcg_constant_ptr(cb->userp)));
203
tcg_debug_assert(db->is_jmp == DISAS_NEXT); /* no early exit */
71
+ tcg_temp_free_i32(cpu_index);
204
@@ -XXX,XX +XXX,XX @@ void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
205
#endif
206
}
207
208
-static inline void translator_maybe_page_protect(DisasContextBase *dcbase,
209
- target_ulong pc, size_t len)
210
+static void *translator_access(CPUArchState *env, DisasContextBase *db,
211
+ target_ulong pc, size_t len)
212
{
213
-#ifdef CONFIG_USER_ONLY
214
- target_ulong end = pc + len - 1;
215
+ void *host;
216
+ target_ulong base, end;
217
+ TranslationBlock *tb;
218
219
- if (end > dcbase->page_protect_end) {
220
- translator_page_protect(dcbase, end);
221
+ tb = db->tb;
222
+
223
+ /* Use slow path if first page is MMIO. */
224
+ if (unlikely(tb->page_addr[0] == -1)) {
225
+ return NULL;
226
}
227
+
228
+ end = pc + len - 1;
229
+ if (likely(is_same_page(db, end))) {
230
+ host = db->host_addr[0];
231
+ base = db->pc_first;
232
+ } else {
233
+ host = db->host_addr[1];
234
+ base = TARGET_PAGE_ALIGN(db->pc_first);
235
+ if (host == NULL) {
236
+ tb->page_addr[1] =
237
+ get_page_addr_code_hostp(env, base, &db->host_addr[1]);
238
+#ifdef CONFIG_USER_ONLY
239
+ page_protect(end);
240
#endif
241
+ /* We cannot handle MMIO as second page. */
242
+ assert(tb->page_addr[1] != -1);
243
+ host = db->host_addr[1];
244
+ }
245
+
246
+ /* Use slow path when crossing pages. */
247
+ if (is_same_page(db, pc)) {
248
+ return NULL;
249
+ }
250
+ }
251
+
252
+ tcg_debug_assert(pc >= base);
253
+ return host + (pc - base);
254
}
255
256
-#define GEN_TRANSLATOR_LD(fullname, type, load_fn, swap_fn) \
257
- type fullname ## _swap(CPUArchState *env, DisasContextBase *dcbase, \
258
- abi_ptr pc, bool do_swap) \
259
- { \
260
- translator_maybe_page_protect(dcbase, pc, sizeof(type)); \
261
- type ret = load_fn(env, pc); \
262
- if (do_swap) { \
263
- ret = swap_fn(ret); \
264
- } \
265
- plugin_insn_append(pc, &ret, sizeof(ret)); \
266
- return ret; \
267
+uint8_t translator_ldub(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
268
+{
269
+ uint8_t ret;
270
+ void *p = translator_access(env, db, pc, sizeof(ret));
271
+
272
+ if (p) {
273
+ plugin_insn_append(pc, p, sizeof(ret));
274
+ return ldub_p(p);
275
}
276
+ ret = cpu_ldub_code(env, pc);
277
+ plugin_insn_append(pc, &ret, sizeof(ret));
278
+ return ret;
279
+}
280
281
-FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
282
+uint16_t translator_lduw(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
283
+{
284
+ uint16_t ret, plug;
285
+ void *p = translator_access(env, db, pc, sizeof(ret));
286
287
-#undef GEN_TRANSLATOR_LD
288
+ if (p) {
289
+ plugin_insn_append(pc, p, sizeof(ret));
290
+ return lduw_p(p);
291
+ }
292
+ ret = cpu_lduw_code(env, pc);
293
+ plug = tswap16(ret);
294
+ plugin_insn_append(pc, &plug, sizeof(ret));
295
+ return ret;
296
+}
72
+}
297
+
73
+
298
+uint32_t translator_ldl(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
74
+static void gen_inline_cb(struct qemu_plugin_dyn_cb *cb)
299
+{
75
+{
300
+ uint32_t ret, plug;
76
+ GArray *arr = cb->inline_insn.entry.score->data;
301
+ void *p = translator_access(env, db, pc, sizeof(ret));
77
+ size_t offset = cb->inline_insn.entry.offset;
78
+ TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
79
+ TCGv_i64 val = tcg_temp_ebb_new_i64();
80
+ TCGv_ptr ptr = tcg_temp_ebb_new_ptr();
302
+
81
+
303
+ if (p) {
82
+ tcg_gen_ld_i32(cpu_index, tcg_env,
304
+ plugin_insn_append(pc, p, sizeof(ret));
83
+ -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
305
+ return ldl_p(p);
84
+ tcg_gen_muli_i32(cpu_index, cpu_index, g_array_get_element_size(arr));
306
+ }
85
+ tcg_gen_ext_i32_ptr(ptr, cpu_index);
307
+ ret = cpu_ldl_code(env, pc);
86
+ tcg_temp_free_i32(cpu_index);
308
+ plug = tswap32(ret);
87
+
309
+ plugin_insn_append(pc, &plug, sizeof(ret));
88
+ tcg_gen_addi_ptr(ptr, ptr, (intptr_t)arr->data);
310
+ return ret;
89
+ tcg_gen_ld_i64(val, ptr, offset);
90
+ tcg_gen_addi_i64(val, val, cb->inline_insn.imm);
91
+ tcg_gen_st_i64(val, ptr, offset);
92
+
93
+ tcg_temp_free_i64(val);
94
+ tcg_temp_free_ptr(ptr);
311
+}
95
+}
312
+
96
+
313
+uint64_t translator_ldq(CPUArchState *env, DisasContextBase *db, abi_ptr pc)
97
/* #define DEBUG_PLUGIN_GEN_OPS */
314
+{
98
static void pr_ops(void)
315
+ uint64_t ret, plug;
99
{
316
+ void *p = translator_access(env, db, pc, sizeof(ret));
100
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
101
{
102
enum plugin_gen_from from = op->args[0];
103
struct qemu_plugin_insn *insn = NULL;
104
+ const GArray *cbs;
105
+ int i, n;
106
107
if (insn_idx >= 0) {
108
insn = g_ptr_array_index(plugin_tb->insns, insn_idx);
109
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
110
assert(insn != NULL);
111
gen_disable_mem_helper(plugin_tb, insn);
112
break;
317
+
113
+
318
+ if (p) {
114
+ case PLUGIN_GEN_FROM_TB:
319
+ plugin_insn_append(pc, p, sizeof(ret));
115
+ assert(insn == NULL);
320
+ return ldq_p(p);
116
+
321
+ }
117
+ cbs = plugin_tb->cbs[PLUGIN_CB_REGULAR];
322
+ ret = cpu_ldq_code(env, pc);
118
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
323
+ plug = tswap64(ret);
119
+ struct qemu_plugin_dyn_cb *cb =
324
+ plugin_insn_append(pc, &plug, sizeof(ret));
120
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
325
+ return ret;
121
+ gen_udata_cb(cb);
326
+}
122
+ }
123
+
124
+ cbs = plugin_tb->cbs[PLUGIN_CB_INLINE];
125
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
126
+ struct qemu_plugin_dyn_cb *cb =
127
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
128
+ gen_inline_cb(cb);
129
+ }
130
+ break;
131
+
132
default:
133
g_assert_not_reached();
134
}
135
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
136
enum plugin_gen_cb type = op->args[1];
137
138
switch (from) {
139
- case PLUGIN_GEN_FROM_TB:
140
- {
141
- g_assert(insn_idx == -1);
142
-
143
- switch (type) {
144
- case PLUGIN_GEN_CB_UDATA:
145
- plugin_gen_tb_udata(plugin_tb, op);
146
- break;
147
- case PLUGIN_GEN_CB_UDATA_R:
148
- plugin_gen_tb_udata_r(plugin_tb, op);
149
- break;
150
- case PLUGIN_GEN_CB_INLINE:
151
- plugin_gen_tb_inline(plugin_tb, op);
152
- break;
153
- default:
154
- g_assert_not_reached();
155
- }
156
- break;
157
- }
158
case PLUGIN_GEN_FROM_INSN:
159
{
160
g_assert(insn_idx >= 0);
161
diff --git a/plugins/api.c b/plugins/api.c
162
index XXXXXXX..XXXXXXX 100644
163
--- a/plugins/api.c
164
+++ b/plugins/api.c
165
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_tb_exec_cb(struct qemu_plugin_tb *tb,
166
void *udata)
167
{
168
if (!tb->mem_only) {
169
- int index = flags == QEMU_PLUGIN_CB_R_REGS ||
170
- flags == QEMU_PLUGIN_CB_RW_REGS ?
171
- PLUGIN_CB_REGULAR_R : PLUGIN_CB_REGULAR;
172
-
173
- plugin_register_dyn_cb__udata(&tb->cbs[index],
174
+ plugin_register_dyn_cb__udata(&tb->cbs[PLUGIN_CB_REGULAR],
175
cb, flags, udata);
176
}
177
}
327
--
178
--
328
2.34.1
179
2.34.1
diff view generated by jsdifflib
1
Pass these along to translator_loop -- pc may be used instead
1
Delay test of plugin_tb->mem_helper until the inject pass.
2
of tb->pc, and host_pc is currently unused. Adjust all targets
3
at one time.
4
2
5
Acked-by: Alistair Francis <alistair.francis@wdc.com>
3
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
5
---
10
include/exec/exec-all.h | 1 -
6
accel/tcg/plugin-gen.c | 37 ++++++++++++++++---------------------
11
include/exec/translator.h | 24 ++++++++++++++++++++----
7
1 file changed, 16 insertions(+), 21 deletions(-)
12
accel/tcg/translate-all.c | 6 ++++--
13
accel/tcg/translator.c | 9 +++++----
14
target/alpha/translate.c | 5 +++--
15
target/arm/translate.c | 5 +++--
16
target/avr/translate.c | 5 +++--
17
target/cris/translate.c | 5 +++--
18
target/hexagon/translate.c | 6 ++++--
19
target/hppa/translate.c | 5 +++--
20
target/i386/tcg/translate.c | 5 +++--
21
target/loongarch/translate.c | 6 ++++--
22
target/m68k/translate.c | 5 +++--
23
target/microblaze/translate.c | 5 +++--
24
target/mips/tcg/translate.c | 5 +++--
25
target/nios2/translate.c | 5 +++--
26
target/openrisc/translate.c | 6 ++++--
27
target/ppc/translate.c | 5 +++--
28
target/riscv/translate.c | 5 +++--
29
target/rx/translate.c | 5 +++--
30
target/s390x/tcg/translate.c | 5 +++--
31
target/sh4/translate.c | 5 +++--
32
target/sparc/translate.c | 5 +++--
33
target/tricore/translate.c | 6 ++++--
34
target/xtensa/translate.c | 6 ++++--
35
25 files changed, 97 insertions(+), 53 deletions(-)
36
8
37
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
9
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
38
index XXXXXXX..XXXXXXX 100644
10
index XXXXXXX..XXXXXXX 100644
39
--- a/include/exec/exec-all.h
11
--- a/accel/tcg/plugin-gen.c
40
+++ b/include/exec/exec-all.h
12
+++ b/accel/tcg/plugin-gen.c
41
@@ -XXX,XX +XXX,XX @@ typedef ram_addr_t tb_page_addr_t;
13
@@ -XXX,XX +XXX,XX @@ enum plugin_gen_from {
42
#define TB_PAGE_ADDR_FMT RAM_ADDR_FMT
14
PLUGIN_GEN_FROM_INSN,
43
#endif
15
PLUGIN_GEN_FROM_MEM,
44
16
PLUGIN_GEN_AFTER_INSN,
45
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns);
17
+ PLUGIN_GEN_AFTER_TB,
46
void restore_state_to_opc(CPUArchState *env, TranslationBlock *tb,
18
PLUGIN_GEN_N_FROMS,
47
target_ulong *data);
19
};
48
20
49
diff --git a/include/exec/translator.h b/include/exec/translator.h
21
@@ -XXX,XX +XXX,XX @@ static void inject_mem_enable_helper(struct qemu_plugin_tb *ptb,
50
index XXXXXXX..XXXXXXX 100644
22
/* called before finishing a TB with exit_tb, goto_tb or goto_ptr */
51
--- a/include/exec/translator.h
23
void plugin_gen_disable_mem_helpers(void)
52
+++ b/include/exec/translator.h
24
{
53
@@ -XXX,XX +XXX,XX @@
25
- /*
54
#include "exec/translate-all.h"
26
- * We could emit the clearing unconditionally and be done. However, this can
55
#include "tcg/tcg.h"
27
- * be wasteful if for instance plugins don't track memory accesses, or if
56
28
- * most TBs don't use helpers. Instead, emit the clearing iff the TB calls
57
+/**
29
- * helpers that might access guest memory.
58
+ * gen_intermediate_code
30
- *
59
+ * @cpu: cpu context
31
- * Note: we do not reset plugin_tb->mem_helper here; a TB might have several
60
+ * @tb: translation block
32
- * exit points, and we want to emit the clearing from all of them.
61
+ * @max_insns: max number of instructions to translate
33
- */
62
+ * @pc: guest virtual program counter address
34
- if (!tcg_ctx->plugin_tb->mem_helper) {
63
+ * @host_pc: host physical program counter address
35
- return;
64
+ *
36
+ if (tcg_ctx->plugin_insn) {
65
+ * This function must be provided by the target, which should create
37
+ tcg_gen_plugin_cb(PLUGIN_GEN_AFTER_TB);
66
+ * the target-specific DisasContext, and then invoke translator_loop.
38
}
67
+ */
39
- tcg_gen_st_ptr(tcg_constant_ptr(NULL), tcg_env,
68
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
40
- offsetof(CPUState, plugin_mem_cbs) - offsetof(ArchCPU, env));
69
+ target_ulong pc, void *host_pc);
70
71
/**
72
* DisasJumpType:
73
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
74
75
/**
76
* translator_loop:
77
- * @ops: Target-specific operations.
78
- * @db: Disassembly context.
79
* @cpu: Target vCPU.
80
* @tb: Translation block.
81
* @max_insns: Maximum number of insns to translate.
82
+ * @pc: guest virtual program counter address
83
+ * @host_pc: host physical program counter address
84
+ * @ops: Target-specific operations.
85
+ * @db: Disassembly context.
86
*
87
* Generic translator loop.
88
*
89
@@ -XXX,XX +XXX,XX @@ typedef struct TranslatorOps {
90
* - When single-stepping is enabled (system-wide or on the current vCPU).
91
* - When too many instructions have been translated.
92
*/
93
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
94
- CPUState *cpu, TranslationBlock *tb, int max_insns);
95
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
96
+ target_ulong pc, void *host_pc,
97
+ const TranslatorOps *ops, DisasContextBase *db);
98
99
void translator_loop_temp_check(DisasContextBase *db);
100
101
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
102
index XXXXXXX..XXXXXXX 100644
103
--- a/accel/tcg/translate-all.c
104
+++ b/accel/tcg/translate-all.c
105
@@ -XXX,XX +XXX,XX @@
106
107
#include "exec/cputlb.h"
108
#include "exec/translate-all.h"
109
+#include "exec/translator.h"
110
#include "qemu/bitmap.h"
111
#include "qemu/qemu-print.h"
112
#include "qemu/timer.h"
113
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
114
TCGProfile *prof = &tcg_ctx->prof;
115
int64_t ti;
116
#endif
117
+ void *host_pc;
118
119
assert_memory_lock();
120
qemu_thread_jit_write();
121
122
- phys_pc = get_page_addr_code(env, pc);
123
+ phys_pc = get_page_addr_code_hostp(env, pc, &host_pc);
124
125
if (phys_pc == -1) {
126
/* Generate a one-shot TB with 1 insn in it */
127
@@ -XXX,XX +XXX,XX @@ TranslationBlock *tb_gen_code(CPUState *cpu,
128
tcg_func_start(tcg_ctx);
129
130
tcg_ctx->cpu = env_cpu(env);
131
- gen_intermediate_code(cpu, tb, max_insns);
132
+ gen_intermediate_code(cpu, tb, max_insns, pc, host_pc);
133
assert(tb->size != 0);
134
tcg_ctx->cpu = NULL;
135
max_insns = tb->icount;
136
diff --git a/accel/tcg/translator.c b/accel/tcg/translator.c
137
index XXXXXXX..XXXXXXX 100644
138
--- a/accel/tcg/translator.c
139
+++ b/accel/tcg/translator.c
140
@@ -XXX,XX +XXX,XX @@ static inline void translator_page_protect(DisasContextBase *dcbase,
141
#endif
142
}
41
}
143
42
144
-void translator_loop(const TranslatorOps *ops, DisasContextBase *db,
43
static void plugin_gen_insn_udata(const struct qemu_plugin_tb *ptb,
145
- CPUState *cpu, TranslationBlock *tb, int max_insns)
44
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
146
+void translator_loop(CPUState *cpu, TranslationBlock *tb, int max_insns,
45
inject_mem_enable_helper(ptb, insn, begin_op);
147
+ target_ulong pc, void *host_pc,
46
}
148
+ const TranslatorOps *ops, DisasContextBase *db)
47
48
-static void gen_disable_mem_helper(struct qemu_plugin_tb *ptb,
49
- struct qemu_plugin_insn *insn)
50
+static void gen_disable_mem_helper(void)
149
{
51
{
150
uint32_t cflags = tb_cflags(tb);
52
- if (insn->mem_helper) {
151
bool plugin_enabled;
53
- tcg_gen_st_ptr(tcg_constant_ptr(0), tcg_env,
152
54
- offsetof(CPUState, plugin_mem_cbs) -
153
/* Initialize DisasContext */
55
- offsetof(ArchCPU, env));
154
db->tb = tb;
56
- }
155
- db->pc_first = tb->pc;
57
+ tcg_gen_st_ptr(tcg_constant_ptr(0), tcg_env,
156
- db->pc_next = db->pc_first;
58
+ offsetof(CPUState, plugin_mem_cbs) -
157
+ db->pc_first = pc;
59
+ offsetof(ArchCPU, env));
158
+ db->pc_next = pc;
159
db->is_jmp = DISAS_NEXT;
160
db->num_insns = 0;
161
db->max_insns = max_insns;
162
diff --git a/target/alpha/translate.c b/target/alpha/translate.c
163
index XXXXXXX..XXXXXXX 100644
164
--- a/target/alpha/translate.c
165
+++ b/target/alpha/translate.c
166
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps alpha_tr_ops = {
167
.disas_log = alpha_tr_disas_log,
168
};
169
170
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
171
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
172
+ target_ulong pc, void *host_pc)
173
{
174
DisasContext dc;
175
- translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
176
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &alpha_tr_ops, &dc.base);
177
}
60
}
178
61
179
void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
62
static void gen_udata_cb(struct qemu_plugin_dyn_cb *cb)
180
diff --git a/target/arm/translate.c b/target/arm/translate.c
63
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
181
index XXXXXXX..XXXXXXX 100644
64
tcg_ctx->emit_before_op = op;
182
--- a/target/arm/translate.c
65
183
+++ b/target/arm/translate.c
66
switch (from) {
184
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps thumb_translator_ops = {
67
+ case PLUGIN_GEN_AFTER_TB:
185
};
68
+ if (plugin_tb->mem_helper) {
186
69
+ gen_disable_mem_helper();
187
/* generate intermediate code for basic block 'tb'. */
70
+ }
188
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
71
+ break;
189
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
72
+
190
+ target_ulong pc, void *host_pc)
73
case PLUGIN_GEN_AFTER_INSN:
191
{
74
assert(insn != NULL);
192
DisasContext dc = { };
75
- gen_disable_mem_helper(plugin_tb, insn);
193
const TranslatorOps *ops = &arm_translator_ops;
76
+ if (insn->mem_helper) {
194
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
77
+ gen_disable_mem_helper();
195
}
78
+ }
196
#endif
79
break;
197
80
198
- translator_loop(ops, &dc.base, cpu, tb, max_insns);
81
case PLUGIN_GEN_FROM_TB:
199
+ translator_loop(cpu, tb, max_insns, pc, host_pc, ops, &dc.base);
200
}
201
202
void restore_state_to_opc(CPUARMState *env, TranslationBlock *tb,
203
diff --git a/target/avr/translate.c b/target/avr/translate.c
204
index XXXXXXX..XXXXXXX 100644
205
--- a/target/avr/translate.c
206
+++ b/target/avr/translate.c
207
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps avr_tr_ops = {
208
.disas_log = avr_tr_disas_log,
209
};
210
211
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
212
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
213
+ target_ulong pc, void *host_pc)
214
{
215
DisasContext dc = { };
216
- translator_loop(&avr_tr_ops, &dc.base, cs, tb, max_insns);
217
+ translator_loop(cs, tb, max_insns, pc, host_pc, &avr_tr_ops, &dc.base);
218
}
219
220
void restore_state_to_opc(CPUAVRState *env, TranslationBlock *tb,
221
diff --git a/target/cris/translate.c b/target/cris/translate.c
222
index XXXXXXX..XXXXXXX 100644
223
--- a/target/cris/translate.c
224
+++ b/target/cris/translate.c
225
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps cris_tr_ops = {
226
.disas_log = cris_tr_disas_log,
227
};
228
229
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
230
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
231
+ target_ulong pc, void *host_pc)
232
{
233
DisasContext dc;
234
- translator_loop(&cris_tr_ops, &dc.base, cs, tb, max_insns);
235
+ translator_loop(cs, tb, max_insns, pc, host_pc, &cris_tr_ops, &dc.base);
236
}
237
238
void cris_cpu_dump_state(CPUState *cs, FILE *f, int flags)
239
diff --git a/target/hexagon/translate.c b/target/hexagon/translate.c
240
index XXXXXXX..XXXXXXX 100644
241
--- a/target/hexagon/translate.c
242
+++ b/target/hexagon/translate.c
243
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hexagon_tr_ops = {
244
.disas_log = hexagon_tr_disas_log,
245
};
246
247
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
248
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
249
+ target_ulong pc, void *host_pc)
250
{
251
DisasContext ctx;
252
253
- translator_loop(&hexagon_tr_ops, &ctx.base, cs, tb, max_insns);
254
+ translator_loop(cs, tb, max_insns, pc, host_pc,
255
+ &hexagon_tr_ops, &ctx.base);
256
}
257
258
#define NAME_LEN 64
259
diff --git a/target/hppa/translate.c b/target/hppa/translate.c
260
index XXXXXXX..XXXXXXX 100644
261
--- a/target/hppa/translate.c
262
+++ b/target/hppa/translate.c
263
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps hppa_tr_ops = {
264
.disas_log = hppa_tr_disas_log,
265
};
266
267
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
268
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
269
+ target_ulong pc, void *host_pc)
270
{
271
DisasContext ctx;
272
- translator_loop(&hppa_tr_ops, &ctx.base, cs, tb, max_insns);
273
+ translator_loop(cs, tb, max_insns, pc, host_pc, &hppa_tr_ops, &ctx.base);
274
}
275
276
void restore_state_to_opc(CPUHPPAState *env, TranslationBlock *tb,
277
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
278
index XXXXXXX..XXXXXXX 100644
279
--- a/target/i386/tcg/translate.c
280
+++ b/target/i386/tcg/translate.c
281
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps i386_tr_ops = {
282
};
283
284
/* generate intermediate code for basic block 'tb'. */
285
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
286
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
287
+ target_ulong pc, void *host_pc)
288
{
289
DisasContext dc;
290
291
- translator_loop(&i386_tr_ops, &dc.base, cpu, tb, max_insns);
292
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &i386_tr_ops, &dc.base);
293
}
294
295
void restore_state_to_opc(CPUX86State *env, TranslationBlock *tb,
296
diff --git a/target/loongarch/translate.c b/target/loongarch/translate.c
297
index XXXXXXX..XXXXXXX 100644
298
--- a/target/loongarch/translate.c
299
+++ b/target/loongarch/translate.c
300
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps loongarch_tr_ops = {
301
.disas_log = loongarch_tr_disas_log,
302
};
303
304
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
305
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
306
+ target_ulong pc, void *host_pc)
307
{
308
DisasContext ctx;
309
310
- translator_loop(&loongarch_tr_ops, &ctx.base, cs, tb, max_insns);
311
+ translator_loop(cs, tb, max_insns, pc, host_pc,
312
+ &loongarch_tr_ops, &ctx.base);
313
}
314
315
void loongarch_translate_init(void)
316
diff --git a/target/m68k/translate.c b/target/m68k/translate.c
317
index XXXXXXX..XXXXXXX 100644
318
--- a/target/m68k/translate.c
319
+++ b/target/m68k/translate.c
320
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps m68k_tr_ops = {
321
.disas_log = m68k_tr_disas_log,
322
};
323
324
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
325
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
326
+ target_ulong pc, void *host_pc)
327
{
328
DisasContext dc;
329
- translator_loop(&m68k_tr_ops, &dc.base, cpu, tb, max_insns);
330
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &m68k_tr_ops, &dc.base);
331
}
332
333
static double floatx80_to_double(CPUM68KState *env, uint16_t high, uint64_t low)
334
diff --git a/target/microblaze/translate.c b/target/microblaze/translate.c
335
index XXXXXXX..XXXXXXX 100644
336
--- a/target/microblaze/translate.c
337
+++ b/target/microblaze/translate.c
338
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mb_tr_ops = {
339
.disas_log = mb_tr_disas_log,
340
};
341
342
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
343
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
344
+ target_ulong pc, void *host_pc)
345
{
346
DisasContext dc;
347
- translator_loop(&mb_tr_ops, &dc.base, cpu, tb, max_insns);
348
+ translator_loop(cpu, tb, max_insns, pc, host_pc, &mb_tr_ops, &dc.base);
349
}
350
351
void mb_cpu_dump_state(CPUState *cs, FILE *f, int flags)
352
diff --git a/target/mips/tcg/translate.c b/target/mips/tcg/translate.c
353
index XXXXXXX..XXXXXXX 100644
354
--- a/target/mips/tcg/translate.c
355
+++ b/target/mips/tcg/translate.c
356
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps mips_tr_ops = {
357
.disas_log = mips_tr_disas_log,
358
};
359
360
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
361
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
362
+ target_ulong pc, void *host_pc)
363
{
364
DisasContext ctx;
365
366
- translator_loop(&mips_tr_ops, &ctx.base, cs, tb, max_insns);
367
+ translator_loop(cs, tb, max_insns, pc, host_pc, &mips_tr_ops, &ctx.base);
368
}
369
370
void mips_tcg_init(void)
371
diff --git a/target/nios2/translate.c b/target/nios2/translate.c
372
index XXXXXXX..XXXXXXX 100644
373
--- a/target/nios2/translate.c
374
+++ b/target/nios2/translate.c
375
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps nios2_tr_ops = {
376
.disas_log = nios2_tr_disas_log,
377
};
378
379
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
380
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
381
+ target_ulong pc, void *host_pc)
382
{
383
DisasContext dc;
384
- translator_loop(&nios2_tr_ops, &dc.base, cs, tb, max_insns);
385
+ translator_loop(cs, tb, max_insns, pc, host_pc, &nios2_tr_ops, &dc.base);
386
}
387
388
void nios2_cpu_dump_state(CPUState *cs, FILE *f, int flags)
389
diff --git a/target/openrisc/translate.c b/target/openrisc/translate.c
390
index XXXXXXX..XXXXXXX 100644
391
--- a/target/openrisc/translate.c
392
+++ b/target/openrisc/translate.c
393
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps openrisc_tr_ops = {
394
.disas_log = openrisc_tr_disas_log,
395
};
396
397
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
398
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
399
+ target_ulong pc, void *host_pc)
400
{
401
DisasContext ctx;
402
403
- translator_loop(&openrisc_tr_ops, &ctx.base, cs, tb, max_insns);
404
+ translator_loop(cs, tb, max_insns, pc, host_pc,
405
+ &openrisc_tr_ops, &ctx.base);
406
}
407
408
void openrisc_cpu_dump_state(CPUState *cs, FILE *f, int flags)
409
diff --git a/target/ppc/translate.c b/target/ppc/translate.c
410
index XXXXXXX..XXXXXXX 100644
411
--- a/target/ppc/translate.c
412
+++ b/target/ppc/translate.c
413
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps ppc_tr_ops = {
414
.disas_log = ppc_tr_disas_log,
415
};
416
417
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
418
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
419
+ target_ulong pc, void *host_pc)
420
{
421
DisasContext ctx;
422
423
- translator_loop(&ppc_tr_ops, &ctx.base, cs, tb, max_insns);
424
+ translator_loop(cs, tb, max_insns, pc, host_pc, &ppc_tr_ops, &ctx.base);
425
}
426
427
void restore_state_to_opc(CPUPPCState *env, TranslationBlock *tb,
428
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
429
index XXXXXXX..XXXXXXX 100644
430
--- a/target/riscv/translate.c
431
+++ b/target/riscv/translate.c
432
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps riscv_tr_ops = {
433
.disas_log = riscv_tr_disas_log,
434
};
435
436
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
437
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
438
+ target_ulong pc, void *host_pc)
439
{
440
DisasContext ctx;
441
442
- translator_loop(&riscv_tr_ops, &ctx.base, cs, tb, max_insns);
443
+ translator_loop(cs, tb, max_insns, pc, host_pc, &riscv_tr_ops, &ctx.base);
444
}
445
446
void riscv_translate_init(void)
447
diff --git a/target/rx/translate.c b/target/rx/translate.c
448
index XXXXXXX..XXXXXXX 100644
449
--- a/target/rx/translate.c
450
+++ b/target/rx/translate.c
451
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps rx_tr_ops = {
452
.disas_log = rx_tr_disas_log,
453
};
454
455
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
456
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
457
+ target_ulong pc, void *host_pc)
458
{
459
DisasContext dc;
460
461
- translator_loop(&rx_tr_ops, &dc.base, cs, tb, max_insns);
462
+ translator_loop(cs, tb, max_insns, pc, host_pc, &rx_tr_ops, &dc.base);
463
}
464
465
void restore_state_to_opc(CPURXState *env, TranslationBlock *tb,
466
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
467
index XXXXXXX..XXXXXXX 100644
468
--- a/target/s390x/tcg/translate.c
469
+++ b/target/s390x/tcg/translate.c
470
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps s390x_tr_ops = {
471
.disas_log = s390x_tr_disas_log,
472
};
473
474
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
475
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
476
+ target_ulong pc, void *host_pc)
477
{
478
DisasContext dc;
479
480
- translator_loop(&s390x_tr_ops, &dc.base, cs, tb, max_insns);
481
+ translator_loop(cs, tb, max_insns, pc, host_pc, &s390x_tr_ops, &dc.base);
482
}
483
484
void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb,
485
diff --git a/target/sh4/translate.c b/target/sh4/translate.c
486
index XXXXXXX..XXXXXXX 100644
487
--- a/target/sh4/translate.c
488
+++ b/target/sh4/translate.c
489
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sh4_tr_ops = {
490
.disas_log = sh4_tr_disas_log,
491
};
492
493
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
494
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
495
+ target_ulong pc, void *host_pc)
496
{
497
DisasContext ctx;
498
499
- translator_loop(&sh4_tr_ops, &ctx.base, cs, tb, max_insns);
500
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base);
501
}
502
503
void restore_state_to_opc(CPUSH4State *env, TranslationBlock *tb,
504
diff --git a/target/sparc/translate.c b/target/sparc/translate.c
505
index XXXXXXX..XXXXXXX 100644
506
--- a/target/sparc/translate.c
507
+++ b/target/sparc/translate.c
508
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps sparc_tr_ops = {
509
.disas_log = sparc_tr_disas_log,
510
};
511
512
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
513
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
514
+ target_ulong pc, void *host_pc)
515
{
516
DisasContext dc = {};
517
518
- translator_loop(&sparc_tr_ops, &dc.base, cs, tb, max_insns);
519
+ translator_loop(cs, tb, max_insns, pc, host_pc, &sparc_tr_ops, &dc.base);
520
}
521
522
void sparc_tcg_init(void)
523
diff --git a/target/tricore/translate.c b/target/tricore/translate.c
524
index XXXXXXX..XXXXXXX 100644
525
--- a/target/tricore/translate.c
526
+++ b/target/tricore/translate.c
527
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps tricore_tr_ops = {
528
};
529
530
531
-void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
532
+void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns,
533
+ target_ulong pc, void *host_pc)
534
{
535
DisasContext ctx;
536
- translator_loop(&tricore_tr_ops, &ctx.base, cs, tb, max_insns);
537
+ translator_loop(cs, tb, max_insns, pc, host_pc,
538
+ &tricore_tr_ops, &ctx.base);
539
}
540
541
void
542
diff --git a/target/xtensa/translate.c b/target/xtensa/translate.c
543
index XXXXXXX..XXXXXXX 100644
544
--- a/target/xtensa/translate.c
545
+++ b/target/xtensa/translate.c
546
@@ -XXX,XX +XXX,XX @@ static const TranslatorOps xtensa_translator_ops = {
547
.disas_log = xtensa_tr_disas_log,
548
};
549
550
-void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
551
+void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns,
552
+ target_ulong pc, void *host_pc)
553
{
554
DisasContext dc = {};
555
- translator_loop(&xtensa_translator_ops, &dc.base, cpu, tb, max_insns);
556
+ translator_loop(cpu, tb, max_insns, pc, host_pc,
557
+ &xtensa_translator_ops, &dc.base);
558
}
559
560
void xtensa_cpu_dump_state(CPUState *cs, FILE *f, int flags)
561
--
82
--
562
2.34.1
83
2.34.1
diff view generated by jsdifflib
1
We're about to start validating PAGE_EXEC, which means
1
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
2
that we've got to mark the commpage executable. We had
3
been placing the commpage outside of reserved_va, which
4
was incorrect and lead to an abort.
5
6
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
---
3
---
10
linux-user/arm/target_cpu.h | 4 ++--
4
include/qemu/plugin.h | 1 -
11
linux-user/elfload.c | 6 +++++-
5
accel/tcg/plugin-gen.c | 286 ++++++++++-------------------------------
12
2 files changed, 7 insertions(+), 3 deletions(-)
6
plugins/api.c | 8 +-
7
3 files changed, 67 insertions(+), 228 deletions(-)
13
8
14
diff --git a/linux-user/arm/target_cpu.h b/linux-user/arm/target_cpu.h
9
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
15
index XXXXXXX..XXXXXXX 100644
10
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/arm/target_cpu.h
11
--- a/include/qemu/plugin.h
17
+++ b/linux-user/arm/target_cpu.h
12
+++ b/include/qemu/plugin.h
18
@@ -XXX,XX +XXX,XX @@ static inline unsigned long arm_max_reserved_va(CPUState *cs)
13
@@ -XXX,XX +XXX,XX @@ enum plugin_dyn_cb_type {
19
} else {
14
20
/*
15
enum plugin_dyn_cb_subtype {
21
* We need to be able to map the commpage.
16
PLUGIN_CB_REGULAR,
22
- * See validate_guest_space in linux-user/elfload.c.
17
- PLUGIN_CB_REGULAR_R,
23
+ * See init_guest_commpage in linux-user/elfload.c.
18
PLUGIN_CB_INLINE,
24
*/
19
PLUGIN_N_CB_SUBTYPES,
25
- return 0xffff0000ul;
20
};
26
+ return 0xfffffffful;
21
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
22
index XXXXXXX..XXXXXXX 100644
23
--- a/accel/tcg/plugin-gen.c
24
+++ b/accel/tcg/plugin-gen.c
25
@@ -XXX,XX +XXX,XX @@ void HELPER(plugin_vcpu_mem_cb)(unsigned int vcpu_index,
26
void *userdata)
27
{ }
28
29
-static void gen_empty_udata_cb(void (*gen_helper)(TCGv_i32, TCGv_ptr))
30
-{
31
- TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
32
- TCGv_ptr udata = tcg_temp_ebb_new_ptr();
33
-
34
- tcg_gen_movi_ptr(udata, 0);
35
- tcg_gen_ld_i32(cpu_index, tcg_env,
36
- -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
37
- gen_helper(cpu_index, udata);
38
-
39
- tcg_temp_free_ptr(udata);
40
- tcg_temp_free_i32(cpu_index);
41
-}
42
-
43
-static void gen_empty_udata_cb_no_wg(void)
44
-{
45
- gen_empty_udata_cb(gen_helper_plugin_vcpu_udata_cb_no_wg);
46
-}
47
-
48
-static void gen_empty_udata_cb_no_rwg(void)
49
-{
50
- gen_empty_udata_cb(gen_helper_plugin_vcpu_udata_cb_no_rwg);
51
-}
52
-
53
/*
54
* For now we only support addi_i64.
55
* When we support more ops, we can generate one empty inline cb for each.
56
@@ -XXX,XX +XXX,XX @@ static void gen_empty_mem_cb(TCGv_i64 addr, uint32_t info)
57
tcg_temp_free_i32(cpu_index);
58
}
59
60
-/*
61
- * Share the same function for enable/disable. When enabling, the NULL
62
- * pointer will be overwritten later.
63
- */
64
-static void gen_empty_mem_helper(void)
65
-{
66
- TCGv_ptr ptr = tcg_temp_ebb_new_ptr();
67
-
68
- tcg_gen_movi_ptr(ptr, 0);
69
- tcg_gen_st_ptr(ptr, tcg_env, offsetof(CPUState, plugin_mem_cbs) -
70
- offsetof(ArchCPU, env));
71
- tcg_temp_free_ptr(ptr);
72
-}
73
-
74
static void gen_plugin_cb_start(enum plugin_gen_from from,
75
enum plugin_gen_cb type, unsigned wr)
76
{
77
tcg_gen_plugin_cb_start(from, type, wr);
78
}
79
80
-static void gen_wrapped(enum plugin_gen_from from,
81
- enum plugin_gen_cb type, void (*func)(void))
82
-{
83
- gen_plugin_cb_start(from, type, 0);
84
- func();
85
- tcg_gen_plugin_cb_end();
86
-}
87
-
88
static void plugin_gen_empty_callback(enum plugin_gen_from from)
89
{
90
switch (from) {
91
case PLUGIN_GEN_AFTER_INSN:
92
case PLUGIN_GEN_FROM_TB:
93
- tcg_gen_plugin_cb(from);
94
- break;
95
case PLUGIN_GEN_FROM_INSN:
96
- /*
97
- * Note: plugin_gen_inject() relies on ENABLE_MEM_HELPER being
98
- * the first callback of an instruction
99
- */
100
- gen_wrapped(from, PLUGIN_GEN_ENABLE_MEM_HELPER,
101
- gen_empty_mem_helper);
102
- gen_wrapped(from, PLUGIN_GEN_CB_UDATA, gen_empty_udata_cb_no_rwg);
103
- gen_wrapped(from, PLUGIN_GEN_CB_UDATA_R, gen_empty_udata_cb_no_wg);
104
- gen_wrapped(from, PLUGIN_GEN_CB_INLINE, gen_empty_inline_cb);
105
+ tcg_gen_plugin_cb(from);
106
break;
107
default:
108
g_assert_not_reached();
109
@@ -XXX,XX +XXX,XX @@ static TCGOp *copy_mul_i32(TCGOp **begin_op, TCGOp *op, uint32_t v)
110
return op;
111
}
112
113
-static TCGOp *copy_st_ptr(TCGOp **begin_op, TCGOp *op)
114
-{
115
- if (UINTPTR_MAX == UINT32_MAX) {
116
- /* st_i32 */
117
- op = copy_op(begin_op, op, INDEX_op_st_i32);
118
- } else {
119
- /* st_i64 */
120
- op = copy_st_i64(begin_op, op);
121
- }
122
- return op;
123
-}
124
-
125
static TCGOp *copy_call(TCGOp **begin_op, TCGOp *op, void *func, int *cb_idx)
126
{
127
TCGOp *old_op;
128
@@ -XXX,XX +XXX,XX @@ static TCGOp *copy_call(TCGOp **begin_op, TCGOp *op, void *func, int *cb_idx)
129
return op;
130
}
131
132
-/*
133
- * When we append/replace ops here we are sensitive to changing patterns of
134
- * TCGOps generated by the tcg_gen_FOO calls when we generated the
135
- * empty callbacks. This will assert very quickly in a debug build as
136
- * we assert the ops we are replacing are the correct ones.
137
- */
138
-static TCGOp *append_udata_cb(const struct qemu_plugin_dyn_cb *cb,
139
- TCGOp *begin_op, TCGOp *op, int *cb_idx)
140
-{
141
- /* const_ptr */
142
- op = copy_const_ptr(&begin_op, op, cb->userp);
143
-
144
- /* copy the ld_i32, but note that we only have to copy it once */
145
- if (*cb_idx == -1) {
146
- op = copy_op(&begin_op, op, INDEX_op_ld_i32);
147
- } else {
148
- begin_op = QTAILQ_NEXT(begin_op, link);
149
- tcg_debug_assert(begin_op && begin_op->opc == INDEX_op_ld_i32);
150
- }
151
-
152
- /* call */
153
- op = copy_call(&begin_op, op, cb->regular.f.vcpu_udata, cb_idx);
154
-
155
- return op;
156
-}
157
-
158
static TCGOp *append_inline_cb(const struct qemu_plugin_dyn_cb *cb,
159
TCGOp *begin_op, TCGOp *op,
160
int *unused)
161
@@ -XXX,XX +XXX,XX @@ typedef TCGOp *(*inject_fn)(const struct qemu_plugin_dyn_cb *cb,
162
TCGOp *begin_op, TCGOp *op, int *intp);
163
typedef bool (*op_ok_fn)(const TCGOp *op, const struct qemu_plugin_dyn_cb *cb);
164
165
-static bool op_ok(const TCGOp *op, const struct qemu_plugin_dyn_cb *cb)
166
-{
167
- return true;
168
-}
169
-
170
static bool op_rw(const TCGOp *op, const struct qemu_plugin_dyn_cb *cb)
171
{
172
int w;
173
@@ -XXX,XX +XXX,XX @@ static void inject_cb_type(const GArray *cbs, TCGOp *begin_op,
174
rm_ops_range(begin_op, end_op);
175
}
176
177
-static void
178
-inject_udata_cb(const GArray *cbs, TCGOp *begin_op)
179
-{
180
- inject_cb_type(cbs, begin_op, append_udata_cb, op_ok);
181
-}
182
-
183
static void
184
inject_inline_cb(const GArray *cbs, TCGOp *begin_op, op_ok_fn ok)
185
{
186
@@ -XXX,XX +XXX,XX @@ inject_mem_cb(const GArray *cbs, TCGOp *begin_op)
187
inject_cb_type(cbs, begin_op, append_mem_cb, op_rw);
188
}
189
190
-/* we could change the ops in place, but we can reuse more code by copying */
191
-static void inject_mem_helper(TCGOp *begin_op, GArray *arr)
192
-{
193
- TCGOp *orig_op = begin_op;
194
- TCGOp *end_op;
195
- TCGOp *op;
196
-
197
- end_op = find_op(begin_op, INDEX_op_plugin_cb_end);
198
- tcg_debug_assert(end_op);
199
-
200
- /* const ptr */
201
- op = copy_const_ptr(&begin_op, end_op, arr);
202
-
203
- /* st_ptr */
204
- op = copy_st_ptr(&begin_op, op);
205
-
206
- rm_ops_range(orig_op, end_op);
207
-}
208
-
209
-/*
210
- * Tracking memory accesses performed from helpers requires extra work.
211
- * If an instruction is emulated with helpers, we do two things:
212
- * (1) copy the CB descriptors, and keep track of it so that they can be
213
- * freed later on, and (2) point CPUState.plugin_mem_cbs to the descriptors, so
214
- * that we can read them at run-time (i.e. when the helper executes).
215
- * This run-time access is performed from qemu_plugin_vcpu_mem_cb.
216
- *
217
- * Note that plugin_gen_disable_mem_helpers undoes (2). Since it
218
- * is possible that the code we generate after the instruction is
219
- * dead, we also add checks before generating tb_exit etc.
220
- */
221
-static void inject_mem_enable_helper(struct qemu_plugin_tb *ptb,
222
- struct qemu_plugin_insn *plugin_insn,
223
- TCGOp *begin_op)
224
-{
225
- GArray *cbs[2];
226
- GArray *arr;
227
- size_t n_cbs, i;
228
-
229
- cbs[0] = plugin_insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR];
230
- cbs[1] = plugin_insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
231
-
232
- n_cbs = 0;
233
- for (i = 0; i < ARRAY_SIZE(cbs); i++) {
234
- n_cbs += cbs[i]->len;
235
- }
236
-
237
- plugin_insn->mem_helper = plugin_insn->calls_helpers && n_cbs;
238
- if (likely(!plugin_insn->mem_helper)) {
239
- rm_ops(begin_op);
240
- return;
241
- }
242
- ptb->mem_helper = true;
243
-
244
- arr = g_array_sized_new(false, false,
245
- sizeof(struct qemu_plugin_dyn_cb), n_cbs);
246
-
247
- for (i = 0; i < ARRAY_SIZE(cbs); i++) {
248
- g_array_append_vals(arr, cbs[i]->data, cbs[i]->len);
249
- }
250
-
251
- qemu_plugin_add_dyn_cb_arr(arr);
252
- inject_mem_helper(begin_op, arr);
253
-}
254
-
255
/* called before finishing a TB with exit_tb, goto_tb or goto_ptr */
256
void plugin_gen_disable_mem_helpers(void)
257
{
258
@@ -XXX,XX +XXX,XX @@ void plugin_gen_disable_mem_helpers(void)
27
}
259
}
28
}
260
}
29
#define MAX_RESERVED_VA arm_max_reserved_va
261
30
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
262
-static void plugin_gen_insn_udata(const struct qemu_plugin_tb *ptb,
263
- TCGOp *begin_op, int insn_idx)
264
-{
265
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
266
-
267
- inject_udata_cb(insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR], begin_op);
268
-}
269
-
270
-static void plugin_gen_insn_udata_r(const struct qemu_plugin_tb *ptb,
271
- TCGOp *begin_op, int insn_idx)
272
-{
273
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
274
-
275
- inject_udata_cb(insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR_R], begin_op);
276
-}
277
-
278
-static void plugin_gen_insn_inline(const struct qemu_plugin_tb *ptb,
279
- TCGOp *begin_op, int insn_idx)
280
-{
281
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
282
- inject_inline_cb(insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_INLINE],
283
- begin_op, op_ok);
284
-}
285
-
286
static void plugin_gen_mem_regular(const struct qemu_plugin_tb *ptb,
287
TCGOp *begin_op, int insn_idx)
288
{
289
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_mem_inline(const struct qemu_plugin_tb *ptb,
290
inject_inline_cb(cbs, begin_op, op_rw);
291
}
292
293
-static void plugin_gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
294
- TCGOp *begin_op, int insn_idx)
295
+static void gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
296
+ struct qemu_plugin_insn *insn)
297
{
298
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
299
- inject_mem_enable_helper(ptb, insn, begin_op);
300
+ GArray *cbs[2];
301
+ GArray *arr;
302
+ size_t n_cbs;
303
+
304
+ /*
305
+ * Tracking memory accesses performed from helpers requires extra work.
306
+ * If an instruction is emulated with helpers, we do two things:
307
+ * (1) copy the CB descriptors, and keep track of it so that they can be
308
+ * freed later on, and (2) point CPUState.plugin_mem_cbs to the
309
+ * descriptors, so that we can read them at run-time
310
+ * (i.e. when the helper executes).
311
+ * This run-time access is performed from qemu_plugin_vcpu_mem_cb.
312
+ *
313
+ * Note that plugin_gen_disable_mem_helpers undoes (2). Since it
314
+ * is possible that the code we generate after the instruction is
315
+ * dead, we also add checks before generating tb_exit etc.
316
+ */
317
+ if (!insn->calls_helpers) {
318
+ return;
319
+ }
320
+
321
+ cbs[0] = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR];
322
+ cbs[1] = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
323
+ n_cbs = cbs[0]->len + cbs[1]->len;
324
+
325
+ if (n_cbs == 0) {
326
+ insn->mem_helper = false;
327
+ return;
328
+ }
329
+ insn->mem_helper = true;
330
+ ptb->mem_helper = true;
331
+
332
+ arr = g_array_sized_new(false, false,
333
+ sizeof(struct qemu_plugin_dyn_cb), n_cbs);
334
+ g_array_append_vals(arr, cbs[0]->data, cbs[0]->len);
335
+ g_array_append_vals(arr, cbs[1]->data, cbs[1]->len);
336
+
337
+ qemu_plugin_add_dyn_cb_arr(arr);
338
+
339
+ tcg_gen_st_ptr(tcg_constant_ptr((intptr_t)arr), tcg_env,
340
+ offsetof(CPUState, plugin_mem_cbs) -
341
+ offsetof(ArchCPU, env));
342
}
343
344
static void gen_disable_mem_helper(void)
345
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
346
}
347
break;
348
349
+ case PLUGIN_GEN_FROM_INSN:
350
+ assert(insn != NULL);
351
+
352
+ gen_enable_mem_helper(plugin_tb, insn);
353
+
354
+ cbs = insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR];
355
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
356
+ struct qemu_plugin_dyn_cb *cb =
357
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
358
+ gen_udata_cb(cb);
359
+ }
360
+
361
+ cbs = insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_INLINE];
362
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
363
+ struct qemu_plugin_dyn_cb *cb =
364
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
365
+ gen_inline_cb(cb);
366
+ }
367
+ break;
368
+
369
default:
370
g_assert_not_reached();
371
}
372
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
373
enum plugin_gen_cb type = op->args[1];
374
375
switch (from) {
376
- case PLUGIN_GEN_FROM_INSN:
377
- {
378
- g_assert(insn_idx >= 0);
379
-
380
- switch (type) {
381
- case PLUGIN_GEN_CB_UDATA:
382
- plugin_gen_insn_udata(plugin_tb, op, insn_idx);
383
- break;
384
- case PLUGIN_GEN_CB_UDATA_R:
385
- plugin_gen_insn_udata_r(plugin_tb, op, insn_idx);
386
- break;
387
- case PLUGIN_GEN_CB_INLINE:
388
- plugin_gen_insn_inline(plugin_tb, op, insn_idx);
389
- break;
390
- case PLUGIN_GEN_ENABLE_MEM_HELPER:
391
- plugin_gen_enable_mem_helper(plugin_tb, op, insn_idx);
392
- break;
393
- default:
394
- g_assert_not_reached();
395
- }
396
- break;
397
- }
398
case PLUGIN_GEN_FROM_MEM:
399
{
400
g_assert(insn_idx >= 0);
401
diff --git a/plugins/api.c b/plugins/api.c
31
index XXXXXXX..XXXXXXX 100644
402
index XXXXXXX..XXXXXXX 100644
32
--- a/linux-user/elfload.c
403
--- a/plugins/api.c
33
+++ b/linux-user/elfload.c
404
+++ b/plugins/api.c
34
@@ -XXX,XX +XXX,XX @@ enum {
405
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_insn_exec_cb(struct qemu_plugin_insn *insn,
35
406
void *udata)
36
static bool init_guest_commpage(void)
407
{
37
{
408
if (!insn->mem_only) {
38
- void *want = g2h_untagged(HI_COMMPAGE & -qemu_host_page_size);
409
- int index = flags == QEMU_PLUGIN_CB_R_REGS ||
39
+ abi_ptr commpage = HI_COMMPAGE & -qemu_host_page_size;
410
- flags == QEMU_PLUGIN_CB_RW_REGS ?
40
+ void *want = g2h_untagged(commpage);
411
- PLUGIN_CB_REGULAR_R : PLUGIN_CB_REGULAR;
41
void *addr = mmap(want, qemu_host_page_size, PROT_READ | PROT_WRITE,
412
-
42
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
413
- plugin_register_dyn_cb__udata(&insn->cbs[PLUGIN_CB_INSN][index],
43
414
- cb, flags, udata);
44
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
415
+ plugin_register_dyn_cb__udata(
45
perror("Protecting guest commpage");
416
+ &insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR], cb, flags, udata);
46
exit(EXIT_FAILURE);
47
}
417
}
48
+
418
}
49
+ page_set_flags(commpage, commpage + qemu_host_page_size,
50
+ PAGE_READ | PAGE_EXEC | PAGE_VALID);
51
return true;
52
}
53
419
54
--
420
--
55
2.34.1
421
2.34.1
diff view generated by jsdifflib
1
We're about to start validating PAGE_EXEC, which means that we've
1
Introduce a new plugin_mem_cb op to hold the address temp
2
got to mark page zero executable. We had been special casing this
2
and meminfo computed by tcg-op-ldst.c. Because this now
3
entirely within translate.
3
has its own opcode, we no longer need PLUGIN_GEN_FROM_MEM.
4
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
7
---
9
linux-user/elfload.c | 34 +++++++++++++++++++++++++++++++---
8
include/exec/plugin-gen.h | 4 -
10
1 file changed, 31 insertions(+), 3 deletions(-)
9
include/tcg/tcg-op-common.h | 1 +
10
include/tcg/tcg-opc.h | 1 +
11
accel/tcg/plugin-gen.c | 408 ++++--------------------------------
12
tcg/tcg-op-ldst.c | 6 +-
13
tcg/tcg-op.c | 5 +
14
6 files changed, 54 insertions(+), 371 deletions(-)
11
15
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
16
diff --git a/include/exec/plugin-gen.h b/include/exec/plugin-gen.h
13
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
14
--- a/linux-user/elfload.c
18
--- a/include/exec/plugin-gen.h
15
+++ b/linux-user/elfload.c
19
+++ b/include/exec/plugin-gen.h
16
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
20
@@ -XXX,XX +XXX,XX @@ void plugin_gen_insn_start(CPUState *cpu, const struct DisasContextBase *db);
17
regs->gr[31] = infop->entry;
21
void plugin_gen_insn_end(void);
22
23
void plugin_gen_disable_mem_helpers(void);
24
-void plugin_gen_empty_mem_callback(TCGv_i64 addr, uint32_t info);
25
26
#else /* !CONFIG_PLUGIN */
27
28
@@ -XXX,XX +XXX,XX @@ static inline void plugin_gen_tb_end(CPUState *cpu, size_t num_insns)
29
static inline void plugin_gen_disable_mem_helpers(void)
30
{ }
31
32
-static inline void plugin_gen_empty_mem_callback(TCGv_i64 addr, uint32_t info)
33
-{ }
34
-
35
#endif /* CONFIG_PLUGIN */
36
37
#endif /* QEMU_PLUGIN_GEN_H */
38
diff --git a/include/tcg/tcg-op-common.h b/include/tcg/tcg-op-common.h
39
index XXXXXXX..XXXXXXX 100644
40
--- a/include/tcg/tcg-op-common.h
41
+++ b/include/tcg/tcg-op-common.h
42
@@ -XXX,XX +XXX,XX @@ void tcg_gen_goto_tb(unsigned idx);
43
void tcg_gen_lookup_and_goto_ptr(void);
44
45
void tcg_gen_plugin_cb(unsigned from);
46
+void tcg_gen_plugin_mem_cb(TCGv_i64 addr, unsigned meminfo);
47
void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr);
48
void tcg_gen_plugin_cb_end(void);
49
50
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/include/tcg/tcg-opc.h
53
+++ b/include/tcg/tcg-opc.h
54
@@ -XXX,XX +XXX,XX @@ DEF(goto_tb, 0, 0, 1, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
55
DEF(goto_ptr, 0, 1, 0, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
56
57
DEF(plugin_cb, 0, 0, 1, TCG_OPF_NOT_PRESENT)
58
+DEF(plugin_mem_cb, 0, 1, 1, TCG_OPF_NOT_PRESENT)
59
DEF(plugin_cb_start, 0, 0, 3, TCG_OPF_NOT_PRESENT)
60
DEF(plugin_cb_end, 0, 0, 0, TCG_OPF_NOT_PRESENT)
61
62
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
63
index XXXXXXX..XXXXXXX 100644
64
--- a/accel/tcg/plugin-gen.c
65
+++ b/accel/tcg/plugin-gen.c
66
@@ -XXX,XX +XXX,XX @@
67
enum plugin_gen_from {
68
PLUGIN_GEN_FROM_TB,
69
PLUGIN_GEN_FROM_INSN,
70
- PLUGIN_GEN_FROM_MEM,
71
PLUGIN_GEN_AFTER_INSN,
72
PLUGIN_GEN_AFTER_TB,
73
PLUGIN_GEN_N_FROMS,
74
@@ -XXX,XX +XXX,XX @@ void HELPER(plugin_vcpu_mem_cb)(unsigned int vcpu_index,
75
void *userdata)
76
{ }
77
78
-/*
79
- * For now we only support addi_i64.
80
- * When we support more ops, we can generate one empty inline cb for each.
81
- */
82
-static void gen_empty_inline_cb(void)
83
-{
84
- TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
85
- TCGv_ptr cpu_index_as_ptr = tcg_temp_ebb_new_ptr();
86
- TCGv_i64 val = tcg_temp_ebb_new_i64();
87
- TCGv_ptr ptr = tcg_temp_ebb_new_ptr();
88
-
89
- tcg_gen_ld_i32(cpu_index, tcg_env,
90
- -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
91
- /* second operand will be replaced by immediate value */
92
- tcg_gen_mul_i32(cpu_index, cpu_index, cpu_index);
93
- tcg_gen_ext_i32_ptr(cpu_index_as_ptr, cpu_index);
94
-
95
- tcg_gen_movi_ptr(ptr, 0);
96
- tcg_gen_add_ptr(ptr, ptr, cpu_index_as_ptr);
97
- tcg_gen_ld_i64(val, ptr, 0);
98
- /* second operand will be replaced by immediate value */
99
- tcg_gen_add_i64(val, val, val);
100
-
101
- tcg_gen_st_i64(val, ptr, 0);
102
- tcg_temp_free_ptr(ptr);
103
- tcg_temp_free_i64(val);
104
- tcg_temp_free_ptr(cpu_index_as_ptr);
105
- tcg_temp_free_i32(cpu_index);
106
-}
107
-
108
-static void gen_empty_mem_cb(TCGv_i64 addr, uint32_t info)
109
-{
110
- TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
111
- TCGv_i32 meminfo = tcg_temp_ebb_new_i32();
112
- TCGv_ptr udata = tcg_temp_ebb_new_ptr();
113
-
114
- tcg_gen_movi_i32(meminfo, info);
115
- tcg_gen_movi_ptr(udata, 0);
116
- tcg_gen_ld_i32(cpu_index, tcg_env,
117
- -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
118
-
119
- gen_helper_plugin_vcpu_mem_cb(cpu_index, meminfo, addr, udata);
120
-
121
- tcg_temp_free_ptr(udata);
122
- tcg_temp_free_i32(meminfo);
123
- tcg_temp_free_i32(cpu_index);
124
-}
125
-
126
-static void gen_plugin_cb_start(enum plugin_gen_from from,
127
- enum plugin_gen_cb type, unsigned wr)
128
-{
129
- tcg_gen_plugin_cb_start(from, type, wr);
130
-}
131
-
132
static void plugin_gen_empty_callback(enum plugin_gen_from from)
133
{
134
switch (from) {
135
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_empty_callback(enum plugin_gen_from from)
136
}
18
}
137
}
19
138
20
+#define LO_COMMPAGE 0
139
-void plugin_gen_empty_mem_callback(TCGv_i64 addr, uint32_t info)
140
-{
141
- enum qemu_plugin_mem_rw rw = get_plugin_meminfo_rw(info);
142
-
143
- gen_plugin_cb_start(PLUGIN_GEN_FROM_MEM, PLUGIN_GEN_CB_MEM, rw);
144
- gen_empty_mem_cb(addr, info);
145
- tcg_gen_plugin_cb_end();
146
-
147
- gen_plugin_cb_start(PLUGIN_GEN_FROM_MEM, PLUGIN_GEN_CB_INLINE, rw);
148
- gen_empty_inline_cb();
149
- tcg_gen_plugin_cb_end();
150
-}
151
-
152
-static TCGOp *find_op(TCGOp *op, TCGOpcode opc)
153
-{
154
- while (op) {
155
- if (op->opc == opc) {
156
- return op;
157
- }
158
- op = QTAILQ_NEXT(op, link);
159
- }
160
- return NULL;
161
-}
162
-
163
-static TCGOp *rm_ops_range(TCGOp *begin, TCGOp *end)
164
-{
165
- TCGOp *ret = QTAILQ_NEXT(end, link);
166
-
167
- QTAILQ_REMOVE_SEVERAL(&tcg_ctx->ops, begin, end, link);
168
- return ret;
169
-}
170
-
171
-/* remove all ops until (and including) plugin_cb_end */
172
-static TCGOp *rm_ops(TCGOp *op)
173
-{
174
- TCGOp *end_op = find_op(op, INDEX_op_plugin_cb_end);
175
-
176
- tcg_debug_assert(end_op);
177
- return rm_ops_range(op, end_op);
178
-}
179
-
180
-static TCGOp *copy_op_nocheck(TCGOp **begin_op, TCGOp *op)
181
-{
182
- TCGOp *old_op = QTAILQ_NEXT(*begin_op, link);
183
- unsigned nargs = old_op->nargs;
184
-
185
- *begin_op = old_op;
186
- op = tcg_op_insert_after(tcg_ctx, op, old_op->opc, nargs);
187
- memcpy(op->args, old_op->args, sizeof(op->args[0]) * nargs);
188
-
189
- return op;
190
-}
191
-
192
-static TCGOp *copy_op(TCGOp **begin_op, TCGOp *op, TCGOpcode opc)
193
-{
194
- op = copy_op_nocheck(begin_op, op);
195
- tcg_debug_assert((*begin_op)->opc == opc);
196
- return op;
197
-}
198
-
199
-static TCGOp *copy_const_ptr(TCGOp **begin_op, TCGOp *op, void *ptr)
200
-{
201
- if (UINTPTR_MAX == UINT32_MAX) {
202
- /* mov_i32 */
203
- op = copy_op(begin_op, op, INDEX_op_mov_i32);
204
- op->args[1] = tcgv_i32_arg(tcg_constant_i32((uintptr_t)ptr));
205
- } else {
206
- /* mov_i64 */
207
- op = copy_op(begin_op, op, INDEX_op_mov_i64);
208
- op->args[1] = tcgv_i64_arg(tcg_constant_i64((uintptr_t)ptr));
209
- }
210
- return op;
211
-}
212
-
213
-static TCGOp *copy_ld_i32(TCGOp **begin_op, TCGOp *op)
214
-{
215
- return copy_op(begin_op, op, INDEX_op_ld_i32);
216
-}
217
-
218
-static TCGOp *copy_ext_i32_ptr(TCGOp **begin_op, TCGOp *op)
219
-{
220
- if (UINTPTR_MAX == UINT32_MAX) {
221
- op = copy_op(begin_op, op, INDEX_op_mov_i32);
222
- } else {
223
- op = copy_op(begin_op, op, INDEX_op_ext_i32_i64);
224
- }
225
- return op;
226
-}
227
-
228
-static TCGOp *copy_add_ptr(TCGOp **begin_op, TCGOp *op)
229
-{
230
- if (UINTPTR_MAX == UINT32_MAX) {
231
- op = copy_op(begin_op, op, INDEX_op_add_i32);
232
- } else {
233
- op = copy_op(begin_op, op, INDEX_op_add_i64);
234
- }
235
- return op;
236
-}
237
-
238
-static TCGOp *copy_ld_i64(TCGOp **begin_op, TCGOp *op)
239
-{
240
- if (TCG_TARGET_REG_BITS == 32) {
241
- /* 2x ld_i32 */
242
- op = copy_ld_i32(begin_op, op);
243
- op = copy_ld_i32(begin_op, op);
244
- } else {
245
- /* ld_i64 */
246
- op = copy_op(begin_op, op, INDEX_op_ld_i64);
247
- }
248
- return op;
249
-}
250
-
251
-static TCGOp *copy_st_i64(TCGOp **begin_op, TCGOp *op)
252
-{
253
- if (TCG_TARGET_REG_BITS == 32) {
254
- /* 2x st_i32 */
255
- op = copy_op(begin_op, op, INDEX_op_st_i32);
256
- op = copy_op(begin_op, op, INDEX_op_st_i32);
257
- } else {
258
- /* st_i64 */
259
- op = copy_op(begin_op, op, INDEX_op_st_i64);
260
- }
261
- return op;
262
-}
263
-
264
-static TCGOp *copy_add_i64(TCGOp **begin_op, TCGOp *op, uint64_t v)
265
-{
266
- if (TCG_TARGET_REG_BITS == 32) {
267
- /* all 32-bit backends must implement add2_i32 */
268
- g_assert(TCG_TARGET_HAS_add2_i32);
269
- op = copy_op(begin_op, op, INDEX_op_add2_i32);
270
- op->args[4] = tcgv_i32_arg(tcg_constant_i32(v));
271
- op->args[5] = tcgv_i32_arg(tcg_constant_i32(v >> 32));
272
- } else {
273
- op = copy_op(begin_op, op, INDEX_op_add_i64);
274
- op->args[2] = tcgv_i64_arg(tcg_constant_i64(v));
275
- }
276
- return op;
277
-}
278
-
279
-static TCGOp *copy_mul_i32(TCGOp **begin_op, TCGOp *op, uint32_t v)
280
-{
281
- op = copy_op(begin_op, op, INDEX_op_mul_i32);
282
- op->args[2] = tcgv_i32_arg(tcg_constant_i32(v));
283
- return op;
284
-}
285
-
286
-static TCGOp *copy_call(TCGOp **begin_op, TCGOp *op, void *func, int *cb_idx)
287
-{
288
- TCGOp *old_op;
289
- int func_idx;
290
-
291
- /* copy all ops until the call */
292
- do {
293
- op = copy_op_nocheck(begin_op, op);
294
- } while (op->opc != INDEX_op_call);
295
-
296
- /* fill in the op call */
297
- old_op = *begin_op;
298
- TCGOP_CALLI(op) = TCGOP_CALLI(old_op);
299
- TCGOP_CALLO(op) = TCGOP_CALLO(old_op);
300
- tcg_debug_assert(op->life == 0);
301
-
302
- func_idx = TCGOP_CALLO(op) + TCGOP_CALLI(op);
303
- *cb_idx = func_idx;
304
- op->args[func_idx] = (uintptr_t)func;
305
-
306
- return op;
307
-}
308
-
309
-static TCGOp *append_inline_cb(const struct qemu_plugin_dyn_cb *cb,
310
- TCGOp *begin_op, TCGOp *op,
311
- int *unused)
312
-{
313
- char *ptr = cb->inline_insn.entry.score->data->data;
314
- size_t elem_size = g_array_get_element_size(
315
- cb->inline_insn.entry.score->data);
316
- size_t offset = cb->inline_insn.entry.offset;
317
-
318
- op = copy_ld_i32(&begin_op, op);
319
- op = copy_mul_i32(&begin_op, op, elem_size);
320
- op = copy_ext_i32_ptr(&begin_op, op);
321
- op = copy_const_ptr(&begin_op, op, ptr + offset);
322
- op = copy_add_ptr(&begin_op, op);
323
- op = copy_ld_i64(&begin_op, op);
324
- op = copy_add_i64(&begin_op, op, cb->inline_insn.imm);
325
- op = copy_st_i64(&begin_op, op);
326
- return op;
327
-}
328
-
329
-static TCGOp *append_mem_cb(const struct qemu_plugin_dyn_cb *cb,
330
- TCGOp *begin_op, TCGOp *op, int *cb_idx)
331
-{
332
- enum plugin_gen_cb type = begin_op->args[1];
333
-
334
- tcg_debug_assert(type == PLUGIN_GEN_CB_MEM);
335
-
336
- /* const_i32 == mov_i32 ("info", so it remains as is) */
337
- op = copy_op(&begin_op, op, INDEX_op_mov_i32);
338
-
339
- /* const_ptr */
340
- op = copy_const_ptr(&begin_op, op, cb->userp);
341
-
342
- /* copy the ld_i32, but note that we only have to copy it once */
343
- if (*cb_idx == -1) {
344
- op = copy_op(&begin_op, op, INDEX_op_ld_i32);
345
- } else {
346
- begin_op = QTAILQ_NEXT(begin_op, link);
347
- tcg_debug_assert(begin_op && begin_op->opc == INDEX_op_ld_i32);
348
- }
349
-
350
- if (type == PLUGIN_GEN_CB_MEM) {
351
- /* call */
352
- op = copy_call(&begin_op, op, cb->regular.f.vcpu_udata, cb_idx);
353
- }
354
-
355
- return op;
356
-}
357
-
358
-typedef TCGOp *(*inject_fn)(const struct qemu_plugin_dyn_cb *cb,
359
- TCGOp *begin_op, TCGOp *op, int *intp);
360
-typedef bool (*op_ok_fn)(const TCGOp *op, const struct qemu_plugin_dyn_cb *cb);
361
-
362
-static bool op_rw(const TCGOp *op, const struct qemu_plugin_dyn_cb *cb)
363
-{
364
- int w;
365
-
366
- w = op->args[2];
367
- return !!(cb->rw & (w + 1));
368
-}
369
-
370
-static void inject_cb_type(const GArray *cbs, TCGOp *begin_op,
371
- inject_fn inject, op_ok_fn ok)
372
-{
373
- TCGOp *end_op;
374
- TCGOp *op;
375
- int cb_idx = -1;
376
- int i;
377
-
378
- if (!cbs || cbs->len == 0) {
379
- rm_ops(begin_op);
380
- return;
381
- }
382
-
383
- end_op = find_op(begin_op, INDEX_op_plugin_cb_end);
384
- tcg_debug_assert(end_op);
385
-
386
- op = end_op;
387
- for (i = 0; i < cbs->len; i++) {
388
- struct qemu_plugin_dyn_cb *cb =
389
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
390
-
391
- if (!ok(begin_op, cb)) {
392
- continue;
393
- }
394
- op = inject(cb, begin_op, op, &cb_idx);
395
- }
396
- rm_ops_range(begin_op, end_op);
397
-}
398
-
399
-static void
400
-inject_inline_cb(const GArray *cbs, TCGOp *begin_op, op_ok_fn ok)
401
-{
402
- inject_cb_type(cbs, begin_op, append_inline_cb, ok);
403
-}
404
-
405
-static void
406
-inject_mem_cb(const GArray *cbs, TCGOp *begin_op)
407
-{
408
- inject_cb_type(cbs, begin_op, append_mem_cb, op_rw);
409
-}
410
-
411
/* called before finishing a TB with exit_tb, goto_tb or goto_ptr */
412
void plugin_gen_disable_mem_helpers(void)
413
{
414
@@ -XXX,XX +XXX,XX @@ void plugin_gen_disable_mem_helpers(void)
415
}
416
}
417
418
-static void plugin_gen_mem_regular(const struct qemu_plugin_tb *ptb,
419
- TCGOp *begin_op, int insn_idx)
420
-{
421
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
422
- inject_mem_cb(insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR], begin_op);
423
-}
424
-
425
-static void plugin_gen_mem_inline(const struct qemu_plugin_tb *ptb,
426
- TCGOp *begin_op, int insn_idx)
427
-{
428
- const GArray *cbs;
429
- struct qemu_plugin_insn *insn = g_ptr_array_index(ptb->insns, insn_idx);
430
-
431
- cbs = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
432
- inject_inline_cb(cbs, begin_op, op_rw);
433
-}
434
-
435
static void gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
436
struct qemu_plugin_insn *insn)
437
{
438
@@ -XXX,XX +XXX,XX @@ static void gen_inline_cb(struct qemu_plugin_dyn_cb *cb)
439
tcg_temp_free_ptr(ptr);
440
}
441
442
+static void gen_mem_cb(struct qemu_plugin_dyn_cb *cb,
443
+ qemu_plugin_meminfo_t meminfo, TCGv_i64 addr)
444
+{
445
+ TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
21
+
446
+
22
+static bool init_guest_commpage(void)
447
+ tcg_gen_ld_i32(cpu_index, tcg_env,
23
+{
448
+ -offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
24
+ void *want = g2h_untagged(LO_COMMPAGE);
449
+ tcg_gen_call4(cb->regular.f.vcpu_mem, cb->regular.info, NULL,
25
+ void *addr = mmap(want, qemu_host_page_size, PROT_NONE,
450
+ tcgv_i32_temp(cpu_index),
26
+ MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
451
+ tcgv_i32_temp(tcg_constant_i32(meminfo)),
27
+
452
+ tcgv_i64_temp(addr),
28
+ if (addr == MAP_FAILED) {
453
+ tcgv_ptr_temp(tcg_constant_ptr(cb->userp)));
29
+ perror("Allocating guest commpage");
454
+ tcg_temp_free_i32(cpu_index);
30
+ exit(EXIT_FAILURE);
31
+ }
32
+ if (addr != want) {
33
+ return false;
34
+ }
35
+
36
+ /*
37
+ * On Linux, page zero is normally marked execute only + gateway.
38
+ * Normal read or write is supposed to fail (thus PROT_NONE above),
39
+ * but specific offsets have kernel code mapped to raise permissions
40
+ * and implement syscalls. Here, simply mark the page executable.
41
+ * Special case the entry points during translation (see do_page_zero).
42
+ */
43
+ page_set_flags(LO_COMMPAGE, LO_COMMPAGE + TARGET_PAGE_SIZE,
44
+ PAGE_EXEC | PAGE_VALID);
45
+ return true;
46
+}
455
+}
47
+
456
+
48
#endif /* TARGET_HPPA */
457
/* #define DEBUG_PLUGIN_GEN_OPS */
49
458
static void pr_ops(void)
50
#ifdef TARGET_XTENSA
459
{
51
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
460
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
461
break;
462
}
463
464
- case INDEX_op_plugin_cb_start:
465
+ case INDEX_op_plugin_mem_cb:
466
{
467
- enum plugin_gen_from from = op->args[0];
468
- enum plugin_gen_cb type = op->args[1];
469
+ TCGv_i64 addr = temp_tcgv_i64(arg_temp(op->args[0]));
470
+ qemu_plugin_meminfo_t meminfo = op->args[1];
471
+ struct qemu_plugin_insn *insn;
472
+ const GArray *cbs;
473
+ int i, n, rw;
474
475
- switch (from) {
476
- case PLUGIN_GEN_FROM_MEM:
477
- {
478
- g_assert(insn_idx >= 0);
479
+ assert(insn_idx >= 0);
480
+ insn = g_ptr_array_index(plugin_tb->insns, insn_idx);
481
+ rw = qemu_plugin_mem_is_store(meminfo) ? 2 : 1;
482
483
- switch (type) {
484
- case PLUGIN_GEN_CB_MEM:
485
- plugin_gen_mem_regular(plugin_tb, op, insn_idx);
486
- break;
487
- case PLUGIN_GEN_CB_INLINE:
488
- plugin_gen_mem_inline(plugin_tb, op, insn_idx);
489
- break;
490
- default:
491
- g_assert_not_reached();
492
+ tcg_ctx->emit_before_op = op;
493
+
494
+ cbs = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR];
495
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
496
+ struct qemu_plugin_dyn_cb *cb =
497
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
498
+ if (cb->rw & rw) {
499
+ gen_mem_cb(cb, meminfo, addr);
500
}
501
+ }
502
503
- break;
504
- }
505
- default:
506
- g_assert_not_reached();
507
+ cbs = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
508
+ for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
509
+ struct qemu_plugin_dyn_cb *cb =
510
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
511
+ if (cb->rw & rw) {
512
+ gen_inline_cb(cb);
513
+ }
514
}
515
+
516
+ tcg_ctx->emit_before_op = NULL;
517
+ tcg_op_remove(tcg_ctx, op);
518
break;
519
}
520
+
521
default:
522
/* plugins don't care about any other ops */
523
break;
524
diff --git a/tcg/tcg-op-ldst.c b/tcg/tcg-op-ldst.c
525
index XXXXXXX..XXXXXXX 100644
526
--- a/tcg/tcg-op-ldst.c
527
+++ b/tcg/tcg-op-ldst.c
528
@@ -XXX,XX +XXX,XX @@ plugin_gen_mem_callbacks(TCGv_i64 copy_addr, TCGTemp *orig_addr, MemOpIdx oi,
529
copy_addr = tcg_temp_ebb_new_i64();
530
tcg_gen_extu_i32_i64(copy_addr, temp_tcgv_i32(orig_addr));
531
}
532
- plugin_gen_empty_mem_callback(copy_addr, info);
533
+ tcg_gen_plugin_mem_cb(copy_addr, info);
534
tcg_temp_free_i64(copy_addr);
535
} else {
536
if (copy_addr) {
537
- plugin_gen_empty_mem_callback(copy_addr, info);
538
+ tcg_gen_plugin_mem_cb(copy_addr, info);
539
tcg_temp_free_i64(copy_addr);
540
} else {
541
- plugin_gen_empty_mem_callback(temp_tcgv_i64(orig_addr), info);
542
+ tcg_gen_plugin_mem_cb(temp_tcgv_i64(orig_addr), info);
543
}
544
}
545
}
546
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
547
index XXXXXXX..XXXXXXX 100644
548
--- a/tcg/tcg-op.c
549
+++ b/tcg/tcg-op.c
550
@@ -XXX,XX +XXX,XX @@ void tcg_gen_plugin_cb(unsigned from)
551
tcg_gen_op1(INDEX_op_plugin_cb, from);
52
}
552
}
53
553
54
#if defined(HI_COMMPAGE)
554
+void tcg_gen_plugin_mem_cb(TCGv_i64 addr, unsigned meminfo)
55
-#define LO_COMMPAGE 0
555
+{
56
+#define LO_COMMPAGE -1
556
+ tcg_gen_op2(INDEX_op_plugin_mem_cb, tcgv_i64_arg(addr), meminfo);
57
#elif defined(LO_COMMPAGE)
557
+}
58
#define HI_COMMPAGE 0
558
+
59
#else
559
void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr)
60
#define HI_COMMPAGE 0
560
{
61
-#define LO_COMMPAGE 0
561
tcg_gen_op3(INDEX_op_plugin_cb_start, from, type, wr);
62
+#define LO_COMMPAGE -1
63
#define init_guest_commpage() true
64
#endif
65
66
@@ -XXX,XX +XXX,XX @@ static void pgb_static(const char *image_name, abi_ulong orig_loaddr,
67
} else {
68
offset = -(HI_COMMPAGE & -align);
69
}
70
- } else if (LO_COMMPAGE != 0) {
71
+ } else if (LO_COMMPAGE != -1) {
72
loaddr = MIN(loaddr, LO_COMMPAGE & -align);
73
}
74
75
--
562
--
76
2.34.1
563
2.34.1
diff view generated by jsdifflib
1
There is no need to go through cc->tcg_ops when
1
These placeholder helpers are no longer required.
2
we know what value that must have.
3
2
4
Reviewed-by: Michael Rolnik <mrolnik@gmail.com>
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
5
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
5
---
8
target/avr/helper.c | 5 ++---
6
accel/tcg/plugin-helpers.h | 5 -----
9
1 file changed, 2 insertions(+), 3 deletions(-)
7
include/exec/helper-gen-common.h | 4 ----
8
include/exec/helper-proto-common.h | 4 ----
9
accel/tcg/plugin-gen.c | 20 --------------------
10
4 files changed, 33 deletions(-)
11
delete mode 100644 accel/tcg/plugin-helpers.h
10
12
11
diff --git a/target/avr/helper.c b/target/avr/helper.c
13
diff --git a/accel/tcg/plugin-helpers.h b/accel/tcg/plugin-helpers.h
14
deleted file mode 100644
15
index XXXXXXX..XXXXXXX
16
--- a/accel/tcg/plugin-helpers.h
17
+++ /dev/null
18
@@ -XXX,XX +XXX,XX @@
19
-#ifdef CONFIG_PLUGIN
20
-DEF_HELPER_FLAGS_2(plugin_vcpu_udata_cb_no_wg, TCG_CALL_NO_WG | TCG_CALL_PLUGIN, void, i32, ptr)
21
-DEF_HELPER_FLAGS_2(plugin_vcpu_udata_cb_no_rwg, TCG_CALL_NO_RWG | TCG_CALL_PLUGIN, void, i32, ptr)
22
-DEF_HELPER_FLAGS_4(plugin_vcpu_mem_cb, TCG_CALL_NO_RWG | TCG_CALL_PLUGIN, void, i32, i32, i64, ptr)
23
-#endif
24
diff --git a/include/exec/helper-gen-common.h b/include/exec/helper-gen-common.h
12
index XXXXXXX..XXXXXXX 100644
25
index XXXXXXX..XXXXXXX 100644
13
--- a/target/avr/helper.c
26
--- a/include/exec/helper-gen-common.h
14
+++ b/target/avr/helper.c
27
+++ b/include/exec/helper-gen-common.h
15
@@ -XXX,XX +XXX,XX @@
28
@@ -XXX,XX +XXX,XX @@
16
bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
29
#include "exec/helper-gen.h.inc"
30
#undef HELPER_H
31
32
-#define HELPER_H "accel/tcg/plugin-helpers.h"
33
-#include "exec/helper-gen.h.inc"
34
-#undef HELPER_H
35
-
36
#endif /* HELPER_GEN_COMMON_H */
37
diff --git a/include/exec/helper-proto-common.h b/include/exec/helper-proto-common.h
38
index XXXXXXX..XXXXXXX 100644
39
--- a/include/exec/helper-proto-common.h
40
+++ b/include/exec/helper-proto-common.h
41
@@ -XXX,XX +XXX,XX @@
42
#include "exec/helper-proto.h.inc"
43
#undef HELPER_H
44
45
-#define HELPER_H "accel/tcg/plugin-helpers.h"
46
-#include "exec/helper-proto.h.inc"
47
-#undef HELPER_H
48
-
49
#endif /* HELPER_PROTO_COMMON_H */
50
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
51
index XXXXXXX..XXXXXXX 100644
52
--- a/accel/tcg/plugin-gen.c
53
+++ b/accel/tcg/plugin-gen.c
54
@@ -XXX,XX +XXX,XX @@
55
#include "exec/exec-all.h"
56
#include "exec/plugin-gen.h"
57
#include "exec/translator.h"
58
-#include "exec/helper-proto-common.h"
59
-
60
-#define HELPER_H "accel/tcg/plugin-helpers.h"
61
-#include "exec/helper-info.c.inc"
62
-#undef HELPER_H
63
64
/*
65
* plugin_cb_start TCG op args[]:
66
@@ -XXX,XX +XXX,XX @@ enum plugin_gen_cb {
67
PLUGIN_GEN_N_CBS,
68
};
69
70
-/*
71
- * These helpers are stubs that get dynamically switched out for calls
72
- * direct to the plugin if they are subscribed to.
73
- */
74
-void HELPER(plugin_vcpu_udata_cb_no_wg)(uint32_t cpu_index, void *udata)
75
-{ }
76
-
77
-void HELPER(plugin_vcpu_udata_cb_no_rwg)(uint32_t cpu_index, void *udata)
78
-{ }
79
-
80
-void HELPER(plugin_vcpu_mem_cb)(unsigned int vcpu_index,
81
- qemu_plugin_meminfo_t info, uint64_t vaddr,
82
- void *userdata)
83
-{ }
84
-
85
static void plugin_gen_empty_callback(enum plugin_gen_from from)
17
{
86
{
18
bool ret = false;
87
switch (from) {
19
- CPUClass *cc = CPU_GET_CLASS(cs);
20
AVRCPU *cpu = AVR_CPU(cs);
21
CPUAVRState *env = &cpu->env;
22
23
if (interrupt_request & CPU_INTERRUPT_RESET) {
24
if (cpu_interrupts_enabled(env)) {
25
cs->exception_index = EXCP_RESET;
26
- cc->tcg_ops->do_interrupt(cs);
27
+ avr_cpu_do_interrupt(cs);
28
29
cs->interrupt_request &= ~CPU_INTERRUPT_RESET;
30
31
@@ -XXX,XX +XXX,XX @@ bool avr_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
32
if (cpu_interrupts_enabled(env) && env->intsrc != 0) {
33
int index = ctz32(env->intsrc);
34
cs->exception_index = EXCP_INT(index);
35
- cc->tcg_ops->do_interrupt(cs);
36
+ avr_cpu_do_interrupt(cs);
37
38
env->intsrc &= env->intsrc - 1; /* clear the interrupt */
39
if (!env->intsrc) {
40
--
88
--
41
2.34.1
89
2.34.1
42
90
43
91
diff view generated by jsdifflib
Deleted patch
1
We're about to start validating PAGE_EXEC, which means that we've
2
got to mark the vsyscall page executable. We had been special
3
casing this entirely within translate.
4
1
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
9
linux-user/elfload.c | 23 +++++++++++++++++++++++
10
1 file changed, 23 insertions(+)
11
12
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/linux-user/elfload.c
15
+++ b/linux-user/elfload.c
16
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
17
(*regs)[26] = tswapreg(env->segs[R_GS].selector & 0xffff);
18
}
19
20
+#if ULONG_MAX >= TARGET_VSYSCALL_PAGE
21
+#define INIT_GUEST_COMMPAGE
22
+static bool init_guest_commpage(void)
23
+{
24
+ /*
25
+ * The vsyscall page is at a high negative address aka kernel space,
26
+ * which means that we cannot actually allocate it with target_mmap.
27
+ * We still should be able to use page_set_flags, unless the user
28
+ * has specified -R reserved_va, which would trigger an assert().
29
+ */
30
+ if (reserved_va != 0 &&
31
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE >= reserved_va) {
32
+ error_report("Cannot allocate vsyscall page");
33
+ exit(EXIT_FAILURE);
34
+ }
35
+ page_set_flags(TARGET_VSYSCALL_PAGE,
36
+ TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE,
37
+ PAGE_EXEC | PAGE_VALID);
38
+ return true;
39
+}
40
+#endif
41
#else
42
43
#define ELF_START_MMAP 0x80000000
44
@@ -XXX,XX +XXX,XX @@ static abi_ulong create_elf_tables(abi_ulong p, int argc, int envc,
45
#else
46
#define HI_COMMPAGE 0
47
#define LO_COMMPAGE -1
48
+#ifndef INIT_GUEST_COMMPAGE
49
#define init_guest_commpage() true
50
#endif
51
+#endif
52
53
static void pgb_fail_in_use(const char *image_name)
54
{
55
--
56
2.34.1
diff view generated by jsdifflib
Deleted patch
1
Map the stack executable if required by default or on demand.
2
1
3
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
---
7
include/elf.h | 1 +
8
linux-user/qemu.h | 1 +
9
linux-user/elfload.c | 19 ++++++++++++++++++-
10
3 files changed, 20 insertions(+), 1 deletion(-)
11
12
diff --git a/include/elf.h b/include/elf.h
13
index XXXXXXX..XXXXXXX 100644
14
--- a/include/elf.h
15
+++ b/include/elf.h
16
@@ -XXX,XX +XXX,XX @@ typedef int64_t Elf64_Sxword;
17
#define PT_LOPROC 0x70000000
18
#define PT_HIPROC 0x7fffffff
19
20
+#define PT_GNU_STACK (PT_LOOS + 0x474e551)
21
#define PT_GNU_PROPERTY (PT_LOOS + 0x474e553)
22
23
#define PT_MIPS_REGINFO 0x70000000
24
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
25
index XXXXXXX..XXXXXXX 100644
26
--- a/linux-user/qemu.h
27
+++ b/linux-user/qemu.h
28
@@ -XXX,XX +XXX,XX @@ struct image_info {
29
uint32_t elf_flags;
30
int personality;
31
abi_ulong alignment;
32
+ bool exec_stack;
33
34
/* Generic semihosting knows about these pointers. */
35
abi_ulong arg_strings; /* strings for argv */
36
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/linux-user/elfload.c
39
+++ b/linux-user/elfload.c
40
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
41
#define ELF_ARCH EM_386
42
43
#define ELF_PLATFORM get_elf_platform()
44
+#define EXSTACK_DEFAULT true
45
46
static const char *get_elf_platform(void)
47
{
48
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUX86State *en
49
50
#define ELF_ARCH EM_ARM
51
#define ELF_CLASS ELFCLASS32
52
+#define EXSTACK_DEFAULT true
53
54
static inline void init_thread(struct target_pt_regs *regs,
55
struct image_info *infop)
56
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
57
#else
58
59
#define ELF_CLASS ELFCLASS32
60
+#define EXSTACK_DEFAULT true
61
62
#endif
63
64
@@ -XXX,XX +XXX,XX @@ static void elf_core_copy_regs(target_elf_gregset_t *regs, const CPUPPCState *en
65
66
#define ELF_CLASS ELFCLASS64
67
#define ELF_ARCH EM_LOONGARCH
68
+#define EXSTACK_DEFAULT true
69
70
#define elf_check_arch(x) ((x) == EM_LOONGARCH)
71
72
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap(void)
73
#define ELF_CLASS ELFCLASS32
74
#endif
75
#define ELF_ARCH EM_MIPS
76
+#define EXSTACK_DEFAULT true
77
78
#ifdef TARGET_ABI_MIPSN32
79
#define elf_check_abi(x) ((x) & EF_MIPS_ABI2)
80
@@ -XXX,XX +XXX,XX @@ static inline void init_thread(struct target_pt_regs *regs,
81
#define bswaptls(ptr) bswap32s(ptr)
82
#endif
83
84
+#ifndef EXSTACK_DEFAULT
85
+#define EXSTACK_DEFAULT false
86
+#endif
87
+
88
#include "elf.h"
89
90
/* We must delay the following stanzas until after "elf.h". */
91
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
92
struct image_info *info)
93
{
94
abi_ulong size, error, guard;
95
+ int prot;
96
97
size = guest_stack_size;
98
if (size < STACK_LOWER_LIMIT) {
99
@@ -XXX,XX +XXX,XX @@ static abi_ulong setup_arg_pages(struct linux_binprm *bprm,
100
guard = qemu_real_host_page_size();
101
}
102
103
- error = target_mmap(0, size + guard, PROT_READ | PROT_WRITE,
104
+ prot = PROT_READ | PROT_WRITE;
105
+ if (info->exec_stack) {
106
+ prot |= PROT_EXEC;
107
+ }
108
+ error = target_mmap(0, size + guard, prot,
109
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
110
if (error == -1) {
111
perror("mmap stack");
112
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
113
*/
114
loaddr = -1, hiaddr = 0;
115
info->alignment = 0;
116
+ info->exec_stack = EXSTACK_DEFAULT;
117
for (i = 0; i < ehdr->e_phnum; ++i) {
118
struct elf_phdr *eppnt = phdr + i;
119
if (eppnt->p_type == PT_LOAD) {
120
@@ -XXX,XX +XXX,XX @@ static void load_elf_image(const char *image_name, int image_fd,
121
if (!parse_elf_properties(image_fd, info, eppnt, bprm_buf, &err)) {
122
goto exit_errmsg;
123
}
124
+ } else if (eppnt->p_type == PT_GNU_STACK) {
125
+ info->exec_stack = eppnt->p_flags & PF_X;
126
}
127
}
128
129
--
130
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
1
3
Currently it's possible to execute pages that do not have PAGE_EXEC
4
if there is an existing translation block. Fix by invalidating TBs
5
that touch the affected pages.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Message-Id: <20220817150506.592862-2-iii@linux.ibm.com>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
11
linux-user/mmap.c | 6 ++++--
12
1 file changed, 4 insertions(+), 2 deletions(-)
13
14
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/mmap.c
17
+++ b/linux-user/mmap.c
18
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
19
goto error;
20
}
21
}
22
+
23
page_set_flags(start, start + len, page_flags);
24
- mmap_unlock();
25
- return 0;
26
+ tb_invalidate_phys_range(start, start + len);
27
+ ret = 0;
28
+
29
error:
30
mmap_unlock();
31
return ret;
32
--
33
2.34.1
diff view generated by jsdifflib
1
These will be useful in properly ending the TB.
1
Since we no longer emit plugin helpers during the initial code
2
translation phase, we don't need to specially mark plugin helpers.
2
3
3
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
4
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
6
---
8
target/riscv/translate.c | 10 +++++++++-
7
include/tcg/tcg.h | 2 --
9
1 file changed, 9 insertions(+), 1 deletion(-)
8
plugins/core.c | 10 ++++------
9
tcg/tcg.c | 4 +---
10
3 files changed, 5 insertions(+), 11 deletions(-)
10
11
11
diff --git a/target/riscv/translate.c b/target/riscv/translate.c
12
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
12
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
13
--- a/target/riscv/translate.c
14
--- a/include/tcg/tcg.h
14
+++ b/target/riscv/translate.c
15
+++ b/include/tcg/tcg.h
15
@@ -XXX,XX +XXX,XX @@ static uint32_t opcode_at(DisasContextBase *dcbase, target_ulong pc)
16
@@ -XXX,XX +XXX,XX @@ typedef TCGv_ptr TCGv_env;
16
/* Include decoders for factored-out extensions */
17
#define TCG_CALL_NO_SIDE_EFFECTS 0x0004
17
#include "decode-XVentanaCondOps.c.inc"
18
/* Helper is G_NORETURN. */
18
19
#define TCG_CALL_NO_RETURN 0x0008
19
+/* The specification allows for longer insns, but not supported by qemu. */
20
-/* Helper is part of Plugins. */
20
+#define MAX_INSN_LEN 4
21
-#define TCG_CALL_PLUGIN 0x0010
21
+
22
22
+static inline int insn_len(uint16_t first_word)
23
/* convenience version of most used call flags */
23
+{
24
#define TCG_CALL_NO_RWG TCG_CALL_NO_READ_GLOBALS
24
+ return (first_word & 3) == 3 ? 4 : 2;
25
diff --git a/plugins/core.c b/plugins/core.c
25
+}
26
index XXXXXXX..XXXXXXX 100644
26
+
27
--- a/plugins/core.c
27
static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
28
+++ b/plugins/core.c
29
@@ -XXX,XX +XXX,XX @@ void plugin_register_dyn_cb__udata(GArray **arr,
30
void *udata)
28
{
31
{
29
/*
32
static TCGHelperInfo info[3] = {
30
@@ -XXX,XX +XXX,XX @@ static void decode_opc(CPURISCVState *env, DisasContext *ctx, uint16_t opcode)
33
- [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG | TCG_CALL_PLUGIN,
31
};
34
- [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG | TCG_CALL_PLUGIN,
32
35
- [QEMU_PLUGIN_CB_RW_REGS].flags = TCG_CALL_PLUGIN,
33
/* Check for compressed insn */
36
+ [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG,
34
- if (extract16(opcode, 0, 2) != 3) {
37
+ [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG,
35
+ if (insn_len(opcode) == 2) {
38
/*
36
if (!has_ext(ctx, RVC)) {
39
* Match qemu_plugin_vcpu_udata_cb_t:
37
gen_exception_illegal(ctx);
40
* void (*)(uint32_t, void *)
38
} else {
41
@@ -XXX,XX +XXX,XX @@ void plugin_register_vcpu_mem_cb(GArray **arr,
42
!__builtin_types_compatible_p(qemu_plugin_meminfo_t, int32_t));
43
44
static TCGHelperInfo info[3] = {
45
- [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG | TCG_CALL_PLUGIN,
46
- [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG | TCG_CALL_PLUGIN,
47
- [QEMU_PLUGIN_CB_RW_REGS].flags = TCG_CALL_PLUGIN,
48
+ [QEMU_PLUGIN_CB_NO_REGS].flags = TCG_CALL_NO_RWG,
49
+ [QEMU_PLUGIN_CB_R_REGS].flags = TCG_CALL_NO_WG,
50
/*
51
* Match qemu_plugin_vcpu_mem_cb_t:
52
* void (*)(uint32_t, qemu_plugin_meminfo_t, uint64_t, void *)
53
diff --git a/tcg/tcg.c b/tcg/tcg.c
54
index XXXXXXX..XXXXXXX 100644
55
--- a/tcg/tcg.c
56
+++ b/tcg/tcg.c
57
@@ -XXX,XX +XXX,XX @@ static void tcg_gen_callN(void *func, TCGHelperInfo *info,
58
59
#ifdef CONFIG_PLUGIN
60
/* Flag helpers that may affect guest state */
61
- if (tcg_ctx->plugin_insn &&
62
- !(info->flags & TCG_CALL_PLUGIN) &&
63
- !(info->flags & TCG_CALL_NO_SIDE_EFFECTS)) {
64
+ if (tcg_ctx->plugin_insn && !(info->flags & TCG_CALL_NO_SIDE_EFFECTS)) {
65
tcg_ctx->plugin_insn->calls_helpers = true;
66
}
67
#endif
39
--
68
--
40
2.34.1
69
2.34.1
diff view generated by jsdifflib
1
The only user can easily use translator_lduw and
1
These opcodes are no longer used.
2
adjust the type to signed during the return.
3
2
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
3
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
5
---
9
include/exec/translator.h | 1 -
6
include/tcg/tcg-op-common.h | 2 --
10
target/i386/tcg/translate.c | 2 +-
7
include/tcg/tcg-opc.h | 2 --
11
2 files changed, 1 insertion(+), 2 deletions(-)
8
accel/tcg/plugin-gen.c | 18 ------------------
9
tcg/tcg-op.c | 10 ----------
10
4 files changed, 32 deletions(-)
12
11
13
diff --git a/include/exec/translator.h b/include/exec/translator.h
12
diff --git a/include/tcg/tcg-op-common.h b/include/tcg/tcg-op-common.h
14
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
15
--- a/include/exec/translator.h
14
--- a/include/tcg/tcg-op-common.h
16
+++ b/include/exec/translator.h
15
+++ b/include/tcg/tcg-op-common.h
17
@@ -XXX,XX +XXX,XX @@ bool translator_use_goto_tb(DisasContextBase *db, target_ulong dest);
16
@@ -XXX,XX +XXX,XX @@ void tcg_gen_lookup_and_goto_ptr(void);
18
17
19
#define FOR_EACH_TRANSLATOR_LD(F) \
18
void tcg_gen_plugin_cb(unsigned from);
20
F(translator_ldub, uint8_t, cpu_ldub_code, /* no swap */) \
19
void tcg_gen_plugin_mem_cb(TCGv_i64 addr, unsigned meminfo);
21
- F(translator_ldsw, int16_t, cpu_ldsw_code, bswap16) \
20
-void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr);
22
F(translator_lduw, uint16_t, cpu_lduw_code, bswap16) \
21
-void tcg_gen_plugin_cb_end(void);
23
F(translator_ldl, uint32_t, cpu_ldl_code, bswap32) \
22
24
F(translator_ldq, uint64_t, cpu_ldq_code, bswap64)
23
/* 32 bit ops */
25
diff --git a/target/i386/tcg/translate.c b/target/i386/tcg/translate.c
24
25
diff --git a/include/tcg/tcg-opc.h b/include/tcg/tcg-opc.h
26
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/i386/tcg/translate.c
27
--- a/include/tcg/tcg-opc.h
28
+++ b/target/i386/tcg/translate.c
28
+++ b/include/tcg/tcg-opc.h
29
@@ -XXX,XX +XXX,XX @@ static inline uint8_t x86_ldub_code(CPUX86State *env, DisasContext *s)
29
@@ -XXX,XX +XXX,XX @@ DEF(goto_ptr, 0, 1, 0, TCG_OPF_BB_EXIT | TCG_OPF_BB_END)
30
30
31
static inline int16_t x86_ldsw_code(CPUX86State *env, DisasContext *s)
31
DEF(plugin_cb, 0, 0, 1, TCG_OPF_NOT_PRESENT)
32
{
32
DEF(plugin_mem_cb, 0, 1, 1, TCG_OPF_NOT_PRESENT)
33
- return translator_ldsw(env, &s->base, advance_pc(env, s, 2));
33
-DEF(plugin_cb_start, 0, 0, 3, TCG_OPF_NOT_PRESENT)
34
+ return translator_lduw(env, &s->base, advance_pc(env, s, 2));
34
-DEF(plugin_cb_end, 0, 0, 0, TCG_OPF_NOT_PRESENT)
35
36
/* Replicate ld/st ops for 32 and 64-bit guest addresses. */
37
DEF(qemu_ld_a32_i32, 1, 1, 1,
38
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/accel/tcg/plugin-gen.c
41
+++ b/accel/tcg/plugin-gen.c
42
@@ -XXX,XX +XXX,XX @@
43
#include "exec/plugin-gen.h"
44
#include "exec/translator.h"
45
46
-/*
47
- * plugin_cb_start TCG op args[]:
48
- * 0: enum plugin_gen_from
49
- * 1: enum plugin_gen_cb
50
- * 2: set to 1 for mem callback that is a write, 0 otherwise.
51
- */
52
-
53
enum plugin_gen_from {
54
PLUGIN_GEN_FROM_TB,
55
PLUGIN_GEN_FROM_INSN,
56
PLUGIN_GEN_AFTER_INSN,
57
PLUGIN_GEN_AFTER_TB,
58
- PLUGIN_GEN_N_FROMS,
59
-};
60
-
61
-enum plugin_gen_cb {
62
- PLUGIN_GEN_CB_UDATA,
63
- PLUGIN_GEN_CB_UDATA_R,
64
- PLUGIN_GEN_CB_INLINE,
65
- PLUGIN_GEN_CB_MEM,
66
- PLUGIN_GEN_ENABLE_MEM_HELPER,
67
- PLUGIN_GEN_DISABLE_MEM_HELPER,
68
- PLUGIN_GEN_N_CBS,
69
};
70
71
static void plugin_gen_empty_callback(enum plugin_gen_from from)
72
diff --git a/tcg/tcg-op.c b/tcg/tcg-op.c
73
index XXXXXXX..XXXXXXX 100644
74
--- a/tcg/tcg-op.c
75
+++ b/tcg/tcg-op.c
76
@@ -XXX,XX +XXX,XX @@ void tcg_gen_plugin_mem_cb(TCGv_i64 addr, unsigned meminfo)
77
tcg_gen_op2(INDEX_op_plugin_mem_cb, tcgv_i64_arg(addr), meminfo);
35
}
78
}
36
79
37
static inline uint16_t x86_lduw_code(CPUX86State *env, DisasContext *s)
80
-void tcg_gen_plugin_cb_start(unsigned from, unsigned type, unsigned wr)
81
-{
82
- tcg_gen_op3(INDEX_op_plugin_cb_start, from, type, wr);
83
-}
84
-
85
-void tcg_gen_plugin_cb_end(void)
86
-{
87
- tcg_emit_op(INDEX_op_plugin_cb_end, 0);
88
-}
89
-
90
/* 32 bit ops */
91
92
void tcg_gen_discard_i32(TCGv_i32 arg)
38
--
93
--
39
2.34.1
94
2.34.1
diff view generated by jsdifflib
1
The function is not used outside of cpu-exec.c. Move it and
1
We have qemu_plugin_dyn_cb.type to differentiate the various
2
its subroutines up in the file, before the first use.
2
callback types, so we do not need to keep them in separate queues.
3
3
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
4
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
6
---
9
include/exec/exec-all.h | 3 -
7
include/qemu/plugin.h | 35 ++++++----------
10
accel/tcg/cpu-exec.c | 122 ++++++++++++++++++++--------------------
8
accel/tcg/plugin-gen.c | 90 ++++++++++++++++++++++--------------------
11
2 files changed, 61 insertions(+), 64 deletions(-)
9
plugins/api.c | 18 +++------
10
3 files changed, 65 insertions(+), 78 deletions(-)
12
11
13
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
12
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
14
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
15
--- a/include/exec/exec-all.h
14
--- a/include/qemu/plugin.h
16
+++ b/include/exec/exec-all.h
15
+++ b/include/qemu/plugin.h
17
@@ -XXX,XX +XXX,XX @@ void tb_invalidate_phys_addr(AddressSpace *as, hwaddr addr, MemTxAttrs attrs);
16
@@ -XXX,XX +XXX,XX @@ union qemu_plugin_cb_sig {
18
#endif
17
};
19
void tb_flush(CPUState *cpu);
18
20
void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr);
19
enum plugin_dyn_cb_type {
21
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
20
- PLUGIN_CB_INSN,
22
- target_ulong cs_base, uint32_t flags,
21
- PLUGIN_CB_MEM,
23
- uint32_t cflags);
22
- PLUGIN_N_CB_TYPES,
24
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr);
25
26
/* GETPC is the true target of the return instruction that we'll execute. */
27
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/accel/tcg/cpu-exec.c
30
+++ b/accel/tcg/cpu-exec.c
31
@@ -XXX,XX +XXX,XX @@ uint32_t curr_cflags(CPUState *cpu)
32
return cflags;
33
}
34
35
+struct tb_desc {
36
+ target_ulong pc;
37
+ target_ulong cs_base;
38
+ CPUArchState *env;
39
+ tb_page_addr_t phys_page1;
40
+ uint32_t flags;
41
+ uint32_t cflags;
42
+ uint32_t trace_vcpu_dstate;
43
+};
44
+
45
+static bool tb_lookup_cmp(const void *p, const void *d)
46
+{
47
+ const TranslationBlock *tb = p;
48
+ const struct tb_desc *desc = d;
49
+
50
+ if (tb->pc == desc->pc &&
51
+ tb->page_addr[0] == desc->phys_page1 &&
52
+ tb->cs_base == desc->cs_base &&
53
+ tb->flags == desc->flags &&
54
+ tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
55
+ tb_cflags(tb) == desc->cflags) {
56
+ /* check next page if needed */
57
+ if (tb->page_addr[1] == -1) {
58
+ return true;
59
+ } else {
60
+ tb_page_addr_t phys_page2;
61
+ target_ulong virt_page2;
62
+
63
+ virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
64
+ phys_page2 = get_page_addr_code(desc->env, virt_page2);
65
+ if (tb->page_addr[1] == phys_page2) {
66
+ return true;
67
+ }
68
+ }
69
+ }
70
+ return false;
71
+}
72
+
73
+static TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
74
+ target_ulong cs_base, uint32_t flags,
75
+ uint32_t cflags)
76
+{
77
+ tb_page_addr_t phys_pc;
78
+ struct tb_desc desc;
79
+ uint32_t h;
80
+
81
+ desc.env = cpu->env_ptr;
82
+ desc.cs_base = cs_base;
83
+ desc.flags = flags;
84
+ desc.cflags = cflags;
85
+ desc.trace_vcpu_dstate = *cpu->trace_dstate;
86
+ desc.pc = pc;
87
+ phys_pc = get_page_addr_code(desc.env, pc);
88
+ if (phys_pc == -1) {
89
+ return NULL;
90
+ }
91
+ desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
92
+ h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
93
+ return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
94
+}
95
+
96
/* Might cause an exception, so have a longjmp destination ready */
97
static inline TranslationBlock *tb_lookup(CPUState *cpu, target_ulong pc,
98
target_ulong cs_base,
99
@@ -XXX,XX +XXX,XX @@ void cpu_exec_step_atomic(CPUState *cpu)
100
end_exclusive();
101
}
102
103
-struct tb_desc {
104
- target_ulong pc;
105
- target_ulong cs_base;
106
- CPUArchState *env;
107
- tb_page_addr_t phys_page1;
108
- uint32_t flags;
109
- uint32_t cflags;
110
- uint32_t trace_vcpu_dstate;
111
-};
23
-};
112
-
24
-
113
-static bool tb_lookup_cmp(const void *p, const void *d)
25
-enum plugin_dyn_cb_subtype {
114
-{
26
PLUGIN_CB_REGULAR,
115
- const TranslationBlock *tb = p;
27
PLUGIN_CB_INLINE,
116
- const struct tb_desc *desc = d;
28
- PLUGIN_N_CB_SUBTYPES,
117
-
29
};
118
- if (tb->pc == desc->pc &&
30
119
- tb->page_addr[0] == desc->phys_page1 &&
31
/*
120
- tb->cs_base == desc->cs_base &&
32
@@ -XXX,XX +XXX,XX @@ enum plugin_dyn_cb_subtype {
121
- tb->flags == desc->flags &&
33
*/
122
- tb->trace_vcpu_dstate == desc->trace_vcpu_dstate &&
34
struct qemu_plugin_dyn_cb {
123
- tb_cflags(tb) == desc->cflags) {
35
void *userp;
124
- /* check next page if needed */
36
- enum plugin_dyn_cb_subtype type;
125
- if (tb->page_addr[1] == -1) {
37
+ enum plugin_dyn_cb_type type;
126
- return true;
38
/* @rw applies to mem callbacks only (both regular and inline) */
127
- } else {
39
enum qemu_plugin_mem_rw rw;
128
- tb_page_addr_t phys_page2;
40
/* fields specific to each dyn_cb type go here */
129
- target_ulong virt_page2;
41
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_insn {
130
-
42
GByteArray *data;
131
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
43
uint64_t vaddr;
132
- phys_page2 = get_page_addr_code(desc->env, virt_page2);
44
void *haddr;
133
- if (tb->page_addr[1] == phys_page2) {
45
- GArray *cbs[PLUGIN_N_CB_TYPES][PLUGIN_N_CB_SUBTYPES];
134
- return true;
46
+ GArray *insn_cbs;
135
- }
47
+ GArray *mem_cbs;
48
bool calls_helpers;
49
50
/* if set, the instruction calls helpers that might access guest memory */
51
@@ -XXX,XX +XXX,XX @@ static inline void qemu_plugin_insn_cleanup_fn(gpointer data)
52
53
static inline struct qemu_plugin_insn *qemu_plugin_insn_alloc(void)
54
{
55
- int i, j;
56
struct qemu_plugin_insn *insn = g_new0(struct qemu_plugin_insn, 1);
57
- insn->data = g_byte_array_sized_new(4);
58
59
- for (i = 0; i < PLUGIN_N_CB_TYPES; i++) {
60
- for (j = 0; j < PLUGIN_N_CB_SUBTYPES; j++) {
61
- insn->cbs[i][j] = g_array_new(false, false,
62
- sizeof(struct qemu_plugin_dyn_cb));
136
- }
63
- }
137
- }
64
- }
138
- return false;
65
+ insn->data = g_byte_array_sized_new(4);
139
-}
66
return insn;
67
}
68
69
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_tb {
70
/* if set, the TB calls helpers that might access guest memory */
71
bool mem_helper;
72
73
- GArray *cbs[PLUGIN_N_CB_SUBTYPES];
74
+ GArray *cbs;
75
};
76
77
/**
78
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_insn *qemu_plugin_tb_insn_get(struct qemu_plugin_tb *tb,
79
uint64_t pc)
80
{
81
struct qemu_plugin_insn *insn;
82
- int i, j;
83
84
if (unlikely(tb->n == tb->insns->len)) {
85
struct qemu_plugin_insn *new_insn = qemu_plugin_insn_alloc();
86
g_ptr_array_add(tb->insns, new_insn);
87
}
88
+
89
insn = g_ptr_array_index(tb->insns, tb->n++);
90
g_byte_array_set_size(insn->data, 0);
91
insn->calls_helpers = false;
92
insn->mem_helper = false;
93
insn->vaddr = pc;
140
-
94
-
141
-TranslationBlock *tb_htable_lookup(CPUState *cpu, target_ulong pc,
95
- for (i = 0; i < PLUGIN_N_CB_TYPES; i++) {
142
- target_ulong cs_base, uint32_t flags,
96
- for (j = 0; j < PLUGIN_N_CB_SUBTYPES; j++) {
143
- uint32_t cflags)
97
- g_array_set_size(insn->cbs[i][j], 0);
144
-{
98
- }
145
- tb_page_addr_t phys_pc;
99
+ if (insn->insn_cbs) {
146
- struct tb_desc desc;
100
+ g_array_set_size(insn->insn_cbs, 0);
147
- uint32_t h;
101
+ }
102
+ if (insn->mem_cbs) {
103
+ g_array_set_size(insn->mem_cbs, 0);
104
}
105
106
return insn;
107
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
108
index XXXXXXX..XXXXXXX 100644
109
--- a/accel/tcg/plugin-gen.c
110
+++ b/accel/tcg/plugin-gen.c
111
@@ -XXX,XX +XXX,XX @@ void plugin_gen_disable_mem_helpers(void)
112
static void gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
113
struct qemu_plugin_insn *insn)
114
{
115
- GArray *cbs[2];
116
GArray *arr;
117
- size_t n_cbs;
118
+ size_t len;
119
120
/*
121
* Tracking memory accesses performed from helpers requires extra work.
122
@@ -XXX,XX +XXX,XX @@ static void gen_enable_mem_helper(struct qemu_plugin_tb *ptb,
123
return;
124
}
125
126
- cbs[0] = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR];
127
- cbs[1] = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
128
- n_cbs = cbs[0]->len + cbs[1]->len;
148
-
129
-
149
- desc.env = cpu->env_ptr;
130
- if (n_cbs == 0) {
150
- desc.cs_base = cs_base;
131
+ if (!insn->mem_cbs || !insn->mem_cbs->len) {
151
- desc.flags = flags;
132
insn->mem_helper = false;
152
- desc.cflags = cflags;
133
return;
153
- desc.trace_vcpu_dstate = *cpu->trace_dstate;
134
}
154
- desc.pc = pc;
135
insn->mem_helper = true;
155
- phys_pc = get_page_addr_code(desc.env, pc);
136
ptb->mem_helper = true;
156
- if (phys_pc == -1) {
137
157
- return NULL;
138
+ /*
158
- }
139
+ * TODO: It seems like we should be able to use ref/unref
159
- desc.phys_page1 = phys_pc & TARGET_PAGE_MASK;
140
+ * to avoid needing to actually copy this array.
160
- h = tb_hash_func(phys_pc, pc, flags, cflags, *cpu->trace_dstate);
141
+ * Alternately, perhaps we could allocate new memory adjacent
161
- return qht_lookup_custom(&tb_ctx.htable, &desc, h, tb_lookup_cmp);
142
+ * to the TranslationBlock itself, so that we do not have to
162
-}
143
+ * actively manage the lifetime after this.
144
+ */
145
+ len = insn->mem_cbs->len;
146
arr = g_array_sized_new(false, false,
147
- sizeof(struct qemu_plugin_dyn_cb), n_cbs);
148
- g_array_append_vals(arr, cbs[0]->data, cbs[0]->len);
149
- g_array_append_vals(arr, cbs[1]->data, cbs[1]->len);
163
-
150
-
164
void tb_set_jmp_target(TranslationBlock *tb, int n, uintptr_t addr)
151
+ sizeof(struct qemu_plugin_dyn_cb), len);
165
{
152
+ memcpy(arr->data, insn->mem_cbs->data,
166
if (TCG_TARGET_HAS_direct_jump) {
153
+ len * sizeof(struct qemu_plugin_dyn_cb));
154
qemu_plugin_add_dyn_cb_arr(arr);
155
156
tcg_gen_st_ptr(tcg_constant_ptr((intptr_t)arr), tcg_env,
157
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
158
case PLUGIN_GEN_FROM_TB:
159
assert(insn == NULL);
160
161
- cbs = plugin_tb->cbs[PLUGIN_CB_REGULAR];
162
+ cbs = plugin_tb->cbs;
163
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
164
struct qemu_plugin_dyn_cb *cb =
165
&g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
166
- gen_udata_cb(cb);
167
- }
168
169
- cbs = plugin_tb->cbs[PLUGIN_CB_INLINE];
170
- for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
171
- struct qemu_plugin_dyn_cb *cb =
172
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
173
- gen_inline_cb(cb);
174
+ switch (cb->type) {
175
+ case PLUGIN_CB_REGULAR:
176
+ gen_udata_cb(cb);
177
+ break;
178
+ case PLUGIN_CB_INLINE:
179
+ gen_inline_cb(cb);
180
+ break;
181
+ default:
182
+ g_assert_not_reached();
183
+ }
184
}
185
break;
186
187
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
188
189
gen_enable_mem_helper(plugin_tb, insn);
190
191
- cbs = insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR];
192
+ cbs = insn->insn_cbs;
193
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
194
struct qemu_plugin_dyn_cb *cb =
195
&g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
196
- gen_udata_cb(cb);
197
- }
198
199
- cbs = insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_INLINE];
200
- for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
201
- struct qemu_plugin_dyn_cb *cb =
202
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
203
- gen_inline_cb(cb);
204
+ switch (cb->type) {
205
+ case PLUGIN_CB_REGULAR:
206
+ gen_udata_cb(cb);
207
+ break;
208
+ case PLUGIN_CB_INLINE:
209
+ gen_inline_cb(cb);
210
+ break;
211
+ default:
212
+ g_assert_not_reached();
213
+ }
214
}
215
break;
216
217
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
218
219
tcg_ctx->emit_before_op = op;
220
221
- cbs = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR];
222
+ cbs = insn->mem_cbs;
223
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
224
struct qemu_plugin_dyn_cb *cb =
225
&g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
226
- if (cb->rw & rw) {
227
- gen_mem_cb(cb, meminfo, addr);
228
- }
229
- }
230
231
- cbs = insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE];
232
- for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
233
- struct qemu_plugin_dyn_cb *cb =
234
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
235
if (cb->rw & rw) {
236
- gen_inline_cb(cb);
237
+ switch (cb->type) {
238
+ case PLUGIN_CB_REGULAR:
239
+ gen_mem_cb(cb, meminfo, addr);
240
+ break;
241
+ case PLUGIN_CB_INLINE:
242
+ gen_inline_cb(cb);
243
+ break;
244
+ default:
245
+ g_assert_not_reached();
246
+ }
247
}
248
}
249
250
@@ -XXX,XX +XXX,XX @@ bool plugin_gen_tb_start(CPUState *cpu, const DisasContextBase *db,
251
252
if (test_bit(QEMU_PLUGIN_EV_VCPU_TB_TRANS, cpu->plugin_state->event_mask)) {
253
struct qemu_plugin_tb *ptb = tcg_ctx->plugin_tb;
254
- int i;
255
256
/* reset callbacks */
257
- for (i = 0; i < PLUGIN_N_CB_SUBTYPES; i++) {
258
- if (ptb->cbs[i]) {
259
- g_array_set_size(ptb->cbs[i], 0);
260
- }
261
+ if (ptb->cbs) {
262
+ g_array_set_size(ptb->cbs, 0);
263
}
264
ptb->n = 0;
265
266
diff --git a/plugins/api.c b/plugins/api.c
267
index XXXXXXX..XXXXXXX 100644
268
--- a/plugins/api.c
269
+++ b/plugins/api.c
270
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_tb_exec_cb(struct qemu_plugin_tb *tb,
271
void *udata)
272
{
273
if (!tb->mem_only) {
274
- plugin_register_dyn_cb__udata(&tb->cbs[PLUGIN_CB_REGULAR],
275
- cb, flags, udata);
276
+ plugin_register_dyn_cb__udata(&tb->cbs, cb, flags, udata);
277
}
278
}
279
280
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_tb_exec_inline_per_vcpu(
281
uint64_t imm)
282
{
283
if (!tb->mem_only) {
284
- plugin_register_inline_op_on_entry(
285
- &tb->cbs[PLUGIN_CB_INLINE], 0, op, entry, imm);
286
+ plugin_register_inline_op_on_entry(&tb->cbs, 0, op, entry, imm);
287
}
288
}
289
290
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_insn_exec_cb(struct qemu_plugin_insn *insn,
291
void *udata)
292
{
293
if (!insn->mem_only) {
294
- plugin_register_dyn_cb__udata(
295
- &insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_REGULAR], cb, flags, udata);
296
+ plugin_register_dyn_cb__udata(&insn->insn_cbs, cb, flags, udata);
297
}
298
}
299
300
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_insn_exec_inline_per_vcpu(
301
uint64_t imm)
302
{
303
if (!insn->mem_only) {
304
- plugin_register_inline_op_on_entry(
305
- &insn->cbs[PLUGIN_CB_INSN][PLUGIN_CB_INLINE], 0, op, entry, imm);
306
+ plugin_register_inline_op_on_entry(&insn->insn_cbs, 0, op, entry, imm);
307
}
308
}
309
310
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_mem_cb(struct qemu_plugin_insn *insn,
311
enum qemu_plugin_mem_rw rw,
312
void *udata)
313
{
314
- plugin_register_vcpu_mem_cb(&insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_REGULAR],
315
- cb, flags, rw, udata);
316
+ plugin_register_vcpu_mem_cb(&insn->mem_cbs, cb, flags, rw, udata);
317
}
318
319
void qemu_plugin_register_vcpu_mem_inline_per_vcpu(
320
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_register_vcpu_mem_inline_per_vcpu(
321
qemu_plugin_u64 entry,
322
uint64_t imm)
323
{
324
- plugin_register_inline_op_on_entry(
325
- &insn->cbs[PLUGIN_CB_MEM][PLUGIN_CB_INLINE], rw, op, entry, imm);
326
+ plugin_register_inline_op_on_entry(&insn->mem_cbs, rw, op, entry, imm);
327
}
328
329
void qemu_plugin_register_vcpu_tb_trans_cb(qemu_plugin_id_t id,
167
--
330
--
168
2.34.1
331
2.34.1
diff view generated by jsdifflib
1
We're about to start validating PAGE_EXEC, which means
1
Use different enumerators for vcpu_udata and vcpu_mem callbacks.
2
that we've got to put this code into a section that is
3
both writable and executable.
4
2
5
Note that this test did not run on hardware beforehand either.
3
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
6
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
---
5
---
11
tests/tcg/i386/test-i386.c | 2 +-
6
include/qemu/plugin.h | 1 +
12
1 file changed, 1 insertion(+), 1 deletion(-)
7
accel/tcg/plugin-gen.c | 2 +-
8
plugins/core.c | 4 ++--
9
3 files changed, 4 insertions(+), 3 deletions(-)
13
10
14
diff --git a/tests/tcg/i386/test-i386.c b/tests/tcg/i386/test-i386.c
11
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
15
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
16
--- a/tests/tcg/i386/test-i386.c
13
--- a/include/qemu/plugin.h
17
+++ b/tests/tcg/i386/test-i386.c
14
+++ b/include/qemu/plugin.h
18
@@ -XXX,XX +XXX,XX @@ uint8_t code[] = {
15
@@ -XXX,XX +XXX,XX @@ union qemu_plugin_cb_sig {
19
0xc3, /* ret */
16
17
enum plugin_dyn_cb_type {
18
PLUGIN_CB_REGULAR,
19
+ PLUGIN_CB_MEM_REGULAR,
20
PLUGIN_CB_INLINE,
20
};
21
};
21
22
22
-asm(".section \".data\"\n"
23
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
23
+asm(".section \".data_x\",\"awx\"\n"
24
index XXXXXXX..XXXXXXX 100644
24
"smc_code2:\n"
25
--- a/accel/tcg/plugin-gen.c
25
"movl 4(%esp), %eax\n"
26
+++ b/accel/tcg/plugin-gen.c
26
"movl %eax, smc_patch_addr2 + 1\n"
27
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
28
29
if (cb->rw & rw) {
30
switch (cb->type) {
31
- case PLUGIN_CB_REGULAR:
32
+ case PLUGIN_CB_MEM_REGULAR:
33
gen_mem_cb(cb, meminfo, addr);
34
break;
35
case PLUGIN_CB_INLINE:
36
diff --git a/plugins/core.c b/plugins/core.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/plugins/core.c
39
+++ b/plugins/core.c
40
@@ -XXX,XX +XXX,XX @@ void plugin_register_vcpu_mem_cb(GArray **arr,
41
42
struct qemu_plugin_dyn_cb *dyn_cb = plugin_get_dyn_cb(arr);
43
dyn_cb->userp = udata;
44
- dyn_cb->type = PLUGIN_CB_REGULAR;
45
+ dyn_cb->type = PLUGIN_CB_MEM_REGULAR;
46
dyn_cb->rw = rw;
47
dyn_cb->regular.f.vcpu_mem = cb;
48
49
@@ -XXX,XX +XXX,XX @@ void qemu_plugin_vcpu_mem_cb(CPUState *cpu, uint64_t vaddr,
50
break;
51
}
52
switch (cb->type) {
53
- case PLUGIN_CB_REGULAR:
54
+ case PLUGIN_CB_MEM_REGULAR:
55
cb->regular.f.vcpu_mem(cpu->cpu_index, make_plugin_meminfo(oi, rw),
56
vaddr, cb->userp);
57
break;
27
--
58
--
28
2.34.1
59
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
2
1
3
Introduce a function that checks whether a given address is on the same
4
page as where disassembly started. Having it improves readability of
5
the following patches.
6
7
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
8
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Message-Id: <20220811095534.241224-3-iii@linux.ibm.com>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
[rth: Make the DisasContextBase parameter const.]
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
---
14
include/exec/translator.h | 10 ++++++++++
15
1 file changed, 10 insertions(+)
16
17
diff --git a/include/exec/translator.h b/include/exec/translator.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/exec/translator.h
20
+++ b/include/exec/translator.h
21
@@ -XXX,XX +XXX,XX @@ FOR_EACH_TRANSLATOR_LD(GEN_TRANSLATOR_LD)
22
23
#undef GEN_TRANSLATOR_LD
24
25
+/*
26
+ * Return whether addr is on the same page as where disassembly started.
27
+ * Translators can use this to enforce the rule that only single-insn
28
+ * translation blocks are allowed to cross page boundaries.
29
+ */
30
+static inline bool is_same_page(const DisasContextBase *db, target_ulong addr)
31
+{
32
+ return ((addr ^ db->pc_first) & TARGET_PAGE_MASK) == 0;
33
+}
34
+
35
#endif /* EXEC__TRANSLATOR_H */
36
--
37
2.34.1
diff view generated by jsdifflib
1
The base qemu_ram_addr_from_host function is already in
1
The DEBUG_PLUGIN_GEN_OPS ifdef is replaced with "-d op_plugin".
2
softmmu/physmem.c; move the nofail version to be adjacent.
2
The second pr_ops call can be obtained with "-d op".
3
3
4
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
4
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
---
6
---
9
include/exec/cpu-common.h | 1 +
7
include/qemu/log.h | 1 +
10
accel/tcg/cputlb.c | 12 ------------
8
include/tcg/tcg.h | 1 +
11
softmmu/physmem.c | 12 ++++++++++++
9
accel/tcg/plugin-gen.c | 67 +++++++-----------------------------------
12
3 files changed, 13 insertions(+), 12 deletions(-)
10
tcg/tcg.c | 29 +++++++++++++++++-
11
util/log.c | 4 +++
12
5 files changed, 45 insertions(+), 57 deletions(-)
13
13
14
diff --git a/include/exec/cpu-common.h b/include/exec/cpu-common.h
14
diff --git a/include/qemu/log.h b/include/qemu/log.h
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/include/exec/cpu-common.h
16
--- a/include/qemu/log.h
17
+++ b/include/exec/cpu-common.h
17
+++ b/include/qemu/log.h
18
@@ -XXX,XX +XXX,XX @@ typedef uintptr_t ram_addr_t;
18
@@ -XXX,XX +XXX,XX @@ bool qemu_log_separate(void);
19
void qemu_ram_remap(ram_addr_t addr, ram_addr_t length);
19
#define LOG_STRACE (1 << 19)
20
/* This should not be used by devices. */
20
#define LOG_PER_THREAD (1 << 20)
21
ram_addr_t qemu_ram_addr_from_host(void *ptr);
21
#define CPU_LOG_TB_VPU (1 << 21)
22
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr);
22
+#define LOG_TB_OP_PLUGIN (1 << 22)
23
RAMBlock *qemu_ram_block_by_name(const char *name);
23
24
RAMBlock *qemu_ram_block_from_host(void *ptr, bool round_offset,
24
/* Lock/unlock output. */
25
ram_addr_t *offset);
25
26
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
26
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
27
index XXXXXXX..XXXXXXX 100644
27
index XXXXXXX..XXXXXXX 100644
28
--- a/accel/tcg/cputlb.c
28
--- a/include/tcg/tcg.h
29
+++ b/accel/tcg/cputlb.c
29
+++ b/include/tcg/tcg.h
30
@@ -XXX,XX +XXX,XX @@ void tlb_set_page(CPUState *cpu, target_ulong vaddr,
30
@@ -XXX,XX +XXX,XX @@ static inline const TCGOpcode *tcg_swap_vecop_list(const TCGOpcode *n)
31
prot, mmu_idx, size);
32
}
31
}
33
32
34
-static inline ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
33
bool tcg_can_emit_vecop_list(const TCGOpcode *, TCGType, unsigned);
34
+void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs);
35
36
#endif /* TCG_H */
37
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
38
index XXXXXXX..XXXXXXX 100644
39
--- a/accel/tcg/plugin-gen.c
40
+++ b/accel/tcg/plugin-gen.c
41
@@ -XXX,XX +XXX,XX @@
42
*/
43
#include "qemu/osdep.h"
44
#include "qemu/plugin.h"
45
+#include "qemu/log.h"
46
#include "cpu.h"
47
#include "tcg/tcg.h"
48
#include "tcg/tcg-temp-internal.h"
49
@@ -XXX,XX +XXX,XX @@ static void gen_mem_cb(struct qemu_plugin_dyn_cb *cb,
50
tcg_temp_free_i32(cpu_index);
51
}
52
53
-/* #define DEBUG_PLUGIN_GEN_OPS */
54
-static void pr_ops(void)
35
-{
55
-{
36
- ram_addr_t ram_addr;
56
-#ifdef DEBUG_PLUGIN_GEN_OPS
57
- TCGOp *op;
58
- int i = 0;
37
-
59
-
38
- ram_addr = qemu_ram_addr_from_host(ptr);
60
- QTAILQ_FOREACH(op, &tcg_ctx->ops, link) {
39
- if (ram_addr == RAM_ADDR_INVALID) {
61
- const char *name = "";
40
- error_report("Bad ram pointer %p", ptr);
62
- const char *type = "";
41
- abort();
63
-
64
- if (op->opc == INDEX_op_plugin_cb_start) {
65
- switch (op->args[0]) {
66
- case PLUGIN_GEN_FROM_TB:
67
- name = "tb";
68
- break;
69
- case PLUGIN_GEN_FROM_INSN:
70
- name = "insn";
71
- break;
72
- case PLUGIN_GEN_FROM_MEM:
73
- name = "mem";
74
- break;
75
- case PLUGIN_GEN_AFTER_INSN:
76
- name = "after insn";
77
- break;
78
- default:
79
- break;
80
- }
81
- switch (op->args[1]) {
82
- case PLUGIN_GEN_CB_UDATA:
83
- type = "udata";
84
- break;
85
- case PLUGIN_GEN_CB_INLINE:
86
- type = "inline";
87
- break;
88
- case PLUGIN_GEN_CB_MEM:
89
- type = "mem";
90
- break;
91
- case PLUGIN_GEN_ENABLE_MEM_HELPER:
92
- type = "enable mem helper";
93
- break;
94
- case PLUGIN_GEN_DISABLE_MEM_HELPER:
95
- type = "disable mem helper";
96
- break;
97
- default:
98
- break;
99
- }
100
- }
101
- printf("op[%2i]: %s %s %s\n", i, tcg_op_defs[op->opc].name, name, type);
102
- i++;
42
- }
103
- }
43
- return ram_addr;
104
-#endif
44
-}
105
-}
45
-
106
-
46
/*
107
static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
47
* Note: tlb_fill() can trigger a resize of the TLB. This means that all of the
108
{
48
* caller's prior references to the TLB table (e.g. CPUTLBEntry pointers) must
109
TCGOp *op, *next;
49
diff --git a/softmmu/physmem.c b/softmmu/physmem.c
110
int insn_idx = -1;
50
index XXXXXXX..XXXXXXX 100644
111
51
--- a/softmmu/physmem.c
112
- pr_ops();
52
+++ b/softmmu/physmem.c
113
+ if (unlikely(qemu_loglevel_mask(LOG_TB_OP_PLUGIN)
53
@@ -XXX,XX +XXX,XX @@ ram_addr_t qemu_ram_addr_from_host(void *ptr)
114
+ && qemu_log_in_addr_range(plugin_tb->vaddr))) {
54
return block->offset + offset;
115
+ FILE *logfile = qemu_log_trylock();
116
+ if (logfile) {
117
+ fprintf(logfile, "OP before plugin injection:\n");
118
+ tcg_dump_ops(tcg_ctx, logfile, false);
119
+ fprintf(logfile, "\n");
120
+ qemu_log_unlock(logfile);
121
+ }
122
+ }
123
124
/*
125
* While injecting code, we cannot afford to reuse any ebb temps
126
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
127
break;
128
}
129
}
130
- pr_ops();
55
}
131
}
56
132
57
+ram_addr_t qemu_ram_addr_from_host_nofail(void *ptr)
133
bool plugin_gen_tb_start(CPUState *cpu, const DisasContextBase *db,
58
+{
134
diff --git a/tcg/tcg.c b/tcg/tcg.c
59
+ ram_addr_t ram_addr;
135
index XXXXXXX..XXXXXXX 100644
136
--- a/tcg/tcg.c
137
+++ b/tcg/tcg.c
138
@@ -XXX,XX +XXX,XX @@ static const char bswap_flag_name[][6] = {
139
[TCG_BSWAP_IZ | TCG_BSWAP_OS] = "iz,os",
140
};
141
142
+#ifdef CONFIG_PLUGIN
143
+static const char * const plugin_from_name[] = {
144
+ "from-tb",
145
+ "from-insn",
146
+ "after-insn",
147
+ "after-tb",
148
+};
149
+#endif
60
+
150
+
61
+ ram_addr = qemu_ram_addr_from_host(ptr);
151
static inline bool tcg_regset_single(TCGRegSet d)
62
+ if (ram_addr == RAM_ADDR_INVALID) {
152
{
63
+ error_report("Bad ram pointer %p", ptr);
153
return (d & (d - 1)) == 0;
64
+ abort();
154
@@ -XXX,XX +XXX,XX @@ static inline TCGReg tcg_regset_first(TCGRegSet d)
65
+ }
155
#define ne_fprintf(...) \
66
+ return ram_addr;
156
({ int ret_ = fprintf(__VA_ARGS__); ret_ >= 0 ? ret_ : 0; })
67
+}
157
158
-static void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs)
159
+void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs)
160
{
161
char buf[128];
162
TCGOp *op;
163
@@ -XXX,XX +XXX,XX @@ static void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs)
164
i = k = 1;
165
}
166
break;
167
+#ifdef CONFIG_PLUGIN
168
+ case INDEX_op_plugin_cb:
169
+ {
170
+ TCGArg from = op->args[k++];
171
+ const char *name = NULL;
68
+
172
+
69
static MemTxResult flatview_read(FlatView *fv, hwaddr addr,
173
+ if (from < ARRAY_SIZE(plugin_from_name)) {
70
MemTxAttrs attrs, void *buf, hwaddr len);
174
+ name = plugin_from_name[from];
71
static MemTxResult flatview_write(FlatView *fv, hwaddr addr, MemTxAttrs attrs,
175
+ }
176
+ if (name) {
177
+ col += ne_fprintf(f, "%s", name);
178
+ } else {
179
+ col += ne_fprintf(f, "$0x%" TCG_PRIlx, from);
180
+ }
181
+ i = 1;
182
+ }
183
+ break;
184
+#endif
185
default:
186
i = 0;
187
break;
188
diff --git a/util/log.c b/util/log.c
189
index XXXXXXX..XXXXXXX 100644
190
--- a/util/log.c
191
+++ b/util/log.c
192
@@ -XXX,XX +XXX,XX @@ const QEMULogItem qemu_log_items[] = {
193
"show micro ops after optimization" },
194
{ CPU_LOG_TB_OP_IND, "op_ind",
195
"show micro ops before indirect lowering" },
196
+#ifdef CONFIG_PLUGIN
197
+ { LOG_TB_OP_PLUGIN, "op_plugin",
198
+ "show micro ops before plugin injection" },
199
+#endif
200
{ CPU_LOG_INT, "int",
201
"show interrupts/exceptions in short format" },
202
{ CPU_LOG_EXEC, "exec",
72
--
203
--
73
2.34.1
204
2.34.1
diff view generated by jsdifflib
1
From: Ilya Leoshkevich <iii@linux.ibm.com>
1
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
2
3
Right now translator stops right *after* the end of a page, which
4
breaks reporting of fault locations when the last instruction of a
5
multi-insn translation block crosses a page boundary.
6
7
Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-Id: <20220817150506.592862-3-iii@linux.ibm.com>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
3
---
12
target/s390x/tcg/translate.c | 15 +++-
4
accel/tcg/plugin-gen.c | 84 +++++++++++++++++++++---------------------
13
tests/tcg/s390x/noexec.c | 106 +++++++++++++++++++++++
5
1 file changed, 41 insertions(+), 43 deletions(-)
14
tests/tcg/multiarch/noexec.c.inc | 139 +++++++++++++++++++++++++++++++
15
tests/tcg/s390x/Makefile.target | 1 +
16
4 files changed, 257 insertions(+), 4 deletions(-)
17
create mode 100644 tests/tcg/s390x/noexec.c
18
create mode 100644 tests/tcg/multiarch/noexec.c.inc
19
6
20
diff --git a/target/s390x/tcg/translate.c b/target/s390x/tcg/translate.c
7
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
21
index XXXXXXX..XXXXXXX 100644
8
index XXXXXXX..XXXXXXX 100644
22
--- a/target/s390x/tcg/translate.c
9
--- a/accel/tcg/plugin-gen.c
23
+++ b/target/s390x/tcg/translate.c
10
+++ b/accel/tcg/plugin-gen.c
24
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
11
@@ -XXX,XX +XXX,XX @@ static void gen_mem_cb(struct qemu_plugin_dyn_cb *cb,
25
dc->insn_start = tcg_last_op();
12
tcg_temp_free_i32(cpu_index);
26
}
13
}
27
14
28
+static target_ulong get_next_pc(CPUS390XState *env, DisasContext *s,
15
+static void inject_cb(struct qemu_plugin_dyn_cb *cb)
29
+ uint64_t pc)
16
+
30
+{
17
+{
31
+ uint64_t insn = ld_code2(env, s, pc);
18
+ switch (cb->type) {
32
+
19
+ case PLUGIN_CB_REGULAR:
33
+ return pc + get_ilen((insn >> 8) & 0xff);
20
+ gen_udata_cb(cb);
34
+}
21
+ break;
35
+
22
+ case PLUGIN_CB_INLINE:
36
static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
23
+ gen_inline_cb(cb);
37
{
24
+ break;
38
CPUS390XState *env = cs->env_ptr;
25
+ default:
39
@@ -XXX,XX +XXX,XX @@ static void s390x_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
26
+ g_assert_not_reached();
40
41
dc->base.is_jmp = translate_one(env, dc);
42
if (dc->base.is_jmp == DISAS_NEXT) {
43
- uint64_t page_start;
44
-
45
- page_start = dc->base.pc_first & TARGET_PAGE_MASK;
46
- if (dc->base.pc_next - page_start >= TARGET_PAGE_SIZE || dc->ex_value) {
47
+ if (!is_same_page(dcbase, dc->base.pc_next) ||
48
+ !is_same_page(dcbase, get_next_pc(env, dc, dc->base.pc_next)) ||
49
+ dc->ex_value) {
50
dc->base.is_jmp = DISAS_TOO_MANY;
51
}
52
}
53
diff --git a/tests/tcg/s390x/noexec.c b/tests/tcg/s390x/noexec.c
54
new file mode 100644
55
index XXXXXXX..XXXXXXX
56
--- /dev/null
57
+++ b/tests/tcg/s390x/noexec.c
58
@@ -XXX,XX +XXX,XX @@
59
+#include "../multiarch/noexec.c.inc"
60
+
61
+static void *arch_mcontext_pc(const mcontext_t *ctx)
62
+{
63
+ return (void *)ctx->psw.addr;
64
+}
65
+
66
+static int arch_mcontext_arg(const mcontext_t *ctx)
67
+{
68
+ return ctx->gregs[2];
69
+}
70
+
71
+static void arch_flush(void *p, int len)
72
+{
73
+}
74
+
75
+extern char noexec_1[];
76
+extern char noexec_2[];
77
+extern char noexec_end[];
78
+
79
+asm("noexec_1:\n"
80
+ " lgfi %r2,1\n" /* %r2 is 0 on entry, set 1. */
81
+ "noexec_2:\n"
82
+ " lgfi %r2,2\n" /* %r2 is 0/1; set 2. */
83
+ " br %r14\n" /* return */
84
+ "noexec_end:");
85
+
86
+extern char exrl_1[];
87
+extern char exrl_2[];
88
+extern char exrl_end[];
89
+
90
+asm("exrl_1:\n"
91
+ " exrl %r0, exrl_2\n"
92
+ " br %r14\n"
93
+ "exrl_2:\n"
94
+ " lgfi %r2,2\n"
95
+ "exrl_end:");
96
+
97
+int main(void)
98
+{
99
+ struct noexec_test noexec_tests[] = {
100
+ {
101
+ .name = "fallthrough",
102
+ .test_code = noexec_1,
103
+ .test_len = noexec_end - noexec_1,
104
+ .page_ofs = noexec_1 - noexec_2,
105
+ .entry_ofs = noexec_1 - noexec_2,
106
+ .expected_si_ofs = 0,
107
+ .expected_pc_ofs = 0,
108
+ .expected_arg = 1,
109
+ },
110
+ {
111
+ .name = "jump",
112
+ .test_code = noexec_1,
113
+ .test_len = noexec_end - noexec_1,
114
+ .page_ofs = noexec_1 - noexec_2,
115
+ .entry_ofs = 0,
116
+ .expected_si_ofs = 0,
117
+ .expected_pc_ofs = 0,
118
+ .expected_arg = 0,
119
+ },
120
+ {
121
+ .name = "exrl",
122
+ .test_code = exrl_1,
123
+ .test_len = exrl_end - exrl_1,
124
+ .page_ofs = exrl_1 - exrl_2,
125
+ .entry_ofs = exrl_1 - exrl_2,
126
+ .expected_si_ofs = 0,
127
+ .expected_pc_ofs = exrl_1 - exrl_2,
128
+ .expected_arg = 0,
129
+ },
130
+ {
131
+ .name = "fallthrough [cross]",
132
+ .test_code = noexec_1,
133
+ .test_len = noexec_end - noexec_1,
134
+ .page_ofs = noexec_1 - noexec_2 - 2,
135
+ .entry_ofs = noexec_1 - noexec_2 - 2,
136
+ .expected_si_ofs = 0,
137
+ .expected_pc_ofs = -2,
138
+ .expected_arg = 1,
139
+ },
140
+ {
141
+ .name = "jump [cross]",
142
+ .test_code = noexec_1,
143
+ .test_len = noexec_end - noexec_1,
144
+ .page_ofs = noexec_1 - noexec_2 - 2,
145
+ .entry_ofs = -2,
146
+ .expected_si_ofs = 0,
147
+ .expected_pc_ofs = -2,
148
+ .expected_arg = 0,
149
+ },
150
+ {
151
+ .name = "exrl [cross]",
152
+ .test_code = exrl_1,
153
+ .test_len = exrl_end - exrl_1,
154
+ .page_ofs = exrl_1 - exrl_2 - 2,
155
+ .entry_ofs = exrl_1 - exrl_2 - 2,
156
+ .expected_si_ofs = 0,
157
+ .expected_pc_ofs = exrl_1 - exrl_2 - 2,
158
+ .expected_arg = 0,
159
+ },
160
+ };
161
+
162
+ return test_noexec(noexec_tests,
163
+ sizeof(noexec_tests) / sizeof(noexec_tests[0]));
164
+}
165
diff --git a/tests/tcg/multiarch/noexec.c.inc b/tests/tcg/multiarch/noexec.c.inc
166
new file mode 100644
167
index XXXXXXX..XXXXXXX
168
--- /dev/null
169
+++ b/tests/tcg/multiarch/noexec.c.inc
170
@@ -XXX,XX +XXX,XX @@
171
+/*
172
+ * Common code for arch-specific MMU_INST_FETCH fault testing.
173
+ */
174
+
175
+#define _GNU_SOURCE
176
+
177
+#include <assert.h>
178
+#include <signal.h>
179
+#include <stdio.h>
180
+#include <stdlib.h>
181
+#include <string.h>
182
+#include <errno.h>
183
+#include <unistd.h>
184
+#include <sys/mman.h>
185
+#include <sys/ucontext.h>
186
+
187
+/* Forward declarations. */
188
+
189
+static void *arch_mcontext_pc(const mcontext_t *ctx);
190
+static int arch_mcontext_arg(const mcontext_t *ctx);
191
+static void arch_flush(void *p, int len);
192
+
193
+/* Testing infrastructure. */
194
+
195
+struct noexec_test {
196
+ const char *name;
197
+ const char *test_code;
198
+ int test_len;
199
+ int page_ofs;
200
+ int entry_ofs;
201
+ int expected_si_ofs;
202
+ int expected_pc_ofs;
203
+ int expected_arg;
204
+};
205
+
206
+static void *page_base;
207
+static int page_size;
208
+static const struct noexec_test *current_noexec_test;
209
+
210
+static void handle_err(const char *syscall)
211
+{
212
+ printf("[ FAILED ] %s: %s\n", syscall, strerror(errno));
213
+ exit(EXIT_FAILURE);
214
+}
215
+
216
+static void handle_segv(int sig, siginfo_t *info, void *ucontext)
217
+{
218
+ const struct noexec_test *test = current_noexec_test;
219
+ const mcontext_t *mc = &((ucontext_t *)ucontext)->uc_mcontext;
220
+ void *expected_si;
221
+ void *expected_pc;
222
+ void *pc;
223
+ int arg;
224
+
225
+ if (test == NULL) {
226
+ printf("[ FAILED ] unexpected SEGV\n");
227
+ exit(EXIT_FAILURE);
228
+ }
229
+ current_noexec_test = NULL;
230
+
231
+ expected_si = page_base + test->expected_si_ofs;
232
+ if (info->si_addr != expected_si) {
233
+ printf("[ FAILED ] wrong si_addr (%p != %p)\n",
234
+ info->si_addr, expected_si);
235
+ exit(EXIT_FAILURE);
236
+ }
237
+
238
+ pc = arch_mcontext_pc(mc);
239
+ expected_pc = page_base + test->expected_pc_ofs;
240
+ if (pc != expected_pc) {
241
+ printf("[ FAILED ] wrong pc (%p != %p)\n", pc, expected_pc);
242
+ exit(EXIT_FAILURE);
243
+ }
244
+
245
+ arg = arch_mcontext_arg(mc);
246
+ if (arg != test->expected_arg) {
247
+ printf("[ FAILED ] wrong arg (%d != %d)\n", arg, test->expected_arg);
248
+ exit(EXIT_FAILURE);
249
+ }
250
+
251
+ if (mprotect(page_base, page_size,
252
+ PROT_READ | PROT_WRITE | PROT_EXEC) < 0) {
253
+ handle_err("mprotect");
254
+ }
27
+ }
255
+}
28
+}
256
+
29
+
257
+static void test_noexec_1(const struct noexec_test *test)
30
+static void inject_mem_cb(struct qemu_plugin_dyn_cb *cb,
31
+ enum qemu_plugin_mem_rw rw,
32
+ qemu_plugin_meminfo_t meminfo, TCGv_i64 addr)
258
+{
33
+{
259
+ void *start = page_base + test->page_ofs;
34
+ if (cb->rw & rw) {
260
+ void (*fn)(int arg) = page_base + test->entry_ofs;
35
+ switch (cb->type) {
261
+
36
+ case PLUGIN_CB_MEM_REGULAR:
262
+ memcpy(start, test->test_code, test->test_len);
37
+ gen_mem_cb(cb, meminfo, addr);
263
+ arch_flush(start, test->test_len);
38
+ break;
264
+
39
+ default:
265
+ /* Trigger TB creation in order to test invalidation. */
40
+ inject_cb(cb);
266
+ fn(0);
41
+ break;
267
+
42
+ }
268
+ if (mprotect(page_base, page_size, PROT_NONE) < 0) {
269
+ handle_err("mprotect");
270
+ }
43
+ }
271
+
272
+ /* Trigger SEGV and check that handle_segv() ran. */
273
+ current_noexec_test = test;
274
+ fn(0);
275
+ assert(current_noexec_test == NULL);
276
+}
44
+}
277
+
45
+
278
+static int test_noexec(struct noexec_test *tests, size_t n_tests)
46
static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
279
+{
47
{
280
+ struct sigaction act;
48
TCGOp *op, *next;
281
+ size_t i;
49
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
282
+
50
283
+ memset(&act, 0, sizeof(act));
51
cbs = plugin_tb->cbs;
284
+ act.sa_sigaction = handle_segv;
52
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
285
+ act.sa_flags = SA_SIGINFO;
53
- struct qemu_plugin_dyn_cb *cb =
286
+ if (sigaction(SIGSEGV, &act, NULL) < 0) {
54
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
287
+ handle_err("sigaction");
55
-
288
+ }
56
- switch (cb->type) {
289
+
57
- case PLUGIN_CB_REGULAR:
290
+ page_size = getpagesize();
58
- gen_udata_cb(cb);
291
+ page_base = mmap(NULL, 2 * page_size,
59
- break;
292
+ PROT_READ | PROT_WRITE | PROT_EXEC,
60
- case PLUGIN_CB_INLINE:
293
+ MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
61
- gen_inline_cb(cb);
294
+ if (page_base == MAP_FAILED) {
62
- break;
295
+ handle_err("mmap");
63
- default:
296
+ }
64
- g_assert_not_reached();
297
+ page_base += page_size;
65
- }
298
+
66
+ inject_cb(
299
+ for (i = 0; i < n_tests; i++) {
67
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i));
300
+ struct noexec_test *test = &tests[i];
68
}
301
+
69
break;
302
+ printf("[ RUN ] %s\n", test->name);
70
303
+ test_noexec_1(test);
71
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
304
+ printf("[ OK ]\n");
72
305
+ }
73
cbs = insn->insn_cbs;
306
+
74
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
307
+ printf("[ PASSED ]\n");
75
- struct qemu_plugin_dyn_cb *cb =
308
+ return EXIT_SUCCESS;
76
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
309
+}
77
-
310
diff --git a/tests/tcg/s390x/Makefile.target b/tests/tcg/s390x/Makefile.target
78
- switch (cb->type) {
311
index XXXXXXX..XXXXXXX 100644
79
- case PLUGIN_CB_REGULAR:
312
--- a/tests/tcg/s390x/Makefile.target
80
- gen_udata_cb(cb);
313
+++ b/tests/tcg/s390x/Makefile.target
81
- break;
314
@@ -XXX,XX +XXX,XX @@ TESTS+=shift
82
- case PLUGIN_CB_INLINE:
315
TESTS+=trap
83
- gen_inline_cb(cb);
316
TESTS+=signals-s390x
84
- break;
317
TESTS+=branch-relative-long
85
- default:
318
+TESTS+=noexec
86
- g_assert_not_reached();
319
87
- }
320
Z14_TESTS=vfminmax
88
+ inject_cb(
321
vfminmax: LDFLAGS+=-lm
89
+ &g_array_index(cbs, struct qemu_plugin_dyn_cb, i));
90
}
91
break;
92
93
@@ -XXX,XX +XXX,XX @@ static void plugin_gen_inject(struct qemu_plugin_tb *plugin_tb)
94
{
95
TCGv_i64 addr = temp_tcgv_i64(arg_temp(op->args[0]));
96
qemu_plugin_meminfo_t meminfo = op->args[1];
97
+ enum qemu_plugin_mem_rw rw =
98
+ (qemu_plugin_mem_is_store(meminfo)
99
+ ? QEMU_PLUGIN_MEM_W : QEMU_PLUGIN_MEM_R);
100
struct qemu_plugin_insn *insn;
101
const GArray *cbs;
102
- int i, n, rw;
103
+ int i, n;
104
105
assert(insn_idx >= 0);
106
insn = g_ptr_array_index(plugin_tb->insns, insn_idx);
107
- rw = qemu_plugin_mem_is_store(meminfo) ? 2 : 1;
108
109
tcg_ctx->emit_before_op = op;
110
111
cbs = insn->mem_cbs;
112
for (i = 0, n = (cbs ? cbs->len : 0); i < n; i++) {
113
- struct qemu_plugin_dyn_cb *cb =
114
- &g_array_index(cbs, struct qemu_plugin_dyn_cb, i);
115
-
116
- if (cb->rw & rw) {
117
- switch (cb->type) {
118
- case PLUGIN_CB_MEM_REGULAR:
119
- gen_mem_cb(cb, meminfo, addr);
120
- break;
121
- case PLUGIN_CB_INLINE:
122
- gen_inline_cb(cb);
123
- break;
124
- default:
125
- g_assert_not_reached();
126
- }
127
- }
128
+ inject_mem_cb(&g_array_index(cbs, struct qemu_plugin_dyn_cb, i),
129
+ rw, meminfo, addr);
130
}
131
132
tcg_ctx->emit_before_op = NULL;
322
--
133
--
323
2.34.1
134
2.34.1
diff view generated by jsdifflib
1
Simplify the implementation of get_page_addr_code_hostp
1
Merge qemu_plugin_insn_alloc and qemu_plugin_tb_insn_get into
2
by reusing the existing probe_access infrastructure.
2
plugin_gen_insn_start, since it is used nowhere else.
3
3
4
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
4
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
5
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
6
---
8
accel/tcg/cputlb.c | 76 ++++++++++++++++------------------------------
7
include/qemu/plugin.h | 39 ---------------------------------------
9
1 file changed, 26 insertions(+), 50 deletions(-)
8
accel/tcg/plugin-gen.c | 39 ++++++++++++++++++++++++++++++++-------
9
2 files changed, 32 insertions(+), 46 deletions(-)
10
10
11
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
11
diff --git a/include/qemu/plugin.h b/include/qemu/plugin.h
12
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
13
--- a/accel/tcg/cputlb.c
13
--- a/include/qemu/plugin.h
14
+++ b/accel/tcg/cputlb.c
14
+++ b/include/qemu/plugin.h
15
@@ -XXX,XX +XXX,XX @@ static bool victim_tlb_hit(CPUArchState *env, size_t mmu_idx, size_t index,
15
@@ -XXX,XX +XXX,XX @@ static inline void qemu_plugin_insn_cleanup_fn(gpointer data)
16
victim_tlb_hit(env, mmu_idx, index, offsetof(CPUTLBEntry, TY), \
16
g_byte_array_free(insn->data, true);
17
(ADDR) & TARGET_PAGE_MASK)
17
}
18
18
19
-/*
19
-static inline struct qemu_plugin_insn *qemu_plugin_insn_alloc(void)
20
- * Return a ram_addr_t for the virtual address for execution.
20
-{
21
- *
21
- struct qemu_plugin_insn *insn = g_new0(struct qemu_plugin_insn, 1);
22
- * Return -1 if we can't translate and execute from an entire page
22
-
23
- * of RAM. This will force us to execute by loading and translating
23
- insn->data = g_byte_array_sized_new(4);
24
- * one insn at a time, without caching.
24
- return insn;
25
- *
25
-}
26
- * NOTE: This function will trigger an exception if the page is
26
-
27
- * not executable.
27
/* Internal context for this TranslationBlock */
28
struct qemu_plugin_tb {
29
GPtrArray *insns;
30
@@ -XXX,XX +XXX,XX @@ struct qemu_plugin_tb {
31
GArray *cbs;
32
};
33
34
-/**
35
- * qemu_plugin_tb_insn_get(): get next plugin record for translation.
36
- * @tb: the internal tb context
37
- * @pc: address of instruction
28
- */
38
- */
29
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
39
-static inline
30
- void **hostp)
40
-struct qemu_plugin_insn *qemu_plugin_tb_insn_get(struct qemu_plugin_tb *tb,
41
- uint64_t pc)
31
-{
42
-{
32
- uintptr_t mmu_idx = cpu_mmu_index(env, true);
43
- struct qemu_plugin_insn *insn;
33
- uintptr_t index = tlb_index(env, mmu_idx, addr);
34
- CPUTLBEntry *entry = tlb_entry(env, mmu_idx, addr);
35
- void *p;
36
-
44
-
37
- if (unlikely(!tlb_hit(entry->addr_code, addr))) {
45
- if (unlikely(tb->n == tb->insns->len)) {
38
- if (!VICTIM_TLB_HIT(addr_code, addr)) {
46
- struct qemu_plugin_insn *new_insn = qemu_plugin_insn_alloc();
39
- tlb_fill(env_cpu(env), addr, 0, MMU_INST_FETCH, mmu_idx, 0);
47
- g_ptr_array_add(tb->insns, new_insn);
40
- index = tlb_index(env, mmu_idx, addr);
41
- entry = tlb_entry(env, mmu_idx, addr);
42
-
43
- if (unlikely(entry->addr_code & TLB_INVALID_MASK)) {
44
- /*
45
- * The MMU protection covers a smaller range than a target
46
- * page, so we must redo the MMU check for every insn.
47
- */
48
- return -1;
49
- }
50
- }
51
- assert(tlb_hit(entry->addr_code, addr));
52
- }
48
- }
53
-
49
-
54
- if (unlikely(entry->addr_code & TLB_MMIO)) {
50
- insn = g_ptr_array_index(tb->insns, tb->n++);
55
- /* The region is not backed by RAM. */
51
- g_byte_array_set_size(insn->data, 0);
56
- if (hostp) {
52
- insn->calls_helpers = false;
57
- *hostp = NULL;
53
- insn->mem_helper = false;
58
- }
54
- insn->vaddr = pc;
59
- return -1;
55
- if (insn->insn_cbs) {
56
- g_array_set_size(insn->insn_cbs, 0);
57
- }
58
- if (insn->mem_cbs) {
59
- g_array_set_size(insn->mem_cbs, 0);
60
- }
60
- }
61
-
61
-
62
- p = (void *)((uintptr_t)addr + entry->addend);
62
- return insn;
63
- if (hostp) {
64
- *hostp = p;
65
- }
66
- return qemu_ram_addr_from_host_nofail(p);
67
-}
63
-}
68
-
64
-
69
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
65
/**
70
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
66
* struct CPUPluginState - per-CPU state for plugins
67
* @event_mask: plugin event bitmap. Modified only via async work.
68
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
69
index XXXXXXX..XXXXXXX 100644
70
--- a/accel/tcg/plugin-gen.c
71
+++ b/accel/tcg/plugin-gen.c
72
@@ -XXX,XX +XXX,XX @@ bool plugin_gen_tb_start(CPUState *cpu, const DisasContextBase *db,
73
void plugin_gen_insn_start(CPUState *cpu, const DisasContextBase *db)
71
{
74
{
72
@@ -XXX,XX +XXX,XX @@ void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
75
struct qemu_plugin_tb *ptb = tcg_ctx->plugin_tb;
73
return flags ? NULL : host;
76
- struct qemu_plugin_insn *pinsn;
77
+ struct qemu_plugin_insn *insn;
78
+ size_t n = db->num_insns;
79
+ vaddr pc;
80
81
- pinsn = qemu_plugin_tb_insn_get(ptb, db->pc_next);
82
- tcg_ctx->plugin_insn = pinsn;
83
- plugin_gen_empty_callback(PLUGIN_GEN_FROM_INSN);
84
+ assert(n >= 1);
85
+ ptb->n = n;
86
+ if (n <= ptb->insns->len) {
87
+ insn = g_ptr_array_index(ptb->insns, n - 1);
88
+ g_byte_array_set_size(insn->data, 0);
89
+ } else {
90
+ assert(n - 1 == ptb->insns->len);
91
+ insn = g_new0(struct qemu_plugin_insn, 1);
92
+ insn->data = g_byte_array_sized_new(4);
93
+ g_ptr_array_add(ptb->insns, insn);
94
+ }
95
+
96
+ tcg_ctx->plugin_insn = insn;
97
+ insn->calls_helpers = false;
98
+ insn->mem_helper = false;
99
+ if (insn->insn_cbs) {
100
+ g_array_set_size(insn->insn_cbs, 0);
101
+ }
102
+ if (insn->mem_cbs) {
103
+ g_array_set_size(insn->mem_cbs, 0);
104
+ }
105
+
106
+ pc = db->pc_next;
107
+ insn->vaddr = pc;
108
109
/*
110
* Detect page crossing to get the new host address.
111
@@ -XXX,XX +XXX,XX @@ void plugin_gen_insn_start(CPUState *cpu, const DisasContextBase *db)
112
* fetching instructions from a region not backed by RAM.
113
*/
114
if (ptb->haddr1 == NULL) {
115
- pinsn->haddr = NULL;
116
+ insn->haddr = NULL;
117
} else if (is_same_page(db, db->pc_next)) {
118
- pinsn->haddr = ptb->haddr1 + pinsn->vaddr - ptb->vaddr;
119
+ insn->haddr = ptb->haddr1 + pc - ptb->vaddr;
120
} else {
121
if (ptb->vaddr2 == -1) {
122
ptb->vaddr2 = TARGET_PAGE_ALIGN(db->pc_first);
123
get_page_addr_code_hostp(cpu_env(cpu), ptb->vaddr2, &ptb->haddr2);
124
}
125
- pinsn->haddr = ptb->haddr2 + pinsn->vaddr - ptb->vaddr2;
126
+ insn->haddr = ptb->haddr2 + pc - ptb->vaddr2;
127
}
128
+
129
+ plugin_gen_empty_callback(PLUGIN_GEN_FROM_INSN);
74
}
130
}
75
131
76
+/*
132
void plugin_gen_insn_end(void)
77
+ * Return a ram_addr_t for the virtual address for execution.
78
+ *
79
+ * Return -1 if we can't translate and execute from an entire page
80
+ * of RAM. This will force us to execute by loading and translating
81
+ * one insn at a time, without caching.
82
+ *
83
+ * NOTE: This function will trigger an exception if the page is
84
+ * not executable.
85
+ */
86
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
87
+ void **hostp)
88
+{
89
+ void *p;
90
+
91
+ (void)probe_access_internal(env, addr, 1, MMU_INST_FETCH,
92
+ cpu_mmu_index(env, true), false, &p, 0);
93
+ if (p == NULL) {
94
+ return -1;
95
+ }
96
+ if (hostp) {
97
+ *hostp = p;
98
+ }
99
+ return qemu_ram_addr_from_host_nofail(p);
100
+}
101
+
102
#ifdef CONFIG_PLUGIN
103
/*
104
* Perform a TLB lookup and populate the qemu_plugin_hwaddr structure.
105
--
133
--
106
2.34.1
134
2.34.1
diff view generated by jsdifflib
1
The current implementation is a no-op, simply returning addr.
1
Each caller can use tcg_gen_plugin_cb directly.
2
This is incorrect, because we ought to be checking the page
3
permissions for execution.
4
2
5
Make get_page_addr_code inline for both implementations.
3
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
6
7
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
8
Tested-by: Ilya Leoshkevich <iii@linux.ibm.com>
9
Acked-by: Alistair Francis <alistair.francis@wdc.com>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
---
5
---
12
include/exec/exec-all.h | 85 ++++++++++++++---------------------------
6
accel/tcg/plugin-gen.c | 19 +++----------------
13
accel/tcg/cputlb.c | 5 ---
7
1 file changed, 3 insertions(+), 16 deletions(-)
14
accel/tcg/user-exec.c | 14 +++++++
15
3 files changed, 42 insertions(+), 62 deletions(-)
16
8
17
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
9
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
18
index XXXXXXX..XXXXXXX 100644
10
index XXXXXXX..XXXXXXX 100644
19
--- a/include/exec/exec-all.h
11
--- a/accel/tcg/plugin-gen.c
20
+++ b/include/exec/exec-all.h
12
+++ b/accel/tcg/plugin-gen.c
21
@@ -XXX,XX +XXX,XX @@ struct MemoryRegionSection *iotlb_to_section(CPUState *cpu,
13
@@ -XXX,XX +XXX,XX @@ enum plugin_gen_from {
22
hwaddr index, MemTxAttrs attrs);
14
PLUGIN_GEN_AFTER_TB,
23
#endif
15
};
24
16
25
-#if defined(CONFIG_USER_ONLY)
17
-static void plugin_gen_empty_callback(enum plugin_gen_from from)
26
-void mmap_lock(void);
27
-void mmap_unlock(void);
28
-bool have_mmap_lock(void);
29
-
30
/**
31
- * get_page_addr_code() - user-mode version
32
+ * get_page_addr_code_hostp()
33
* @env: CPUArchState
34
* @addr: guest virtual address of guest code
35
*
36
- * Returns @addr.
37
+ * See get_page_addr_code() (full-system version) for documentation on the
38
+ * return value.
39
+ *
40
+ * Sets *@hostp (when @hostp is non-NULL) as follows.
41
+ * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
42
+ * to the host address where @addr's content is kept.
43
+ *
44
+ * Note: this function can trigger an exception.
45
+ */
46
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
47
+ void **hostp);
48
+
49
+/**
50
+ * get_page_addr_code()
51
+ * @env: CPUArchState
52
+ * @addr: guest virtual address of guest code
53
+ *
54
+ * If we cannot translate and execute from the entire RAM page, or if
55
+ * the region is not backed by RAM, returns -1. Otherwise, returns the
56
+ * ram_addr_t corresponding to the guest code at @addr.
57
+ *
58
+ * Note: this function can trigger an exception.
59
*/
60
static inline tb_page_addr_t get_page_addr_code(CPUArchState *env,
61
target_ulong addr)
62
{
63
- return addr;
64
+ return get_page_addr_code_hostp(env, addr, NULL);
65
}
66
67
-/**
68
- * get_page_addr_code_hostp() - user-mode version
69
- * @env: CPUArchState
70
- * @addr: guest virtual address of guest code
71
- *
72
- * Returns @addr.
73
- *
74
- * If @hostp is non-NULL, sets *@hostp to the host address where @addr's content
75
- * is kept.
76
- */
77
-static inline tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env,
78
- target_ulong addr,
79
- void **hostp)
80
-{
18
-{
81
- if (hostp) {
19
- switch (from) {
82
- *hostp = g2h_untagged(addr);
20
- case PLUGIN_GEN_AFTER_INSN:
21
- case PLUGIN_GEN_FROM_TB:
22
- case PLUGIN_GEN_FROM_INSN:
23
- tcg_gen_plugin_cb(from);
24
- break;
25
- default:
26
- g_assert_not_reached();
83
- }
27
- }
84
- return addr;
85
-}
86
+#if defined(CONFIG_USER_ONLY)
87
+void mmap_lock(void);
88
+void mmap_unlock(void);
89
+bool have_mmap_lock(void);
90
91
/**
92
* adjust_signal_pc:
93
@@ -XXX,XX +XXX,XX @@ G_NORETURN void cpu_loop_exit_sigbus(CPUState *cpu, target_ulong addr,
94
static inline void mmap_lock(void) {}
95
static inline void mmap_unlock(void) {}
96
97
-/**
98
- * get_page_addr_code() - full-system version
99
- * @env: CPUArchState
100
- * @addr: guest virtual address of guest code
101
- *
102
- * If we cannot translate and execute from the entire RAM page, or if
103
- * the region is not backed by RAM, returns -1. Otherwise, returns the
104
- * ram_addr_t corresponding to the guest code at @addr.
105
- *
106
- * Note: this function can trigger an exception.
107
- */
108
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr);
109
-
110
-/**
111
- * get_page_addr_code_hostp() - full-system version
112
- * @env: CPUArchState
113
- * @addr: guest virtual address of guest code
114
- *
115
- * See get_page_addr_code() (full-system version) for documentation on the
116
- * return value.
117
- *
118
- * Sets *@hostp (when @hostp is non-NULL) as follows.
119
- * If the return value is -1, sets *@hostp to NULL. Otherwise, sets *@hostp
120
- * to the host address where @addr's content is kept.
121
- *
122
- * Note: this function can trigger an exception.
123
- */
124
-tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
125
- void **hostp);
126
-
127
void tlb_reset_dirty(CPUState *cpu, ram_addr_t start1, ram_addr_t length);
128
void tlb_set_dirty(CPUState *cpu, target_ulong vaddr);
129
130
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
131
index XXXXXXX..XXXXXXX 100644
132
--- a/accel/tcg/cputlb.c
133
+++ b/accel/tcg/cputlb.c
134
@@ -XXX,XX +XXX,XX @@ tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
135
return qemu_ram_addr_from_host_nofail(p);
136
}
137
138
-tb_page_addr_t get_page_addr_code(CPUArchState *env, target_ulong addr)
139
-{
140
- return get_page_addr_code_hostp(env, addr, NULL);
141
-}
28
-}
142
-
29
-
143
static void notdirty_write(CPUState *cpu, vaddr mem_vaddr, unsigned size,
30
/* called before finishing a TB with exit_tb, goto_tb or goto_ptr */
144
CPUIOTLBEntry *iotlbentry, uintptr_t retaddr)
31
void plugin_gen_disable_mem_helpers(void)
145
{
32
{
146
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
33
@@ -XXX,XX +XXX,XX @@ bool plugin_gen_tb_start(CPUState *cpu, const DisasContextBase *db,
147
index XXXXXXX..XXXXXXX 100644
34
ptb->mem_only = mem_only;
148
--- a/accel/tcg/user-exec.c
35
ptb->mem_helper = false;
149
+++ b/accel/tcg/user-exec.c
36
150
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
37
- plugin_gen_empty_callback(PLUGIN_GEN_FROM_TB);
151
return size ? g2h(env_cpu(env), addr) : NULL;
38
+ tcg_gen_plugin_cb(PLUGIN_GEN_FROM_TB);
39
}
40
41
tcg_ctx->plugin_insn = NULL;
42
@@ -XXX,XX +XXX,XX @@ void plugin_gen_insn_start(CPUState *cpu, const DisasContextBase *db)
43
insn->haddr = ptb->haddr2 + pc - ptb->vaddr2;
44
}
45
46
- plugin_gen_empty_callback(PLUGIN_GEN_FROM_INSN);
47
+ tcg_gen_plugin_cb(PLUGIN_GEN_FROM_INSN);
152
}
48
}
153
49
154
+tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env, target_ulong addr,
50
void plugin_gen_insn_end(void)
155
+ void **hostp)
51
{
156
+{
52
- plugin_gen_empty_callback(PLUGIN_GEN_AFTER_INSN);
157
+ int flags;
53
+ tcg_gen_plugin_cb(PLUGIN_GEN_AFTER_INSN);
158
+
54
}
159
+ flags = probe_access_internal(env, addr, 1, MMU_INST_FETCH, false, 0);
160
+ g_assert(flags == 0);
161
+
162
+ if (hostp) {
163
+ *hostp = g2h_untagged(addr);
164
+ }
165
+ return addr;
166
+}
167
+
168
/* The softmmu versions of these helpers are in cputlb.c. */
169
55
170
/*
56
/*
171
--
57
--
172
2.34.1
58
2.34.1
59
60
diff view generated by jsdifflib
1
It was non-obvious to me why we can raise an exception in
1
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
2
the middle of a comparison function, but it works.
3
While nearby, use TARGET_PAGE_ALIGN instead of open-coding.
4
5
Acked-by: Ilya Leoshkevich <iii@linux.ibm.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
---
3
---
8
accel/tcg/cpu-exec.c | 11 ++++++++++-
4
accel/tcg/plugin-gen.c | 31 ++++---------------------------
9
1 file changed, 10 insertions(+), 1 deletion(-)
5
1 file changed, 4 insertions(+), 27 deletions(-)
10
6
11
diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c
7
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
12
index XXXXXXX..XXXXXXX 100644
8
index XXXXXXX..XXXXXXX 100644
13
--- a/accel/tcg/cpu-exec.c
9
--- a/accel/tcg/plugin-gen.c
14
+++ b/accel/tcg/cpu-exec.c
10
+++ b/accel/tcg/plugin-gen.c
15
@@ -XXX,XX +XXX,XX @@ static bool tb_lookup_cmp(const void *p, const void *d)
11
@@ -XXX,XX +XXX,XX @@
16
tb_page_addr_t phys_page2;
12
* Injecting the desired instrumentation could be done with a second
17
target_ulong virt_page2;
13
* translation pass that combined the instrumentation requests, but that
18
14
* would be ugly and inefficient since we would decode the guest code twice.
19
- virt_page2 = (desc->pc & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
15
- * Instead, during TB translation we add "empty" instrumentation calls for all
20
+ /*
16
- * possible instrumentation events, and then once we collect the instrumentation
21
+ * We know that the first page matched, and an otherwise valid TB
17
- * requests from plugins, we either "fill in" those empty events or remove them
22
+ * encountered an incomplete instruction at the end of that page,
18
- * if they have no requests.
23
+ * therefore we know that generating a new TB from the current PC
19
- *
24
+ * must also require reading from the next page -- even if the
20
- * When "filling in" an event we first copy the empty callback's TCG ops. This
25
+ * second pages do not match, and therefore the resulting insn
21
- * might seem unnecessary, but it is done to support an arbitrary number
26
+ * is different for the new TB. Therefore any exception raised
22
- * of callbacks per event. Take for example a regular instruction callback.
27
+ * here by the faulting lookup is not premature.
23
- * We first generate a callback to an empty helper function. Then, if two
28
+ */
24
- * plugins register one callback each for this instruction, we make two copies
29
+ virt_page2 = TARGET_PAGE_ALIGN(desc->pc);
25
- * of the TCG ops generated for the empty callback, substituting the function
30
phys_page2 = get_page_addr_code(desc->env, virt_page2);
26
- * pointer that points to the empty helper function with the plugins' desired
31
if (tb->page_addr[1] == phys_page2) {
27
- * callback functions. After that we remove the empty callback's ops.
32
return true;
28
- *
29
- * Note that the location in TCGOp.args[] of the pointer to a helper function
30
- * varies across different guest and host architectures. Instead of duplicating
31
- * the logic that figures this out, we rely on the fact that the empty
32
- * callbacks point to empty functions that are unique pointers in the program.
33
- * Thus, to find the right location we just have to look for a match in
34
- * TCGOp.args[]. This is the main reason why we first copy an empty callback's
35
- * TCG ops and then fill them in; regardless of whether we have one or many
36
- * callbacks for that event, the logic to add all of them is the same.
37
- *
38
- * When generating more than one callback per event, we make a small
39
- * optimization to avoid generating redundant operations. For instance, for the
40
- * second and all subsequent callbacks of an event, we do not need to reload the
41
- * CPU's index into a TCG temp, since the first callback did it already.
42
+ * Instead, during TB translation we add "plugin_cb" marker opcodes
43
+ * for all possible instrumentation events, and then once we collect the
44
+ * instrumentation requests from plugins, we generate code for those markers
45
+ * or remove them if they have no requests.
46
*/
47
#include "qemu/osdep.h"
48
#include "qemu/plugin.h"
33
--
49
--
34
2.34.1
50
2.34.1
diff view generated by jsdifflib