1
Second ARM pull request of this week; this one has my next
1
Small pile of bug fixes for rc1. I've included my patches to get
2
set of v8M patches and a handful of more minor stuff from
2
our docs building with Sphinx 3, just for convenience...
3
other people.
4
3
5
thanks
6
-- PMM
4
-- PMM
7
5
8
The following changes since commit 8ee5f9b3ecc94e3eb7a8235f4b2c3ec9024807f6:
6
The following changes since commit b149dea55cce97cb226683d06af61984a1c11e96:
9
7
10
Merge remote-tracking branch 'remotes/kevin/tags/for-upstream' into staging (2017-09-07 10:45:18 +0100)
8
Merge remote-tracking branch 'remotes/cschoenebeck/tags/pull-9p-20201102' into staging (2020-11-02 10:57:48 +0000)
11
9
12
are available in the git repository at:
10
are available in the Git repository at:
13
11
14
git://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20170907
12
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20201102
15
13
16
for you to fetch changes up to c99a55d38dd5b5131f3fcbbaf41828a09ee62544:
14
for you to fetch changes up to ffb4fbf90a2f63c9cb33e4bb9f854c79bf04ca4a:
17
15
18
target/arm: Add Jazelle feature (2017-09-07 13:54:55 +0100)
16
tests/qtest/npcm7xx_rng-test: Disable randomness tests (2020-11-02 16:52:18 +0000)
19
17
20
----------------------------------------------------------------
18
----------------------------------------------------------------
21
target-arm:
19
target-arm queue:
22
* cleanups converting to DEFINE_PROP_LINK
20
* target/arm: Fix Neon emulation bugs on big-endian hosts
23
* allwinner-a10: mark as not user-creatable
21
* target/arm: fix handling of HCR.FB
24
* initial patches working towards ARMv8M support
22
* target/arm: fix LORID_EL1 access check
25
* implement generating aborts on memory transaction failures
23
* disas/capstone: Fix monitor disassembly of >32 bytes
26
* make BXJ behave correctly (ie not UNDEF) on ARMv6-and-later
24
* hw/arm/smmuv3: Fix potential integer overflow (CID 1432363)
25
* hw/arm/boot: fix SVE for EL3 direct kernel boot
26
* hw/display/omap_lcdc: Fix potential NULL pointer dereference
27
* hw/display/exynos4210_fimd: Fix potential NULL pointer dereference
28
* target/arm: Get correct MMU index for other-security-state
29
* configure: Test that gio libs from pkg-config work
30
* hw/intc/arm_gicv3_cpuif: Make GIC maintenance interrupts work
31
* docs: Fix building with Sphinx 3
32
* tests/qtest/npcm7xx_rng-test: Disable randomness tests
27
33
28
----------------------------------------------------------------
34
----------------------------------------------------------------
29
Fam Zheng (6):
35
AlexChen (2):
30
armv7m: Convert bitband.source-memory to DEFINE_PROP_LINK
36
hw/display/omap_lcdc: Fix potential NULL pointer dereference
31
armv7m: Convert armv7m.memory to DEFINE_PROP_LINK
37
hw/display/exynos4210_fimd: Fix potential NULL pointer dereference
32
gicv3: Convert to DEFINE_PROP_LINK
33
xlnx_zynqmp: Convert to DEFINE_PROP_LINK
34
xilinx_axienet: Convert to DEFINE_PROP_LINK
35
xilinx_axidma: Convert to DEFINE_PROP_LINK
36
38
37
Peter Maydell (23):
39
Peter Maydell (9):
38
target/arm: Implement ARMv8M's PMSAv8 registers
40
target/arm: Fix float16 pairwise Neon ops on big-endian hosts
39
target/arm: Implement new PMSAv8 behaviour
41
target/arm: Fix VUDOT/VSDOT (scalar) on big-endian hosts
40
target/arm: Add state field, feature bit and migration for v8M secure state
42
disas/capstone: Fix monitor disassembly of >32 bytes
41
target/arm: Register second AddressSpace for secure v8M CPUs
43
target/arm: Get correct MMU index for other-security-state
42
target/arm: Add MMU indexes for secure v8M
44
configure: Test that gio libs from pkg-config work
43
target/arm: Make BASEPRI register banked for v8M
45
hw/intc/arm_gicv3_cpuif: Make GIC maintenance interrupts work
44
target/arm: Make PRIMASK register banked for v8M
46
scripts/kerneldoc: For Sphinx 3 use c:macro for macros with arguments
45
target/arm: Make FAULTMASK register banked for v8M
47
qemu-option-trace.rst.inc: Don't use option:: markup
46
target/arm: Make CONTROL register banked for v8M
48
tests/qtest/npcm7xx_rng-test: Disable randomness tests
47
nvic: Add NS alias SCS region
48
target/arm: Make VTOR register banked for v8M
49
target/arm: Make MPU_MAIR0, MPU_MAIR1 registers banked for v8M
50
target/arm: Make MPU_RBAR, MPU_RLAR banked for v8M
51
target/arm: Make MPU_RNR register banked for v8M
52
target/arm: Make MPU_CTRL register banked for v8M
53
target/arm: Make CCR register banked for v8M
54
target/arm: Make MMFAR banked for v8M
55
target/arm: Make CFSR register banked for v8M
56
target/arm: Move regime_is_secure() to target/arm/internals.h
57
target/arm: Implement BXNS, and banked stack pointers
58
boards.h: Define new flag ignore_memory_transaction_failures
59
hw/arm: Set ignore_memory_transaction_failures for most ARM boards
60
target/arm: Implement new do_transaction_failed hook
61
49
62
Portia Stephens (1):
50
Philippe Mathieu-Daudé (1):
63
target/arm: Add Jazelle feature
51
hw/arm/smmuv3: Fix potential integer overflow (CID 1432363)
64
52
65
Thomas Huth (1):
53
Richard Henderson (11):
66
hw/arm/allwinner-a10: Mark the allwinner-a10 device with user_creatable = false
54
target/arm: Introduce neon_full_reg_offset
55
target/arm: Move neon_element_offset to translate.c
56
target/arm: Use neon_element_offset in neon_load/store_reg
57
target/arm: Use neon_element_offset in vfp_reg_offset
58
target/arm: Add read/write_neon_element32
59
target/arm: Expand read/write_neon_element32 to all MemOp
60
target/arm: Rename neon_load_reg32 to vfp_load_reg32
61
target/arm: Add read/write_neon_element64
62
target/arm: Rename neon_load_reg64 to vfp_load_reg64
63
target/arm: Simplify do_long_3d and do_2scalar_long
64
target/arm: Improve do_prewiden_3d
67
65
68
include/hw/boards.h | 11 ++
66
Rémi Denis-Courmont (3):
69
include/hw/intc/armv7m_nvic.h | 1 +
67
target/arm: fix handling of HCR.FB
70
include/qom/cpu.h | 7 +-
68
target/arm: fix LORID_EL1 access check
71
target/arm/cpu.h | 101 ++++++++++++--
69
hw/arm/boot: fix SVE for EL3 direct kernel boot
72
target/arm/helper.h | 2 +
73
target/arm/internals.h | 36 +++++
74
target/arm/translate.h | 1 +
75
hw/arm/allwinner-a10.c | 2 +
76
hw/arm/armv7m.c | 16 +--
77
hw/arm/aspeed.c | 3 +
78
hw/arm/collie.c | 1 +
79
hw/arm/cubieboard.c | 1 +
80
hw/arm/digic_boards.c | 1 +
81
hw/arm/exynos4_boards.c | 2 +
82
hw/arm/gumstix.c | 2 +
83
hw/arm/highbank.c | 2 +
84
hw/arm/imx25_pdk.c | 1 +
85
hw/arm/integratorcp.c | 1 +
86
hw/arm/kzm.c | 1 +
87
hw/arm/mainstone.c | 1 +
88
hw/arm/musicpal.c | 1 +
89
hw/arm/netduino2.c | 1 +
90
hw/arm/nseries.c | 2 +
91
hw/arm/omap_sx1.c | 2 +
92
hw/arm/palm.c | 1 +
93
hw/arm/raspi.c | 1 +
94
hw/arm/realview.c | 4 +
95
hw/arm/sabrelite.c | 1 +
96
hw/arm/spitz.c | 4 +
97
hw/arm/stellaris.c | 2 +
98
hw/arm/tosa.c | 1 +
99
hw/arm/versatilepb.c | 2 +
100
hw/arm/vexpress.c | 1 +
101
hw/arm/xilinx_zynq.c | 1 +
102
hw/arm/xlnx-ep108.c | 2 +
103
hw/arm/xlnx-zynqmp.c | 7 +-
104
hw/arm/z2.c | 1 +
105
hw/dma/xilinx_axidma.c | 16 +--
106
hw/intc/arm_gicv3_its_kvm.c | 19 +--
107
hw/intc/armv7m_nvic.c | 291 ++++++++++++++++++++++++++++++++------
108
hw/net/xilinx_axienet.c | 16 +--
109
qom/cpu.c | 16 +++
110
target/arm/cpu.c | 88 +++++++++---
111
target/arm/helper.c | 315 +++++++++++++++++++++++++++++++++---------
112
target/arm/machine.c | 105 ++++++++++++--
113
target/arm/op_helper.c | 43 ++++++
114
target/arm/translate.c | 54 +++++++-
115
scripts/device-crash-test | 1 -
116
48 files changed, 978 insertions(+), 213 deletions(-)
117
70
71
docs/qemu-option-trace.rst.inc | 6 +-
72
configure | 10 +-
73
include/hw/intc/arm_gicv3_common.h | 1 -
74
disas/capstone.c | 2 +-
75
hw/arm/boot.c | 3 +
76
hw/arm/smmuv3.c | 3 +-
77
hw/display/exynos4210_fimd.c | 4 +-
78
hw/display/omap_lcdc.c | 10 +-
79
hw/intc/arm_gicv3_cpuif.c | 5 +-
80
target/arm/helper.c | 24 +-
81
target/arm/m_helper.c | 3 +-
82
target/arm/translate.c | 153 +++++++++---
83
target/arm/vec_helper.c | 12 +-
84
tests/qtest/npcm7xx_rng-test.c | 14 +-
85
scripts/kernel-doc | 18 +-
86
target/arm/translate-neon.c.inc | 472 ++++++++++++++++++++-----------------
87
target/arm/translate-vfp.c.inc | 341 +++++++++++----------------
88
17 files changed, 588 insertions(+), 493 deletions(-)
89
diff view generated by jsdifflib
1
Implement the BXNS v8M instruction, which is like BX but will do a
1
From: Richard Henderson <richard.henderson@linaro.org>
2
jump-and-switch-to-NonSecure if the branch target address has bit 0
3
clear.
4
2
5
This is the first piece of code which implements "switch to the
3
This function makes it clear that we're talking about the whole
6
other security state", so the commit also includes the code to
4
register, and not the 32-bit piece at index 0. This fixes a bug
7
switch the stack pointers around, which is the only complicated
5
when running on a big-endian host.
8
part of switching security state.
9
6
10
BLXNS is more complicated than just "BXNS but set the link register",
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
so we leave it for a separate commit.
8
Message-id: 20201030022618.785675-2-richard.henderson@linaro.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/translate.c | 8 ++++++
13
target/arm/translate-neon.c.inc | 44 ++++++++++++++++-----------------
14
target/arm/translate-vfp.c.inc | 2 +-
15
3 files changed, 31 insertions(+), 23 deletions(-)
12
16
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 1503414539-28762-21-git-send-email-peter.maydell@linaro.org
16
---
17
target/arm/cpu.h | 13 +++++++++
18
target/arm/helper.h | 2 ++
19
target/arm/translate.h | 1 +
20
target/arm/helper.c | 79 ++++++++++++++++++++++++++++++++++++++++++++++++++
21
target/arm/machine.c | 2 ++
22
target/arm/translate.c | 42 ++++++++++++++++++++++++++-
23
6 files changed, 138 insertions(+), 1 deletion(-)
24
25
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/cpu.h
28
+++ b/target/arm/cpu.h
29
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
30
} cp15;
31
32
struct {
33
+ /* M profile has up to 4 stack pointers:
34
+ * a Main Stack Pointer and a Process Stack Pointer for each
35
+ * of the Secure and Non-Secure states. (If the CPU doesn't support
36
+ * the security extension then it has only two SPs.)
37
+ * In QEMU we always store the currently active SP in regs[13],
38
+ * and the non-active SP for the current security state in
39
+ * v7m.other_sp. The stack pointers for the inactive security state
40
+ * are stored in other_ss_msp and other_ss_psp.
41
+ * switch_v7m_security_state() is responsible for rearranging them
42
+ * when we change security state.
43
+ */
44
uint32_t other_sp;
45
+ uint32_t other_ss_msp;
46
+ uint32_t other_ss_psp;
47
uint32_t vecbase[2];
48
uint32_t basepri[2];
49
uint32_t control[2];
50
diff --git a/target/arm/helper.h b/target/arm/helper.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/helper.h
53
+++ b/target/arm/helper.h
54
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_1(cpsr_read, i32, env)
55
DEF_HELPER_3(v7m_msr, void, env, i32, i32)
56
DEF_HELPER_2(v7m_mrs, i32, env, i32)
57
58
+DEF_HELPER_2(v7m_bxns, void, env, i32)
59
+
60
DEF_HELPER_4(access_check_cp_reg, void, env, ptr, i32, i32)
61
DEF_HELPER_3(set_cp_reg, void, env, ptr, i32)
62
DEF_HELPER_2(get_cp_reg, i32, env, ptr)
63
diff --git a/target/arm/translate.h b/target/arm/translate.h
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/translate.h
66
+++ b/target/arm/translate.h
67
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
68
int vec_len;
69
int vec_stride;
70
bool v7m_handler_mode;
71
+ bool v8m_secure; /* true if v8M and we're in Secure mode */
72
/* Immediate value in AArch32 SVC insn; must be set if is_jmp == DISAS_SWI
73
* so that top level loop can generate correct syndrome information.
74
*/
75
diff --git a/target/arm/helper.c b/target/arm/helper.c
76
index XXXXXXX..XXXXXXX 100644
77
--- a/target/arm/helper.c
78
+++ b/target/arm/helper.c
79
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
80
return 0;
81
}
82
83
+void HELPER(v7m_bxns)(CPUARMState *env, uint32_t dest)
84
+{
85
+ /* translate.c should never generate calls here in user-only mode */
86
+ g_assert_not_reached();
87
+}
88
+
89
void switch_mode(CPUARMState *env, int mode)
90
{
91
ARMCPU *cpu = arm_env_get_cpu(env);
92
@@ -XXX,XX +XXX,XX @@ static uint32_t v7m_pop(CPUARMState *env)
93
return val;
94
}
95
96
+/* Return true if we're using the process stack pointer (not the MSP) */
97
+static bool v7m_using_psp(CPUARMState *env)
98
+{
99
+ /* Handler mode always uses the main stack; for thread mode
100
+ * the CONTROL.SPSEL bit determines the answer.
101
+ * Note that in v7M it is not possible to be in Handler mode with
102
+ * CONTROL.SPSEL non-zero, but in v8M it is, so we must check both.
103
+ */
104
+ return !arm_v7m_is_handler_mode(env) &&
105
+ env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK;
106
+}
107
+
108
/* Switch to V7M main or process stack pointer. */
109
static void switch_v7m_sp(CPUARMState *env, bool new_spsel)
110
{
111
@@ -XXX,XX +XXX,XX @@ static void switch_v7m_sp(CPUARMState *env, bool new_spsel)
112
}
113
}
114
115
+/* Switch M profile security state between NS and S */
116
+static void switch_v7m_security_state(CPUARMState *env, bool new_secstate)
117
+{
118
+ uint32_t new_ss_msp, new_ss_psp;
119
+
120
+ if (env->v7m.secure == new_secstate) {
121
+ return;
122
+ }
123
+
124
+ /* All the banked state is accessed by looking at env->v7m.secure
125
+ * except for the stack pointer; rearrange the SP appropriately.
126
+ */
127
+ new_ss_msp = env->v7m.other_ss_msp;
128
+ new_ss_psp = env->v7m.other_ss_psp;
129
+
130
+ if (v7m_using_psp(env)) {
131
+ env->v7m.other_ss_psp = env->regs[13];
132
+ env->v7m.other_ss_msp = env->v7m.other_sp;
133
+ } else {
134
+ env->v7m.other_ss_msp = env->regs[13];
135
+ env->v7m.other_ss_psp = env->v7m.other_sp;
136
+ }
137
+
138
+ env->v7m.secure = new_secstate;
139
+
140
+ if (v7m_using_psp(env)) {
141
+ env->regs[13] = new_ss_psp;
142
+ env->v7m.other_sp = new_ss_msp;
143
+ } else {
144
+ env->regs[13] = new_ss_msp;
145
+ env->v7m.other_sp = new_ss_psp;
146
+ }
147
+}
148
+
149
+void HELPER(v7m_bxns)(CPUARMState *env, uint32_t dest)
150
+{
151
+ /* Handle v7M BXNS:
152
+ * - if the return value is a magic value, do exception return (like BX)
153
+ * - otherwise bit 0 of the return value is the target security state
154
+ */
155
+ if (dest >= 0xff000000) {
156
+ /* This is an exception return magic value; put it where
157
+ * do_v7m_exception_exit() expects and raise EXCEPTION_EXIT.
158
+ * Note that if we ever add gen_ss_advance() singlestep support to
159
+ * M profile this should count as an "instruction execution complete"
160
+ * event (compare gen_bx_excret_final_code()).
161
+ */
162
+ env->regs[15] = dest & ~1;
163
+ env->thumb = dest & 1;
164
+ HELPER(exception_internal)(env, EXCP_EXCEPTION_EXIT);
165
+ /* notreached */
166
+ }
167
+
168
+ /* translate.c should have made BXNS UNDEF unless we're secure */
169
+ assert(env->v7m.secure);
170
+
171
+ switch_v7m_security_state(env, dest & 1);
172
+ env->thumb = 1;
173
+ env->regs[15] = dest & ~1;
174
+}
175
+
176
static uint32_t arm_v7m_load_vector(ARMCPU *cpu)
177
{
178
CPUState *cs = CPU(cpu);
179
diff --git a/target/arm/machine.c b/target/arm/machine.c
180
index XXXXXXX..XXXXXXX 100644
181
--- a/target/arm/machine.c
182
+++ b/target/arm/machine.c
183
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
184
.needed = m_security_needed,
185
.fields = (VMStateField[]) {
186
VMSTATE_UINT32(env.v7m.secure, ARMCPU),
187
+ VMSTATE_UINT32(env.v7m.other_ss_msp, ARMCPU),
188
+ VMSTATE_UINT32(env.v7m.other_ss_psp, ARMCPU),
189
VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
190
VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
191
VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
192
diff --git a/target/arm/translate.c b/target/arm/translate.c
17
diff --git a/target/arm/translate.c b/target/arm/translate.c
193
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
194
--- a/target/arm/translate.c
19
--- a/target/arm/translate.c
195
+++ b/target/arm/translate.c
20
+++ b/target/arm/translate.c
196
@@ -XXX,XX +XXX,XX @@ static inline void gen_bx_excret_final_code(DisasContext *s)
21
@@ -XXX,XX +XXX,XX @@ static inline void gen_hlt(DisasContext *s, int imm)
197
gen_exception_internal(EXCP_EXCEPTION_EXIT);
22
unallocated_encoding(s);
198
}
23
}
199
24
200
+static inline void gen_bxns(DisasContext *s, int rm)
25
+/*
26
+ * Return the offset of a "full" NEON Dreg.
27
+ */
28
+static long neon_full_reg_offset(unsigned reg)
201
+{
29
+{
202
+ TCGv_i32 var = load_reg(s, rm);
30
+ return offsetof(CPUARMState, vfp.zregs[reg >> 1].d[reg & 1]);
203
+
204
+ /* The bxns helper may raise an EXCEPTION_EXIT exception, so in theory
205
+ * we need to sync state before calling it, but:
206
+ * - we don't need to do gen_set_pc_im() because the bxns helper will
207
+ * always set the PC itself
208
+ * - we don't need to do gen_set_condexec() because BXNS is UNPREDICTABLE
209
+ * unless it's outside an IT block or the last insn in an IT block,
210
+ * so we know that condexec == 0 (already set at the top of the TB)
211
+ * is correct in the non-UNPREDICTABLE cases, and we can choose
212
+ * "zeroes the IT bits" as our UNPREDICTABLE behaviour otherwise.
213
+ */
214
+ gen_helper_v7m_bxns(cpu_env, var);
215
+ tcg_temp_free_i32(var);
216
+ s->is_jmp = DISAS_EXIT;
217
+}
31
+}
218
+
32
+
219
/* Variant of store_reg which uses branch&exchange logic when storing
33
static inline long vfp_reg_offset(bool dp, unsigned reg)
220
to r15 in ARM architecture v7 and above. The source must be a temporary
34
{
221
and will be marked as dead. */
35
if (dp) {
222
@@ -XXX,XX +XXX,XX @@ static void disas_thumb_insn(CPUARMState *env, DisasContext *s)
36
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
223
*/
37
index XXXXXXX..XXXXXXX 100644
224
bool link = insn & (1 << 7);
38
--- a/target/arm/translate-neon.c.inc
225
39
+++ b/target/arm/translate-neon.c.inc
226
- if (insn & 7) {
40
@@ -XXX,XX +XXX,XX @@ neon_element_offset(int reg, int element, MemOp size)
227
+ if (insn & 3) {
41
ofs ^= 8 - element_size;
228
goto undef;
42
}
229
}
43
#endif
230
if (link) {
44
- return neon_reg_offset(reg, 0) + ofs;
231
ARCH(5);
45
+ return neon_full_reg_offset(reg) + ofs;
232
}
46
}
233
+ if ((insn & 4)) {
47
234
+ /* BXNS/BLXNS: only exists for v8M with the
48
static void neon_load_element(TCGv_i32 var, int reg, int ele, MemOp mop)
235
+ * security extensions, and always UNDEF if NonSecure.
49
@@ -XXX,XX +XXX,XX @@ static bool trans_VLD_all_lanes(DisasContext *s, arg_VLD_all_lanes *a)
236
+ * We don't implement these in the user-only mode
50
* We cannot write 16 bytes at once because the
237
+ * either (in theory you can use them from Secure User
51
* destination is unaligned.
238
+ * mode but they are too tied in to system emulation.)
52
*/
239
+ */
53
- tcg_gen_gvec_dup_i32(size, neon_reg_offset(vd, 0),
240
+ if (!s->v8m_secure || IS_USER_ONLY) {
54
+ tcg_gen_gvec_dup_i32(size, neon_full_reg_offset(vd),
241
+ goto undef;
55
8, 8, tmp);
242
+ }
56
- tcg_gen_gvec_mov(0, neon_reg_offset(vd + 1, 0),
243
+ if (link) {
57
- neon_reg_offset(vd, 0), 8, 8);
244
+ /* BLXNS: not yet implemented */
58
+ tcg_gen_gvec_mov(0, neon_full_reg_offset(vd + 1),
245
+ goto undef;
59
+ neon_full_reg_offset(vd), 8, 8);
246
+ } else {
60
} else {
247
+ gen_bxns(s, rm);
61
- tcg_gen_gvec_dup_i32(size, neon_reg_offset(vd, 0),
248
+ }
62
+ tcg_gen_gvec_dup_i32(size, neon_full_reg_offset(vd),
249
+ break;
63
vec_size, vec_size, tmp);
250
+ }
64
}
251
+ /* BLX/BX */
65
tcg_gen_addi_i32(addr, addr, 1 << size);
252
tmp = load_reg(s, rm);
66
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDST_single(DisasContext *s, arg_VLDST_single *a)
253
if (link) {
67
static bool do_3same(DisasContext *s, arg_3same *a, GVecGen3Fn fn)
254
val = (uint32_t)s->pc | 1;
68
{
255
@@ -XXX,XX +XXX,XX @@ void gen_intermediate_code(CPUState *cs, TranslationBlock *tb)
69
int vec_size = a->q ? 16 : 8;
256
dc->vec_stride = ARM_TBFLAG_VECSTRIDE(tb->flags);
70
- int rd_ofs = neon_reg_offset(a->vd, 0);
257
dc->c15_cpar = ARM_TBFLAG_XSCALE_CPAR(tb->flags);
71
- int rn_ofs = neon_reg_offset(a->vn, 0);
258
dc->v7m_handler_mode = ARM_TBFLAG_HANDLER(tb->flags);
72
- int rm_ofs = neon_reg_offset(a->vm, 0);
259
+ dc->v8m_secure = arm_feature(env, ARM_FEATURE_M_SECURITY) &&
73
+ int rd_ofs = neon_full_reg_offset(a->vd);
260
+ regime_is_secure(env, dc->mmu_idx);
74
+ int rn_ofs = neon_full_reg_offset(a->vn);
261
dc->cp_regs = cpu->cp_regs;
75
+ int rm_ofs = neon_full_reg_offset(a->vm);
262
dc->features = env->features;
76
77
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
78
return false;
79
@@ -XXX,XX +XXX,XX @@ static bool do_vector_2sh(DisasContext *s, arg_2reg_shift *a, GVecGen2iFn *fn)
80
{
81
/* Handle a 2-reg-shift insn which can be vectorized. */
82
int vec_size = a->q ? 16 : 8;
83
- int rd_ofs = neon_reg_offset(a->vd, 0);
84
- int rm_ofs = neon_reg_offset(a->vm, 0);
85
+ int rd_ofs = neon_full_reg_offset(a->vd);
86
+ int rm_ofs = neon_full_reg_offset(a->vm);
87
88
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
89
return false;
90
@@ -XXX,XX +XXX,XX @@ static bool do_fp_2sh(DisasContext *s, arg_2reg_shift *a,
91
{
92
/* FP operations in 2-reg-and-shift group */
93
int vec_size = a->q ? 16 : 8;
94
- int rd_ofs = neon_reg_offset(a->vd, 0);
95
- int rm_ofs = neon_reg_offset(a->vm, 0);
96
+ int rd_ofs = neon_full_reg_offset(a->vd);
97
+ int rm_ofs = neon_full_reg_offset(a->vm);
98
TCGv_ptr fpst;
99
100
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
101
@@ -XXX,XX +XXX,XX @@ static bool do_1reg_imm(DisasContext *s, arg_1reg_imm *a,
102
return true;
103
}
104
105
- reg_ofs = neon_reg_offset(a->vd, 0);
106
+ reg_ofs = neon_full_reg_offset(a->vd);
107
vec_size = a->q ? 16 : 8;
108
imm = asimd_imm_const(a->imm, a->cmode, a->op);
109
110
@@ -XXX,XX +XXX,XX @@ static bool trans_VMULL_P_3d(DisasContext *s, arg_3diff *a)
111
return true;
112
}
113
114
- tcg_gen_gvec_3_ool(neon_reg_offset(a->vd, 0),
115
- neon_reg_offset(a->vn, 0),
116
- neon_reg_offset(a->vm, 0),
117
+ tcg_gen_gvec_3_ool(neon_full_reg_offset(a->vd),
118
+ neon_full_reg_offset(a->vn),
119
+ neon_full_reg_offset(a->vm),
120
16, 16, 0, fn_gvec);
121
return true;
122
}
123
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar_fp_vec(DisasContext *s, arg_2scalar *a,
124
{
125
/* Two registers and a scalar, using gvec */
126
int vec_size = a->q ? 16 : 8;
127
- int rd_ofs = neon_reg_offset(a->vd, 0);
128
- int rn_ofs = neon_reg_offset(a->vn, 0);
129
+ int rd_ofs = neon_full_reg_offset(a->vd);
130
+ int rn_ofs = neon_full_reg_offset(a->vn);
131
int rm_ofs;
132
int idx;
133
TCGv_ptr fpstatus;
134
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar_fp_vec(DisasContext *s, arg_2scalar *a,
135
/* a->vm is M:Vm, which encodes both register and index */
136
idx = extract32(a->vm, a->size + 2, 2);
137
a->vm = extract32(a->vm, 0, a->size + 2);
138
- rm_ofs = neon_reg_offset(a->vm, 0);
139
+ rm_ofs = neon_full_reg_offset(a->vm);
140
141
fpstatus = fpstatus_ptr(a->size == 1 ? FPST_STD_F16 : FPST_STD);
142
tcg_gen_gvec_3_ptr(rd_ofs, rn_ofs, rm_ofs, fpstatus,
143
@@ -XXX,XX +XXX,XX @@ static bool trans_VDUP_scalar(DisasContext *s, arg_VDUP_scalar *a)
144
return true;
145
}
146
147
- tcg_gen_gvec_dup_mem(a->size, neon_reg_offset(a->vd, 0),
148
+ tcg_gen_gvec_dup_mem(a->size, neon_full_reg_offset(a->vd),
149
neon_element_offset(a->vm, a->index, a->size),
150
a->q ? 16 : 8, a->q ? 16 : 8);
151
return true;
152
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_F32_F16(DisasContext *s, arg_2misc *a)
153
static bool do_2misc_vec(DisasContext *s, arg_2misc *a, GVecGen2Fn *fn)
154
{
155
int vec_size = a->q ? 16 : 8;
156
- int rd_ofs = neon_reg_offset(a->vd, 0);
157
- int rm_ofs = neon_reg_offset(a->vm, 0);
158
+ int rd_ofs = neon_full_reg_offset(a->vd);
159
+ int rm_ofs = neon_full_reg_offset(a->vm);
160
161
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
162
return false;
163
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
164
index XXXXXXX..XXXXXXX 100644
165
--- a/target/arm/translate-vfp.c.inc
166
+++ b/target/arm/translate-vfp.c.inc
167
@@ -XXX,XX +XXX,XX @@ static bool trans_VDUP(DisasContext *s, arg_VDUP *a)
168
}
169
170
tmp = load_reg(s, a->rt);
171
- tcg_gen_gvec_dup_i32(size, neon_reg_offset(a->vn, 0),
172
+ tcg_gen_gvec_dup_i32(size, neon_full_reg_offset(a->vn),
173
vec_size, vec_size, tmp);
174
tcg_temp_free_i32(tmp);
263
175
264
--
176
--
265
2.7.4
177
2.20.1
266
178
267
179
diff view generated by jsdifflib
1
From: Portia Stephens <portia.stephens@xilinx.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This adds a feature bit indicating support of the (trivial) Jazelle
3
This will shortly have users outside of translate-neon.c.inc.
4
implementation if ARM_FEATURE_V6 is set or if the processor is arm926
5
or arm1026. This fixes the issue that any BXJ instruction will
6
result in an illegal_op. BXJ instructions will now check if the
7
architecture supports ARM_FEATURE_JAZELLE.
8
4
9
Signed-off-by: Portia Stephens <portia.stephens@xilinx.com>
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Alistair Francis <alistair.francis@xilinx.com>
6
Message-id: 20201030022618.785675-3-richard.henderson@linaro.org
11
Message-id: 20170905211232.11092-1-portia.stephens@xilinx.com
12
[PMM: edited commit message and comment text a bit]
13
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
---
9
---
16
target/arm/cpu.h | 1 +
10
target/arm/translate.c | 20 ++++++++++++++++++++
17
target/arm/cpu.c | 3 +++
11
target/arm/translate-neon.c.inc | 19 -------------------
18
target/arm/translate.c | 2 +-
12
2 files changed, 20 insertions(+), 19 deletions(-)
19
3 files changed, 5 insertions(+), 1 deletion(-)
20
13
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
22
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
25
@@ -XXX,XX +XXX,XX @@ enum arm_features {
26
ARM_FEATURE_PMU, /* has PMU support */
27
ARM_FEATURE_VBAR, /* has cp15 VBAR */
28
ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
29
+ ARM_FEATURE_JAZELLE, /* has (trivial) Jazelle implementation */
30
};
31
32
static inline int arm_feature(CPUARMState *env, int feature)
33
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
34
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/cpu.c
36
+++ b/target/arm/cpu.c
37
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
38
}
39
if (arm_feature(env, ARM_FEATURE_V6)) {
40
set_feature(env, ARM_FEATURE_V5);
41
+ set_feature(env, ARM_FEATURE_JAZELLE);
42
if (!arm_feature(env, ARM_FEATURE_M)) {
43
set_feature(env, ARM_FEATURE_AUXCR);
44
}
45
@@ -XXX,XX +XXX,XX @@ static void arm926_initfn(Object *obj)
46
set_feature(&cpu->env, ARM_FEATURE_VFP);
47
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
48
set_feature(&cpu->env, ARM_FEATURE_CACHE_TEST_CLEAN);
49
+ set_feature(&cpu->env, ARM_FEATURE_JAZELLE);
50
cpu->midr = 0x41069265;
51
cpu->reset_fpsid = 0x41011090;
52
cpu->ctr = 0x1dd20d2;
53
@@ -XXX,XX +XXX,XX @@ static void arm1026_initfn(Object *obj)
54
set_feature(&cpu->env, ARM_FEATURE_AUXCR);
55
set_feature(&cpu->env, ARM_FEATURE_DUMMY_C15_REGS);
56
set_feature(&cpu->env, ARM_FEATURE_CACHE_TEST_CLEAN);
57
+ set_feature(&cpu->env, ARM_FEATURE_JAZELLE);
58
cpu->midr = 0x4106a262;
59
cpu->reset_fpsid = 0x410110a0;
60
cpu->ctr = 0x1dd20d2;
61
diff --git a/target/arm/translate.c b/target/arm/translate.c
14
diff --git a/target/arm/translate.c b/target/arm/translate.c
62
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate.c
16
--- a/target/arm/translate.c
64
+++ b/target/arm/translate.c
17
+++ b/target/arm/translate.c
65
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ static long neon_full_reg_offset(unsigned reg)
66
#define ENABLE_ARCH_5 arm_dc_feature(s, ARM_FEATURE_V5)
19
return offsetof(CPUARMState, vfp.zregs[reg >> 1].d[reg & 1]);
67
/* currently all emulated v5 cores are also v5TE, so don't bother */
20
}
68
#define ENABLE_ARCH_5TE arm_dc_feature(s, ARM_FEATURE_V5)
21
69
-#define ENABLE_ARCH_5J 0
22
+/*
70
+#define ENABLE_ARCH_5J arm_dc_feature(s, ARM_FEATURE_JAZELLE)
23
+ * Return the offset of a 2**SIZE piece of a NEON register, at index ELE,
71
#define ENABLE_ARCH_6 arm_dc_feature(s, ARM_FEATURE_V6)
24
+ * where 0 is the least significant end of the register.
72
#define ENABLE_ARCH_6K arm_dc_feature(s, ARM_FEATURE_V6K)
25
+ */
73
#define ENABLE_ARCH_6T2 arm_dc_feature(s, ARM_FEATURE_THUMB2)
26
+static long neon_element_offset(int reg, int element, MemOp size)
27
+{
28
+ int element_size = 1 << size;
29
+ int ofs = element * element_size;
30
+#ifdef HOST_WORDS_BIGENDIAN
31
+ /*
32
+ * Calculate the offset assuming fully little-endian,
33
+ * then XOR to account for the order of the 8-byte units.
34
+ */
35
+ if (element_size < 8) {
36
+ ofs ^= 8 - element_size;
37
+ }
38
+#endif
39
+ return neon_full_reg_offset(reg) + ofs;
40
+}
41
+
42
static inline long vfp_reg_offset(bool dp, unsigned reg)
43
{
44
if (dp) {
45
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
46
index XXXXXXX..XXXXXXX 100644
47
--- a/target/arm/translate-neon.c.inc
48
+++ b/target/arm/translate-neon.c.inc
49
@@ -XXX,XX +XXX,XX @@ static inline int neon_3same_fp_size(DisasContext *s, int x)
50
#include "decode-neon-ls.c.inc"
51
#include "decode-neon-shared.c.inc"
52
53
-/* Return the offset of a 2**SIZE piece of a NEON register, at index ELE,
54
- * where 0 is the least significant end of the register.
55
- */
56
-static inline long
57
-neon_element_offset(int reg, int element, MemOp size)
58
-{
59
- int element_size = 1 << size;
60
- int ofs = element * element_size;
61
-#ifdef HOST_WORDS_BIGENDIAN
62
- /* Calculate the offset assuming fully little-endian,
63
- * then XOR to account for the order of the 8-byte units.
64
- */
65
- if (element_size < 8) {
66
- ofs ^= 8 - element_size;
67
- }
68
-#endif
69
- return neon_full_reg_offset(reg) + ofs;
70
-}
71
-
72
static void neon_load_element(TCGv_i32 var, int reg, int ele, MemOp mop)
73
{
74
long offset = neon_element_offset(reg, ele, mop & MO_SIZE);
74
--
75
--
75
2.7.4
76
2.20.1
76
77
77
78
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
These are the only users of neon_reg_offset, so remove that.
4
Message-id: 20170905131149.10669-4-famz@redhat.com
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20201030022618.785675-4-richard.henderson@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
9
---
8
hw/intc/arm_gicv3_its_kvm.c | 19 +++++++------------
10
target/arm/translate.c | 14 ++------------
9
1 file changed, 7 insertions(+), 12 deletions(-)
11
1 file changed, 2 insertions(+), 12 deletions(-)
10
12
11
diff --git a/hw/intc/arm_gicv3_its_kvm.c b/hw/intc/arm_gicv3_its_kvm.c
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
12
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
13
--- a/hw/intc/arm_gicv3_its_kvm.c
15
--- a/target/arm/translate.c
14
+++ b/hw/intc/arm_gicv3_its_kvm.c
16
+++ b/target/arm/translate.c
15
@@ -XXX,XX +XXX,XX @@ static void kvm_arm_its_realize(DeviceState *dev, Error **errp)
17
@@ -XXX,XX +XXX,XX @@ static inline long vfp_reg_offset(bool dp, unsigned reg)
16
qemu_add_vm_change_state_handler(vm_change_state_handler, s);
18
}
17
}
19
}
18
20
19
-static void kvm_arm_its_init(Object *obj)
21
-/* Return the offset of a 32-bit piece of a NEON register.
22
- zero is the least significant end of the register. */
23
-static inline long
24
-neon_reg_offset (int reg, int n)
20
-{
25
-{
21
- GICv3ITSState *s = KVM_ARM_ITS(obj);
26
- int sreg;
22
-
27
- sreg = reg * 2 + n;
23
- object_property_add_link(obj, "parent-gicv3",
28
- return vfp_reg_offset(0, sreg);
24
- "kvm-arm-gicv3", (Object **)&s->gicv3,
25
- object_property_allow_set_link,
26
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
27
- &error_abort);
28
-}
29
-}
29
-
30
-
30
/**
31
static TCGv_i32 neon_load_reg(int reg, int pass)
31
* kvm_arm_its_pre_save - handles the saving of ITS registers.
32
{
32
* ITS tables are flushed into guest RAM separately and earlier,
33
TCGv_i32 tmp = tcg_temp_new_i32();
33
@@ -XXX,XX +XXX,XX @@ static void kvm_arm_its_post_load(GICv3ITSState *s)
34
- tcg_gen_ld_i32(tmp, cpu_env, neon_reg_offset(reg, pass));
34
GITS_CTLR, &s->ctlr, true, &error_abort);
35
+ tcg_gen_ld_i32(tmp, cpu_env, neon_element_offset(reg, pass, MO_32));
36
return tmp;
35
}
37
}
36
38
37
+static Property kvm_arm_its_props[] = {
39
static void neon_store_reg(int reg, int pass, TCGv_i32 var)
38
+ DEFINE_PROP_LINK("parent-gicv3", GICv3ITSState, gicv3, "kvm-arm-gicv3",
39
+ GICv3State *),
40
+ DEFINE_PROP_END_OF_LIST(),
41
+};
42
+
43
static void kvm_arm_its_class_init(ObjectClass *klass, void *data)
44
{
40
{
45
DeviceClass *dc = DEVICE_CLASS(klass);
41
- tcg_gen_st_i32(var, cpu_env, neon_reg_offset(reg, pass));
46
GICv3ITSCommonClass *icc = ARM_GICV3_ITS_COMMON_CLASS(klass);
42
+ tcg_gen_st_i32(var, cpu_env, neon_element_offset(reg, pass, MO_32));
47
43
tcg_temp_free_i32(var);
48
dc->realize = kvm_arm_its_realize;
44
}
49
+ dc->props = kvm_arm_its_props;
50
icc->send_msi = kvm_its_send_msi;
51
icc->pre_save = kvm_arm_its_pre_save;
52
icc->post_load = kvm_arm_its_post_load;
53
@@ -XXX,XX +XXX,XX @@ static const TypeInfo kvm_arm_its_info = {
54
.name = TYPE_KVM_ARM_ITS,
55
.parent = TYPE_ARM_GICV3_ITS_COMMON,
56
.instance_size = sizeof(GICv3ITSState),
57
- .instance_init = kvm_arm_its_init,
58
.class_init = kvm_arm_its_class_init,
59
};
60
45
61
--
46
--
62
2.7.4
47
2.20.1
63
48
64
49
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
This seems a bit more readable than using offsetof CPU_DoubleU.
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
5
Message-id: 20170905131149.10669-6-famz@redhat.com
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20201030022618.785675-5-richard.henderson@linaro.org
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
9
---
9
hw/net/xilinx_axienet.c | 16 ++++------------
10
target/arm/translate.c | 13 ++++---------
10
1 file changed, 4 insertions(+), 12 deletions(-)
11
1 file changed, 4 insertions(+), 9 deletions(-)
11
12
12
diff --git a/hw/net/xilinx_axienet.c b/hw/net/xilinx_axienet.c
13
diff --git a/target/arm/translate.c b/target/arm/translate.c
13
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/net/xilinx_axienet.c
15
--- a/target/arm/translate.c
15
+++ b/hw/net/xilinx_axienet.c
16
+++ b/target/arm/translate.c
16
@@ -XXX,XX +XXX,XX @@ static void xilinx_enet_init(Object *obj)
17
@@ -XXX,XX +XXX,XX @@ static long neon_element_offset(int reg, int element, MemOp size)
17
XilinxAXIEnet *s = XILINX_AXI_ENET(obj);
18
return neon_full_reg_offset(reg) + ofs;
18
SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
19
}
19
20
20
- object_property_add_link(obj, "axistream-connected", TYPE_STREAM_SLAVE,
21
-static inline long vfp_reg_offset(bool dp, unsigned reg)
21
- (Object **) &s->tx_data_dev,
22
+/* Return the offset of a VFP Dreg (dp = true) or VFP Sreg (dp = false). */
22
- qdev_prop_allow_set_link_before_realize,
23
+static long vfp_reg_offset(bool dp, unsigned reg)
23
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
24
{
24
- &error_abort);
25
if (dp) {
25
- object_property_add_link(obj, "axistream-control-connected",
26
- return offsetof(CPUARMState, vfp.zregs[reg >> 1].d[reg & 1]);
26
- TYPE_STREAM_SLAVE,
27
+ return neon_element_offset(reg, 0, MO_64);
27
- (Object **) &s->tx_control_dev,
28
} else {
28
- qdev_prop_allow_set_link_before_realize,
29
- long ofs = offsetof(CPUARMState, vfp.zregs[reg >> 2].d[(reg >> 1) & 1]);
29
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
30
- if (reg & 1) {
30
- &error_abort);
31
- ofs += offsetof(CPU_DoubleU, l.upper);
31
-
32
- } else {
32
object_initialize(&s->rx_data_dev, sizeof(s->rx_data_dev),
33
- ofs += offsetof(CPU_DoubleU, l.lower);
33
TYPE_XILINX_AXI_ENET_DATA_STREAM);
34
- }
34
object_initialize(&s->rx_control_dev, sizeof(s->rx_control_dev),
35
- return ofs;
35
@@ -XXX,XX +XXX,XX @@ static Property xilinx_enet_properties[] = {
36
+ return neon_element_offset(reg >> 1, reg & 1, MO_32);
36
DEFINE_PROP_UINT32("rxmem", XilinxAXIEnet, c_rxmem, 0x1000),
37
}
37
DEFINE_PROP_UINT32("txmem", XilinxAXIEnet, c_txmem, 0x1000),
38
}
38
DEFINE_NIC_PROPERTIES(XilinxAXIEnet, conf),
39
+ DEFINE_PROP_LINK("axistream-connected", XilinxAXIEnet,
40
+ tx_data_dev, TYPE_STREAM_SLAVE, StreamSlave *),
41
+ DEFINE_PROP_LINK("axistream-control-connected", XilinxAXIEnet,
42
+ tx_control_dev, TYPE_STREAM_SLAVE, StreamSlave *),
43
DEFINE_PROP_END_OF_LIST(),
44
};
45
39
46
--
40
--
47
2.7.4
41
2.20.1
48
42
49
43
diff view generated by jsdifflib
1
For v8M the range 0xe002e000..0xe002efff is an alias region which
1
From: Richard Henderson <richard.henderson@linaro.org>
2
for secure accesses behaves like a NonSecure access to the main
3
SCS region. (For nonsecure accesses including when the security
4
extension is not implemented, it is RAZ/WI.)
5
2
3
Model these off the aa64 read/write_vec_element functions.
4
Use it within translate-neon.c.inc. The new functions do
5
not allocate or free temps, so this rearranges the calling
6
code a bit.
7
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20201030022618.785675-6-richard.henderson@linaro.org
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Message-id: 1503414539-28762-11-git-send-email-peter.maydell@linaro.org
8
---
12
---
9
include/hw/intc/armv7m_nvic.h | 1 +
13
target/arm/translate.c | 26 ++++
10
hw/intc/armv7m_nvic.c | 66 ++++++++++++++++++++++++++++++++++++++++++-
14
target/arm/translate-neon.c.inc | 256 ++++++++++++++++++++------------
11
2 files changed, 66 insertions(+), 1 deletion(-)
15
2 files changed, 183 insertions(+), 99 deletions(-)
12
16
13
diff --git a/include/hw/intc/armv7m_nvic.h b/include/hw/intc/armv7m_nvic.h
17
diff --git a/target/arm/translate.c b/target/arm/translate.c
14
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
15
--- a/include/hw/intc/armv7m_nvic.h
19
--- a/target/arm/translate.c
16
+++ b/include/hw/intc/armv7m_nvic.h
20
+++ b/target/arm/translate.c
17
@@ -XXX,XX +XXX,XX @@ typedef struct NVICState {
21
@@ -XXX,XX +XXX,XX @@ static inline void neon_store_reg32(TCGv_i32 var, int reg)
18
int exception_prio; /* group prio of the highest prio active exception */
22
tcg_gen_st_i32(var, cpu_env, vfp_reg_offset(false, reg));
19
23
}
20
MemoryRegion sysregmem;
24
21
+ MemoryRegion sysreg_ns_mem;
25
+static void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp size)
22
MemoryRegion container;
23
24
uint32_t num_irq;
25
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
26
index XXXXXXX..XXXXXXX 100644
27
--- a/hw/intc/armv7m_nvic.c
28
+++ b/hw/intc/armv7m_nvic.c
29
@@ -XXX,XX +XXX,XX @@ static const MemoryRegionOps nvic_sysreg_ops = {
30
.endianness = DEVICE_NATIVE_ENDIAN,
31
};
32
33
+static MemTxResult nvic_sysreg_ns_write(void *opaque, hwaddr addr,
34
+ uint64_t value, unsigned size,
35
+ MemTxAttrs attrs)
36
+{
26
+{
37
+ if (attrs.secure) {
27
+ long off = neon_element_offset(reg, ele, size);
38
+ /* S accesses to the alias act like NS accesses to the real region */
28
+
39
+ attrs.secure = 0;
29
+ switch (size) {
40
+ return nvic_sysreg_write(opaque, addr, value, size, attrs);
30
+ case MO_32:
41
+ } else {
31
+ tcg_gen_ld_i32(dest, cpu_env, off);
42
+ /* NS attrs are RAZ/WI for privileged, and BusFault for user */
32
+ break;
43
+ if (attrs.user) {
33
+ default:
44
+ return MEMTX_ERROR;
34
+ g_assert_not_reached();
45
+ }
46
+ return MEMTX_OK;
47
+ }
35
+ }
48
+}
36
+}
49
+
37
+
50
+static MemTxResult nvic_sysreg_ns_read(void *opaque, hwaddr addr,
38
+static void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp size)
51
+ uint64_t *data, unsigned size,
52
+ MemTxAttrs attrs)
53
+{
39
+{
54
+ if (attrs.secure) {
40
+ long off = neon_element_offset(reg, ele, size);
55
+ /* S accesses to the alias act like NS accesses to the real region */
41
+
56
+ attrs.secure = 0;
42
+ switch (size) {
57
+ return nvic_sysreg_read(opaque, addr, data, size, attrs);
43
+ case MO_32:
58
+ } else {
44
+ tcg_gen_st_i32(src, cpu_env, off);
59
+ /* NS attrs are RAZ/WI for privileged, and BusFault for user */
45
+ break;
60
+ if (attrs.user) {
46
+ default:
61
+ return MEMTX_ERROR;
47
+ g_assert_not_reached();
62
+ }
63
+ *data = 0;
64
+ return MEMTX_OK;
65
+ }
48
+ }
66
+}
49
+}
67
+
50
+
68
+static const MemoryRegionOps nvic_sysreg_ns_ops = {
51
static TCGv_ptr vfp_reg_ptr(bool dp, int reg)
69
+ .read_with_attrs = nvic_sysreg_ns_read,
70
+ .write_with_attrs = nvic_sysreg_ns_write,
71
+ .endianness = DEVICE_NATIVE_ENDIAN,
72
+};
73
+
74
static int nvic_post_load(void *opaque, int version_id)
75
{
52
{
76
NVICState *s = opaque;
53
TCGv_ptr ret = tcg_temp_new_ptr();
77
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
54
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
78
NVICState *s = NVIC(dev);
55
index XXXXXXX..XXXXXXX 100644
79
SysBusDevice *systick_sbd;
56
--- a/target/arm/translate-neon.c.inc
80
Error *err = NULL;
57
+++ b/target/arm/translate-neon.c.inc
81
+ int regionlen;
58
@@ -XXX,XX +XXX,XX @@ static bool do_3same_pair(DisasContext *s, arg_3same *a, NeonGenTwoOpFn *fn)
82
59
* early. Since Q is 0 there are always just two passes, so instead
83
s->cpu = ARM_CPU(qemu_get_cpu(0));
60
* of a complicated loop over each pass we just unroll.
84
assert(s->cpu);
85
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
86
* 0xd00..0xd3c - SCS registers
87
* 0xd40..0xeff - Reserved or Not implemented
88
* 0xf00 - STIR
89
+ *
90
+ * Some registers within this space are banked between security states.
91
+ * In v8M there is a second range 0xe002e000..0xe002efff which is the
92
+ * NonSecure alias SCS; secure accesses to this behave like NS accesses
93
+ * to the main SCS range, and non-secure accesses (including when
94
+ * the security extension is not implemented) are RAZ/WI.
95
+ * Note that both the main SCS range and the alias range are defined
96
+ * to be exempt from memory attribution (R_BLJT) and so the memory
97
+ * transaction attribute always matches the current CPU security
98
+ * state (attrs.secure == env->v7m.secure). In the nvic_sysreg_ns_ops
99
+ * wrappers we change attrs.secure to indicate the NS access; so
100
+ * generally code determining which banked register to use should
101
+ * use attrs.secure; code determining actual behaviour of the system
102
+ * should use env->v7m.secure.
103
*/
61
*/
104
- memory_region_init(&s->container, OBJECT(s), "nvic", 0x1000);
62
- tmp = neon_load_reg(a->vn, 0);
105
+ regionlen = arm_feature(&s->cpu->env, ARM_FEATURE_V8) ? 0x21000 : 0x1000;
63
- tmp2 = neon_load_reg(a->vn, 1);
106
+ memory_region_init(&s->container, OBJECT(s), "nvic", regionlen);
64
+ tmp = tcg_temp_new_i32();
107
/* The system register region goes at the bottom of the priority
65
+ tmp2 = tcg_temp_new_i32();
108
* stack as it covers the whole page.
66
+ tmp3 = tcg_temp_new_i32();
67
+
68
+ read_neon_element32(tmp, a->vn, 0, MO_32);
69
+ read_neon_element32(tmp2, a->vn, 1, MO_32);
70
fn(tmp, tmp, tmp2);
71
- tcg_temp_free_i32(tmp2);
72
73
- tmp3 = neon_load_reg(a->vm, 0);
74
- tmp2 = neon_load_reg(a->vm, 1);
75
+ read_neon_element32(tmp3, a->vm, 0, MO_32);
76
+ read_neon_element32(tmp2, a->vm, 1, MO_32);
77
fn(tmp3, tmp3, tmp2);
78
- tcg_temp_free_i32(tmp2);
79
80
- neon_store_reg(a->vd, 0, tmp);
81
- neon_store_reg(a->vd, 1, tmp3);
82
+ write_neon_element32(tmp, a->vd, 0, MO_32);
83
+ write_neon_element32(tmp3, a->vd, 1, MO_32);
84
+
85
+ tcg_temp_free_i32(tmp);
86
+ tcg_temp_free_i32(tmp2);
87
+ tcg_temp_free_i32(tmp3);
88
return true;
89
}
90
91
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_env_32(DisasContext *s, arg_2reg_shift *a,
92
* 2-reg-and-shift operations, size < 3 case, where the
93
* helper needs to be passed cpu_env.
109
*/
94
*/
110
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
95
- TCGv_i32 constimm;
111
sysbus_mmio_get_region(systick_sbd, 0),
96
+ TCGv_i32 constimm, tmp;
112
1);
97
int pass;
113
98
114
+ if (arm_feature(&s->cpu->env, ARM_FEATURE_V8)) {
99
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
115
+ memory_region_init_io(&s->sysreg_ns_mem, OBJECT(s),
100
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_env_32(DisasContext *s, arg_2reg_shift *a,
116
+ &nvic_sysreg_ns_ops, s,
101
* by immediate using the variable shift operations.
117
+ "nvic_sysregs_ns", 0x1000);
102
*/
118
+ memory_region_add_subregion(&s->container, 0x20000, &s->sysreg_ns_mem);
103
constimm = tcg_const_i32(dup_const(a->size, a->shift));
119
+ }
104
+ tmp = tcg_temp_new_i32();
120
+
105
121
sysbus_init_mmio(SYS_BUS_DEVICE(dev), &s->container);
106
for (pass = 0; pass < (a->q ? 4 : 2); pass++) {
122
}
107
- TCGv_i32 tmp = neon_load_reg(a->vm, pass);
123
108
+ read_neon_element32(tmp, a->vm, pass, MO_32);
109
fn(tmp, cpu_env, tmp, constimm);
110
- neon_store_reg(a->vd, pass, tmp);
111
+ write_neon_element32(tmp, a->vd, pass, MO_32);
112
}
113
+ tcg_temp_free_i32(tmp);
114
tcg_temp_free_i32(constimm);
115
return true;
116
}
117
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_narrow_64(DisasContext *s, arg_2reg_shift *a,
118
constimm = tcg_const_i64(-a->shift);
119
rm1 = tcg_temp_new_i64();
120
rm2 = tcg_temp_new_i64();
121
+ rd = tcg_temp_new_i32();
122
123
/* Load both inputs first to avoid potential overwrite if rm == rd */
124
neon_load_reg64(rm1, a->vm);
125
neon_load_reg64(rm2, a->vm + 1);
126
127
shiftfn(rm1, rm1, constimm);
128
- rd = tcg_temp_new_i32();
129
narrowfn(rd, cpu_env, rm1);
130
- neon_store_reg(a->vd, 0, rd);
131
+ write_neon_element32(rd, a->vd, 0, MO_32);
132
133
shiftfn(rm2, rm2, constimm);
134
- rd = tcg_temp_new_i32();
135
narrowfn(rd, cpu_env, rm2);
136
- neon_store_reg(a->vd, 1, rd);
137
+ write_neon_element32(rd, a->vd, 1, MO_32);
138
139
+ tcg_temp_free_i32(rd);
140
tcg_temp_free_i64(rm1);
141
tcg_temp_free_i64(rm2);
142
tcg_temp_free_i64(constimm);
143
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_narrow_32(DisasContext *s, arg_2reg_shift *a,
144
constimm = tcg_const_i32(imm);
145
146
/* Load all inputs first to avoid potential overwrite */
147
- rm1 = neon_load_reg(a->vm, 0);
148
- rm2 = neon_load_reg(a->vm, 1);
149
- rm3 = neon_load_reg(a->vm + 1, 0);
150
- rm4 = neon_load_reg(a->vm + 1, 1);
151
+ rm1 = tcg_temp_new_i32();
152
+ rm2 = tcg_temp_new_i32();
153
+ rm3 = tcg_temp_new_i32();
154
+ rm4 = tcg_temp_new_i32();
155
+ read_neon_element32(rm1, a->vm, 0, MO_32);
156
+ read_neon_element32(rm2, a->vm, 1, MO_32);
157
+ read_neon_element32(rm3, a->vm, 2, MO_32);
158
+ read_neon_element32(rm4, a->vm, 3, MO_32);
159
rtmp = tcg_temp_new_i64();
160
161
shiftfn(rm1, rm1, constimm);
162
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_narrow_32(DisasContext *s, arg_2reg_shift *a,
163
tcg_temp_free_i32(rm2);
164
165
narrowfn(rm1, cpu_env, rtmp);
166
- neon_store_reg(a->vd, 0, rm1);
167
+ write_neon_element32(rm1, a->vd, 0, MO_32);
168
+ tcg_temp_free_i32(rm1);
169
170
shiftfn(rm3, rm3, constimm);
171
shiftfn(rm4, rm4, constimm);
172
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_narrow_32(DisasContext *s, arg_2reg_shift *a,
173
174
narrowfn(rm3, cpu_env, rtmp);
175
tcg_temp_free_i64(rtmp);
176
- neon_store_reg(a->vd, 1, rm3);
177
+ write_neon_element32(rm3, a->vd, 1, MO_32);
178
+ tcg_temp_free_i32(rm3);
179
return true;
180
}
181
182
@@ -XXX,XX +XXX,XX @@ static bool do_vshll_2sh(DisasContext *s, arg_2reg_shift *a,
183
widen_mask = dup_const(a->size + 1, widen_mask);
184
}
185
186
- rm0 = neon_load_reg(a->vm, 0);
187
- rm1 = neon_load_reg(a->vm, 1);
188
+ rm0 = tcg_temp_new_i32();
189
+ rm1 = tcg_temp_new_i32();
190
+ read_neon_element32(rm0, a->vm, 0, MO_32);
191
+ read_neon_element32(rm1, a->vm, 1, MO_32);
192
tmp = tcg_temp_new_i64();
193
194
widenfn(tmp, rm0);
195
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
196
if (src1_wide) {
197
neon_load_reg64(rn0_64, a->vn);
198
} else {
199
- TCGv_i32 tmp = neon_load_reg(a->vn, 0);
200
+ TCGv_i32 tmp = tcg_temp_new_i32();
201
+ read_neon_element32(tmp, a->vn, 0, MO_32);
202
widenfn(rn0_64, tmp);
203
tcg_temp_free_i32(tmp);
204
}
205
- rm = neon_load_reg(a->vm, 0);
206
+ rm = tcg_temp_new_i32();
207
+ read_neon_element32(rm, a->vm, 0, MO_32);
208
209
widenfn(rm_64, rm);
210
tcg_temp_free_i32(rm);
211
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
212
if (src1_wide) {
213
neon_load_reg64(rn1_64, a->vn + 1);
214
} else {
215
- TCGv_i32 tmp = neon_load_reg(a->vn, 1);
216
+ TCGv_i32 tmp = tcg_temp_new_i32();
217
+ read_neon_element32(tmp, a->vn, 1, MO_32);
218
widenfn(rn1_64, tmp);
219
tcg_temp_free_i32(tmp);
220
}
221
- rm = neon_load_reg(a->vm, 1);
222
+ rm = tcg_temp_new_i32();
223
+ read_neon_element32(rm, a->vm, 1, MO_32);
224
225
neon_store_reg64(rn0_64, a->vd);
226
227
@@ -XXX,XX +XXX,XX @@ static bool do_narrow_3d(DisasContext *s, arg_3diff *a,
228
229
narrowfn(rd1, rn_64);
230
231
- neon_store_reg(a->vd, 0, rd0);
232
- neon_store_reg(a->vd, 1, rd1);
233
+ write_neon_element32(rd0, a->vd, 0, MO_32);
234
+ write_neon_element32(rd1, a->vd, 1, MO_32);
235
236
+ tcg_temp_free_i32(rd0);
237
+ tcg_temp_free_i32(rd1);
238
tcg_temp_free_i64(rn_64);
239
tcg_temp_free_i64(rm_64);
240
241
@@ -XXX,XX +XXX,XX @@ static bool do_long_3d(DisasContext *s, arg_3diff *a,
242
rd0 = tcg_temp_new_i64();
243
rd1 = tcg_temp_new_i64();
244
245
- rn = neon_load_reg(a->vn, 0);
246
- rm = neon_load_reg(a->vm, 0);
247
+ rn = tcg_temp_new_i32();
248
+ rm = tcg_temp_new_i32();
249
+ read_neon_element32(rn, a->vn, 0, MO_32);
250
+ read_neon_element32(rm, a->vm, 0, MO_32);
251
opfn(rd0, rn, rm);
252
- tcg_temp_free_i32(rn);
253
- tcg_temp_free_i32(rm);
254
255
- rn = neon_load_reg(a->vn, 1);
256
- rm = neon_load_reg(a->vm, 1);
257
+ read_neon_element32(rn, a->vn, 1, MO_32);
258
+ read_neon_element32(rm, a->vm, 1, MO_32);
259
opfn(rd1, rn, rm);
260
tcg_temp_free_i32(rn);
261
tcg_temp_free_i32(rm);
262
@@ -XXX,XX +XXX,XX @@ static void gen_neon_dup_high16(TCGv_i32 var)
263
264
static inline TCGv_i32 neon_get_scalar(int size, int reg)
265
{
266
- TCGv_i32 tmp;
267
- if (size == 1) {
268
- tmp = neon_load_reg(reg & 7, reg >> 4);
269
+ TCGv_i32 tmp = tcg_temp_new_i32();
270
+ if (size == MO_16) {
271
+ read_neon_element32(tmp, reg & 7, reg >> 4, MO_32);
272
if (reg & 8) {
273
gen_neon_dup_high16(tmp);
274
} else {
275
gen_neon_dup_low16(tmp);
276
}
277
} else {
278
- tmp = neon_load_reg(reg & 15, reg >> 4);
279
+ read_neon_element32(tmp, reg & 15, reg >> 4, MO_32);
280
}
281
return tmp;
282
}
283
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar(DisasContext *s, arg_2scalar *a,
284
* perform an accumulation operation of that result into the
285
* destination.
286
*/
287
- TCGv_i32 scalar;
288
+ TCGv_i32 scalar, tmp;
289
int pass;
290
291
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
292
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar(DisasContext *s, arg_2scalar *a,
293
}
294
295
scalar = neon_get_scalar(a->size, a->vm);
296
+ tmp = tcg_temp_new_i32();
297
298
for (pass = 0; pass < (a->q ? 4 : 2); pass++) {
299
- TCGv_i32 tmp = neon_load_reg(a->vn, pass);
300
+ read_neon_element32(tmp, a->vn, pass, MO_32);
301
opfn(tmp, tmp, scalar);
302
if (accfn) {
303
- TCGv_i32 rd = neon_load_reg(a->vd, pass);
304
+ TCGv_i32 rd = tcg_temp_new_i32();
305
+ read_neon_element32(rd, a->vd, pass, MO_32);
306
accfn(tmp, rd, tmp);
307
tcg_temp_free_i32(rd);
308
}
309
- neon_store_reg(a->vd, pass, tmp);
310
+ write_neon_element32(tmp, a->vd, pass, MO_32);
311
}
312
+ tcg_temp_free_i32(tmp);
313
tcg_temp_free_i32(scalar);
314
return true;
315
}
316
@@ -XXX,XX +XXX,XX @@ static bool do_vqrdmlah_2sc(DisasContext *s, arg_2scalar *a,
317
* performs a kind of fused op-then-accumulate using a helper
318
* function that takes all of rd, rn and the scalar at once.
319
*/
320
- TCGv_i32 scalar;
321
+ TCGv_i32 scalar, rn, rd;
322
int pass;
323
324
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
325
@@ -XXX,XX +XXX,XX @@ static bool do_vqrdmlah_2sc(DisasContext *s, arg_2scalar *a,
326
}
327
328
scalar = neon_get_scalar(a->size, a->vm);
329
+ rn = tcg_temp_new_i32();
330
+ rd = tcg_temp_new_i32();
331
332
for (pass = 0; pass < (a->q ? 4 : 2); pass++) {
333
- TCGv_i32 rn = neon_load_reg(a->vn, pass);
334
- TCGv_i32 rd = neon_load_reg(a->vd, pass);
335
+ read_neon_element32(rn, a->vn, pass, MO_32);
336
+ read_neon_element32(rd, a->vd, pass, MO_32);
337
opfn(rd, cpu_env, rn, scalar, rd);
338
- tcg_temp_free_i32(rn);
339
- neon_store_reg(a->vd, pass, rd);
340
+ write_neon_element32(rd, a->vd, pass, MO_32);
341
}
342
+ tcg_temp_free_i32(rn);
343
+ tcg_temp_free_i32(rd);
344
tcg_temp_free_i32(scalar);
345
346
return true;
347
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar_long(DisasContext *s, arg_2scalar *a,
348
scalar = neon_get_scalar(a->size, a->vm);
349
350
/* Load all inputs before writing any outputs, in case of overlap */
351
- rn = neon_load_reg(a->vn, 0);
352
+ rn = tcg_temp_new_i32();
353
+ read_neon_element32(rn, a->vn, 0, MO_32);
354
rn0_64 = tcg_temp_new_i64();
355
opfn(rn0_64, rn, scalar);
356
- tcg_temp_free_i32(rn);
357
358
- rn = neon_load_reg(a->vn, 1);
359
+ read_neon_element32(rn, a->vn, 1, MO_32);
360
rn1_64 = tcg_temp_new_i64();
361
opfn(rn1_64, rn, scalar);
362
tcg_temp_free_i32(rn);
363
@@ -XXX,XX +XXX,XX @@ static bool trans_VTBL(DisasContext *s, arg_VTBL *a)
364
return false;
365
}
366
n <<= 3;
367
+ tmp = tcg_temp_new_i32();
368
if (a->op) {
369
- tmp = neon_load_reg(a->vd, 0);
370
+ read_neon_element32(tmp, a->vd, 0, MO_32);
371
} else {
372
- tmp = tcg_temp_new_i32();
373
tcg_gen_movi_i32(tmp, 0);
374
}
375
- tmp2 = neon_load_reg(a->vm, 0);
376
+ tmp2 = tcg_temp_new_i32();
377
+ read_neon_element32(tmp2, a->vm, 0, MO_32);
378
ptr1 = vfp_reg_ptr(true, a->vn);
379
tmp4 = tcg_const_i32(n);
380
gen_helper_neon_tbl(tmp2, tmp2, tmp, ptr1, tmp4);
381
- tcg_temp_free_i32(tmp);
382
+
383
if (a->op) {
384
- tmp = neon_load_reg(a->vd, 1);
385
+ read_neon_element32(tmp, a->vd, 1, MO_32);
386
} else {
387
- tmp = tcg_temp_new_i32();
388
tcg_gen_movi_i32(tmp, 0);
389
}
390
- tmp3 = neon_load_reg(a->vm, 1);
391
+ tmp3 = tcg_temp_new_i32();
392
+ read_neon_element32(tmp3, a->vm, 1, MO_32);
393
gen_helper_neon_tbl(tmp3, tmp3, tmp, ptr1, tmp4);
394
+ tcg_temp_free_i32(tmp);
395
tcg_temp_free_i32(tmp4);
396
tcg_temp_free_ptr(ptr1);
397
- neon_store_reg(a->vd, 0, tmp2);
398
- neon_store_reg(a->vd, 1, tmp3);
399
- tcg_temp_free_i32(tmp);
400
+
401
+ write_neon_element32(tmp2, a->vd, 0, MO_32);
402
+ write_neon_element32(tmp3, a->vd, 1, MO_32);
403
+ tcg_temp_free_i32(tmp2);
404
+ tcg_temp_free_i32(tmp3);
405
return true;
406
}
407
408
@@ -XXX,XX +XXX,XX @@ static bool trans_VDUP_scalar(DisasContext *s, arg_VDUP_scalar *a)
409
static bool trans_VREV64(DisasContext *s, arg_VREV64 *a)
410
{
411
int pass, half;
412
+ TCGv_i32 tmp[2];
413
414
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
415
return false;
416
@@ -XXX,XX +XXX,XX @@ static bool trans_VREV64(DisasContext *s, arg_VREV64 *a)
417
return true;
418
}
419
420
- for (pass = 0; pass < (a->q ? 2 : 1); pass++) {
421
- TCGv_i32 tmp[2];
422
+ tmp[0] = tcg_temp_new_i32();
423
+ tmp[1] = tcg_temp_new_i32();
424
425
+ for (pass = 0; pass < (a->q ? 2 : 1); pass++) {
426
for (half = 0; half < 2; half++) {
427
- tmp[half] = neon_load_reg(a->vm, pass * 2 + half);
428
+ read_neon_element32(tmp[half], a->vm, pass * 2 + half, MO_32);
429
switch (a->size) {
430
case 0:
431
tcg_gen_bswap32_i32(tmp[half], tmp[half]);
432
@@ -XXX,XX +XXX,XX @@ static bool trans_VREV64(DisasContext *s, arg_VREV64 *a)
433
g_assert_not_reached();
434
}
435
}
436
- neon_store_reg(a->vd, pass * 2, tmp[1]);
437
- neon_store_reg(a->vd, pass * 2 + 1, tmp[0]);
438
+ write_neon_element32(tmp[1], a->vd, pass * 2, MO_32);
439
+ write_neon_element32(tmp[0], a->vd, pass * 2 + 1, MO_32);
440
}
441
+
442
+ tcg_temp_free_i32(tmp[0]);
443
+ tcg_temp_free_i32(tmp[1]);
444
return true;
445
}
446
447
@@ -XXX,XX +XXX,XX @@ static bool do_2misc_pairwise(DisasContext *s, arg_2misc *a,
448
rm0_64 = tcg_temp_new_i64();
449
rm1_64 = tcg_temp_new_i64();
450
rd_64 = tcg_temp_new_i64();
451
- tmp = neon_load_reg(a->vm, pass * 2);
452
+
453
+ tmp = tcg_temp_new_i32();
454
+ read_neon_element32(tmp, a->vm, pass * 2, MO_32);
455
widenfn(rm0_64, tmp);
456
- tcg_temp_free_i32(tmp);
457
- tmp = neon_load_reg(a->vm, pass * 2 + 1);
458
+ read_neon_element32(tmp, a->vm, pass * 2 + 1, MO_32);
459
widenfn(rm1_64, tmp);
460
tcg_temp_free_i32(tmp);
461
+
462
opfn(rd_64, rm0_64, rm1_64);
463
tcg_temp_free_i64(rm0_64);
464
tcg_temp_free_i64(rm1_64);
465
@@ -XXX,XX +XXX,XX @@ static bool do_vmovn(DisasContext *s, arg_2misc *a,
466
narrowfn(rd0, cpu_env, rm);
467
neon_load_reg64(rm, a->vm + 1);
468
narrowfn(rd1, cpu_env, rm);
469
- neon_store_reg(a->vd, 0, rd0);
470
- neon_store_reg(a->vd, 1, rd1);
471
+ write_neon_element32(rd0, a->vd, 0, MO_32);
472
+ write_neon_element32(rd1, a->vd, 1, MO_32);
473
+ tcg_temp_free_i32(rd0);
474
+ tcg_temp_free_i32(rd1);
475
tcg_temp_free_i64(rm);
476
return true;
477
}
478
@@ -XXX,XX +XXX,XX @@ static bool trans_VSHLL(DisasContext *s, arg_2misc *a)
479
}
480
481
rd = tcg_temp_new_i64();
482
+ rm0 = tcg_temp_new_i32();
483
+ rm1 = tcg_temp_new_i32();
484
485
- rm0 = neon_load_reg(a->vm, 0);
486
- rm1 = neon_load_reg(a->vm, 1);
487
+ read_neon_element32(rm0, a->vm, 0, MO_32);
488
+ read_neon_element32(rm1, a->vm, 1, MO_32);
489
490
widenfn(rd, rm0);
491
tcg_gen_shli_i64(rd, rd, 8 << a->size);
492
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_F16_F32(DisasContext *s, arg_2misc *a)
493
494
fpst = fpstatus_ptr(FPST_STD);
495
ahp = get_ahp_flag();
496
- tmp = neon_load_reg(a->vm, 0);
497
+ tmp = tcg_temp_new_i32();
498
+ read_neon_element32(tmp, a->vm, 0, MO_32);
499
gen_helper_vfp_fcvt_f32_to_f16(tmp, tmp, fpst, ahp);
500
- tmp2 = neon_load_reg(a->vm, 1);
501
+ tmp2 = tcg_temp_new_i32();
502
+ read_neon_element32(tmp2, a->vm, 1, MO_32);
503
gen_helper_vfp_fcvt_f32_to_f16(tmp2, tmp2, fpst, ahp);
504
tcg_gen_shli_i32(tmp2, tmp2, 16);
505
tcg_gen_or_i32(tmp2, tmp2, tmp);
506
- tcg_temp_free_i32(tmp);
507
- tmp = neon_load_reg(a->vm, 2);
508
+ read_neon_element32(tmp, a->vm, 2, MO_32);
509
gen_helper_vfp_fcvt_f32_to_f16(tmp, tmp, fpst, ahp);
510
- tmp3 = neon_load_reg(a->vm, 3);
511
- neon_store_reg(a->vd, 0, tmp2);
512
+ tmp3 = tcg_temp_new_i32();
513
+ read_neon_element32(tmp3, a->vm, 3, MO_32);
514
+ write_neon_element32(tmp2, a->vd, 0, MO_32);
515
+ tcg_temp_free_i32(tmp2);
516
gen_helper_vfp_fcvt_f32_to_f16(tmp3, tmp3, fpst, ahp);
517
tcg_gen_shli_i32(tmp3, tmp3, 16);
518
tcg_gen_or_i32(tmp3, tmp3, tmp);
519
- neon_store_reg(a->vd, 1, tmp3);
520
+ write_neon_element32(tmp3, a->vd, 1, MO_32);
521
+ tcg_temp_free_i32(tmp3);
522
tcg_temp_free_i32(tmp);
523
tcg_temp_free_i32(ahp);
524
tcg_temp_free_ptr(fpst);
525
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_F32_F16(DisasContext *s, arg_2misc *a)
526
fpst = fpstatus_ptr(FPST_STD);
527
ahp = get_ahp_flag();
528
tmp3 = tcg_temp_new_i32();
529
- tmp = neon_load_reg(a->vm, 0);
530
- tmp2 = neon_load_reg(a->vm, 1);
531
+ tmp2 = tcg_temp_new_i32();
532
+ tmp = tcg_temp_new_i32();
533
+ read_neon_element32(tmp, a->vm, 0, MO_32);
534
+ read_neon_element32(tmp2, a->vm, 1, MO_32);
535
tcg_gen_ext16u_i32(tmp3, tmp);
536
gen_helper_vfp_fcvt_f16_to_f32(tmp3, tmp3, fpst, ahp);
537
- neon_store_reg(a->vd, 0, tmp3);
538
+ write_neon_element32(tmp3, a->vd, 0, MO_32);
539
tcg_gen_shri_i32(tmp, tmp, 16);
540
gen_helper_vfp_fcvt_f16_to_f32(tmp, tmp, fpst, ahp);
541
- neon_store_reg(a->vd, 1, tmp);
542
- tmp3 = tcg_temp_new_i32();
543
+ write_neon_element32(tmp, a->vd, 1, MO_32);
544
+ tcg_temp_free_i32(tmp);
545
tcg_gen_ext16u_i32(tmp3, tmp2);
546
gen_helper_vfp_fcvt_f16_to_f32(tmp3, tmp3, fpst, ahp);
547
- neon_store_reg(a->vd, 2, tmp3);
548
+ write_neon_element32(tmp3, a->vd, 2, MO_32);
549
+ tcg_temp_free_i32(tmp3);
550
tcg_gen_shri_i32(tmp2, tmp2, 16);
551
gen_helper_vfp_fcvt_f16_to_f32(tmp2, tmp2, fpst, ahp);
552
- neon_store_reg(a->vd, 3, tmp2);
553
+ write_neon_element32(tmp2, a->vd, 3, MO_32);
554
+ tcg_temp_free_i32(tmp2);
555
tcg_temp_free_i32(ahp);
556
tcg_temp_free_ptr(fpst);
557
558
@@ -XXX,XX +XXX,XX @@ DO_2M_CRYPTO(SHA256SU0, aa32_sha2, 2)
559
560
static bool do_2misc(DisasContext *s, arg_2misc *a, NeonGenOneOpFn *fn)
561
{
562
+ TCGv_i32 tmp;
563
int pass;
564
565
/* Handle a 2-reg-misc operation by iterating 32 bits at a time */
566
@@ -XXX,XX +XXX,XX @@ static bool do_2misc(DisasContext *s, arg_2misc *a, NeonGenOneOpFn *fn)
567
return true;
568
}
569
570
+ tmp = tcg_temp_new_i32();
571
for (pass = 0; pass < (a->q ? 4 : 2); pass++) {
572
- TCGv_i32 tmp = neon_load_reg(a->vm, pass);
573
+ read_neon_element32(tmp, a->vm, pass, MO_32);
574
fn(tmp, tmp);
575
- neon_store_reg(a->vd, pass, tmp);
576
+ write_neon_element32(tmp, a->vd, pass, MO_32);
577
}
578
+ tcg_temp_free_i32(tmp);
579
580
return true;
581
}
582
@@ -XXX,XX +XXX,XX @@ static bool trans_VTRN(DisasContext *s, arg_2misc *a)
583
return true;
584
}
585
586
- if (a->size == 2) {
587
+ tmp = tcg_temp_new_i32();
588
+ tmp2 = tcg_temp_new_i32();
589
+ if (a->size == MO_32) {
590
for (pass = 0; pass < (a->q ? 4 : 2); pass += 2) {
591
- tmp = neon_load_reg(a->vm, pass);
592
- tmp2 = neon_load_reg(a->vd, pass + 1);
593
- neon_store_reg(a->vm, pass, tmp2);
594
- neon_store_reg(a->vd, pass + 1, tmp);
595
+ read_neon_element32(tmp, a->vm, pass, MO_32);
596
+ read_neon_element32(tmp2, a->vd, pass + 1, MO_32);
597
+ write_neon_element32(tmp2, a->vm, pass, MO_32);
598
+ write_neon_element32(tmp, a->vd, pass + 1, MO_32);
599
}
600
} else {
601
for (pass = 0; pass < (a->q ? 4 : 2); pass++) {
602
- tmp = neon_load_reg(a->vm, pass);
603
- tmp2 = neon_load_reg(a->vd, pass);
604
- if (a->size == 0) {
605
+ read_neon_element32(tmp, a->vm, pass, MO_32);
606
+ read_neon_element32(tmp2, a->vd, pass, MO_32);
607
+ if (a->size == MO_8) {
608
gen_neon_trn_u8(tmp, tmp2);
609
} else {
610
gen_neon_trn_u16(tmp, tmp2);
611
}
612
- neon_store_reg(a->vm, pass, tmp2);
613
- neon_store_reg(a->vd, pass, tmp);
614
+ write_neon_element32(tmp2, a->vm, pass, MO_32);
615
+ write_neon_element32(tmp, a->vd, pass, MO_32);
616
}
617
}
618
+ tcg_temp_free_i32(tmp);
619
+ tcg_temp_free_i32(tmp2);
620
return true;
621
}
124
--
622
--
125
2.7.4
623
2.20.1
126
624
127
625
diff view generated by jsdifflib
1
Make the CONTROL register banked if v8M security extensions are enabled.
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We can then use this to improve VMOV (scalar to gp) and
4
VMOV (gp to scalar) so that we simply perform the memory
5
operation that we wanted, rather than inserting or
6
extracting from a 32-bit quantity.
7
8
These were the last uses of neon_load/store_reg, so remove them.
9
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20201030022618.785675-7-richard.henderson@linaro.org
12
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 1503414539-28762-10-git-send-email-peter.maydell@linaro.org
6
---
14
---
7
target/arm/cpu.h | 5 +++--
15
target/arm/translate.c | 50 +++++++++++++-----------
8
target/arm/helper.c | 21 +++++++++++----------
16
target/arm/translate-vfp.c.inc | 71 +++++-----------------------------
9
target/arm/machine.c | 3 ++-
17
2 files changed, 37 insertions(+), 84 deletions(-)
10
target/arm/translate.c | 2 +-
18
11
4 files changed, 17 insertions(+), 14 deletions(-)
12
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.h
16
+++ b/target/arm/cpu.h
17
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
18
uint32_t other_sp;
19
uint32_t vecbase;
20
uint32_t basepri[2];
21
- uint32_t control;
22
+ uint32_t control[2];
23
uint32_t ccr; /* Configuration and Control */
24
uint32_t cfsr; /* Configurable Fault Status */
25
uint32_t hfsr; /* HardFault Status */
26
@@ -XXX,XX +XXX,XX @@ static inline bool arm_v7m_is_handler_mode(CPUARMState *env)
27
static inline int arm_current_el(CPUARMState *env)
28
{
29
if (arm_feature(env, ARM_FEATURE_M)) {
30
- return arm_v7m_is_handler_mode(env) || !(env->v7m.control & 1);
31
+ return arm_v7m_is_handler_mode(env) ||
32
+ !(env->v7m.control[env->v7m.secure] & 1);
33
}
34
35
if (is_a64(env)) {
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/helper.c
39
+++ b/target/arm/helper.c
40
@@ -XXX,XX +XXX,XX @@ static uint32_t v7m_pop(CPUARMState *env)
41
static void switch_v7m_sp(CPUARMState *env, bool new_spsel)
42
{
43
uint32_t tmp;
44
- bool old_spsel = env->v7m.control & R_V7M_CONTROL_SPSEL_MASK;
45
+ uint32_t old_control = env->v7m.control[env->v7m.secure];
46
+ bool old_spsel = old_control & R_V7M_CONTROL_SPSEL_MASK;
47
48
if (old_spsel != new_spsel) {
49
tmp = env->v7m.other_sp;
50
env->v7m.other_sp = env->regs[13];
51
env->regs[13] = tmp;
52
53
- env->v7m.control = deposit32(env->v7m.control,
54
+ env->v7m.control[env->v7m.secure] = deposit32(old_control,
55
R_V7M_CONTROL_SPSEL_SHIFT,
56
R_V7M_CONTROL_SPSEL_LENGTH, new_spsel);
57
}
58
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
59
}
60
61
lr = 0xfffffff1;
62
- if (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) {
63
+ if (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK) {
64
lr |= 4;
65
}
66
if (!arm_v7m_is_handler_mode(env)) {
67
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
68
return xpsr_read(env) & mask;
69
break;
70
case 20: /* CONTROL */
71
- return env->v7m.control;
72
+ return env->v7m.control[env->v7m.secure];
73
}
74
75
if (el == 0) {
76
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
77
78
switch (reg) {
79
case 8: /* MSP */
80
- return (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) ?
81
+ return (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK) ?
82
env->v7m.other_sp : env->regs[13];
83
case 9: /* PSP */
84
- return (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) ?
85
+ return (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK) ?
86
env->regs[13] : env->v7m.other_sp;
87
case 16: /* PRIMASK */
88
return env->v7m.primask[env->v7m.secure];
89
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
90
}
91
break;
92
case 8: /* MSP */
93
- if (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) {
94
+ if (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK) {
95
env->v7m.other_sp = val;
96
} else {
97
env->regs[13] = val;
98
}
99
break;
100
case 9: /* PSP */
101
- if (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) {
102
+ if (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_SPSEL_MASK) {
103
env->regs[13] = val;
104
} else {
105
env->v7m.other_sp = val;
106
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
107
if (!arm_v7m_is_handler_mode(env)) {
108
switch_v7m_sp(env, (val & R_V7M_CONTROL_SPSEL_MASK) != 0);
109
}
110
- env->v7m.control &= ~R_V7M_CONTROL_NPRIV_MASK;
111
- env->v7m.control |= val & R_V7M_CONTROL_NPRIV_MASK;
112
+ env->v7m.control[env->v7m.secure] &= ~R_V7M_CONTROL_NPRIV_MASK;
113
+ env->v7m.control[env->v7m.secure] |= val & R_V7M_CONTROL_NPRIV_MASK;
114
break;
115
default:
116
qemu_log_mask(LOG_GUEST_ERROR, "Attempt to write unknown special"
117
diff --git a/target/arm/machine.c b/target/arm/machine.c
118
index XXXXXXX..XXXXXXX 100644
119
--- a/target/arm/machine.c
120
+++ b/target/arm/machine.c
121
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
122
.fields = (VMStateField[]) {
123
VMSTATE_UINT32(env.v7m.vecbase, ARMCPU),
124
VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
125
- VMSTATE_UINT32(env.v7m.control, ARMCPU),
126
+ VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
127
VMSTATE_UINT32(env.v7m.ccr, ARMCPU),
128
VMSTATE_UINT32(env.v7m.cfsr, ARMCPU),
129
VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
130
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
131
VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
132
VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
133
VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
134
+ VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
135
VMSTATE_END_OF_LIST()
136
}
137
};
138
diff --git a/target/arm/translate.c b/target/arm/translate.c
19
diff --git a/target/arm/translate.c b/target/arm/translate.c
139
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
140
--- a/target/arm/translate.c
21
--- a/target/arm/translate.c
141
+++ b/target/arm/translate.c
22
+++ b/target/arm/translate.c
142
@@ -XXX,XX +XXX,XX @@ void arm_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
23
@@ -XXX,XX +XXX,XX @@ static long neon_full_reg_offset(unsigned reg)
143
if (xpsr & XPSR_EXCP) {
24
* Return the offset of a 2**SIZE piece of a NEON register, at index ELE,
144
mode = "handler";
25
* where 0 is the least significant end of the register.
145
} else {
26
*/
146
- if (env->v7m.control & R_V7M_CONTROL_NPRIV_MASK) {
27
-static long neon_element_offset(int reg, int element, MemOp size)
147
+ if (env->v7m.control[env->v7m.secure] & R_V7M_CONTROL_NPRIV_MASK) {
28
+static long neon_element_offset(int reg, int element, MemOp memop)
148
mode = "unpriv-thread";
29
{
149
} else {
30
- int element_size = 1 << size;
150
mode = "priv-thread";
31
+ int element_size = 1 << (memop & MO_SIZE);
32
int ofs = element * element_size;
33
#ifdef HOST_WORDS_BIGENDIAN
34
/*
35
@@ -XXX,XX +XXX,XX @@ static long vfp_reg_offset(bool dp, unsigned reg)
36
}
37
}
38
39
-static TCGv_i32 neon_load_reg(int reg, int pass)
40
-{
41
- TCGv_i32 tmp = tcg_temp_new_i32();
42
- tcg_gen_ld_i32(tmp, cpu_env, neon_element_offset(reg, pass, MO_32));
43
- return tmp;
44
-}
45
-
46
-static void neon_store_reg(int reg, int pass, TCGv_i32 var)
47
-{
48
- tcg_gen_st_i32(var, cpu_env, neon_element_offset(reg, pass, MO_32));
49
- tcg_temp_free_i32(var);
50
-}
51
-
52
static inline void neon_load_reg64(TCGv_i64 var, int reg)
53
{
54
tcg_gen_ld_i64(var, cpu_env, vfp_reg_offset(1, reg));
55
@@ -XXX,XX +XXX,XX @@ static inline void neon_store_reg32(TCGv_i32 var, int reg)
56
tcg_gen_st_i32(var, cpu_env, vfp_reg_offset(false, reg));
57
}
58
59
-static void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp size)
60
+static void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp memop)
61
{
62
- long off = neon_element_offset(reg, ele, size);
63
+ long off = neon_element_offset(reg, ele, memop);
64
65
- switch (size) {
66
- case MO_32:
67
+ switch (memop) {
68
+ case MO_SB:
69
+ tcg_gen_ld8s_i32(dest, cpu_env, off);
70
+ break;
71
+ case MO_UB:
72
+ tcg_gen_ld8u_i32(dest, cpu_env, off);
73
+ break;
74
+ case MO_SW:
75
+ tcg_gen_ld16s_i32(dest, cpu_env, off);
76
+ break;
77
+ case MO_UW:
78
+ tcg_gen_ld16u_i32(dest, cpu_env, off);
79
+ break;
80
+ case MO_UL:
81
+ case MO_SL:
82
tcg_gen_ld_i32(dest, cpu_env, off);
83
break;
84
default:
85
@@ -XXX,XX +XXX,XX @@ static void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp size)
86
}
87
}
88
89
-static void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp size)
90
+static void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp memop)
91
{
92
- long off = neon_element_offset(reg, ele, size);
93
+ long off = neon_element_offset(reg, ele, memop);
94
95
- switch (size) {
96
+ switch (memop) {
97
+ case MO_8:
98
+ tcg_gen_st8_i32(src, cpu_env, off);
99
+ break;
100
+ case MO_16:
101
+ tcg_gen_st16_i32(src, cpu_env, off);
102
+ break;
103
case MO_32:
104
tcg_gen_st_i32(src, cpu_env, off);
105
break;
106
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
107
index XXXXXXX..XXXXXXX 100644
108
--- a/target/arm/translate-vfp.c.inc
109
+++ b/target/arm/translate-vfp.c.inc
110
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_to_gp(DisasContext *s, arg_VMOV_to_gp *a)
111
{
112
/* VMOV scalar to general purpose register */
113
TCGv_i32 tmp;
114
- int pass;
115
- uint32_t offset;
116
117
- /* SIZE == 2 is a VFP instruction; otherwise NEON. */
118
- if (a->size == 2
119
+ /* SIZE == MO_32 is a VFP instruction; otherwise NEON. */
120
+ if (a->size == MO_32
121
? !dc_isar_feature(aa32_fpsp_v2, s)
122
: !arm_dc_feature(s, ARM_FEATURE_NEON)) {
123
return false;
124
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_to_gp(DisasContext *s, arg_VMOV_to_gp *a)
125
return false;
126
}
127
128
- offset = a->index << a->size;
129
- pass = extract32(offset, 2, 1);
130
- offset = extract32(offset, 0, 2) * 8;
131
-
132
if (!vfp_access_check(s)) {
133
return true;
134
}
135
136
- tmp = neon_load_reg(a->vn, pass);
137
- switch (a->size) {
138
- case 0:
139
- if (offset) {
140
- tcg_gen_shri_i32(tmp, tmp, offset);
141
- }
142
- if (a->u) {
143
- gen_uxtb(tmp);
144
- } else {
145
- gen_sxtb(tmp);
146
- }
147
- break;
148
- case 1:
149
- if (a->u) {
150
- if (offset) {
151
- tcg_gen_shri_i32(tmp, tmp, 16);
152
- } else {
153
- gen_uxth(tmp);
154
- }
155
- } else {
156
- if (offset) {
157
- tcg_gen_sari_i32(tmp, tmp, 16);
158
- } else {
159
- gen_sxth(tmp);
160
- }
161
- }
162
- break;
163
- case 2:
164
- break;
165
- }
166
+ tmp = tcg_temp_new_i32();
167
+ read_neon_element32(tmp, a->vn, a->index, a->size | (a->u ? 0 : MO_SIGN));
168
store_reg(s, a->rt, tmp);
169
170
return true;
171
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_to_gp(DisasContext *s, arg_VMOV_to_gp *a)
172
static bool trans_VMOV_from_gp(DisasContext *s, arg_VMOV_from_gp *a)
173
{
174
/* VMOV general purpose register to scalar */
175
- TCGv_i32 tmp, tmp2;
176
- int pass;
177
- uint32_t offset;
178
+ TCGv_i32 tmp;
179
180
- /* SIZE == 2 is a VFP instruction; otherwise NEON. */
181
- if (a->size == 2
182
+ /* SIZE == MO_32 is a VFP instruction; otherwise NEON. */
183
+ if (a->size == MO_32
184
? !dc_isar_feature(aa32_fpsp_v2, s)
185
: !arm_dc_feature(s, ARM_FEATURE_NEON)) {
186
return false;
187
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_from_gp(DisasContext *s, arg_VMOV_from_gp *a)
188
return false;
189
}
190
191
- offset = a->index << a->size;
192
- pass = extract32(offset, 2, 1);
193
- offset = extract32(offset, 0, 2) * 8;
194
-
195
if (!vfp_access_check(s)) {
196
return true;
197
}
198
199
tmp = load_reg(s, a->rt);
200
- switch (a->size) {
201
- case 0:
202
- tmp2 = neon_load_reg(a->vn, pass);
203
- tcg_gen_deposit_i32(tmp, tmp2, tmp, offset, 8);
204
- tcg_temp_free_i32(tmp2);
205
- break;
206
- case 1:
207
- tmp2 = neon_load_reg(a->vn, pass);
208
- tcg_gen_deposit_i32(tmp, tmp2, tmp, offset, 16);
209
- tcg_temp_free_i32(tmp2);
210
- break;
211
- case 2:
212
- break;
213
- }
214
- neon_store_reg(a->vn, pass, tmp);
215
+ write_neon_element32(tmp, a->vn, a->index, a->size);
216
+ tcg_temp_free_i32(tmp);
217
218
return true;
219
}
151
--
220
--
152
2.7.4
221
2.20.1
153
222
154
223
diff view generated by jsdifflib
1
As the first step in implementing ARM v8M's security extension:
1
From: Richard Henderson <richard.henderson@linaro.org>
2
* add a new feature bit ARM_FEATURE_M_SECURITY
3
* add the CPU state field that indicates whether the CPU is
4
currently in the secure state
5
* add a migration subsection for this new state
6
(we will add the Secure copies of banked register state
7
to this subsection in later patches)
8
* add a #define for the one new-in-v8M exception type
9
* make the CPU debug log print S/NS status
10
2
3
The only uses of this function are for loading VFP
4
single-precision values, and nothing to do with NEON.
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201030022618.785675-8-richard.henderson@linaro.org
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 1503414539-28762-4-git-send-email-peter.maydell@linaro.org
14
---
10
---
15
target/arm/cpu.h | 3 +++
11
target/arm/translate.c | 4 +-
16
target/arm/cpu.c | 4 ++++
12
target/arm/translate-vfp.c.inc | 184 ++++++++++++++++-----------------
17
target/arm/machine.c | 20 ++++++++++++++++++++
13
2 files changed, 94 insertions(+), 94 deletions(-)
18
target/arm/translate.c | 8 +++++++-
19
4 files changed, 34 insertions(+), 1 deletion(-)
20
14
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
22
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
25
@@ -XXX,XX +XXX,XX @@
26
#define ARMV7M_EXCP_MEM 4
27
#define ARMV7M_EXCP_BUS 5
28
#define ARMV7M_EXCP_USAGE 6
29
+#define ARMV7M_EXCP_SECURE 7
30
#define ARMV7M_EXCP_SVC 11
31
#define ARMV7M_EXCP_DEBUG 12
32
#define ARMV7M_EXCP_PENDSV 14
33
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
34
int exception;
35
uint32_t primask;
36
uint32_t faultmask;
37
+ uint32_t secure; /* Is CPU in Secure state? (not guest visible) */
38
} v7m;
39
40
/* Information associated with an exception about to be taken:
41
@@ -XXX,XX +XXX,XX @@ enum arm_features {
42
ARM_FEATURE_THUMB_DSP, /* DSP insns supported in the Thumb encodings */
43
ARM_FEATURE_PMU, /* has PMU support */
44
ARM_FEATURE_VBAR, /* has cp15 VBAR */
45
+ ARM_FEATURE_M_SECURITY, /* M profile Security Extension */
46
};
47
48
static inline int arm_feature(CPUARMState *env, int feature)
49
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
50
index XXXXXXX..XXXXXXX 100644
51
--- a/target/arm/cpu.c
52
+++ b/target/arm/cpu.c
53
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
54
uint32_t initial_pc; /* Loaded from 0x4 */
55
uint8_t *rom;
56
57
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
58
+ env->v7m.secure = true;
59
+ }
60
+
61
/* The reset value of this bit is IMPDEF, but ARM recommends
62
* that it resets to 1, so QEMU always does that rather than making
63
* it dependent on CPU model.
64
diff --git a/target/arm/machine.c b/target/arm/machine.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/machine.c
67
+++ b/target/arm/machine.c
68
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav8 = {
69
}
70
};
71
72
+static bool m_security_needed(void *opaque)
73
+{
74
+ ARMCPU *cpu = opaque;
75
+ CPUARMState *env = &cpu->env;
76
+
77
+ return arm_feature(env, ARM_FEATURE_M_SECURITY);
78
+}
79
+
80
+static const VMStateDescription vmstate_m_security = {
81
+ .name = "cpu/m-security",
82
+ .version_id = 1,
83
+ .minimum_version_id = 1,
84
+ .needed = m_security_needed,
85
+ .fields = (VMStateField[]) {
86
+ VMSTATE_UINT32(env.v7m.secure, ARMCPU),
87
+ VMSTATE_END_OF_LIST()
88
+ }
89
+};
90
+
91
static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
92
VMStateField *field)
93
{
94
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_arm_cpu = {
95
&vmstate_pmsav7_rnr,
96
&vmstate_pmsav7,
97
&vmstate_pmsav8,
98
+ &vmstate_m_security,
99
NULL
100
}
101
};
102
diff --git a/target/arm/translate.c b/target/arm/translate.c
15
diff --git a/target/arm/translate.c b/target/arm/translate.c
103
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
104
--- a/target/arm/translate.c
17
--- a/target/arm/translate.c
105
+++ b/target/arm/translate.c
18
+++ b/target/arm/translate.c
106
@@ -XXX,XX +XXX,XX @@ void arm_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
19
@@ -XXX,XX +XXX,XX @@ static inline void neon_store_reg64(TCGv_i64 var, int reg)
107
if (arm_feature(env, ARM_FEATURE_M)) {
20
tcg_gen_st_i64(var, cpu_env, vfp_reg_offset(1, reg));
108
uint32_t xpsr = xpsr_read(env);
21
}
109
const char *mode;
22
110
+ const char *ns_status = "";
23
-static inline void neon_load_reg32(TCGv_i32 var, int reg)
111
+
24
+static inline void vfp_load_reg32(TCGv_i32 var, int reg)
112
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
25
{
113
+ ns_status = env->v7m.secure ? "S " : "NS ";
26
tcg_gen_ld_i32(var, cpu_env, vfp_reg_offset(false, reg));
114
+ }
27
}
115
28
116
if (xpsr & XPSR_EXCP) {
29
-static inline void neon_store_reg32(TCGv_i32 var, int reg)
117
mode = "handler";
30
+static inline void vfp_store_reg32(TCGv_i32 var, int reg)
118
@@ -XXX,XX +XXX,XX @@ void arm_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
31
{
32
tcg_gen_st_i32(var, cpu_env, vfp_reg_offset(false, reg));
33
}
34
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
35
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/translate-vfp.c.inc
37
+++ b/target/arm/translate-vfp.c.inc
38
@@ -XXX,XX +XXX,XX @@ static bool trans_VSEL(DisasContext *s, arg_VSEL *a)
39
frn = tcg_temp_new_i32();
40
frm = tcg_temp_new_i32();
41
dest = tcg_temp_new_i32();
42
- neon_load_reg32(frn, rn);
43
- neon_load_reg32(frm, rm);
44
+ vfp_load_reg32(frn, rn);
45
+ vfp_load_reg32(frm, rm);
46
switch (a->cc) {
47
case 0: /* eq: Z */
48
tcg_gen_movcond_i32(TCG_COND_EQ, dest, cpu_ZF, zero,
49
@@ -XXX,XX +XXX,XX @@ static bool trans_VSEL(DisasContext *s, arg_VSEL *a)
50
if (sz == 1) {
51
tcg_gen_andi_i32(dest, dest, 0xffff);
52
}
53
- neon_store_reg32(dest, rd);
54
+ vfp_store_reg32(dest, rd);
55
tcg_temp_free_i32(frn);
56
tcg_temp_free_i32(frm);
57
tcg_temp_free_i32(dest);
58
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINT(DisasContext *s, arg_VRINT *a)
59
TCGv_i32 tcg_res;
60
tcg_op = tcg_temp_new_i32();
61
tcg_res = tcg_temp_new_i32();
62
- neon_load_reg32(tcg_op, rm);
63
+ vfp_load_reg32(tcg_op, rm);
64
if (sz == 1) {
65
gen_helper_rinth(tcg_res, tcg_op, fpst);
66
} else {
67
gen_helper_rints(tcg_res, tcg_op, fpst);
68
}
69
- neon_store_reg32(tcg_res, rd);
70
+ vfp_store_reg32(tcg_res, rd);
71
tcg_temp_free_i32(tcg_op);
72
tcg_temp_free_i32(tcg_res);
73
}
74
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT(DisasContext *s, arg_VCVT *a)
75
gen_helper_vfp_tould(tcg_res, tcg_double, tcg_shift, fpst);
76
}
77
tcg_gen_extrl_i64_i32(tcg_tmp, tcg_res);
78
- neon_store_reg32(tcg_tmp, rd);
79
+ vfp_store_reg32(tcg_tmp, rd);
80
tcg_temp_free_i32(tcg_tmp);
81
tcg_temp_free_i64(tcg_res);
82
tcg_temp_free_i64(tcg_double);
83
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT(DisasContext *s, arg_VCVT *a)
84
TCGv_i32 tcg_single, tcg_res;
85
tcg_single = tcg_temp_new_i32();
86
tcg_res = tcg_temp_new_i32();
87
- neon_load_reg32(tcg_single, rm);
88
+ vfp_load_reg32(tcg_single, rm);
89
if (sz == 1) {
90
if (is_signed) {
91
gen_helper_vfp_toslh(tcg_res, tcg_single, tcg_shift, fpst);
92
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT(DisasContext *s, arg_VCVT *a)
93
gen_helper_vfp_touls(tcg_res, tcg_single, tcg_shift, fpst);
119
}
94
}
120
}
95
}
121
96
- neon_store_reg32(tcg_res, rd);
122
- cpu_fprintf(f, "XPSR=%08x %c%c%c%c %c %s\n",
97
+ vfp_store_reg32(tcg_res, rd);
123
+ cpu_fprintf(f, "XPSR=%08x %c%c%c%c %c %s%s\n",
98
tcg_temp_free_i32(tcg_res);
124
xpsr,
99
tcg_temp_free_i32(tcg_single);
125
xpsr & XPSR_N ? 'N' : '-',
100
}
126
xpsr & XPSR_Z ? 'Z' : '-',
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_half(DisasContext *s, arg_VMOV_single *a)
127
xpsr & XPSR_C ? 'C' : '-',
102
if (a->l) {
128
xpsr & XPSR_V ? 'V' : '-',
103
/* VFP to general purpose register */
129
xpsr & XPSR_T ? 'T' : 'A',
104
tmp = tcg_temp_new_i32();
130
+ ns_status,
105
- neon_load_reg32(tmp, a->vn);
131
mode);
106
+ vfp_load_reg32(tmp, a->vn);
132
} else {
107
tcg_gen_andi_i32(tmp, tmp, 0xffff);
133
uint32_t psr = cpsr_read(env);
108
store_reg(s, a->rt, tmp);
109
} else {
110
/* general purpose register to VFP */
111
tmp = load_reg(s, a->rt);
112
tcg_gen_andi_i32(tmp, tmp, 0xffff);
113
- neon_store_reg32(tmp, a->vn);
114
+ vfp_store_reg32(tmp, a->vn);
115
tcg_temp_free_i32(tmp);
116
}
117
118
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_single(DisasContext *s, arg_VMOV_single *a)
119
if (a->l) {
120
/* VFP to general purpose register */
121
tmp = tcg_temp_new_i32();
122
- neon_load_reg32(tmp, a->vn);
123
+ vfp_load_reg32(tmp, a->vn);
124
if (a->rt == 15) {
125
/* Set the 4 flag bits in the CPSR. */
126
gen_set_nzcv(tmp);
127
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_single(DisasContext *s, arg_VMOV_single *a)
128
} else {
129
/* general purpose register to VFP */
130
tmp = load_reg(s, a->rt);
131
- neon_store_reg32(tmp, a->vn);
132
+ vfp_store_reg32(tmp, a->vn);
133
tcg_temp_free_i32(tmp);
134
}
135
136
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_64_sp(DisasContext *s, arg_VMOV_64_sp *a)
137
if (a->op) {
138
/* fpreg to gpreg */
139
tmp = tcg_temp_new_i32();
140
- neon_load_reg32(tmp, a->vm);
141
+ vfp_load_reg32(tmp, a->vm);
142
store_reg(s, a->rt, tmp);
143
tmp = tcg_temp_new_i32();
144
- neon_load_reg32(tmp, a->vm + 1);
145
+ vfp_load_reg32(tmp, a->vm + 1);
146
store_reg(s, a->rt2, tmp);
147
} else {
148
/* gpreg to fpreg */
149
tmp = load_reg(s, a->rt);
150
- neon_store_reg32(tmp, a->vm);
151
+ vfp_store_reg32(tmp, a->vm);
152
tcg_temp_free_i32(tmp);
153
tmp = load_reg(s, a->rt2);
154
- neon_store_reg32(tmp, a->vm + 1);
155
+ vfp_store_reg32(tmp, a->vm + 1);
156
tcg_temp_free_i32(tmp);
157
}
158
159
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_64_dp(DisasContext *s, arg_VMOV_64_dp *a)
160
if (a->op) {
161
/* fpreg to gpreg */
162
tmp = tcg_temp_new_i32();
163
- neon_load_reg32(tmp, a->vm * 2);
164
+ vfp_load_reg32(tmp, a->vm * 2);
165
store_reg(s, a->rt, tmp);
166
tmp = tcg_temp_new_i32();
167
- neon_load_reg32(tmp, a->vm * 2 + 1);
168
+ vfp_load_reg32(tmp, a->vm * 2 + 1);
169
store_reg(s, a->rt2, tmp);
170
} else {
171
/* gpreg to fpreg */
172
tmp = load_reg(s, a->rt);
173
- neon_store_reg32(tmp, a->vm * 2);
174
+ vfp_store_reg32(tmp, a->vm * 2);
175
tcg_temp_free_i32(tmp);
176
tmp = load_reg(s, a->rt2);
177
- neon_store_reg32(tmp, a->vm * 2 + 1);
178
+ vfp_store_reg32(tmp, a->vm * 2 + 1);
179
tcg_temp_free_i32(tmp);
180
}
181
182
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDR_VSTR_hp(DisasContext *s, arg_VLDR_VSTR_sp *a)
183
tmp = tcg_temp_new_i32();
184
if (a->l) {
185
gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
186
- neon_store_reg32(tmp, a->vd);
187
+ vfp_store_reg32(tmp, a->vd);
188
} else {
189
- neon_load_reg32(tmp, a->vd);
190
+ vfp_load_reg32(tmp, a->vd);
191
gen_aa32_st16(s, tmp, addr, get_mem_index(s));
192
}
193
tcg_temp_free_i32(tmp);
194
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDR_VSTR_sp(DisasContext *s, arg_VLDR_VSTR_sp *a)
195
tmp = tcg_temp_new_i32();
196
if (a->l) {
197
gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
198
- neon_store_reg32(tmp, a->vd);
199
+ vfp_store_reg32(tmp, a->vd);
200
} else {
201
- neon_load_reg32(tmp, a->vd);
202
+ vfp_load_reg32(tmp, a->vd);
203
gen_aa32_st32(s, tmp, addr, get_mem_index(s));
204
}
205
tcg_temp_free_i32(tmp);
206
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDM_VSTM_sp(DisasContext *s, arg_VLDM_VSTM_sp *a)
207
if (a->l) {
208
/* load */
209
gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
210
- neon_store_reg32(tmp, a->vd + i);
211
+ vfp_store_reg32(tmp, a->vd + i);
212
} else {
213
/* store */
214
- neon_load_reg32(tmp, a->vd + i);
215
+ vfp_load_reg32(tmp, a->vd + i);
216
gen_aa32_st32(s, tmp, addr, get_mem_index(s));
217
}
218
tcg_gen_addi_i32(addr, addr, offset);
219
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_3op_sp(DisasContext *s, VFPGen3OpSPFn *fn,
220
fd = tcg_temp_new_i32();
221
fpst = fpstatus_ptr(FPST_FPCR);
222
223
- neon_load_reg32(f0, vn);
224
- neon_load_reg32(f1, vm);
225
+ vfp_load_reg32(f0, vn);
226
+ vfp_load_reg32(f1, vm);
227
228
for (;;) {
229
if (reads_vd) {
230
- neon_load_reg32(fd, vd);
231
+ vfp_load_reg32(fd, vd);
232
}
233
fn(fd, f0, f1, fpst);
234
- neon_store_reg32(fd, vd);
235
+ vfp_store_reg32(fd, vd);
236
237
if (veclen == 0) {
238
break;
239
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_3op_sp(DisasContext *s, VFPGen3OpSPFn *fn,
240
veclen--;
241
vd = vfp_advance_sreg(vd, delta_d);
242
vn = vfp_advance_sreg(vn, delta_d);
243
- neon_load_reg32(f0, vn);
244
+ vfp_load_reg32(f0, vn);
245
if (delta_m) {
246
vm = vfp_advance_sreg(vm, delta_m);
247
- neon_load_reg32(f1, vm);
248
+ vfp_load_reg32(f1, vm);
249
}
250
}
251
252
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_3op_hp(DisasContext *s, VFPGen3OpSPFn *fn,
253
fd = tcg_temp_new_i32();
254
fpst = fpstatus_ptr(FPST_FPCR_F16);
255
256
- neon_load_reg32(f0, vn);
257
- neon_load_reg32(f1, vm);
258
+ vfp_load_reg32(f0, vn);
259
+ vfp_load_reg32(f1, vm);
260
261
if (reads_vd) {
262
- neon_load_reg32(fd, vd);
263
+ vfp_load_reg32(fd, vd);
264
}
265
fn(fd, f0, f1, fpst);
266
- neon_store_reg32(fd, vd);
267
+ vfp_store_reg32(fd, vd);
268
269
tcg_temp_free_i32(f0);
270
tcg_temp_free_i32(f1);
271
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_sp(DisasContext *s, VFPGen2OpSPFn *fn, int vd, int vm)
272
f0 = tcg_temp_new_i32();
273
fd = tcg_temp_new_i32();
274
275
- neon_load_reg32(f0, vm);
276
+ vfp_load_reg32(f0, vm);
277
278
for (;;) {
279
fn(fd, f0);
280
- neon_store_reg32(fd, vd);
281
+ vfp_store_reg32(fd, vd);
282
283
if (veclen == 0) {
284
break;
285
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_sp(DisasContext *s, VFPGen2OpSPFn *fn, int vd, int vm)
286
/* single source one-many */
287
while (veclen--) {
288
vd = vfp_advance_sreg(vd, delta_d);
289
- neon_store_reg32(fd, vd);
290
+ vfp_store_reg32(fd, vd);
291
}
292
break;
293
}
294
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_sp(DisasContext *s, VFPGen2OpSPFn *fn, int vd, int vm)
295
veclen--;
296
vd = vfp_advance_sreg(vd, delta_d);
297
vm = vfp_advance_sreg(vm, delta_m);
298
- neon_load_reg32(f0, vm);
299
+ vfp_load_reg32(f0, vm);
300
}
301
302
tcg_temp_free_i32(f0);
303
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_hp(DisasContext *s, VFPGen2OpSPFn *fn, int vd, int vm)
304
}
305
306
f0 = tcg_temp_new_i32();
307
- neon_load_reg32(f0, vm);
308
+ vfp_load_reg32(f0, vm);
309
fn(f0, f0);
310
- neon_store_reg32(f0, vd);
311
+ vfp_store_reg32(f0, vd);
312
tcg_temp_free_i32(f0);
313
314
return true;
315
@@ -XXX,XX +XXX,XX @@ static bool do_vfm_hp(DisasContext *s, arg_VFMA_sp *a, bool neg_n, bool neg_d)
316
vm = tcg_temp_new_i32();
317
vd = tcg_temp_new_i32();
318
319
- neon_load_reg32(vn, a->vn);
320
- neon_load_reg32(vm, a->vm);
321
+ vfp_load_reg32(vn, a->vn);
322
+ vfp_load_reg32(vm, a->vm);
323
if (neg_n) {
324
/* VFNMS, VFMS */
325
gen_helper_vfp_negh(vn, vn);
326
}
327
- neon_load_reg32(vd, a->vd);
328
+ vfp_load_reg32(vd, a->vd);
329
if (neg_d) {
330
/* VFNMA, VFNMS */
331
gen_helper_vfp_negh(vd, vd);
332
}
333
fpst = fpstatus_ptr(FPST_FPCR_F16);
334
gen_helper_vfp_muladdh(vd, vn, vm, vd, fpst);
335
- neon_store_reg32(vd, a->vd);
336
+ vfp_store_reg32(vd, a->vd);
337
338
tcg_temp_free_ptr(fpst);
339
tcg_temp_free_i32(vn);
340
@@ -XXX,XX +XXX,XX @@ static bool do_vfm_sp(DisasContext *s, arg_VFMA_sp *a, bool neg_n, bool neg_d)
341
vm = tcg_temp_new_i32();
342
vd = tcg_temp_new_i32();
343
344
- neon_load_reg32(vn, a->vn);
345
- neon_load_reg32(vm, a->vm);
346
+ vfp_load_reg32(vn, a->vn);
347
+ vfp_load_reg32(vm, a->vm);
348
if (neg_n) {
349
/* VFNMS, VFMS */
350
gen_helper_vfp_negs(vn, vn);
351
}
352
- neon_load_reg32(vd, a->vd);
353
+ vfp_load_reg32(vd, a->vd);
354
if (neg_d) {
355
/* VFNMA, VFNMS */
356
gen_helper_vfp_negs(vd, vd);
357
}
358
fpst = fpstatus_ptr(FPST_FPCR);
359
gen_helper_vfp_muladds(vd, vn, vm, vd, fpst);
360
- neon_store_reg32(vd, a->vd);
361
+ vfp_store_reg32(vd, a->vd);
362
363
tcg_temp_free_ptr(fpst);
364
tcg_temp_free_i32(vn);
365
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_imm_hp(DisasContext *s, arg_VMOV_imm_sp *a)
366
}
367
368
fd = tcg_const_i32(vfp_expand_imm(MO_16, a->imm));
369
- neon_store_reg32(fd, a->vd);
370
+ vfp_store_reg32(fd, a->vd);
371
tcg_temp_free_i32(fd);
372
return true;
373
}
374
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_imm_sp(DisasContext *s, arg_VMOV_imm_sp *a)
375
fd = tcg_const_i32(vfp_expand_imm(MO_32, a->imm));
376
377
for (;;) {
378
- neon_store_reg32(fd, vd);
379
+ vfp_store_reg32(fd, vd);
380
381
if (veclen == 0) {
382
break;
383
@@ -XXX,XX +XXX,XX @@ static bool trans_VCMP_hp(DisasContext *s, arg_VCMP_sp *a)
384
vd = tcg_temp_new_i32();
385
vm = tcg_temp_new_i32();
386
387
- neon_load_reg32(vd, a->vd);
388
+ vfp_load_reg32(vd, a->vd);
389
if (a->z) {
390
tcg_gen_movi_i32(vm, 0);
391
} else {
392
- neon_load_reg32(vm, a->vm);
393
+ vfp_load_reg32(vm, a->vm);
394
}
395
396
if (a->e) {
397
@@ -XXX,XX +XXX,XX @@ static bool trans_VCMP_sp(DisasContext *s, arg_VCMP_sp *a)
398
vd = tcg_temp_new_i32();
399
vm = tcg_temp_new_i32();
400
401
- neon_load_reg32(vd, a->vd);
402
+ vfp_load_reg32(vd, a->vd);
403
if (a->z) {
404
tcg_gen_movi_i32(vm, 0);
405
} else {
406
- neon_load_reg32(vm, a->vm);
407
+ vfp_load_reg32(vm, a->vm);
408
}
409
410
if (a->e) {
411
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_f32_f16(DisasContext *s, arg_VCVT_f32_f16 *a)
412
/* The T bit tells us if we want the low or high 16 bits of Vm */
413
tcg_gen_ld16u_i32(tmp, cpu_env, vfp_f16_offset(a->vm, a->t));
414
gen_helper_vfp_fcvt_f16_to_f32(tmp, tmp, fpst, ahp_mode);
415
- neon_store_reg32(tmp, a->vd);
416
+ vfp_store_reg32(tmp, a->vd);
417
tcg_temp_free_i32(ahp_mode);
418
tcg_temp_free_ptr(fpst);
419
tcg_temp_free_i32(tmp);
420
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_f16_f32(DisasContext *s, arg_VCVT_f16_f32 *a)
421
ahp_mode = get_ahp_flag();
422
tmp = tcg_temp_new_i32();
423
424
- neon_load_reg32(tmp, a->vm);
425
+ vfp_load_reg32(tmp, a->vm);
426
gen_helper_vfp_fcvt_f32_to_f16(tmp, tmp, fpst, ahp_mode);
427
tcg_gen_st16_i32(tmp, cpu_env, vfp_f16_offset(a->vd, a->t));
428
tcg_temp_free_i32(ahp_mode);
429
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTR_hp(DisasContext *s, arg_VRINTR_sp *a)
430
}
431
432
tmp = tcg_temp_new_i32();
433
- neon_load_reg32(tmp, a->vm);
434
+ vfp_load_reg32(tmp, a->vm);
435
fpst = fpstatus_ptr(FPST_FPCR_F16);
436
gen_helper_rinth(tmp, tmp, fpst);
437
- neon_store_reg32(tmp, a->vd);
438
+ vfp_store_reg32(tmp, a->vd);
439
tcg_temp_free_ptr(fpst);
440
tcg_temp_free_i32(tmp);
441
return true;
442
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTR_sp(DisasContext *s, arg_VRINTR_sp *a)
443
}
444
445
tmp = tcg_temp_new_i32();
446
- neon_load_reg32(tmp, a->vm);
447
+ vfp_load_reg32(tmp, a->vm);
448
fpst = fpstatus_ptr(FPST_FPCR);
449
gen_helper_rints(tmp, tmp, fpst);
450
- neon_store_reg32(tmp, a->vd);
451
+ vfp_store_reg32(tmp, a->vd);
452
tcg_temp_free_ptr(fpst);
453
tcg_temp_free_i32(tmp);
454
return true;
455
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTZ_hp(DisasContext *s, arg_VRINTZ_sp *a)
456
}
457
458
tmp = tcg_temp_new_i32();
459
- neon_load_reg32(tmp, a->vm);
460
+ vfp_load_reg32(tmp, a->vm);
461
fpst = fpstatus_ptr(FPST_FPCR_F16);
462
tcg_rmode = tcg_const_i32(float_round_to_zero);
463
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
464
gen_helper_rinth(tmp, tmp, fpst);
465
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
466
- neon_store_reg32(tmp, a->vd);
467
+ vfp_store_reg32(tmp, a->vd);
468
tcg_temp_free_ptr(fpst);
469
tcg_temp_free_i32(tcg_rmode);
470
tcg_temp_free_i32(tmp);
471
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTZ_sp(DisasContext *s, arg_VRINTZ_sp *a)
472
}
473
474
tmp = tcg_temp_new_i32();
475
- neon_load_reg32(tmp, a->vm);
476
+ vfp_load_reg32(tmp, a->vm);
477
fpst = fpstatus_ptr(FPST_FPCR);
478
tcg_rmode = tcg_const_i32(float_round_to_zero);
479
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
480
gen_helper_rints(tmp, tmp, fpst);
481
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
482
- neon_store_reg32(tmp, a->vd);
483
+ vfp_store_reg32(tmp, a->vd);
484
tcg_temp_free_ptr(fpst);
485
tcg_temp_free_i32(tcg_rmode);
486
tcg_temp_free_i32(tmp);
487
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTX_hp(DisasContext *s, arg_VRINTX_sp *a)
488
}
489
490
tmp = tcg_temp_new_i32();
491
- neon_load_reg32(tmp, a->vm);
492
+ vfp_load_reg32(tmp, a->vm);
493
fpst = fpstatus_ptr(FPST_FPCR_F16);
494
gen_helper_rinth_exact(tmp, tmp, fpst);
495
- neon_store_reg32(tmp, a->vd);
496
+ vfp_store_reg32(tmp, a->vd);
497
tcg_temp_free_ptr(fpst);
498
tcg_temp_free_i32(tmp);
499
return true;
500
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTX_sp(DisasContext *s, arg_VRINTX_sp *a)
501
}
502
503
tmp = tcg_temp_new_i32();
504
- neon_load_reg32(tmp, a->vm);
505
+ vfp_load_reg32(tmp, a->vm);
506
fpst = fpstatus_ptr(FPST_FPCR);
507
gen_helper_rints_exact(tmp, tmp, fpst);
508
- neon_store_reg32(tmp, a->vd);
509
+ vfp_store_reg32(tmp, a->vd);
510
tcg_temp_free_ptr(fpst);
511
tcg_temp_free_i32(tmp);
512
return true;
513
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_sp(DisasContext *s, arg_VCVT_sp *a)
514
515
vm = tcg_temp_new_i32();
516
vd = tcg_temp_new_i64();
517
- neon_load_reg32(vm, a->vm);
518
+ vfp_load_reg32(vm, a->vm);
519
gen_helper_vfp_fcvtds(vd, vm, cpu_env);
520
neon_store_reg64(vd, a->vd);
521
tcg_temp_free_i32(vm);
522
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_dp(DisasContext *s, arg_VCVT_dp *a)
523
vm = tcg_temp_new_i64();
524
neon_load_reg64(vm, a->vm);
525
gen_helper_vfp_fcvtsd(vd, vm, cpu_env);
526
- neon_store_reg32(vd, a->vd);
527
+ vfp_store_reg32(vd, a->vd);
528
tcg_temp_free_i32(vd);
529
tcg_temp_free_i64(vm);
530
return true;
531
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_hp(DisasContext *s, arg_VCVT_int_sp *a)
532
}
533
534
vm = tcg_temp_new_i32();
535
- neon_load_reg32(vm, a->vm);
536
+ vfp_load_reg32(vm, a->vm);
537
fpst = fpstatus_ptr(FPST_FPCR_F16);
538
if (a->s) {
539
/* i32 -> f16 */
540
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_hp(DisasContext *s, arg_VCVT_int_sp *a)
541
/* u32 -> f16 */
542
gen_helper_vfp_uitoh(vm, vm, fpst);
543
}
544
- neon_store_reg32(vm, a->vd);
545
+ vfp_store_reg32(vm, a->vd);
546
tcg_temp_free_i32(vm);
547
tcg_temp_free_ptr(fpst);
548
return true;
549
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_sp(DisasContext *s, arg_VCVT_int_sp *a)
550
}
551
552
vm = tcg_temp_new_i32();
553
- neon_load_reg32(vm, a->vm);
554
+ vfp_load_reg32(vm, a->vm);
555
fpst = fpstatus_ptr(FPST_FPCR);
556
if (a->s) {
557
/* i32 -> f32 */
558
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_sp(DisasContext *s, arg_VCVT_int_sp *a)
559
/* u32 -> f32 */
560
gen_helper_vfp_uitos(vm, vm, fpst);
561
}
562
- neon_store_reg32(vm, a->vd);
563
+ vfp_store_reg32(vm, a->vd);
564
tcg_temp_free_i32(vm);
565
tcg_temp_free_ptr(fpst);
566
return true;
567
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_dp(DisasContext *s, arg_VCVT_int_dp *a)
568
569
vm = tcg_temp_new_i32();
570
vd = tcg_temp_new_i64();
571
- neon_load_reg32(vm, a->vm);
572
+ vfp_load_reg32(vm, a->vm);
573
fpst = fpstatus_ptr(FPST_FPCR);
574
if (a->s) {
575
/* i32 -> f64 */
576
@@ -XXX,XX +XXX,XX @@ static bool trans_VJCVT(DisasContext *s, arg_VJCVT *a)
577
vd = tcg_temp_new_i32();
578
neon_load_reg64(vm, a->vm);
579
gen_helper_vjcvt(vd, vm, cpu_env);
580
- neon_store_reg32(vd, a->vd);
581
+ vfp_store_reg32(vd, a->vd);
582
tcg_temp_free_i64(vm);
583
tcg_temp_free_i32(vd);
584
return true;
585
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_hp(DisasContext *s, arg_VCVT_fix_sp *a)
586
frac_bits = (a->opc & 1) ? (32 - a->imm) : (16 - a->imm);
587
588
vd = tcg_temp_new_i32();
589
- neon_load_reg32(vd, a->vd);
590
+ vfp_load_reg32(vd, a->vd);
591
592
fpst = fpstatus_ptr(FPST_FPCR_F16);
593
shift = tcg_const_i32(frac_bits);
594
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_hp(DisasContext *s, arg_VCVT_fix_sp *a)
595
g_assert_not_reached();
596
}
597
598
- neon_store_reg32(vd, a->vd);
599
+ vfp_store_reg32(vd, a->vd);
600
tcg_temp_free_i32(vd);
601
tcg_temp_free_i32(shift);
602
tcg_temp_free_ptr(fpst);
603
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_sp(DisasContext *s, arg_VCVT_fix_sp *a)
604
frac_bits = (a->opc & 1) ? (32 - a->imm) : (16 - a->imm);
605
606
vd = tcg_temp_new_i32();
607
- neon_load_reg32(vd, a->vd);
608
+ vfp_load_reg32(vd, a->vd);
609
610
fpst = fpstatus_ptr(FPST_FPCR);
611
shift = tcg_const_i32(frac_bits);
612
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_sp(DisasContext *s, arg_VCVT_fix_sp *a)
613
g_assert_not_reached();
614
}
615
616
- neon_store_reg32(vd, a->vd);
617
+ vfp_store_reg32(vd, a->vd);
618
tcg_temp_free_i32(vd);
619
tcg_temp_free_i32(shift);
620
tcg_temp_free_ptr(fpst);
621
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_hp_int(DisasContext *s, arg_VCVT_sp_int *a)
622
623
fpst = fpstatus_ptr(FPST_FPCR_F16);
624
vm = tcg_temp_new_i32();
625
- neon_load_reg32(vm, a->vm);
626
+ vfp_load_reg32(vm, a->vm);
627
628
if (a->s) {
629
if (a->rz) {
630
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_hp_int(DisasContext *s, arg_VCVT_sp_int *a)
631
gen_helper_vfp_touih(vm, vm, fpst);
632
}
633
}
634
- neon_store_reg32(vm, a->vd);
635
+ vfp_store_reg32(vm, a->vd);
636
tcg_temp_free_i32(vm);
637
tcg_temp_free_ptr(fpst);
638
return true;
639
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_sp_int(DisasContext *s, arg_VCVT_sp_int *a)
640
641
fpst = fpstatus_ptr(FPST_FPCR);
642
vm = tcg_temp_new_i32();
643
- neon_load_reg32(vm, a->vm);
644
+ vfp_load_reg32(vm, a->vm);
645
646
if (a->s) {
647
if (a->rz) {
648
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_sp_int(DisasContext *s, arg_VCVT_sp_int *a)
649
gen_helper_vfp_touis(vm, vm, fpst);
650
}
651
}
652
- neon_store_reg32(vm, a->vd);
653
+ vfp_store_reg32(vm, a->vd);
654
tcg_temp_free_i32(vm);
655
tcg_temp_free_ptr(fpst);
656
return true;
657
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_dp_int(DisasContext *s, arg_VCVT_dp_int *a)
658
gen_helper_vfp_touid(vd, vm, fpst);
659
}
660
}
661
- neon_store_reg32(vd, a->vd);
662
+ vfp_store_reg32(vd, a->vd);
663
tcg_temp_free_i32(vd);
664
tcg_temp_free_i64(vm);
665
tcg_temp_free_ptr(fpst);
666
@@ -XXX,XX +XXX,XX @@ static bool trans_VINS(DisasContext *s, arg_VINS *a)
667
/* Insert low half of Vm into high half of Vd */
668
rm = tcg_temp_new_i32();
669
rd = tcg_temp_new_i32();
670
- neon_load_reg32(rm, a->vm);
671
- neon_load_reg32(rd, a->vd);
672
+ vfp_load_reg32(rm, a->vm);
673
+ vfp_load_reg32(rd, a->vd);
674
tcg_gen_deposit_i32(rd, rd, rm, 16, 16);
675
- neon_store_reg32(rd, a->vd);
676
+ vfp_store_reg32(rd, a->vd);
677
tcg_temp_free_i32(rm);
678
tcg_temp_free_i32(rd);
679
return true;
680
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOVX(DisasContext *s, arg_VINS *a)
681
682
/* Set Vd to high half of Vm */
683
rm = tcg_temp_new_i32();
684
- neon_load_reg32(rm, a->vm);
685
+ vfp_load_reg32(rm, a->vm);
686
tcg_gen_shri_i32(rm, rm, 16);
687
- neon_store_reg32(rm, a->vd);
688
+ vfp_store_reg32(rm, a->vd);
689
tcg_temp_free_i32(rm);
690
return true;
691
}
134
--
692
--
135
2.7.4
693
2.20.1
136
694
137
695
diff view generated by jsdifflib
1
Move the regime_is_secure() utility function to internals.h;
1
From: Richard Henderson <richard.henderson@linaro.org>
2
we are going to want to call it from translate.c.
2
3
3
Replace all uses of neon_load/store_reg64 within translate-neon.c.inc.
4
5
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20201030022618.785675-9-richard.henderson@linaro.org
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 1503414539-28762-20-git-send-email-peter.maydell@linaro.org
7
---
9
---
8
target/arm/internals.h | 26 ++++++++++++++++++++++++++
10
target/arm/translate.c | 26 +++++++++
9
target/arm/helper.c | 26 --------------------------
11
target/arm/translate-neon.c.inc | 94 ++++++++++++++++-----------------
10
2 files changed, 26 insertions(+), 26 deletions(-)
12
2 files changed, 73 insertions(+), 47 deletions(-)
11
13
12
diff --git a/target/arm/internals.h b/target/arm/internals.h
14
diff --git a/target/arm/translate.c b/target/arm/translate.c
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/internals.h
16
--- a/target/arm/translate.c
15
+++ b/target/arm/internals.h
17
+++ b/target/arm/translate.c
16
@@ -XXX,XX +XXX,XX @@ static inline void arm_call_el_change_hook(ARMCPU *cpu)
18
@@ -XXX,XX +XXX,XX @@ static void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp memop)
17
}
19
}
18
}
20
}
19
21
20
+/* Return true if this address translation regime is secure */
22
+static void read_neon_element64(TCGv_i64 dest, int reg, int ele, MemOp memop)
21
+static inline bool regime_is_secure(CPUARMState *env, ARMMMUIdx mmu_idx)
22
+{
23
+{
23
+ switch (mmu_idx) {
24
+ long off = neon_element_offset(reg, ele, memop);
24
+ case ARMMMUIdx_S12NSE0:
25
+
25
+ case ARMMMUIdx_S12NSE1:
26
+ switch (memop) {
26
+ case ARMMMUIdx_S1NSE0:
27
+ case MO_Q:
27
+ case ARMMMUIdx_S1NSE1:
28
+ tcg_gen_ld_i64(dest, cpu_env, off);
28
+ case ARMMMUIdx_S1E2:
29
+ break;
29
+ case ARMMMUIdx_S2NS:
30
+ case ARMMMUIdx_MPriv:
31
+ case ARMMMUIdx_MNegPri:
32
+ case ARMMMUIdx_MUser:
33
+ return false;
34
+ case ARMMMUIdx_S1E3:
35
+ case ARMMMUIdx_S1SE0:
36
+ case ARMMMUIdx_S1SE1:
37
+ case ARMMMUIdx_MSPriv:
38
+ case ARMMMUIdx_MSNegPri:
39
+ case ARMMMUIdx_MSUser:
40
+ return true;
41
+ default:
30
+ default:
42
+ g_assert_not_reached();
31
+ g_assert_not_reached();
43
+ }
32
+ }
44
+}
33
+}
45
+
34
+
46
#endif
35
static void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp memop)
47
diff --git a/target/arm/helper.c b/target/arm/helper.c
36
{
37
long off = neon_element_offset(reg, ele, memop);
38
@@ -XXX,XX +XXX,XX @@ static void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp memop)
39
}
40
}
41
42
+static void write_neon_element64(TCGv_i64 src, int reg, int ele, MemOp memop)
43
+{
44
+ long off = neon_element_offset(reg, ele, memop);
45
+
46
+ switch (memop) {
47
+ case MO_64:
48
+ tcg_gen_st_i64(src, cpu_env, off);
49
+ break;
50
+ default:
51
+ g_assert_not_reached();
52
+ }
53
+}
54
+
55
static TCGv_ptr vfp_reg_ptr(bool dp, int reg)
56
{
57
TCGv_ptr ret = tcg_temp_new_ptr();
58
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
48
index XXXXXXX..XXXXXXX 100644
59
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/helper.c
60
--- a/target/arm/translate-neon.c.inc
50
+++ b/target/arm/helper.c
61
+++ b/target/arm/translate-neon.c.inc
51
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
62
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_env_64(DisasContext *s, arg_2reg_shift *a,
52
}
63
for (pass = 0; pass < a->q + 1; pass++) {
64
TCGv_i64 tmp = tcg_temp_new_i64();
65
66
- neon_load_reg64(tmp, a->vm + pass);
67
+ read_neon_element64(tmp, a->vm, pass, MO_64);
68
fn(tmp, cpu_env, tmp, constimm);
69
- neon_store_reg64(tmp, a->vd + pass);
70
+ write_neon_element64(tmp, a->vd, pass, MO_64);
71
tcg_temp_free_i64(tmp);
72
}
73
tcg_temp_free_i64(constimm);
74
@@ -XXX,XX +XXX,XX @@ static bool do_2shift_narrow_64(DisasContext *s, arg_2reg_shift *a,
75
rd = tcg_temp_new_i32();
76
77
/* Load both inputs first to avoid potential overwrite if rm == rd */
78
- neon_load_reg64(rm1, a->vm);
79
- neon_load_reg64(rm2, a->vm + 1);
80
+ read_neon_element64(rm1, a->vm, 0, MO_64);
81
+ read_neon_element64(rm2, a->vm, 1, MO_64);
82
83
shiftfn(rm1, rm1, constimm);
84
narrowfn(rd, cpu_env, rm1);
85
@@ -XXX,XX +XXX,XX @@ static bool do_vshll_2sh(DisasContext *s, arg_2reg_shift *a,
86
tcg_gen_shli_i64(tmp, tmp, a->shift);
87
tcg_gen_andi_i64(tmp, tmp, ~widen_mask);
88
}
89
- neon_store_reg64(tmp, a->vd);
90
+ write_neon_element64(tmp, a->vd, 0, MO_64);
91
92
widenfn(tmp, rm1);
93
tcg_temp_free_i32(rm1);
94
@@ -XXX,XX +XXX,XX @@ static bool do_vshll_2sh(DisasContext *s, arg_2reg_shift *a,
95
tcg_gen_shli_i64(tmp, tmp, a->shift);
96
tcg_gen_andi_i64(tmp, tmp, ~widen_mask);
97
}
98
- neon_store_reg64(tmp, a->vd + 1);
99
+ write_neon_element64(tmp, a->vd, 1, MO_64);
100
tcg_temp_free_i64(tmp);
101
return true;
53
}
102
}
54
103
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
55
-/* Return true if this address translation regime is secure */
104
rm_64 = tcg_temp_new_i64();
56
-static inline bool regime_is_secure(CPUARMState *env, ARMMMUIdx mmu_idx)
105
57
-{
106
if (src1_wide) {
58
- switch (mmu_idx) {
107
- neon_load_reg64(rn0_64, a->vn);
59
- case ARMMMUIdx_S12NSE0:
108
+ read_neon_element64(rn0_64, a->vn, 0, MO_64);
60
- case ARMMMUIdx_S12NSE1:
109
} else {
61
- case ARMMMUIdx_S1NSE0:
110
TCGv_i32 tmp = tcg_temp_new_i32();
62
- case ARMMMUIdx_S1NSE1:
111
read_neon_element32(tmp, a->vn, 0, MO_32);
63
- case ARMMMUIdx_S1E2:
112
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
64
- case ARMMMUIdx_S2NS:
113
* avoid incorrect results if a narrow input overlaps with the result.
65
- case ARMMMUIdx_MPriv:
114
*/
66
- case ARMMMUIdx_MNegPri:
115
if (src1_wide) {
67
- case ARMMMUIdx_MUser:
116
- neon_load_reg64(rn1_64, a->vn + 1);
68
- return false;
117
+ read_neon_element64(rn1_64, a->vn, 1, MO_64);
69
- case ARMMMUIdx_S1E3:
118
} else {
70
- case ARMMMUIdx_S1SE0:
119
TCGv_i32 tmp = tcg_temp_new_i32();
71
- case ARMMMUIdx_S1SE1:
120
read_neon_element32(tmp, a->vn, 1, MO_32);
72
- case ARMMMUIdx_MSPriv:
121
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
73
- case ARMMMUIdx_MSNegPri:
122
rm = tcg_temp_new_i32();
74
- case ARMMMUIdx_MSUser:
123
read_neon_element32(rm, a->vm, 1, MO_32);
75
- return true;
124
76
- default:
125
- neon_store_reg64(rn0_64, a->vd);
77
- g_assert_not_reached();
126
+ write_neon_element64(rn0_64, a->vd, 0, MO_64);
78
- }
127
79
-}
128
widenfn(rm_64, rm);
80
-
129
tcg_temp_free_i32(rm);
81
/* Return the SCTLR value which controls this address translation regime */
130
opfn(rn1_64, rn1_64, rm_64);
82
static inline uint32_t regime_sctlr(CPUARMState *env, ARMMMUIdx mmu_idx)
131
- neon_store_reg64(rn1_64, a->vd + 1);
83
{
132
+ write_neon_element64(rn1_64, a->vd, 1, MO_64);
133
134
tcg_temp_free_i64(rn0_64);
135
tcg_temp_free_i64(rn1_64);
136
@@ -XXX,XX +XXX,XX @@ static bool do_narrow_3d(DisasContext *s, arg_3diff *a,
137
rd0 = tcg_temp_new_i32();
138
rd1 = tcg_temp_new_i32();
139
140
- neon_load_reg64(rn_64, a->vn);
141
- neon_load_reg64(rm_64, a->vm);
142
+ read_neon_element64(rn_64, a->vn, 0, MO_64);
143
+ read_neon_element64(rm_64, a->vm, 0, MO_64);
144
145
opfn(rn_64, rn_64, rm_64);
146
147
narrowfn(rd0, rn_64);
148
149
- neon_load_reg64(rn_64, a->vn + 1);
150
- neon_load_reg64(rm_64, a->vm + 1);
151
+ read_neon_element64(rn_64, a->vn, 1, MO_64);
152
+ read_neon_element64(rm_64, a->vm, 1, MO_64);
153
154
opfn(rn_64, rn_64, rm_64);
155
156
@@ -XXX,XX +XXX,XX @@ static bool do_long_3d(DisasContext *s, arg_3diff *a,
157
/* Don't store results until after all loads: they might overlap */
158
if (accfn) {
159
tmp = tcg_temp_new_i64();
160
- neon_load_reg64(tmp, a->vd);
161
+ read_neon_element64(tmp, a->vd, 0, MO_64);
162
accfn(tmp, tmp, rd0);
163
- neon_store_reg64(tmp, a->vd);
164
- neon_load_reg64(tmp, a->vd + 1);
165
+ write_neon_element64(tmp, a->vd, 0, MO_64);
166
+ read_neon_element64(tmp, a->vd, 1, MO_64);
167
accfn(tmp, tmp, rd1);
168
- neon_store_reg64(tmp, a->vd + 1);
169
+ write_neon_element64(tmp, a->vd, 1, MO_64);
170
tcg_temp_free_i64(tmp);
171
} else {
172
- neon_store_reg64(rd0, a->vd);
173
- neon_store_reg64(rd1, a->vd + 1);
174
+ write_neon_element64(rd0, a->vd, 0, MO_64);
175
+ write_neon_element64(rd1, a->vd, 1, MO_64);
176
}
177
178
tcg_temp_free_i64(rd0);
179
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar_long(DisasContext *s, arg_2scalar *a,
180
181
if (accfn) {
182
TCGv_i64 t64 = tcg_temp_new_i64();
183
- neon_load_reg64(t64, a->vd);
184
+ read_neon_element64(t64, a->vd, 0, MO_64);
185
accfn(t64, t64, rn0_64);
186
- neon_store_reg64(t64, a->vd);
187
- neon_load_reg64(t64, a->vd + 1);
188
+ write_neon_element64(t64, a->vd, 0, MO_64);
189
+ read_neon_element64(t64, a->vd, 1, MO_64);
190
accfn(t64, t64, rn1_64);
191
- neon_store_reg64(t64, a->vd + 1);
192
+ write_neon_element64(t64, a->vd, 1, MO_64);
193
tcg_temp_free_i64(t64);
194
} else {
195
- neon_store_reg64(rn0_64, a->vd);
196
- neon_store_reg64(rn1_64, a->vd + 1);
197
+ write_neon_element64(rn0_64, a->vd, 0, MO_64);
198
+ write_neon_element64(rn1_64, a->vd, 1, MO_64);
199
}
200
tcg_temp_free_i64(rn0_64);
201
tcg_temp_free_i64(rn1_64);
202
@@ -XXX,XX +XXX,XX @@ static bool trans_VEXT(DisasContext *s, arg_VEXT *a)
203
right = tcg_temp_new_i64();
204
dest = tcg_temp_new_i64();
205
206
- neon_load_reg64(right, a->vn);
207
- neon_load_reg64(left, a->vm);
208
+ read_neon_element64(right, a->vn, 0, MO_64);
209
+ read_neon_element64(left, a->vm, 0, MO_64);
210
tcg_gen_extract2_i64(dest, right, left, a->imm * 8);
211
- neon_store_reg64(dest, a->vd);
212
+ write_neon_element64(dest, a->vd, 0, MO_64);
213
214
tcg_temp_free_i64(left);
215
tcg_temp_free_i64(right);
216
@@ -XXX,XX +XXX,XX @@ static bool trans_VEXT(DisasContext *s, arg_VEXT *a)
217
destright = tcg_temp_new_i64();
218
219
if (a->imm < 8) {
220
- neon_load_reg64(right, a->vn);
221
- neon_load_reg64(middle, a->vn + 1);
222
+ read_neon_element64(right, a->vn, 0, MO_64);
223
+ read_neon_element64(middle, a->vn, 1, MO_64);
224
tcg_gen_extract2_i64(destright, right, middle, a->imm * 8);
225
- neon_load_reg64(left, a->vm);
226
+ read_neon_element64(left, a->vm, 0, MO_64);
227
tcg_gen_extract2_i64(destleft, middle, left, a->imm * 8);
228
} else {
229
- neon_load_reg64(right, a->vn + 1);
230
- neon_load_reg64(middle, a->vm);
231
+ read_neon_element64(right, a->vn, 1, MO_64);
232
+ read_neon_element64(middle, a->vm, 0, MO_64);
233
tcg_gen_extract2_i64(destright, right, middle, (a->imm - 8) * 8);
234
- neon_load_reg64(left, a->vm + 1);
235
+ read_neon_element64(left, a->vm, 1, MO_64);
236
tcg_gen_extract2_i64(destleft, middle, left, (a->imm - 8) * 8);
237
}
238
239
- neon_store_reg64(destright, a->vd);
240
- neon_store_reg64(destleft, a->vd + 1);
241
+ write_neon_element64(destright, a->vd, 0, MO_64);
242
+ write_neon_element64(destleft, a->vd, 1, MO_64);
243
244
tcg_temp_free_i64(destright);
245
tcg_temp_free_i64(destleft);
246
@@ -XXX,XX +XXX,XX @@ static bool do_2misc_pairwise(DisasContext *s, arg_2misc *a,
247
248
if (accfn) {
249
TCGv_i64 tmp64 = tcg_temp_new_i64();
250
- neon_load_reg64(tmp64, a->vd + pass);
251
+ read_neon_element64(tmp64, a->vd, pass, MO_64);
252
accfn(rd_64, tmp64, rd_64);
253
tcg_temp_free_i64(tmp64);
254
}
255
- neon_store_reg64(rd_64, a->vd + pass);
256
+ write_neon_element64(rd_64, a->vd, pass, MO_64);
257
tcg_temp_free_i64(rd_64);
258
}
259
return true;
260
@@ -XXX,XX +XXX,XX @@ static bool do_vmovn(DisasContext *s, arg_2misc *a,
261
rd0 = tcg_temp_new_i32();
262
rd1 = tcg_temp_new_i32();
263
264
- neon_load_reg64(rm, a->vm);
265
+ read_neon_element64(rm, a->vm, 0, MO_64);
266
narrowfn(rd0, cpu_env, rm);
267
- neon_load_reg64(rm, a->vm + 1);
268
+ read_neon_element64(rm, a->vm, 1, MO_64);
269
narrowfn(rd1, cpu_env, rm);
270
write_neon_element32(rd0, a->vd, 0, MO_32);
271
write_neon_element32(rd1, a->vd, 1, MO_32);
272
@@ -XXX,XX +XXX,XX @@ static bool trans_VSHLL(DisasContext *s, arg_2misc *a)
273
274
widenfn(rd, rm0);
275
tcg_gen_shli_i64(rd, rd, 8 << a->size);
276
- neon_store_reg64(rd, a->vd);
277
+ write_neon_element64(rd, a->vd, 0, MO_64);
278
widenfn(rd, rm1);
279
tcg_gen_shli_i64(rd, rd, 8 << a->size);
280
- neon_store_reg64(rd, a->vd + 1);
281
+ write_neon_element64(rd, a->vd, 1, MO_64);
282
283
tcg_temp_free_i64(rd);
284
tcg_temp_free_i32(rm0);
285
@@ -XXX,XX +XXX,XX @@ static bool trans_VSWP(DisasContext *s, arg_2misc *a)
286
rm = tcg_temp_new_i64();
287
rd = tcg_temp_new_i64();
288
for (pass = 0; pass < (a->q ? 2 : 1); pass++) {
289
- neon_load_reg64(rm, a->vm + pass);
290
- neon_load_reg64(rd, a->vd + pass);
291
- neon_store_reg64(rm, a->vd + pass);
292
- neon_store_reg64(rd, a->vm + pass);
293
+ read_neon_element64(rm, a->vm, pass, MO_64);
294
+ read_neon_element64(rd, a->vd, pass, MO_64);
295
+ write_neon_element64(rm, a->vd, pass, MO_64);
296
+ write_neon_element64(rd, a->vm, pass, MO_64);
297
}
298
tcg_temp_free_i64(rm);
299
tcg_temp_free_i64(rd);
84
--
300
--
85
2.7.4
301
2.20.1
86
302
87
303
diff view generated by jsdifflib
1
Set the MachineClass flag ignore_memory_transaction_failures
1
From: Richard Henderson <richard.henderson@linaro.org>
2
for almost all ARM boards. This means they retain the legacy
2
3
behaviour that accesses to unimplemented addresses will RAZ/WI
3
The only uses of this function are for loading VFP
4
rather than aborting, when a subsequent commit adds support
4
double-precision values, and nothing to do with NEON.
5
for external aborts.
5
6
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
The exceptions are:
7
Message-id: 20201030022618.785675-10-richard.henderson@linaro.org
8
* virt -- we know that guests won't try to prod devices
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
that we don't describe in the device tree or ACPI tables
10
* mps2 -- this board was written to use unimplemented-device
11
for all the ranges with devices we don't yet handle
12
13
New boards should not set the flag, but instead be written
14
like the mps2.
15
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Reviewed-by: Alistair Francis <alistair.francis@xilinx.com>
18
Message-id: 1504626814-23124-3-git-send-email-peter.maydell@linaro.org
19
For the Xilinx boards:
20
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
21
---
10
---
22
hw/arm/aspeed.c | 3 +++
11
target/arm/translate.c | 8 ++--
23
hw/arm/collie.c | 1 +
12
target/arm/translate-vfp.c.inc | 84 +++++++++++++++++-----------------
24
hw/arm/cubieboard.c | 1 +
13
2 files changed, 46 insertions(+), 46 deletions(-)
25
hw/arm/digic_boards.c | 1 +
14
26
hw/arm/exynos4_boards.c | 2 ++
15
diff --git a/target/arm/translate.c b/target/arm/translate.c
27
hw/arm/gumstix.c | 2 ++
28
hw/arm/highbank.c | 2 ++
29
hw/arm/imx25_pdk.c | 1 +
30
hw/arm/integratorcp.c | 1 +
31
hw/arm/kzm.c | 1 +
32
hw/arm/mainstone.c | 1 +
33
hw/arm/musicpal.c | 1 +
34
hw/arm/netduino2.c | 1 +
35
hw/arm/nseries.c | 2 ++
36
hw/arm/omap_sx1.c | 2 ++
37
hw/arm/palm.c | 1 +
38
hw/arm/raspi.c | 1 +
39
hw/arm/realview.c | 4 ++++
40
hw/arm/sabrelite.c | 1 +
41
hw/arm/spitz.c | 4 ++++
42
hw/arm/stellaris.c | 2 ++
43
hw/arm/tosa.c | 1 +
44
hw/arm/versatilepb.c | 2 ++
45
hw/arm/vexpress.c | 1 +
46
hw/arm/xilinx_zynq.c | 1 +
47
hw/arm/xlnx-ep108.c | 2 ++
48
hw/arm/z2.c | 1 +
49
27 files changed, 43 insertions(+)
50
51
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
52
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
53
--- a/hw/arm/aspeed.c
17
--- a/target/arm/translate.c
54
+++ b/hw/arm/aspeed.c
18
+++ b/target/arm/translate.c
55
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_class_init(ObjectClass *oc, void *data)
19
@@ -XXX,XX +XXX,XX @@ static long vfp_reg_offset(bool dp, unsigned reg)
56
mc->no_floppy = 1;
20
}
57
mc->no_cdrom = 1;
58
mc->no_parallel = 1;
59
+ mc->ignore_memory_transaction_failures = true;
60
}
21
}
61
22
62
static const TypeInfo palmetto_bmc_type = {
23
-static inline void neon_load_reg64(TCGv_i64 var, int reg)
63
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_class_init(ObjectClass *oc, void *data)
24
+static inline void vfp_load_reg64(TCGv_i64 var, int reg)
64
mc->no_floppy = 1;
25
{
65
mc->no_cdrom = 1;
26
- tcg_gen_ld_i64(var, cpu_env, vfp_reg_offset(1, reg));
66
mc->no_parallel = 1;
27
+ tcg_gen_ld_i64(var, cpu_env, vfp_reg_offset(true, reg));
67
+ mc->ignore_memory_transaction_failures = true;
68
}
28
}
69
29
70
static const TypeInfo ast2500_evb_type = {
30
-static inline void neon_store_reg64(TCGv_i64 var, int reg)
71
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_class_init(ObjectClass *oc, void *data)
31
+static inline void vfp_store_reg64(TCGv_i64 var, int reg)
72
mc->no_floppy = 1;
32
{
73
mc->no_cdrom = 1;
33
- tcg_gen_st_i64(var, cpu_env, vfp_reg_offset(1, reg));
74
mc->no_parallel = 1;
34
+ tcg_gen_st_i64(var, cpu_env, vfp_reg_offset(true, reg));
75
+ mc->ignore_memory_transaction_failures = true;
76
}
35
}
77
36
78
static const TypeInfo romulus_bmc_type = {
37
static inline void vfp_load_reg32(TCGv_i32 var, int reg)
79
diff --git a/hw/arm/collie.c b/hw/arm/collie.c
38
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
80
index XXXXXXX..XXXXXXX 100644
39
index XXXXXXX..XXXXXXX 100644
81
--- a/hw/arm/collie.c
40
--- a/target/arm/translate-vfp.c.inc
82
+++ b/hw/arm/collie.c
41
+++ b/target/arm/translate-vfp.c.inc
83
@@ -XXX,XX +XXX,XX @@ static void collie_machine_init(MachineClass *mc)
42
@@ -XXX,XX +XXX,XX @@ static bool trans_VSEL(DisasContext *s, arg_VSEL *a)
84
{
43
tcg_gen_ext_i32_i64(nf, cpu_NF);
85
mc->desc = "Sharp SL-5500 (Collie) PDA (SA-1110)";
44
tcg_gen_ext_i32_i64(vf, cpu_VF);
86
mc->init = collie_init;
45
87
+ mc->ignore_memory_transaction_failures = true;
46
- neon_load_reg64(frn, rn);
88
}
47
- neon_load_reg64(frm, rm);
89
48
+ vfp_load_reg64(frn, rn);
90
DEFINE_MACHINE("collie", collie_machine_init)
49
+ vfp_load_reg64(frm, rm);
91
diff --git a/hw/arm/cubieboard.c b/hw/arm/cubieboard.c
50
switch (a->cc) {
92
index XXXXXXX..XXXXXXX 100644
51
case 0: /* eq: Z */
93
--- a/hw/arm/cubieboard.c
52
tcg_gen_movcond_i64(TCG_COND_EQ, dest, zf, zero,
94
+++ b/hw/arm/cubieboard.c
53
@@ -XXX,XX +XXX,XX @@ static bool trans_VSEL(DisasContext *s, arg_VSEL *a)
95
@@ -XXX,XX +XXX,XX @@ static void cubieboard_machine_init(MachineClass *mc)
54
tcg_temp_free_i64(tmp);
96
mc->init = cubieboard_init;
55
break;
97
mc->block_default_type = IF_IDE;
56
}
98
mc->units_per_default_bus = 1;
57
- neon_store_reg64(dest, rd);
99
+ mc->ignore_memory_transaction_failures = true;
58
+ vfp_store_reg64(dest, rd);
100
}
59
tcg_temp_free_i64(frn);
101
60
tcg_temp_free_i64(frm);
102
DEFINE_MACHINE("cubieboard", cubieboard_machine_init)
61
tcg_temp_free_i64(dest);
103
diff --git a/hw/arm/digic_boards.c b/hw/arm/digic_boards.c
62
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINT(DisasContext *s, arg_VRINT *a)
104
index XXXXXXX..XXXXXXX 100644
63
TCGv_i64 tcg_res;
105
--- a/hw/arm/digic_boards.c
64
tcg_op = tcg_temp_new_i64();
106
+++ b/hw/arm/digic_boards.c
65
tcg_res = tcg_temp_new_i64();
107
@@ -XXX,XX +XXX,XX @@ static void canon_a1100_machine_init(MachineClass *mc)
66
- neon_load_reg64(tcg_op, rm);
108
{
67
+ vfp_load_reg64(tcg_op, rm);
109
mc->desc = "Canon PowerShot A1100 IS";
68
gen_helper_rintd(tcg_res, tcg_op, fpst);
110
mc->init = &canon_a1100_init;
69
- neon_store_reg64(tcg_res, rd);
111
+ mc->ignore_memory_transaction_failures = true;
70
+ vfp_store_reg64(tcg_res, rd);
112
}
71
tcg_temp_free_i64(tcg_op);
113
72
tcg_temp_free_i64(tcg_res);
114
DEFINE_MACHINE("canon-a1100", canon_a1100_machine_init)
73
} else {
115
diff --git a/hw/arm/exynos4_boards.c b/hw/arm/exynos4_boards.c
74
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT(DisasContext *s, arg_VCVT *a)
116
index XXXXXXX..XXXXXXX 100644
75
tcg_double = tcg_temp_new_i64();
117
--- a/hw/arm/exynos4_boards.c
76
tcg_res = tcg_temp_new_i64();
118
+++ b/hw/arm/exynos4_boards.c
77
tcg_tmp = tcg_temp_new_i32();
119
@@ -XXX,XX +XXX,XX @@ static void nuri_class_init(ObjectClass *oc, void *data)
78
- neon_load_reg64(tcg_double, rm);
120
mc->desc = "Samsung NURI board (Exynos4210)";
79
+ vfp_load_reg64(tcg_double, rm);
121
mc->init = nuri_init;
80
if (is_signed) {
122
mc->max_cpus = EXYNOS4210_NCPUS;
81
gen_helper_vfp_tosld(tcg_res, tcg_double, tcg_shift, fpst);
123
+ mc->ignore_memory_transaction_failures = true;
82
} else {
124
}
83
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDR_VSTR_dp(DisasContext *s, arg_VLDR_VSTR_dp *a)
125
84
tmp = tcg_temp_new_i64();
126
static const TypeInfo nuri_type = {
85
if (a->l) {
127
@@ -XXX,XX +XXX,XX @@ static void smdkc210_class_init(ObjectClass *oc, void *data)
86
gen_aa32_ld64(s, tmp, addr, get_mem_index(s));
128
mc->desc = "Samsung SMDKC210 board (Exynos4210)";
87
- neon_store_reg64(tmp, a->vd);
129
mc->init = smdkc210_init;
88
+ vfp_store_reg64(tmp, a->vd);
130
mc->max_cpus = EXYNOS4210_NCPUS;
89
} else {
131
+ mc->ignore_memory_transaction_failures = true;
90
- neon_load_reg64(tmp, a->vd);
132
}
91
+ vfp_load_reg64(tmp, a->vd);
133
92
gen_aa32_st64(s, tmp, addr, get_mem_index(s));
134
static const TypeInfo smdkc210_type = {
93
}
135
diff --git a/hw/arm/gumstix.c b/hw/arm/gumstix.c
94
tcg_temp_free_i64(tmp);
136
index XXXXXXX..XXXXXXX 100644
95
@@ -XXX,XX +XXX,XX @@ static bool trans_VLDM_VSTM_dp(DisasContext *s, arg_VLDM_VSTM_dp *a)
137
--- a/hw/arm/gumstix.c
96
if (a->l) {
138
+++ b/hw/arm/gumstix.c
97
/* load */
139
@@ -XXX,XX +XXX,XX @@ static void connex_class_init(ObjectClass *oc, void *data)
98
gen_aa32_ld64(s, tmp, addr, get_mem_index(s));
140
99
- neon_store_reg64(tmp, a->vd + i);
141
mc->desc = "Gumstix Connex (PXA255)";
100
+ vfp_store_reg64(tmp, a->vd + i);
142
mc->init = connex_init;
101
} else {
143
+ mc->ignore_memory_transaction_failures = true;
102
/* store */
144
}
103
- neon_load_reg64(tmp, a->vd + i);
145
104
+ vfp_load_reg64(tmp, a->vd + i);
146
static const TypeInfo connex_type = {
105
gen_aa32_st64(s, tmp, addr, get_mem_index(s));
147
@@ -XXX,XX +XXX,XX @@ static void verdex_class_init(ObjectClass *oc, void *data)
106
}
148
107
tcg_gen_addi_i32(addr, addr, offset);
149
mc->desc = "Gumstix Verdex (PXA270)";
108
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_3op_dp(DisasContext *s, VFPGen3OpDPFn *fn,
150
mc->init = verdex_init;
109
fd = tcg_temp_new_i64();
151
+ mc->ignore_memory_transaction_failures = true;
110
fpst = fpstatus_ptr(FPST_FPCR);
152
}
111
153
112
- neon_load_reg64(f0, vn);
154
static const TypeInfo verdex_type = {
113
- neon_load_reg64(f1, vm);
155
diff --git a/hw/arm/highbank.c b/hw/arm/highbank.c
114
+ vfp_load_reg64(f0, vn);
156
index XXXXXXX..XXXXXXX 100644
115
+ vfp_load_reg64(f1, vm);
157
--- a/hw/arm/highbank.c
116
158
+++ b/hw/arm/highbank.c
117
for (;;) {
159
@@ -XXX,XX +XXX,XX @@ static void highbank_class_init(ObjectClass *oc, void *data)
118
if (reads_vd) {
160
mc->block_default_type = IF_IDE;
119
- neon_load_reg64(fd, vd);
161
mc->units_per_default_bus = 1;
120
+ vfp_load_reg64(fd, vd);
162
mc->max_cpus = 4;
121
}
163
+ mc->ignore_memory_transaction_failures = true;
122
fn(fd, f0, f1, fpst);
164
}
123
- neon_store_reg64(fd, vd);
165
124
+ vfp_store_reg64(fd, vd);
166
static const TypeInfo highbank_type = {
125
167
@@ -XXX,XX +XXX,XX @@ static void midway_class_init(ObjectClass *oc, void *data)
126
if (veclen == 0) {
168
mc->block_default_type = IF_IDE;
127
break;
169
mc->units_per_default_bus = 1;
128
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_3op_dp(DisasContext *s, VFPGen3OpDPFn *fn,
170
mc->max_cpus = 4;
129
veclen--;
171
+ mc->ignore_memory_transaction_failures = true;
130
vd = vfp_advance_dreg(vd, delta_d);
172
}
131
vn = vfp_advance_dreg(vn, delta_d);
173
132
- neon_load_reg64(f0, vn);
174
static const TypeInfo midway_type = {
133
+ vfp_load_reg64(f0, vn);
175
diff --git a/hw/arm/imx25_pdk.c b/hw/arm/imx25_pdk.c
134
if (delta_m) {
176
index XXXXXXX..XXXXXXX 100644
135
vm = vfp_advance_dreg(vm, delta_m);
177
--- a/hw/arm/imx25_pdk.c
136
- neon_load_reg64(f1, vm);
178
+++ b/hw/arm/imx25_pdk.c
137
+ vfp_load_reg64(f1, vm);
179
@@ -XXX,XX +XXX,XX @@ static void imx25_pdk_machine_init(MachineClass *mc)
138
}
180
{
139
}
181
mc->desc = "ARM i.MX25 PDK board (ARM926)";
140
182
mc->init = imx25_pdk_init;
141
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_dp(DisasContext *s, VFPGen2OpDPFn *fn, int vd, int vm)
183
+ mc->ignore_memory_transaction_failures = true;
142
f0 = tcg_temp_new_i64();
184
}
143
fd = tcg_temp_new_i64();
185
144
186
DEFINE_MACHINE("imx25-pdk", imx25_pdk_machine_init)
145
- neon_load_reg64(f0, vm);
187
diff --git a/hw/arm/integratorcp.c b/hw/arm/integratorcp.c
146
+ vfp_load_reg64(f0, vm);
188
index XXXXXXX..XXXXXXX 100644
147
189
--- a/hw/arm/integratorcp.c
148
for (;;) {
190
+++ b/hw/arm/integratorcp.c
149
fn(fd, f0);
191
@@ -XXX,XX +XXX,XX @@ static void integratorcp_machine_init(MachineClass *mc)
150
- neon_store_reg64(fd, vd);
192
{
151
+ vfp_store_reg64(fd, vd);
193
mc->desc = "ARM Integrator/CP (ARM926EJ-S)";
152
194
mc->init = integratorcp_init;
153
if (veclen == 0) {
195
+ mc->ignore_memory_transaction_failures = true;
154
break;
196
}
155
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_dp(DisasContext *s, VFPGen2OpDPFn *fn, int vd, int vm)
197
156
/* single source one-many */
198
DEFINE_MACHINE("integratorcp", integratorcp_machine_init)
157
while (veclen--) {
199
diff --git a/hw/arm/kzm.c b/hw/arm/kzm.c
158
vd = vfp_advance_dreg(vd, delta_d);
200
index XXXXXXX..XXXXXXX 100644
159
- neon_store_reg64(fd, vd);
201
--- a/hw/arm/kzm.c
160
+ vfp_store_reg64(fd, vd);
202
+++ b/hw/arm/kzm.c
161
}
203
@@ -XXX,XX +XXX,XX @@ static void kzm_machine_init(MachineClass *mc)
162
break;
204
{
163
}
205
mc->desc = "ARM KZM Emulation Baseboard (ARM1136)";
164
@@ -XXX,XX +XXX,XX @@ static bool do_vfp_2op_dp(DisasContext *s, VFPGen2OpDPFn *fn, int vd, int vm)
206
mc->init = kzm_init;
165
veclen--;
207
+ mc->ignore_memory_transaction_failures = true;
166
vd = vfp_advance_dreg(vd, delta_d);
208
}
167
vd = vfp_advance_dreg(vm, delta_m);
209
168
- neon_load_reg64(f0, vm);
210
DEFINE_MACHINE("kzm", kzm_machine_init)
169
+ vfp_load_reg64(f0, vm);
211
diff --git a/hw/arm/mainstone.c b/hw/arm/mainstone.c
170
}
212
index XXXXXXX..XXXXXXX 100644
171
213
--- a/hw/arm/mainstone.c
172
tcg_temp_free_i64(f0);
214
+++ b/hw/arm/mainstone.c
173
@@ -XXX,XX +XXX,XX @@ static bool do_vfm_dp(DisasContext *s, arg_VFMA_dp *a, bool neg_n, bool neg_d)
215
@@ -XXX,XX +XXX,XX @@ static void mainstone2_machine_init(MachineClass *mc)
174
vm = tcg_temp_new_i64();
216
{
175
vd = tcg_temp_new_i64();
217
mc->desc = "Mainstone II (PXA27x)";
176
218
mc->init = mainstone_init;
177
- neon_load_reg64(vn, a->vn);
219
+ mc->ignore_memory_transaction_failures = true;
178
- neon_load_reg64(vm, a->vm);
220
}
179
+ vfp_load_reg64(vn, a->vn);
221
180
+ vfp_load_reg64(vm, a->vm);
222
DEFINE_MACHINE("mainstone", mainstone2_machine_init)
181
if (neg_n) {
223
diff --git a/hw/arm/musicpal.c b/hw/arm/musicpal.c
182
/* VFNMS, VFMS */
224
index XXXXXXX..XXXXXXX 100644
183
gen_helper_vfp_negd(vn, vn);
225
--- a/hw/arm/musicpal.c
184
}
226
+++ b/hw/arm/musicpal.c
185
- neon_load_reg64(vd, a->vd);
227
@@ -XXX,XX +XXX,XX @@ static void musicpal_machine_init(MachineClass *mc)
186
+ vfp_load_reg64(vd, a->vd);
228
{
187
if (neg_d) {
229
mc->desc = "Marvell 88w8618 / MusicPal (ARM926EJ-S)";
188
/* VFNMA, VFNMS */
230
mc->init = musicpal_init;
189
gen_helper_vfp_negd(vd, vd);
231
+ mc->ignore_memory_transaction_failures = true;
190
}
232
}
191
fpst = fpstatus_ptr(FPST_FPCR);
233
192
gen_helper_vfp_muladdd(vd, vn, vm, vd, fpst);
234
DEFINE_MACHINE("musicpal", musicpal_machine_init)
193
- neon_store_reg64(vd, a->vd);
235
diff --git a/hw/arm/netduino2.c b/hw/arm/netduino2.c
194
+ vfp_store_reg64(vd, a->vd);
236
index XXXXXXX..XXXXXXX 100644
195
237
--- a/hw/arm/netduino2.c
196
tcg_temp_free_ptr(fpst);
238
+++ b/hw/arm/netduino2.c
197
tcg_temp_free_i64(vn);
239
@@ -XXX,XX +XXX,XX @@ static void netduino2_machine_init(MachineClass *mc)
198
@@ -XXX,XX +XXX,XX @@ static bool trans_VMOV_imm_dp(DisasContext *s, arg_VMOV_imm_dp *a)
240
{
199
fd = tcg_const_i64(vfp_expand_imm(MO_64, a->imm));
241
mc->desc = "Netduino 2 Machine";
200
242
mc->init = netduino2_init;
201
for (;;) {
243
+ mc->ignore_memory_transaction_failures = true;
202
- neon_store_reg64(fd, vd);
244
}
203
+ vfp_store_reg64(fd, vd);
245
204
246
DEFINE_MACHINE("netduino2", netduino2_machine_init)
205
if (veclen == 0) {
247
diff --git a/hw/arm/nseries.c b/hw/arm/nseries.c
206
break;
248
index XXXXXXX..XXXXXXX 100644
207
@@ -XXX,XX +XXX,XX @@ static bool trans_VCMP_dp(DisasContext *s, arg_VCMP_dp *a)
249
--- a/hw/arm/nseries.c
208
vd = tcg_temp_new_i64();
250
+++ b/hw/arm/nseries.c
209
vm = tcg_temp_new_i64();
251
@@ -XXX,XX +XXX,XX @@ static void n800_class_init(ObjectClass *oc, void *data)
210
252
mc->desc = "Nokia N800 tablet aka. RX-34 (OMAP2420)";
211
- neon_load_reg64(vd, a->vd);
253
mc->init = n800_init;
212
+ vfp_load_reg64(vd, a->vd);
254
mc->default_boot_order = "";
213
if (a->z) {
255
+ mc->ignore_memory_transaction_failures = true;
214
tcg_gen_movi_i64(vm, 0);
256
}
215
} else {
257
216
- neon_load_reg64(vm, a->vm);
258
static const TypeInfo n800_type = {
217
+ vfp_load_reg64(vm, a->vm);
259
@@ -XXX,XX +XXX,XX @@ static void n810_class_init(ObjectClass *oc, void *data)
218
}
260
mc->desc = "Nokia N810 tablet aka. RX-44 (OMAP2420)";
219
261
mc->init = n810_init;
220
if (a->e) {
262
mc->default_boot_order = "";
221
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_f64_f16(DisasContext *s, arg_VCVT_f64_f16 *a)
263
+ mc->ignore_memory_transaction_failures = true;
222
tcg_gen_ld16u_i32(tmp, cpu_env, vfp_f16_offset(a->vm, a->t));
264
}
223
vd = tcg_temp_new_i64();
265
224
gen_helper_vfp_fcvt_f16_to_f64(vd, tmp, fpst, ahp_mode);
266
static const TypeInfo n810_type = {
225
- neon_store_reg64(vd, a->vd);
267
diff --git a/hw/arm/omap_sx1.c b/hw/arm/omap_sx1.c
226
+ vfp_store_reg64(vd, a->vd);
268
index XXXXXXX..XXXXXXX 100644
227
tcg_temp_free_i32(ahp_mode);
269
--- a/hw/arm/omap_sx1.c
228
tcg_temp_free_ptr(fpst);
270
+++ b/hw/arm/omap_sx1.c
229
tcg_temp_free_i32(tmp);
271
@@ -XXX,XX +XXX,XX @@ static void sx1_machine_v2_class_init(ObjectClass *oc, void *data)
230
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_f16_f64(DisasContext *s, arg_VCVT_f16_f64 *a)
272
231
tmp = tcg_temp_new_i32();
273
mc->desc = "Siemens SX1 (OMAP310) V2";
232
vm = tcg_temp_new_i64();
274
mc->init = sx1_init_v2;
233
275
+ mc->ignore_memory_transaction_failures = true;
234
- neon_load_reg64(vm, a->vm);
276
}
235
+ vfp_load_reg64(vm, a->vm);
277
236
gen_helper_vfp_fcvt_f64_to_f16(tmp, vm, fpst, ahp_mode);
278
static const TypeInfo sx1_machine_v2_type = {
237
tcg_temp_free_i64(vm);
279
@@ -XXX,XX +XXX,XX @@ static void sx1_machine_v1_class_init(ObjectClass *oc, void *data)
238
tcg_gen_st16_i32(tmp, cpu_env, vfp_f16_offset(a->vd, a->t));
280
239
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTR_dp(DisasContext *s, arg_VRINTR_dp *a)
281
mc->desc = "Siemens SX1 (OMAP310) V1";
240
}
282
mc->init = sx1_init_v1;
241
283
+ mc->ignore_memory_transaction_failures = true;
242
tmp = tcg_temp_new_i64();
284
}
243
- neon_load_reg64(tmp, a->vm);
285
244
+ vfp_load_reg64(tmp, a->vm);
286
static const TypeInfo sx1_machine_v1_type = {
245
fpst = fpstatus_ptr(FPST_FPCR);
287
diff --git a/hw/arm/palm.c b/hw/arm/palm.c
246
gen_helper_rintd(tmp, tmp, fpst);
288
index XXXXXXX..XXXXXXX 100644
247
- neon_store_reg64(tmp, a->vd);
289
--- a/hw/arm/palm.c
248
+ vfp_store_reg64(tmp, a->vd);
290
+++ b/hw/arm/palm.c
249
tcg_temp_free_ptr(fpst);
291
@@ -XXX,XX +XXX,XX @@ static void palmte_machine_init(MachineClass *mc)
250
tcg_temp_free_i64(tmp);
292
{
251
return true;
293
mc->desc = "Palm Tungsten|E aka. Cheetah PDA (OMAP310)";
252
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTZ_dp(DisasContext *s, arg_VRINTZ_dp *a)
294
mc->init = palmte_init;
253
}
295
+ mc->ignore_memory_transaction_failures = true;
254
296
}
255
tmp = tcg_temp_new_i64();
297
256
- neon_load_reg64(tmp, a->vm);
298
DEFINE_MACHINE("cheetah", palmte_machine_init)
257
+ vfp_load_reg64(tmp, a->vm);
299
diff --git a/hw/arm/raspi.c b/hw/arm/raspi.c
258
fpst = fpstatus_ptr(FPST_FPCR);
300
index XXXXXXX..XXXXXXX 100644
259
tcg_rmode = tcg_const_i32(float_round_to_zero);
301
--- a/hw/arm/raspi.c
260
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
302
+++ b/hw/arm/raspi.c
261
gen_helper_rintd(tmp, tmp, fpst);
303
@@ -XXX,XX +XXX,XX @@ static void raspi2_machine_init(MachineClass *mc)
262
gen_helper_set_rmode(tcg_rmode, tcg_rmode, fpst);
304
mc->no_cdrom = 1;
263
- neon_store_reg64(tmp, a->vd);
305
mc->max_cpus = BCM2836_NCPUS;
264
+ vfp_store_reg64(tmp, a->vd);
306
mc->default_ram_size = 1024 * 1024 * 1024;
265
tcg_temp_free_ptr(fpst);
307
+ mc->ignore_memory_transaction_failures = true;
266
tcg_temp_free_i64(tmp);
308
};
267
tcg_temp_free_i32(tcg_rmode);
309
DEFINE_MACHINE("raspi2", raspi2_machine_init)
268
@@ -XXX,XX +XXX,XX @@ static bool trans_VRINTX_dp(DisasContext *s, arg_VRINTX_dp *a)
310
diff --git a/hw/arm/realview.c b/hw/arm/realview.c
269
}
311
index XXXXXXX..XXXXXXX 100644
270
312
--- a/hw/arm/realview.c
271
tmp = tcg_temp_new_i64();
313
+++ b/hw/arm/realview.c
272
- neon_load_reg64(tmp, a->vm);
314
@@ -XXX,XX +XXX,XX @@ static void realview_eb_class_init(ObjectClass *oc, void *data)
273
+ vfp_load_reg64(tmp, a->vm);
315
mc->desc = "ARM RealView Emulation Baseboard (ARM926EJ-S)";
274
fpst = fpstatus_ptr(FPST_FPCR);
316
mc->init = realview_eb_init;
275
gen_helper_rintd_exact(tmp, tmp, fpst);
317
mc->block_default_type = IF_SCSI;
276
- neon_store_reg64(tmp, a->vd);
318
+ mc->ignore_memory_transaction_failures = true;
277
+ vfp_store_reg64(tmp, a->vd);
319
}
278
tcg_temp_free_ptr(fpst);
320
279
tcg_temp_free_i64(tmp);
321
static const TypeInfo realview_eb_type = {
280
return true;
322
@@ -XXX,XX +XXX,XX @@ static void realview_eb_mpcore_class_init(ObjectClass *oc, void *data)
281
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_sp(DisasContext *s, arg_VCVT_sp *a)
323
mc->init = realview_eb_mpcore_init;
282
vd = tcg_temp_new_i64();
324
mc->block_default_type = IF_SCSI;
283
vfp_load_reg32(vm, a->vm);
325
mc->max_cpus = 4;
284
gen_helper_vfp_fcvtds(vd, vm, cpu_env);
326
+ mc->ignore_memory_transaction_failures = true;
285
- neon_store_reg64(vd, a->vd);
327
}
286
+ vfp_store_reg64(vd, a->vd);
328
287
tcg_temp_free_i32(vm);
329
static const TypeInfo realview_eb_mpcore_type = {
288
tcg_temp_free_i64(vd);
330
@@ -XXX,XX +XXX,XX @@ static void realview_pb_a8_class_init(ObjectClass *oc, void *data)
289
return true;
331
290
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_dp(DisasContext *s, arg_VCVT_dp *a)
332
mc->desc = "ARM RealView Platform Baseboard for Cortex-A8";
291
333
mc->init = realview_pb_a8_init;
292
vd = tcg_temp_new_i32();
334
+ mc->ignore_memory_transaction_failures = true;
293
vm = tcg_temp_new_i64();
335
}
294
- neon_load_reg64(vm, a->vm);
336
295
+ vfp_load_reg64(vm, a->vm);
337
static const TypeInfo realview_pb_a8_type = {
296
gen_helper_vfp_fcvtsd(vd, vm, cpu_env);
338
@@ -XXX,XX +XXX,XX @@ static void realview_pbx_a9_class_init(ObjectClass *oc, void *data)
297
vfp_store_reg32(vd, a->vd);
339
mc->desc = "ARM RealView Platform Baseboard Explore for Cortex-A9";
298
tcg_temp_free_i32(vd);
340
mc->init = realview_pbx_a9_init;
299
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_int_dp(DisasContext *s, arg_VCVT_int_dp *a)
341
mc->max_cpus = 4;
300
/* u32 -> f64 */
342
+ mc->ignore_memory_transaction_failures = true;
301
gen_helper_vfp_uitod(vd, vm, fpst);
343
}
302
}
344
303
- neon_store_reg64(vd, a->vd);
345
static const TypeInfo realview_pbx_a9_type = {
304
+ vfp_store_reg64(vd, a->vd);
346
diff --git a/hw/arm/sabrelite.c b/hw/arm/sabrelite.c
305
tcg_temp_free_i32(vm);
347
index XXXXXXX..XXXXXXX 100644
306
tcg_temp_free_i64(vd);
348
--- a/hw/arm/sabrelite.c
307
tcg_temp_free_ptr(fpst);
349
+++ b/hw/arm/sabrelite.c
308
@@ -XXX,XX +XXX,XX @@ static bool trans_VJCVT(DisasContext *s, arg_VJCVT *a)
350
@@ -XXX,XX +XXX,XX @@ static void sabrelite_machine_init(MachineClass *mc)
309
351
mc->desc = "Freescale i.MX6 Quad SABRE Lite Board (Cortex A9)";
310
vm = tcg_temp_new_i64();
352
mc->init = sabrelite_init;
311
vd = tcg_temp_new_i32();
353
mc->max_cpus = FSL_IMX6_NUM_CPUS;
312
- neon_load_reg64(vm, a->vm);
354
+ mc->ignore_memory_transaction_failures = true;
313
+ vfp_load_reg64(vm, a->vm);
355
}
314
gen_helper_vjcvt(vd, vm, cpu_env);
356
315
vfp_store_reg32(vd, a->vd);
357
DEFINE_MACHINE("sabrelite", sabrelite_machine_init)
316
tcg_temp_free_i64(vm);
358
diff --git a/hw/arm/spitz.c b/hw/arm/spitz.c
317
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_dp(DisasContext *s, arg_VCVT_fix_dp *a)
359
index XXXXXXX..XXXXXXX 100644
318
frac_bits = (a->opc & 1) ? (32 - a->imm) : (16 - a->imm);
360
--- a/hw/arm/spitz.c
319
361
+++ b/hw/arm/spitz.c
320
vd = tcg_temp_new_i64();
362
@@ -XXX,XX +XXX,XX @@ static void akitapda_class_init(ObjectClass *oc, void *data)
321
- neon_load_reg64(vd, a->vd);
363
322
+ vfp_load_reg64(vd, a->vd);
364
mc->desc = "Sharp SL-C1000 (Akita) PDA (PXA270)";
323
365
mc->init = akita_init;
324
fpst = fpstatus_ptr(FPST_FPCR);
366
+ mc->ignore_memory_transaction_failures = true;
325
shift = tcg_const_i32(frac_bits);
367
}
326
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_fix_dp(DisasContext *s, arg_VCVT_fix_dp *a)
368
327
g_assert_not_reached();
369
static const TypeInfo akitapda_type = {
328
}
370
@@ -XXX,XX +XXX,XX @@ static void spitzpda_class_init(ObjectClass *oc, void *data)
329
371
mc->desc = "Sharp SL-C3000 (Spitz) PDA (PXA270)";
330
- neon_store_reg64(vd, a->vd);
372
mc->init = spitz_init;
331
+ vfp_store_reg64(vd, a->vd);
373
mc->block_default_type = IF_IDE;
332
tcg_temp_free_i64(vd);
374
+ mc->ignore_memory_transaction_failures = true;
333
tcg_temp_free_i32(shift);
375
}
334
tcg_temp_free_ptr(fpst);
376
335
@@ -XXX,XX +XXX,XX @@ static bool trans_VCVT_dp_int(DisasContext *s, arg_VCVT_dp_int *a)
377
static const TypeInfo spitzpda_type = {
336
fpst = fpstatus_ptr(FPST_FPCR);
378
@@ -XXX,XX +XXX,XX @@ static void borzoipda_class_init(ObjectClass *oc, void *data)
337
vm = tcg_temp_new_i64();
379
mc->desc = "Sharp SL-C3100 (Borzoi) PDA (PXA270)";
338
vd = tcg_temp_new_i32();
380
mc->init = borzoi_init;
339
- neon_load_reg64(vm, a->vm);
381
mc->block_default_type = IF_IDE;
340
+ vfp_load_reg64(vm, a->vm);
382
+ mc->ignore_memory_transaction_failures = true;
341
383
}
342
if (a->s) {
384
343
if (a->rz) {
385
static const TypeInfo borzoipda_type = {
386
@@ -XXX,XX +XXX,XX @@ static void terrierpda_class_init(ObjectClass *oc, void *data)
387
mc->desc = "Sharp SL-C3200 (Terrier) PDA (PXA270)";
388
mc->init = terrier_init;
389
mc->block_default_type = IF_IDE;
390
+ mc->ignore_memory_transaction_failures = true;
391
}
392
393
static const TypeInfo terrierpda_type = {
394
diff --git a/hw/arm/stellaris.c b/hw/arm/stellaris.c
395
index XXXXXXX..XXXXXXX 100644
396
--- a/hw/arm/stellaris.c
397
+++ b/hw/arm/stellaris.c
398
@@ -XXX,XX +XXX,XX @@ static void lm3s811evb_class_init(ObjectClass *oc, void *data)
399
400
mc->desc = "Stellaris LM3S811EVB";
401
mc->init = lm3s811evb_init;
402
+ mc->ignore_memory_transaction_failures = true;
403
}
404
405
static const TypeInfo lm3s811evb_type = {
406
@@ -XXX,XX +XXX,XX @@ static void lm3s6965evb_class_init(ObjectClass *oc, void *data)
407
408
mc->desc = "Stellaris LM3S6965EVB";
409
mc->init = lm3s6965evb_init;
410
+ mc->ignore_memory_transaction_failures = true;
411
}
412
413
static const TypeInfo lm3s6965evb_type = {
414
diff --git a/hw/arm/tosa.c b/hw/arm/tosa.c
415
index XXXXXXX..XXXXXXX 100644
416
--- a/hw/arm/tosa.c
417
+++ b/hw/arm/tosa.c
418
@@ -XXX,XX +XXX,XX @@ static void tosapda_machine_init(MachineClass *mc)
419
mc->desc = "Sharp SL-6000 (Tosa) PDA (PXA255)";
420
mc->init = tosa_init;
421
mc->block_default_type = IF_IDE;
422
+ mc->ignore_memory_transaction_failures = true;
423
}
424
425
DEFINE_MACHINE("tosa", tosapda_machine_init)
426
diff --git a/hw/arm/versatilepb.c b/hw/arm/versatilepb.c
427
index XXXXXXX..XXXXXXX 100644
428
--- a/hw/arm/versatilepb.c
429
+++ b/hw/arm/versatilepb.c
430
@@ -XXX,XX +XXX,XX @@ static void versatilepb_class_init(ObjectClass *oc, void *data)
431
mc->desc = "ARM Versatile/PB (ARM926EJ-S)";
432
mc->init = vpb_init;
433
mc->block_default_type = IF_SCSI;
434
+ mc->ignore_memory_transaction_failures = true;
435
}
436
437
static const TypeInfo versatilepb_type = {
438
@@ -XXX,XX +XXX,XX @@ static void versatileab_class_init(ObjectClass *oc, void *data)
439
mc->desc = "ARM Versatile/AB (ARM926EJ-S)";
440
mc->init = vab_init;
441
mc->block_default_type = IF_SCSI;
442
+ mc->ignore_memory_transaction_failures = true;
443
}
444
445
static const TypeInfo versatileab_type = {
446
diff --git a/hw/arm/vexpress.c b/hw/arm/vexpress.c
447
index XXXXXXX..XXXXXXX 100644
448
--- a/hw/arm/vexpress.c
449
+++ b/hw/arm/vexpress.c
450
@@ -XXX,XX +XXX,XX @@ static void vexpress_class_init(ObjectClass *oc, void *data)
451
mc->desc = "ARM Versatile Express";
452
mc->init = vexpress_common_init;
453
mc->max_cpus = 4;
454
+ mc->ignore_memory_transaction_failures = true;
455
}
456
457
static void vexpress_a9_class_init(ObjectClass *oc, void *data)
458
diff --git a/hw/arm/xilinx_zynq.c b/hw/arm/xilinx_zynq.c
459
index XXXXXXX..XXXXXXX 100644
460
--- a/hw/arm/xilinx_zynq.c
461
+++ b/hw/arm/xilinx_zynq.c
462
@@ -XXX,XX +XXX,XX @@ static void zynq_machine_init(MachineClass *mc)
463
mc->init = zynq_init;
464
mc->max_cpus = 1;
465
mc->no_sdcard = 1;
466
+ mc->ignore_memory_transaction_failures = true;
467
}
468
469
DEFINE_MACHINE("xilinx-zynq-a9", zynq_machine_init)
470
diff --git a/hw/arm/xlnx-ep108.c b/hw/arm/xlnx-ep108.c
471
index XXXXXXX..XXXXXXX 100644
472
--- a/hw/arm/xlnx-ep108.c
473
+++ b/hw/arm/xlnx-ep108.c
474
@@ -XXX,XX +XXX,XX @@ static void xlnx_ep108_machine_init(MachineClass *mc)
475
mc->init = xlnx_ep108_init;
476
mc->block_default_type = IF_IDE;
477
mc->units_per_default_bus = 1;
478
+ mc->ignore_memory_transaction_failures = true;
479
}
480
481
DEFINE_MACHINE("xlnx-ep108", xlnx_ep108_machine_init)
482
@@ -XXX,XX +XXX,XX @@ static void xlnx_zcu102_machine_init(MachineClass *mc)
483
mc->init = xlnx_ep108_init;
484
mc->block_default_type = IF_IDE;
485
mc->units_per_default_bus = 1;
486
+ mc->ignore_memory_transaction_failures = true;
487
}
488
489
DEFINE_MACHINE("xlnx-zcu102", xlnx_zcu102_machine_init)
490
diff --git a/hw/arm/z2.c b/hw/arm/z2.c
491
index XXXXXXX..XXXXXXX 100644
492
--- a/hw/arm/z2.c
493
+++ b/hw/arm/z2.c
494
@@ -XXX,XX +XXX,XX @@ static void z2_machine_init(MachineClass *mc)
495
{
496
mc->desc = "Zipit Z2 (PXA27x)";
497
mc->init = z2_init;
498
+ mc->ignore_memory_transaction_failures = true;
499
}
500
501
DEFINE_MACHINE("z2", z2_machine_init)
502
--
344
--
503
2.7.4
345
2.20.1
504
346
505
347
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
In both cases, we can sink the write-back and perform
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
the accumulate into the normal destination temps.
5
Message-id: 20170905131149.10669-2-famz@redhat.com
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201030022618.785675-11-richard.henderson@linaro.org
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
10
---
9
hw/arm/armv7m.c | 8 ++------
11
target/arm/translate-neon.c.inc | 23 +++++++++--------------
10
1 file changed, 2 insertions(+), 6 deletions(-)
12
1 file changed, 9 insertions(+), 14 deletions(-)
11
13
12
diff --git a/hw/arm/armv7m.c b/hw/arm/armv7m.c
14
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
13
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/arm/armv7m.c
16
--- a/target/arm/translate-neon.c.inc
15
+++ b/hw/arm/armv7m.c
17
+++ b/target/arm/translate-neon.c.inc
16
@@ -XXX,XX +XXX,XX @@ static void bitband_init(Object *obj)
18
@@ -XXX,XX +XXX,XX @@ static bool do_long_3d(DisasContext *s, arg_3diff *a,
17
BitBandState *s = BITBAND(obj);
19
if (accfn) {
18
SysBusDevice *dev = SYS_BUS_DEVICE(obj);
20
tmp = tcg_temp_new_i64();
19
21
read_neon_element64(tmp, a->vd, 0, MO_64);
20
- object_property_add_link(obj, "source-memory",
22
- accfn(tmp, tmp, rd0);
21
- TYPE_MEMORY_REGION,
23
- write_neon_element64(tmp, a->vd, 0, MO_64);
22
- (Object **)&s->source_memory,
24
+ accfn(rd0, tmp, rd0);
23
- qdev_prop_allow_set_link_before_realize,
25
read_neon_element64(tmp, a->vd, 1, MO_64);
24
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
26
- accfn(tmp, tmp, rd1);
25
- &error_abort);
27
- write_neon_element64(tmp, a->vd, 1, MO_64);
26
memory_region_init_io(&s->iomem, obj, &bitband_ops, s,
28
+ accfn(rd1, tmp, rd1);
27
"bitband", 0x02000000);
29
tcg_temp_free_i64(tmp);
28
sysbus_init_mmio(dev, &s->iomem);
30
- } else {
29
@@ -XXX,XX +XXX,XX @@ void armv7m_load_kernel(ARMCPU *cpu, const char *kernel_filename, int mem_size)
31
- write_neon_element64(rd0, a->vd, 0, MO_64);
30
32
- write_neon_element64(rd1, a->vd, 1, MO_64);
31
static Property bitband_properties[] = {
33
}
32
DEFINE_PROP_UINT32("base", BitBandState, base, 0),
34
33
+ DEFINE_PROP_LINK("source-memory", BitBandState, source_memory,
35
+ write_neon_element64(rd0, a->vd, 0, MO_64);
34
+ TYPE_MEMORY_REGION, MemoryRegion *),
36
+ write_neon_element64(rd1, a->vd, 1, MO_64);
35
DEFINE_PROP_END_OF_LIST(),
37
tcg_temp_free_i64(rd0);
36
};
38
tcg_temp_free_i64(rd1);
37
39
40
@@ -XXX,XX +XXX,XX @@ static bool do_2scalar_long(DisasContext *s, arg_2scalar *a,
41
if (accfn) {
42
TCGv_i64 t64 = tcg_temp_new_i64();
43
read_neon_element64(t64, a->vd, 0, MO_64);
44
- accfn(t64, t64, rn0_64);
45
- write_neon_element64(t64, a->vd, 0, MO_64);
46
+ accfn(rn0_64, t64, rn0_64);
47
read_neon_element64(t64, a->vd, 1, MO_64);
48
- accfn(t64, t64, rn1_64);
49
- write_neon_element64(t64, a->vd, 1, MO_64);
50
+ accfn(rn1_64, t64, rn1_64);
51
tcg_temp_free_i64(t64);
52
- } else {
53
- write_neon_element64(rn0_64, a->vd, 0, MO_64);
54
- write_neon_element64(rn1_64, a->vd, 1, MO_64);
55
}
56
+
57
+ write_neon_element64(rn0_64, a->vd, 0, MO_64);
58
+ write_neon_element64(rn1_64, a->vd, 1, MO_64);
59
tcg_temp_free_i64(rn0_64);
60
tcg_temp_free_i64(rn1_64);
61
return true;
38
--
62
--
39
2.7.4
63
2.20.1
40
64
41
65
diff view generated by jsdifflib
1
Make the MPU_RNR register banked if v8M security extensions are
1
From: Richard Henderson <richard.henderson@linaro.org>
2
enabled.
3
2
3
We can use proper widening loads to extend 32-bit inputs,
4
and skip the "widenfn" step.
5
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201030022618.785675-12-richard.henderson@linaro.org
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 1503414539-28762-15-git-send-email-peter.maydell@linaro.org
7
---
10
---
8
target/arm/cpu.h | 2 +-
11
target/arm/translate.c | 6 +++
9
hw/intc/armv7m_nvic.c | 18 +++++++++---------
12
target/arm/translate-neon.c.inc | 66 ++++++++++++++++++---------------
10
target/arm/cpu.c | 3 ++-
13
2 files changed, 43 insertions(+), 29 deletions(-)
11
target/arm/helper.c | 6 +++---
12
target/arm/machine.c | 13 +++++++++++--
13
5 files changed, 26 insertions(+), 16 deletions(-)
14
14
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/translate.c b/target/arm/translate.c
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.h
17
--- a/target/arm/translate.c
18
+++ b/target/arm/cpu.h
18
+++ b/target/arm/translate.c
19
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
19
@@ -XXX,XX +XXX,XX @@ static void read_neon_element64(TCGv_i64 dest, int reg, int ele, MemOp memop)
20
uint32_t *drbar;
20
long off = neon_element_offset(reg, ele, memop);
21
uint32_t *drsr;
21
22
uint32_t *dracr;
22
switch (memop) {
23
- uint32_t rnr;
23
+ case MO_SL:
24
+ uint32_t rnr[2];
24
+ tcg_gen_ld32s_i64(dest, cpu_env, off);
25
} pmsav7;
25
+ break;
26
26
+ case MO_UL:
27
/* PMSAv8 MPU */
27
+ tcg_gen_ld32u_i64(dest, cpu_env, off);
28
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
28
+ break;
29
case MO_Q:
30
tcg_gen_ld_i64(dest, cpu_env, off);
31
break;
32
diff --git a/target/arm/translate-neon.c.inc b/target/arm/translate-neon.c.inc
29
index XXXXXXX..XXXXXXX 100644
33
index XXXXXXX..XXXXXXX 100644
30
--- a/hw/intc/armv7m_nvic.c
34
--- a/target/arm/translate-neon.c.inc
31
+++ b/hw/intc/armv7m_nvic.c
35
+++ b/target/arm/translate-neon.c.inc
32
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
36
@@ -XXX,XX +XXX,XX @@ static bool trans_Vimm_1r(DisasContext *s, arg_1reg_imm *a)
33
case 0xd94: /* MPU_CTRL */
37
static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
34
return cpu->env.v7m.mpu_ctrl;
38
NeonGenWidenFn *widenfn,
35
case 0xd98: /* MPU_RNR */
39
NeonGenTwo64OpFn *opfn,
36
- return cpu->env.pmsav7.rnr;
40
- bool src1_wide)
37
+ return cpu->env.pmsav7.rnr[attrs.secure];
41
+ int src1_mop, int src2_mop)
38
case 0xd9c: /* MPU_RBAR */
42
{
39
case 0xda4: /* MPU_RBAR_A1 */
43
/* 3-regs different lengths, prewidening case (VADDL/VSUBL/VAADW/VSUBW) */
40
case 0xdac: /* MPU_RBAR_A2 */
44
TCGv_i64 rn0_64, rn1_64, rm_64;
41
case 0xdb4: /* MPU_RBAR_A3 */
45
- TCGv_i32 rm;
42
{
46
43
- int region = cpu->env.pmsav7.rnr;
47
if (!arm_dc_feature(s, ARM_FEATURE_NEON)) {
44
+ int region = cpu->env.pmsav7.rnr[attrs.secure];
48
return false;
45
49
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
46
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
50
return false;
47
/* PMSAv8M handling of the aliases is different from v7M:
48
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
49
case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
50
case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
51
{
52
- int region = cpu->env.pmsav7.rnr;
53
+ int region = cpu->env.pmsav7.rnr[attrs.secure];
54
55
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
56
/* PMSAv8M handling of the aliases is different from v7M:
57
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
58
PRIu32 "/%" PRIu32 "\n",
59
value, cpu->pmsav7_dregion);
60
} else {
61
- cpu->env.pmsav7.rnr = value;
62
+ cpu->env.pmsav7.rnr[attrs.secure] = value;
63
}
64
break;
65
case 0xd9c: /* MPU_RBAR */
66
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
67
*/
68
int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
69
70
- region = cpu->env.pmsav7.rnr;
71
+ region = cpu->env.pmsav7.rnr[attrs.secure];
72
if (aliasno) {
73
region = deposit32(region, 0, 2, aliasno);
74
}
75
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
76
region, cpu->pmsav7_dregion);
77
return;
78
}
79
- cpu->env.pmsav7.rnr = region;
80
+ cpu->env.pmsav7.rnr[attrs.secure] = region;
81
} else {
82
- region = cpu->env.pmsav7.rnr;
83
+ region = cpu->env.pmsav7.rnr[attrs.secure];
84
}
85
86
if (region >= cpu->pmsav7_dregion) {
87
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
88
case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
89
case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
90
{
91
- int region = cpu->env.pmsav7.rnr;
92
+ int region = cpu->env.pmsav7.rnr[attrs.secure];
93
94
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
95
/* PMSAv8M handling of the aliases is different from v7M:
96
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
97
*/
98
int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
99
100
- region = cpu->env.pmsav7.rnr;
101
+ region = cpu->env.pmsav7.rnr[attrs.secure];
102
if (aliasno) {
103
region = deposit32(region, 0, 2, aliasno);
104
}
105
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
106
index XXXXXXX..XXXXXXX 100644
107
--- a/target/arm/cpu.c
108
+++ b/target/arm/cpu.c
109
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
110
sizeof(*env->pmsav7.dracr) * cpu->pmsav7_dregion);
111
}
112
}
113
- env->pmsav7.rnr = 0;
114
+ env->pmsav7.rnr[M_REG_NS] = 0;
115
+ env->pmsav7.rnr[M_REG_S] = 0;
116
env->pmsav8.mair0[M_REG_NS] = 0;
117
env->pmsav8.mair0[M_REG_S] = 0;
118
env->pmsav8.mair1[M_REG_NS] = 0;
119
diff --git a/target/arm/helper.c b/target/arm/helper.c
120
index XXXXXXX..XXXXXXX 100644
121
--- a/target/arm/helper.c
122
+++ b/target/arm/helper.c
123
@@ -XXX,XX +XXX,XX @@ static uint64_t pmsav7_read(CPUARMState *env, const ARMCPRegInfo *ri)
124
return 0;
125
}
51
}
126
52
127
- u32p += env->pmsav7.rnr;
53
- if (!widenfn || !opfn) {
128
+ u32p += env->pmsav7.rnr[M_REG_NS];
54
+ if (!opfn) {
129
return *u32p;
55
/* size == 3 case, which is an entirely different insn group */
56
return false;
57
}
58
59
- if ((a->vd & 1) || (src1_wide && (a->vn & 1))) {
60
+ if ((a->vd & 1) || (src1_mop == MO_Q && (a->vn & 1))) {
61
return false;
62
}
63
64
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
65
rn1_64 = tcg_temp_new_i64();
66
rm_64 = tcg_temp_new_i64();
67
68
- if (src1_wide) {
69
- read_neon_element64(rn0_64, a->vn, 0, MO_64);
70
+ if (src1_mop >= 0) {
71
+ read_neon_element64(rn0_64, a->vn, 0, src1_mop);
72
} else {
73
TCGv_i32 tmp = tcg_temp_new_i32();
74
read_neon_element32(tmp, a->vn, 0, MO_32);
75
widenfn(rn0_64, tmp);
76
tcg_temp_free_i32(tmp);
77
}
78
- rm = tcg_temp_new_i32();
79
- read_neon_element32(rm, a->vm, 0, MO_32);
80
+ if (src2_mop >= 0) {
81
+ read_neon_element64(rm_64, a->vm, 0, src2_mop);
82
+ } else {
83
+ TCGv_i32 tmp = tcg_temp_new_i32();
84
+ read_neon_element32(tmp, a->vm, 0, MO_32);
85
+ widenfn(rm_64, tmp);
86
+ tcg_temp_free_i32(tmp);
87
+ }
88
89
- widenfn(rm_64, rm);
90
- tcg_temp_free_i32(rm);
91
opfn(rn0_64, rn0_64, rm_64);
92
93
/*
94
* Load second pass inputs before storing the first pass result, to
95
* avoid incorrect results if a narrow input overlaps with the result.
96
*/
97
- if (src1_wide) {
98
- read_neon_element64(rn1_64, a->vn, 1, MO_64);
99
+ if (src1_mop >= 0) {
100
+ read_neon_element64(rn1_64, a->vn, 1, src1_mop);
101
} else {
102
TCGv_i32 tmp = tcg_temp_new_i32();
103
read_neon_element32(tmp, a->vn, 1, MO_32);
104
widenfn(rn1_64, tmp);
105
tcg_temp_free_i32(tmp);
106
}
107
- rm = tcg_temp_new_i32();
108
- read_neon_element32(rm, a->vm, 1, MO_32);
109
+ if (src2_mop >= 0) {
110
+ read_neon_element64(rm_64, a->vm, 1, src2_mop);
111
+ } else {
112
+ TCGv_i32 tmp = tcg_temp_new_i32();
113
+ read_neon_element32(tmp, a->vm, 1, MO_32);
114
+ widenfn(rm_64, tmp);
115
+ tcg_temp_free_i32(tmp);
116
+ }
117
118
write_neon_element64(rn0_64, a->vd, 0, MO_64);
119
120
- widenfn(rm_64, rm);
121
- tcg_temp_free_i32(rm);
122
opfn(rn1_64, rn1_64, rm_64);
123
write_neon_element64(rn1_64, a->vd, 1, MO_64);
124
125
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
126
return true;
130
}
127
}
131
128
132
@@ -XXX,XX +XXX,XX @@ static void pmsav7_write(CPUARMState *env, const ARMCPRegInfo *ri,
129
-#define DO_PREWIDEN(INSN, S, EXT, OP, SRC1WIDE) \
133
return;
130
+#define DO_PREWIDEN(INSN, S, OP, SRC1WIDE, SIGN) \
131
static bool trans_##INSN##_3d(DisasContext *s, arg_3diff *a) \
132
{ \
133
static NeonGenWidenFn * const widenfn[] = { \
134
gen_helper_neon_widen_##S##8, \
135
gen_helper_neon_widen_##S##16, \
136
- tcg_gen_##EXT##_i32_i64, \
137
- NULL, \
138
+ NULL, NULL, \
139
}; \
140
static NeonGenTwo64OpFn * const addfn[] = { \
141
gen_helper_neon_##OP##l_u16, \
142
@@ -XXX,XX +XXX,XX @@ static bool do_prewiden_3d(DisasContext *s, arg_3diff *a,
143
tcg_gen_##OP##_i64, \
144
NULL, \
145
}; \
146
- return do_prewiden_3d(s, a, widenfn[a->size], \
147
- addfn[a->size], SRC1WIDE); \
148
+ int narrow_mop = a->size == MO_32 ? MO_32 | SIGN : -1; \
149
+ return do_prewiden_3d(s, a, widenfn[a->size], addfn[a->size], \
150
+ SRC1WIDE ? MO_Q : narrow_mop, \
151
+ narrow_mop); \
134
}
152
}
135
153
136
- u32p += env->pmsav7.rnr;
154
-DO_PREWIDEN(VADDL_S, s, ext, add, false)
137
+ u32p += env->pmsav7.rnr[M_REG_NS];
155
-DO_PREWIDEN(VADDL_U, u, extu, add, false)
138
tlb_flush(CPU(cpu)); /* Mappings may have changed - purge! */
156
-DO_PREWIDEN(VSUBL_S, s, ext, sub, false)
139
*u32p = value;
157
-DO_PREWIDEN(VSUBL_U, u, extu, sub, false)
140
}
158
-DO_PREWIDEN(VADDW_S, s, ext, add, true)
141
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo pmsav7_cp_reginfo[] = {
159
-DO_PREWIDEN(VADDW_U, u, extu, add, true)
142
.resetfn = arm_cp_reset_ignore },
160
-DO_PREWIDEN(VSUBW_S, s, ext, sub, true)
143
{ .name = "RGNR", .cp = 15, .crn = 6, .opc1 = 0, .crm = 2, .opc2 = 0,
161
-DO_PREWIDEN(VSUBW_U, u, extu, sub, true)
144
.access = PL1_RW,
162
+DO_PREWIDEN(VADDL_S, s, add, false, MO_SIGN)
145
- .fieldoffset = offsetof(CPUARMState, pmsav7.rnr),
163
+DO_PREWIDEN(VADDL_U, u, add, false, 0)
146
+ .fieldoffset = offsetof(CPUARMState, pmsav7.rnr[M_REG_NS]),
164
+DO_PREWIDEN(VSUBL_S, s, sub, false, MO_SIGN)
147
.writefn = pmsav7_rgnr_write,
165
+DO_PREWIDEN(VSUBL_U, u, sub, false, 0)
148
.resetfn = arm_cp_reset_ignore },
166
+DO_PREWIDEN(VADDW_S, s, add, true, MO_SIGN)
149
REGINFO_SENTINEL
167
+DO_PREWIDEN(VADDW_U, u, add, true, 0)
150
diff --git a/target/arm/machine.c b/target/arm/machine.c
168
+DO_PREWIDEN(VSUBW_S, s, sub, true, MO_SIGN)
151
index XXXXXXX..XXXXXXX 100644
169
+DO_PREWIDEN(VSUBW_U, u, sub, true, 0)
152
--- a/target/arm/machine.c
170
153
+++ b/target/arm/machine.c
171
static bool do_narrow_3d(DisasContext *s, arg_3diff *a,
154
@@ -XXX,XX +XXX,XX @@ static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
172
NeonGenTwo64OpFn *opfn, NeonGenNarrowFn *narrowfn)
155
{
156
ARMCPU *cpu = opaque;
157
158
- return cpu->env.pmsav7.rnr < cpu->pmsav7_dregion;
159
+ return cpu->env.pmsav7.rnr[M_REG_NS] < cpu->pmsav7_dregion;
160
}
161
162
static const VMStateDescription vmstate_pmsav7 = {
163
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav7_rnr = {
164
.minimum_version_id = 1,
165
.needed = pmsav7_rnr_needed,
166
.fields = (VMStateField[]) {
167
- VMSTATE_UINT32(env.pmsav7.rnr, ARMCPU),
168
+ VMSTATE_UINT32(env.pmsav7.rnr[M_REG_NS], ARMCPU),
169
VMSTATE_END_OF_LIST()
170
}
171
};
172
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav8 = {
173
}
174
};
175
176
+static bool s_rnr_vmstate_validate(void *opaque, int version_id)
177
+{
178
+ ARMCPU *cpu = opaque;
179
+
180
+ return cpu->env.pmsav7.rnr[M_REG_S] < cpu->pmsav7_dregion;
181
+}
182
+
183
static bool m_security_needed(void *opaque)
184
{
185
ARMCPU *cpu = opaque;
186
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
187
0, vmstate_info_uint32, uint32_t),
188
VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
189
0, vmstate_info_uint32, uint32_t),
190
+ VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
191
+ VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
192
VMSTATE_END_OF_LIST()
193
}
194
};
195
--
173
--
196
2.7.4
174
2.20.1
197
175
198
176
diff view generated by jsdifflib
1
Make the MPU_CTRL register banked if v8M security extensions are
1
In the neon_padd/pmax/pmin helpers for float16, a cut-and-paste error
2
enabled.
2
meant we were using the H4() address swizzler macro rather than the
3
H2() which is required for 2-byte data. This had no effect on
4
little-endian hosts but meant we put the result data into the
5
destination Dreg in the wrong order on big-endian hosts.
3
6
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 1503414539-28762-16-git-send-email-peter.maydell@linaro.org
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Message-id: 20201028191712.4910-2-peter.maydell@linaro.org
7
---
11
---
8
target/arm/cpu.h | 2 +-
12
target/arm/vec_helper.c | 8 ++++----
9
hw/intc/armv7m_nvic.c | 9 +++++----
13
1 file changed, 4 insertions(+), 4 deletions(-)
10
target/arm/helper.c | 5 +++--
11
target/arm/machine.c | 3 ++-
12
4 files changed, 11 insertions(+), 8 deletions(-)
13
14
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.h
17
--- a/target/arm/vec_helper.c
17
+++ b/target/arm/cpu.h
18
+++ b/target/arm/vec_helper.c
18
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
19
@@ -XXX,XX +XXX,XX @@ DO_ABA(gvec_uaba_d, uint64_t)
19
uint32_t dfsr; /* Debug Fault Status Register */
20
r2 = float16_##OP(m[H2(0)], m[H2(1)], fpst); \
20
uint32_t mmfar; /* MemManage Fault Address */
21
r3 = float16_##OP(m[H2(2)], m[H2(3)], fpst); \
21
uint32_t bfar; /* BusFault Address */
22
\
22
- unsigned mpu_ctrl; /* MPU_CTRL */
23
- d[H4(0)] = r0; \
23
+ unsigned mpu_ctrl[2]; /* MPU_CTRL */
24
- d[H4(1)] = r1; \
24
int exception;
25
- d[H4(2)] = r2; \
25
uint32_t primask[2];
26
- d[H4(3)] = r3; \
26
uint32_t faultmask[2];
27
+ d[H2(0)] = r0; \
27
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
28
+ d[H2(1)] = r1; \
28
index XXXXXXX..XXXXXXX 100644
29
+ d[H2(2)] = r2; \
29
--- a/hw/intc/armv7m_nvic.c
30
+ d[H2(3)] = r3; \
30
+++ b/hw/intc/armv7m_nvic.c
31
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
32
return cpu->pmsav7_dregion << 8;
33
break;
34
case 0xd94: /* MPU_CTRL */
35
- return cpu->env.v7m.mpu_ctrl;
36
+ return cpu->env.v7m.mpu_ctrl[attrs.secure];
37
case 0xd98: /* MPU_RNR */
38
return cpu->env.pmsav7.rnr[attrs.secure];
39
case 0xd9c: /* MPU_RBAR */
40
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
41
qemu_log_mask(LOG_GUEST_ERROR, "MPU_CTRL: HFNMIENA and !ENABLE is "
42
"UNPREDICTABLE\n");
43
}
44
- cpu->env.v7m.mpu_ctrl = value & (R_V7M_MPU_CTRL_ENABLE_MASK |
45
- R_V7M_MPU_CTRL_HFNMIENA_MASK |
46
- R_V7M_MPU_CTRL_PRIVDEFENA_MASK);
47
+ cpu->env.v7m.mpu_ctrl[attrs.secure]
48
+ = value & (R_V7M_MPU_CTRL_ENABLE_MASK |
49
+ R_V7M_MPU_CTRL_HFNMIENA_MASK |
50
+ R_V7M_MPU_CTRL_PRIVDEFENA_MASK);
51
tlb_flush(CPU(cpu));
52
break;
53
case 0xd98: /* MPU_RNR */
54
diff --git a/target/arm/helper.c b/target/arm/helper.c
55
index XXXXXXX..XXXXXXX 100644
56
--- a/target/arm/helper.c
57
+++ b/target/arm/helper.c
58
@@ -XXX,XX +XXX,XX @@ static inline bool regime_translation_disabled(CPUARMState *env,
59
ARMMMUIdx mmu_idx)
60
{
61
if (arm_feature(env, ARM_FEATURE_M)) {
62
- switch (env->v7m.mpu_ctrl &
63
+ switch (env->v7m.mpu_ctrl[regime_is_secure(env, mmu_idx)] &
64
(R_V7M_MPU_CTRL_ENABLE_MASK | R_V7M_MPU_CTRL_HFNMIENA_MASK)) {
65
case R_V7M_MPU_CTRL_ENABLE_MASK:
66
/* Enabled, but not for HardFault and NMI */
67
@@ -XXX,XX +XXX,XX @@ static bool pmsav7_use_background_region(ARMCPU *cpu,
68
}
31
}
69
32
70
if (arm_feature(env, ARM_FEATURE_M)) {
33
DO_NEON_PAIRWISE(neon_padd, add)
71
- return env->v7m.mpu_ctrl & R_V7M_MPU_CTRL_PRIVDEFENA_MASK;
72
+ return env->v7m.mpu_ctrl[regime_is_secure(env, mmu_idx)]
73
+ & R_V7M_MPU_CTRL_PRIVDEFENA_MASK;
74
} else {
75
return regime_sctlr(env, mmu_idx) & SCTLR_BR;
76
}
77
diff --git a/target/arm/machine.c b/target/arm/machine.c
78
index XXXXXXX..XXXXXXX 100644
79
--- a/target/arm/machine.c
80
+++ b/target/arm/machine.c
81
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
82
VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
83
VMSTATE_UINT32(env.v7m.mmfar, ARMCPU),
84
VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
85
- VMSTATE_UINT32(env.v7m.mpu_ctrl, ARMCPU),
86
+ VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
87
VMSTATE_INT32(env.v7m.exception, ARMCPU),
88
VMSTATE_END_OF_LIST()
89
},
90
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
91
0, vmstate_info_uint32, uint32_t),
92
VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
93
VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
94
+ VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
95
VMSTATE_END_OF_LIST()
96
}
97
};
98
--
34
--
99
2.7.4
35
2.20.1
100
36
101
37
diff view generated by jsdifflib
1
Implement the new do_transaction_failed hook for ARM, which should
1
The helper functions for performing the udot/sdot operations against
2
cause the CPU to take a prefetch abort or data abort.
2
a scalar were not using an address-swizzling macro when converting
3
the index of the scalar element into a pointer into the vm array.
4
This had no effect on little-endian hosts but meant we generated
5
incorrect results on big-endian hosts.
6
7
For these insns, the index is indexing over group of 4 8-bit values,
8
so 32 bits per indexed entity, and H4() is therefore what we want.
9
(For Neon the only possible input indexes are 0 and 1.)
3
10
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
13
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Message-id: 1504626814-23124-4-git-send-email-peter.maydell@linaro.org
14
Message-id: 20201028191712.4910-3-peter.maydell@linaro.org
8
---
15
---
9
target/arm/internals.h | 10 ++++++++++
16
target/arm/vec_helper.c | 4 ++--
10
target/arm/cpu.c | 1 +
17
1 file changed, 2 insertions(+), 2 deletions(-)
11
target/arm/op_helper.c | 43 +++++++++++++++++++++++++++++++++++++++++++
12
3 files changed, 54 insertions(+)
13
18
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
19
diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c
15
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
21
--- a/target/arm/vec_helper.c
17
+++ b/target/arm/internals.h
22
+++ b/target/arm/vec_helper.c
18
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
23
@@ -XXX,XX +XXX,XX @@ void HELPER(gvec_sdot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc)
19
MMUAccessType access_type,
24
intptr_t index = simd_data(desc);
20
int mmu_idx, uintptr_t retaddr);
25
uint32_t *d = vd;
21
26
int8_t *n = vn;
22
+/* arm_cpu_do_transaction_failed: handle a memory system error response
27
- int8_t *m_indexed = (int8_t *)vm + index * 4;
23
+ * (eg "no device/memory present at address") by raising an external abort
28
+ int8_t *m_indexed = (int8_t *)vm + H4(index) * 4;
24
+ * exception
29
25
+ */
30
/* Notice the special case of opr_sz == 8, from aa64/aa32 advsimd.
26
+void arm_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr,
31
* Otherwise opr_sz is a multiple of 16.
27
+ vaddr addr, unsigned size,
32
@@ -XXX,XX +XXX,XX @@ void HELPER(gvec_udot_idx_b)(void *vd, void *vn, void *vm, uint32_t desc)
28
+ MMUAccessType access_type,
33
intptr_t index = simd_data(desc);
29
+ int mmu_idx, MemTxAttrs attrs,
34
uint32_t *d = vd;
30
+ MemTxResult response, uintptr_t retaddr);
35
uint8_t *n = vn;
31
+
36
- uint8_t *m_indexed = (uint8_t *)vm + index * 4;
32
/* Call the EL change hook if one has been registered */
37
+ uint8_t *m_indexed = (uint8_t *)vm + H4(index) * 4;
33
static inline void arm_call_el_change_hook(ARMCPU *cpu)
38
34
{
39
/* Notice the special case of opr_sz == 8, from aa64/aa32 advsimd.
35
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
40
* Otherwise opr_sz is a multiple of 16.
36
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/cpu.c
38
+++ b/target/arm/cpu.c
39
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
40
#else
41
cc->do_interrupt = arm_cpu_do_interrupt;
42
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
43
+ cc->do_transaction_failed = arm_cpu_do_transaction_failed;
44
cc->get_phys_page_attrs_debug = arm_cpu_get_phys_page_attrs_debug;
45
cc->asidx_from_attrs = arm_asidx_from_attrs;
46
cc->vmsd = &vmstate_arm_cpu;
47
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/op_helper.c
50
+++ b/target/arm/op_helper.c
51
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
52
deliver_fault(cpu, vaddr, access_type, fsr, fsc, &fi);
53
}
54
55
+/* arm_cpu_do_transaction_failed: handle a memory system error response
56
+ * (eg "no device/memory present at address") by raising an external abort
57
+ * exception
58
+ */
59
+void arm_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr,
60
+ vaddr addr, unsigned size,
61
+ MMUAccessType access_type,
62
+ int mmu_idx, MemTxAttrs attrs,
63
+ MemTxResult response, uintptr_t retaddr)
64
+{
65
+ ARMCPU *cpu = ARM_CPU(cs);
66
+ CPUARMState *env = &cpu->env;
67
+ uint32_t fsr, fsc;
68
+ ARMMMUFaultInfo fi = {};
69
+ ARMMMUIdx arm_mmu_idx = core_to_arm_mmu_idx(env, mmu_idx);
70
+
71
+ if (retaddr) {
72
+ /* now we have a real cpu fault */
73
+ cpu_restore_state(cs, retaddr);
74
+ }
75
+
76
+ /* The EA bit in syndromes and fault status registers is an
77
+ * IMPDEF classification of external aborts. ARM implementations
78
+ * usually use this to indicate AXI bus Decode error (0) or
79
+ * Slave error (1); in QEMU we follow that.
80
+ */
81
+ fi.ea = (response != MEMTX_DECODE_ERROR);
82
+
83
+ /* The fault status register format depends on whether we're using
84
+ * the LPAE long descriptor format, or the short descriptor format.
85
+ */
86
+ if (arm_s1_regime_using_lpae_format(env, arm_mmu_idx)) {
87
+ /* long descriptor form, STATUS 0b010000: synchronous ext abort */
88
+ fsr = (fi.ea << 12) | (1 << 9) | 0x10;
89
+ } else {
90
+ /* short descriptor form, FSR 0b01000 : synchronous ext abort */
91
+ fsr = (fi.ea << 12) | 0x8;
92
+ }
93
+ fsc = 0x10;
94
+
95
+ deliver_fault(cpu, addr, access_type, fsr, fsc, &fi);
96
+}
97
+
98
#endif /* !defined(CONFIG_USER_ONLY) */
99
100
uint32_t HELPER(add_setq)(CPUARMState *env, uint32_t a, uint32_t b)
101
--
41
--
102
2.7.4
42
2.20.1
103
43
104
44
diff view generated by jsdifflib
1
Implement the behavioural side of the new PMSAv8 specification.
1
From: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
2
2
3
HCR should be applied when NS is set, not when it is cleared.
4
5
Signed-off-by: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 1503414539-28762-3-git-send-email-peter.maydell@linaro.org
6
---
8
---
7
target/arm/helper.c | 111 +++++++++++++++++++++++++++++++++++++++++++++++++++-
9
target/arm/helper.c | 5 ++---
8
1 file changed, 110 insertions(+), 1 deletion(-)
10
1 file changed, 2 insertions(+), 3 deletions(-)
9
11
10
diff --git a/target/arm/helper.c b/target/arm/helper.c
12
diff --git a/target/arm/helper.c b/target/arm/helper.c
11
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
12
--- a/target/arm/helper.c
14
--- a/target/arm/helper.c
13
+++ b/target/arm/helper.c
15
+++ b/target/arm/helper.c
14
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_pmsav7(CPUARMState *env, uint32_t address,
16
@@ -XXX,XX +XXX,XX @@ static void tlbimvaa_is_write(CPUARMState *env, const ARMCPRegInfo *ri,
15
return !(*prot & (1 << access_type));
17
18
/*
19
* Non-IS variants of TLB operations are upgraded to
20
- * IS versions if we are at NS EL1 and HCR_EL2.FB is set to
21
+ * IS versions if we are at EL1 and HCR_EL2.FB is effectively set to
22
* force broadcast of these operations.
23
*/
24
static bool tlb_force_broadcast(CPUARMState *env)
25
{
26
- return (env->cp15.hcr_el2 & HCR_FB) &&
27
- arm_current_el(env) == 1 && arm_is_secure_below_el3(env);
28
+ return arm_current_el(env) == 1 && (arm_hcr_el2_eff(env) & HCR_FB);
16
}
29
}
17
30
18
+static bool get_phys_addr_pmsav8(CPUARMState *env, uint32_t address,
31
static void tlbiall_write(CPUARMState *env, const ARMCPRegInfo *ri,
19
+ MMUAccessType access_type, ARMMMUIdx mmu_idx,
20
+ hwaddr *phys_ptr, int *prot, uint32_t *fsr)
21
+{
22
+ ARMCPU *cpu = arm_env_get_cpu(env);
23
+ bool is_user = regime_is_user(env, mmu_idx);
24
+ int n;
25
+ int matchregion = -1;
26
+ bool hit = false;
27
+
28
+ *phys_ptr = address;
29
+ *prot = 0;
30
+
31
+ /* Unlike the ARM ARM pseudocode, we don't need to check whether this
32
+ * was an exception vector read from the vector table (which is always
33
+ * done using the default system address map), because those accesses
34
+ * are done in arm_v7m_load_vector(), which always does a direct
35
+ * read using address_space_ldl(), rather than going via this function.
36
+ */
37
+ if (regime_translation_disabled(env, mmu_idx)) { /* MPU disabled */
38
+ hit = true;
39
+ } else if (m_is_ppb_region(env, address)) {
40
+ hit = true;
41
+ } else if (pmsav7_use_background_region(cpu, mmu_idx, is_user)) {
42
+ hit = true;
43
+ } else {
44
+ for (n = (int)cpu->pmsav7_dregion - 1; n >= 0; n--) {
45
+ /* region search */
46
+ /* Note that the base address is bits [31:5] from the register
47
+ * with bits [4:0] all zeroes, but the limit address is bits
48
+ * [31:5] from the register with bits [4:0] all ones.
49
+ */
50
+ uint32_t base = env->pmsav8.rbar[n] & ~0x1f;
51
+ uint32_t limit = env->pmsav8.rlar[n] | 0x1f;
52
+
53
+ if (!(env->pmsav8.rlar[n] & 0x1)) {
54
+ /* Region disabled */
55
+ continue;
56
+ }
57
+
58
+ if (address < base || address > limit) {
59
+ continue;
60
+ }
61
+
62
+ if (hit) {
63
+ /* Multiple regions match -- always a failure (unlike
64
+ * PMSAv7 where highest-numbered-region wins)
65
+ */
66
+ *fsr = 0x00d; /* permission fault */
67
+ return true;
68
+ }
69
+
70
+ matchregion = n;
71
+ hit = true;
72
+
73
+ if (base & ~TARGET_PAGE_MASK) {
74
+ qemu_log_mask(LOG_UNIMP,
75
+ "MPU_RBAR[%d]: No support for MPU region base"
76
+ "address of 0x%" PRIx32 ". Minimum alignment is "
77
+ "%d\n",
78
+ n, base, TARGET_PAGE_BITS);
79
+ continue;
80
+ }
81
+ if ((limit + 1) & ~TARGET_PAGE_MASK) {
82
+ qemu_log_mask(LOG_UNIMP,
83
+ "MPU_RBAR[%d]: No support for MPU region limit"
84
+ "address of 0x%" PRIx32 ". Minimum alignment is "
85
+ "%d\n",
86
+ n, limit, TARGET_PAGE_BITS);
87
+ continue;
88
+ }
89
+ }
90
+ }
91
+
92
+ if (!hit) {
93
+ /* background fault */
94
+ *fsr = 0;
95
+ return true;
96
+ }
97
+
98
+ if (matchregion == -1) {
99
+ /* hit using the background region */
100
+ get_phys_addr_pmsav7_default(env, mmu_idx, address, prot);
101
+ } else {
102
+ uint32_t ap = extract32(env->pmsav8.rbar[matchregion], 1, 2);
103
+ uint32_t xn = extract32(env->pmsav8.rbar[matchregion], 0, 1);
104
+
105
+ if (m_is_system_region(env, address)) {
106
+ /* System space is always execute never */
107
+ xn = 1;
108
+ }
109
+
110
+ *prot = simple_ap_to_rw_prot(env, mmu_idx, ap);
111
+ if (*prot && !xn) {
112
+ *prot |= PAGE_EXEC;
113
+ }
114
+ /* We don't need to look the attribute up in the MAIR0/MAIR1
115
+ * registers because that only tells us about cacheability.
116
+ */
117
+ }
118
+
119
+ *fsr = 0x00d; /* Permission fault */
120
+ return !(*prot & (1 << access_type));
121
+}
122
+
123
static bool get_phys_addr_pmsav5(CPUARMState *env, uint32_t address,
124
MMUAccessType access_type, ARMMMUIdx mmu_idx,
125
hwaddr *phys_ptr, int *prot, uint32_t *fsr)
126
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr(CPUARMState *env, target_ulong address,
127
bool ret;
128
*page_size = TARGET_PAGE_SIZE;
129
130
- if (arm_feature(env, ARM_FEATURE_V7)) {
131
+ if (arm_feature(env, ARM_FEATURE_V8)) {
132
+ /* PMSAv8 */
133
+ ret = get_phys_addr_pmsav8(env, address, access_type, mmu_idx,
134
+ phys_ptr, prot, fsr);
135
+ } else if (arm_feature(env, ARM_FEATURE_V7)) {
136
/* PMSAv7 */
137
ret = get_phys_addr_pmsav7(env, address, access_type, mmu_idx,
138
phys_ptr, prot, fsr);
139
--
32
--
140
2.7.4
33
2.20.1
141
34
142
35
diff view generated by jsdifflib
1
Make the CFSR register banked if v8M security extensions are enabled.
1
From: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
2
2
3
Not all the bits in this register are banked: the BFSR
3
Secure mode is not exempted from checking SCR_EL3.TLOR, and in the
4
bits [15:8] are shared between S and NS, and we store them
4
future HCR_EL2.TLOR when S-EL2 is enabled.
5
in the NS copy of the register.
6
5
6
Signed-off-by: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 1503414539-28762-19-git-send-email-peter.maydell@linaro.org
10
---
9
---
11
target/arm/cpu.h | 7 ++++++-
10
target/arm/helper.c | 19 +++++--------------
12
hw/intc/armv7m_nvic.c | 15 +++++++++++++--
11
1 file changed, 5 insertions(+), 14 deletions(-)
13
target/arm/helper.c | 18 +++++++++---------
14
target/arm/machine.c | 3 ++-
15
4 files changed, 30 insertions(+), 13 deletions(-)
16
12
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
uint32_t basepri[2];
23
uint32_t control[2];
24
uint32_t ccr[2]; /* Configuration and Control */
25
- uint32_t cfsr; /* Configurable Fault Status */
26
+ uint32_t cfsr[2]; /* Configurable Fault Status */
27
uint32_t hfsr; /* HardFault Status */
28
uint32_t dfsr; /* Debug Fault Status Register */
29
uint32_t mmfar[2]; /* MemManage Fault Address */
30
@@ -XXX,XX +XXX,XX @@ FIELD(V7M_CFSR, NOCP, 16 + 3, 1)
31
FIELD(V7M_CFSR, UNALIGNED, 16 + 8, 1)
32
FIELD(V7M_CFSR, DIVBYZERO, 16 + 9, 1)
33
34
+/* V7M CFSR bit masks covering all of the subregister bits */
35
+FIELD(V7M_CFSR, MMFSR, 0, 8)
36
+FIELD(V7M_CFSR, BFSR, 8, 8)
37
+FIELD(V7M_CFSR, UFSR, 16, 16)
38
+
39
/* V7M HFSR bits */
40
FIELD(V7M_HFSR, VECTTBL, 1, 1)
41
FIELD(V7M_HFSR, FORCED, 30, 1)
42
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
43
index XXXXXXX..XXXXXXX 100644
44
--- a/hw/intc/armv7m_nvic.c
45
+++ b/hw/intc/armv7m_nvic.c
46
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
47
}
48
return val;
49
case 0xd28: /* Configurable Fault Status. */
50
- return cpu->env.v7m.cfsr;
51
+ /* The BFSR bits [15:8] are shared between security states
52
+ * and we store them in the NS copy
53
+ */
54
+ val = cpu->env.v7m.cfsr[attrs.secure];
55
+ val |= cpu->env.v7m.cfsr[M_REG_NS] & R_V7M_CFSR_BFSR_MASK;
56
+ return val;
57
case 0xd2c: /* Hard Fault Status. */
58
return cpu->env.v7m.hfsr;
59
case 0xd30: /* Debug Fault Status. */
60
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
61
nvic_irq_update(s);
62
break;
63
case 0xd28: /* Configurable Fault Status. */
64
- cpu->env.v7m.cfsr &= ~value; /* W1C */
65
+ cpu->env.v7m.cfsr[attrs.secure] &= ~value; /* W1C */
66
+ if (attrs.secure) {
67
+ /* The BFSR bits [15:8] are shared between security states
68
+ * and we store them in the NS copy.
69
+ */
70
+ cpu->env.v7m.cfsr[M_REG_NS] &= ~(value & R_V7M_CFSR_BFSR_MASK);
71
+ }
72
break;
73
case 0xd2c: /* Hard Fault Status. */
74
cpu->env.v7m.hfsr &= ~value; /* W1C */
75
diff --git a/target/arm/helper.c b/target/arm/helper.c
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
76
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
77
--- a/target/arm/helper.c
15
--- a/target/arm/helper.c
78
+++ b/target/arm/helper.c
16
+++ b/target/arm/helper.c
79
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
17
@@ -XXX,XX +XXX,XX @@ static uint64_t id_aa64pfr0_read(CPUARMState *env, const ARMCPRegInfo *ri)
80
/* Bad exception return: instead of popping the exception
18
#endif
81
* stack, directly take a usage fault on the current stack.
19
82
*/
20
/* Shared logic between LORID and the rest of the LOR* registers.
83
- env->v7m.cfsr |= R_V7M_CFSR_INVPC_MASK;
21
- * Secure state has already been delt with.
84
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_INVPC_MASK;
22
+ * Secure state exclusion has already been dealt with.
85
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE);
23
*/
86
v7m_exception_taken(cpu, type | 0xf0000000);
24
-static CPAccessResult access_lor_ns(CPUARMState *env)
87
qemu_log_mask(CPU_LOG_INT, "...taking UsageFault on existing "
25
+static CPAccessResult access_lor_ns(CPUARMState *env,
88
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
26
+ const ARMCPRegInfo *ri, bool isread)
89
if (return_to_handler != arm_v7m_is_handler_mode(env)) {
27
{
90
/* Take an INVPC UsageFault by pushing the stack again. */
28
int el = arm_current_el(env);
91
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE);
29
92
- env->v7m.cfsr |= R_V7M_CFSR_INVPC_MASK;
30
@@ -XXX,XX +XXX,XX @@ static CPAccessResult access_lor_ns(CPUARMState *env)
93
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_INVPC_MASK;
31
return CP_ACCESS_OK;
94
v7m_push_stack(cpu);
32
}
95
v7m_exception_taken(cpu, type | 0xf0000000);
33
96
qemu_log_mask(CPU_LOG_INT, "...taking UsageFault on new stackframe: "
34
-static CPAccessResult access_lorid(CPUARMState *env, const ARMCPRegInfo *ri,
97
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
35
- bool isread)
98
switch (cs->exception_index) {
36
-{
99
case EXCP_UDEF:
37
- if (arm_is_secure_below_el3(env)) {
100
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE);
38
- /* Access ok in secure mode. */
101
- env->v7m.cfsr |= R_V7M_CFSR_UNDEFINSTR_MASK;
39
- return CP_ACCESS_OK;
102
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_UNDEFINSTR_MASK;
40
- }
103
break;
41
- return access_lor_ns(env);
104
case EXCP_NOCP:
42
-}
105
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE);
43
-
106
- env->v7m.cfsr |= R_V7M_CFSR_NOCP_MASK;
44
static CPAccessResult access_lor_other(CPUARMState *env,
107
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_NOCP_MASK;
45
const ARMCPRegInfo *ri, bool isread)
108
break;
46
{
109
case EXCP_INVSTATE:
47
@@ -XXX,XX +XXX,XX @@ static CPAccessResult access_lor_other(CPUARMState *env,
110
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE);
48
/* Access denied in secure mode. */
111
- env->v7m.cfsr |= R_V7M_CFSR_INVSTATE_MASK;
49
return CP_ACCESS_TRAP;
112
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_INVSTATE_MASK;
113
break;
114
case EXCP_SWI:
115
/* The PC already points to the next instruction. */
116
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
117
case 0x8: /* External Abort */
118
switch (cs->exception_index) {
119
case EXCP_PREFETCH_ABORT:
120
- env->v7m.cfsr |= R_V7M_CFSR_PRECISERR_MASK;
121
+ env->v7m.cfsr[M_REG_NS] |= R_V7M_CFSR_PRECISERR_MASK;
122
qemu_log_mask(CPU_LOG_INT, "...with CFSR.PRECISERR\n");
123
break;
124
case EXCP_DATA_ABORT:
125
- env->v7m.cfsr |=
126
+ env->v7m.cfsr[M_REG_NS] |=
127
(R_V7M_CFSR_IBUSERR_MASK | R_V7M_CFSR_BFARVALID_MASK);
128
env->v7m.bfar = env->exception.vaddress;
129
qemu_log_mask(CPU_LOG_INT,
130
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
131
*/
132
switch (cs->exception_index) {
133
case EXCP_PREFETCH_ABORT:
134
- env->v7m.cfsr |= R_V7M_CFSR_IACCVIOL_MASK;
135
+ env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
136
qemu_log_mask(CPU_LOG_INT, "...with CFSR.IACCVIOL\n");
137
break;
138
case EXCP_DATA_ABORT:
139
- env->v7m.cfsr |=
140
+ env->v7m.cfsr[env->v7m.secure] |=
141
(R_V7M_CFSR_DACCVIOL_MASK | R_V7M_CFSR_MMARVALID_MASK);
142
env->v7m.mmfar[env->v7m.secure] = env->exception.vaddress;
143
qemu_log_mask(CPU_LOG_INT,
144
diff --git a/target/arm/machine.c b/target/arm/machine.c
145
index XXXXXXX..XXXXXXX 100644
146
--- a/target/arm/machine.c
147
+++ b/target/arm/machine.c
148
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
149
VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
150
VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
151
VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
152
- VMSTATE_UINT32(env.v7m.cfsr, ARMCPU),
153
+ VMSTATE_UINT32(env.v7m.cfsr[M_REG_NS], ARMCPU),
154
VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
155
VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
156
VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
157
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
158
VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
159
VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
160
VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
161
+ VMSTATE_UINT32(env.v7m.cfsr[M_REG_S], ARMCPU),
162
VMSTATE_END_OF_LIST()
163
}
50
}
51
- return access_lor_ns(env);
52
+ return access_lor_ns(env, ri, isread);
53
}
54
55
/*
56
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo lor_reginfo[] = {
57
.type = ARM_CP_CONST, .resetvalue = 0 },
58
{ .name = "LORID_EL1", .state = ARM_CP_STATE_AA64,
59
.opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7,
60
- .access = PL1_R, .accessfn = access_lorid,
61
+ .access = PL1_R, .accessfn = access_lor_ns,
62
.type = ARM_CP_CONST, .resetvalue = 0 },
63
REGINFO_SENTINEL
164
};
64
};
165
--
65
--
166
2.7.4
66
2.20.1
167
67
168
68
diff view generated by jsdifflib
1
Make the MMFAR register banked if v8M security extensions are
1
If we're using the capstone disassembler, disassembly of a run of
2
enabled.
2
instructions more than 32 bytes long disassembles the wrong data for
3
instructions beyond the 32 byte mark:
3
4
5
(qemu) xp /16x 0x100
6
0000000000000100: 0x00000005 0x54410001 0x00000001 0x00001000
7
0000000000000110: 0x00000000 0x00000004 0x54410002 0x3c000000
8
0000000000000120: 0x00000000 0x00000004 0x54410009 0x74736574
9
0000000000000130: 0x00000000 0x00000000 0x00000000 0x00000000
10
(qemu) xp /16i 0x100
11
0x00000100: 00000005 andeq r0, r0, r5
12
0x00000104: 54410001 strbpl r0, [r1], #-1
13
0x00000108: 00000001 andeq r0, r0, r1
14
0x0000010c: 00001000 andeq r1, r0, r0
15
0x00000110: 00000000 andeq r0, r0, r0
16
0x00000114: 00000004 andeq r0, r0, r4
17
0x00000118: 54410002 strbpl r0, [r1], #-2
18
0x0000011c: 3c000000 .byte 0x00, 0x00, 0x00, 0x3c
19
0x00000120: 54410001 strbpl r0, [r1], #-1
20
0x00000124: 00000001 andeq r0, r0, r1
21
0x00000128: 00001000 andeq r1, r0, r0
22
0x0000012c: 00000000 andeq r0, r0, r0
23
0x00000130: 00000004 andeq r0, r0, r4
24
0x00000134: 54410002 strbpl r0, [r1], #-2
25
0x00000138: 3c000000 .byte 0x00, 0x00, 0x00, 0x3c
26
0x0000013c: 00000000 andeq r0, r0, r0
27
28
Here the disassembly of 0x120..0x13f is using the data that is in
29
0x104..0x123.
30
31
This is caused by passing the wrong value to the read_memory_func().
32
The intention is that at this point in the loop the 'cap_buf' buffer
33
already contains 'csize' bytes of data for the instruction at guest
34
addr 'pc', and we want to read in an extra 'tsize' bytes. Those
35
extra bytes are therefore at 'pc + csize', not 'pc'. On the first
36
time through the loop 'csize' happens to be zero, so the initial read
37
of 32 bytes into cap_buf is correct and as long as the disassembly
38
never needs to read more data we return the correct information.
39
40
Use the correct guest address in the call to read_memory_func().
41
42
Cc: qemu-stable@nongnu.org
43
Fixes: https://bugs.launchpad.net/qemu/+bug/1900779
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
44
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
45
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
Message-id: 1503414539-28762-18-git-send-email-peter.maydell@linaro.org
46
Message-id: 20201022132445.25039-1-peter.maydell@linaro.org
7
---
47
---
8
target/arm/cpu.h | 2 +-
48
disas/capstone.c | 2 +-
9
hw/intc/armv7m_nvic.c | 4 ++--
49
1 file changed, 1 insertion(+), 1 deletion(-)
10
target/arm/helper.c | 4 ++--
11
target/arm/machine.c | 3 ++-
12
4 files changed, 7 insertions(+), 6 deletions(-)
13
50
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
51
diff --git a/disas/capstone.c b/disas/capstone.c
15
index XXXXXXX..XXXXXXX 100644
52
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.h
53
--- a/disas/capstone.c
17
+++ b/target/arm/cpu.h
54
+++ b/disas/capstone.c
18
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
55
@@ -XXX,XX +XXX,XX @@ bool cap_disas_monitor(disassemble_info *info, uint64_t pc, int count)
19
uint32_t cfsr; /* Configurable Fault Status */
56
20
uint32_t hfsr; /* HardFault Status */
57
/* Make certain that we can make progress. */
21
uint32_t dfsr; /* Debug Fault Status Register */
58
assert(tsize != 0);
22
- uint32_t mmfar; /* MemManage Fault Address */
59
- info->read_memory_func(pc, cap_buf + csize, tsize, info);
23
+ uint32_t mmfar[2]; /* MemManage Fault Address */
60
+ info->read_memory_func(pc + csize, cap_buf + csize, tsize, info);
24
uint32_t bfar; /* BusFault Address */
61
csize += tsize;
25
unsigned mpu_ctrl[2]; /* MPU_CTRL */
62
26
int exception;
63
if (cs_disasm_iter(handle, &cbuf, &csize, &pc, insn)) {
27
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/intc/armv7m_nvic.c
30
+++ b/hw/intc/armv7m_nvic.c
31
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
32
case 0xd30: /* Debug Fault Status. */
33
return cpu->env.v7m.dfsr;
34
case 0xd34: /* MMFAR MemManage Fault Address */
35
- return cpu->env.v7m.mmfar;
36
+ return cpu->env.v7m.mmfar[attrs.secure];
37
case 0xd38: /* Bus Fault Address. */
38
return cpu->env.v7m.bfar;
39
case 0xd3c: /* Aux Fault Status. */
40
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
41
cpu->env.v7m.dfsr &= ~value; /* W1C */
42
break;
43
case 0xd34: /* Mem Manage Address. */
44
- cpu->env.v7m.mmfar = value;
45
+ cpu->env.v7m.mmfar[attrs.secure] = value;
46
return;
47
case 0xd38: /* Bus Fault Address. */
48
cpu->env.v7m.bfar = value;
49
diff --git a/target/arm/helper.c b/target/arm/helper.c
50
index XXXXXXX..XXXXXXX 100644
51
--- a/target/arm/helper.c
52
+++ b/target/arm/helper.c
53
@@ -XXX,XX +XXX,XX @@ void arm_v7m_cpu_do_interrupt(CPUState *cs)
54
case EXCP_DATA_ABORT:
55
env->v7m.cfsr |=
56
(R_V7M_CFSR_DACCVIOL_MASK | R_V7M_CFSR_MMARVALID_MASK);
57
- env->v7m.mmfar = env->exception.vaddress;
58
+ env->v7m.mmfar[env->v7m.secure] = env->exception.vaddress;
59
qemu_log_mask(CPU_LOG_INT,
60
"...with CFSR.DACCVIOL and MMFAR 0x%x\n",
61
- env->v7m.mmfar);
62
+ env->v7m.mmfar[env->v7m.secure]);
63
break;
64
}
65
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM);
66
diff --git a/target/arm/machine.c b/target/arm/machine.c
67
index XXXXXXX..XXXXXXX 100644
68
--- a/target/arm/machine.c
69
+++ b/target/arm/machine.c
70
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
71
VMSTATE_UINT32(env.v7m.cfsr, ARMCPU),
72
VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
73
VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
74
- VMSTATE_UINT32(env.v7m.mmfar, ARMCPU),
75
+ VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
76
VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
77
VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
78
VMSTATE_INT32(env.v7m.exception, ARMCPU),
79
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
80
VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
81
VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
82
VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
83
+ VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
84
VMSTATE_END_OF_LIST()
85
}
86
};
87
--
64
--
88
2.7.4
65
2.20.1
89
66
90
67
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
Use the BIT_ULL() macro to ensure we use 64-bit arithmetic.
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
This fixes the following Coverity issue (OVERFLOW_BEFORE_WIDEN):
5
Message-id: 20170905131149.10669-5-famz@redhat.com
5
6
CID 1432363 (#1 of 1): Unintentional integer overflow:
7
8
overflow_before_widen:
9
Potentially overflowing expression 1 << scale with type int
10
(32 bits, signed) is evaluated using 32-bit arithmetic, and
11
then used in a context that expects an expression of type
12
hwaddr (64 bits, unsigned).
13
14
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
15
Acked-by: Eric Auger <eric.auger@redhat.com>
16
Message-id: 20201030144617.1535064-1-philmd@redhat.com
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
19
---
9
hw/arm/xlnx-zynqmp.c | 7 ++-----
20
hw/arm/smmuv3.c | 3 ++-
10
1 file changed, 2 insertions(+), 5 deletions(-)
21
1 file changed, 2 insertions(+), 1 deletion(-)
11
22
12
diff --git a/hw/arm/xlnx-zynqmp.c b/hw/arm/xlnx-zynqmp.c
23
diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c
13
index XXXXXXX..XXXXXXX 100644
24
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/arm/xlnx-zynqmp.c
25
--- a/hw/arm/smmuv3.c
15
+++ b/hw/arm/xlnx-zynqmp.c
26
+++ b/hw/arm/smmuv3.c
16
@@ -XXX,XX +XXX,XX @@ static void xlnx_zynqmp_init(Object *obj)
27
@@ -XXX,XX +XXX,XX @@
17
&error_abort);
28
*/
29
30
#include "qemu/osdep.h"
31
+#include "qemu/bitops.h"
32
#include "hw/irq.h"
33
#include "hw/sysbus.h"
34
#include "migration/vmstate.h"
35
@@ -XXX,XX +XXX,XX @@ static void smmuv3_s1_range_inval(SMMUState *s, Cmd *cmd)
36
scale = CMD_SCALE(cmd);
37
num = CMD_NUM(cmd);
38
ttl = CMD_TTL(cmd);
39
- num_pages = (num + 1) * (1 << (scale));
40
+ num_pages = (num + 1) * BIT_ULL(scale);
18
}
41
}
19
42
20
- object_property_add_link(obj, "ddr-ram", TYPE_MEMORY_REGION,
43
if (type == SMMU_CMD_TLBI_NH_VA) {
21
- (Object **)&s->ddr_ram,
22
- qdev_prop_allow_set_link_before_realize,
23
- OBJ_PROP_LINK_UNREF_ON_RELEASE, &error_abort);
24
-
25
object_initialize(&s->gic, sizeof(s->gic), gic_class_name());
26
qdev_set_parent_bus(DEVICE(&s->gic), sysbus_get_default());
27
28
@@ -XXX,XX +XXX,XX @@ static Property xlnx_zynqmp_props[] = {
29
DEFINE_PROP_STRING("boot-cpu", XlnxZynqMPState, boot_cpu),
30
DEFINE_PROP_BOOL("secure", XlnxZynqMPState, secure, false),
31
DEFINE_PROP_BOOL("has_rpu", XlnxZynqMPState, has_rpu, false),
32
+ DEFINE_PROP_LINK("ddr-ram", XlnxZynqMPState, ddr_ram, TYPE_MEMORY_REGION,
33
+ MemoryRegion *),
34
DEFINE_PROP_END_OF_LIST()
35
};
36
37
--
44
--
38
2.7.4
45
2.20.1
39
46
40
47
diff view generated by jsdifflib
1
From: Thomas Huth <thuth@redhat.com>
1
From: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
2
2
3
QEMU currently exits unexpectedly when the user accidentially
3
When booting a CPU with EL3 using the -kernel flag, set up CPTR_EL3 so
4
tries to do something like this:
4
that SVE will not trap to EL3.
5
5
6
$ aarch64-softmmu/qemu-system-aarch64 -S -M integratorcp -nographic
6
Signed-off-by: Rémi Denis-Courmont <remi.denis.courmont@huawei.com>
7
QEMU 2.9.93 monitor - type 'help' for more information
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
(qemu) device_add allwinner-a10
8
Message-id: 20201030151541.11976-1-remi@remlab.net
9
Unsupported NIC model: smc91c111
10
11
Exiting just due to a "device_add" should not happen. Looking closer
12
at the the realize and instance_init function of this device also
13
reveals that it is using serial_hds and nd_table directly there, so
14
this device is clearly not creatable by the user and should be marked
15
accordingly.
16
17
Signed-off-by: Thomas Huth <thuth@redhat.com>
18
Reviewed-by: Eduardo Habkost <ehabkost@redhat.com>
19
Message-id: 1503416789-32080-1-git-send-email-thuth@redhat.com
20
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
21
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
22
---
10
---
23
hw/arm/allwinner-a10.c | 2 ++
11
hw/arm/boot.c | 3 +++
24
scripts/device-crash-test | 1 -
12
1 file changed, 3 insertions(+)
25
2 files changed, 2 insertions(+), 1 deletion(-)
26
13
27
diff --git a/hw/arm/allwinner-a10.c b/hw/arm/allwinner-a10.c
14
diff --git a/hw/arm/boot.c b/hw/arm/boot.c
28
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/arm/allwinner-a10.c
16
--- a/hw/arm/boot.c
30
+++ b/hw/arm/allwinner-a10.c
17
+++ b/hw/arm/boot.c
31
@@ -XXX,XX +XXX,XX @@ static void aw_a10_class_init(ObjectClass *oc, void *data)
18
@@ -XXX,XX +XXX,XX @@ static void do_cpu_reset(void *opaque)
32
DeviceClass *dc = DEVICE_CLASS(oc);
19
if (cpu_isar_feature(aa64_mte, cpu)) {
33
20
env->cp15.scr_el3 |= SCR_ATA;
34
dc->realize = aw_a10_realize;
21
}
35
+ /* Reason: Uses serial_hds in realize and nd_table in instance_init */
22
+ if (cpu_isar_feature(aa64_sve, cpu)) {
36
+ dc->user_creatable = false;
23
+ env->cp15.cptr_el[3] |= CPTR_EZ;
37
}
24
+ }
38
25
/* AArch64 kernels never boot in secure mode */
39
static const TypeInfo aw_a10_type_info = {
26
assert(!info->secure_boot);
40
diff --git a/scripts/device-crash-test b/scripts/device-crash-test
27
/* This hook is only supported for AArch32 currently:
41
index XXXXXXX..XXXXXXX 100755
42
--- a/scripts/device-crash-test
43
+++ b/scripts/device-crash-test
44
@@ -XXX,XX +XXX,XX @@ ERROR_WHITELIST = [
45
{'log':r"Device [\w.,-]+ can not be dynamically instantiated"},
46
{'log':r"Platform Bus: Can not fit MMIO region of size "},
47
# other more specific errors we will ignore:
48
- {'device':'allwinner-a10', 'log':"Unsupported NIC model:"},
49
{'device':'.*-spapr-cpu-core', 'log':r"CPU core type should be"},
50
{'log':r"MSI(-X)? is not supported by interrupt controller"},
51
{'log':r"pxb-pcie? devices cannot reside on a PCIe? bus"},
52
--
28
--
53
2.7.4
29
2.20.1
54
30
55
31
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: AlexChen <alex.chen@huawei.com>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
In omap_lcd_interrupts(), the pointer omap_lcd is dereferinced before
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
being check if it is valid, which may lead to NULL pointer dereference.
5
Message-id: 20170905131149.10669-3-famz@redhat.com
5
So move the assignment to surface after checking that the omap_lcd is valid
6
and move surface_bits_per_pixel(surface) to after the surface assignment.
7
8
Reported-by: Euler Robot <euler.robot@huawei.com>
9
Signed-off-by: AlexChen <alex.chen@huawei.com>
10
Message-id: 5F9CDB8A.9000001@huawei.com
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
13
---
9
hw/arm/armv7m.c | 8 ++------
14
hw/display/omap_lcdc.c | 10 +++++++---
10
1 file changed, 2 insertions(+), 6 deletions(-)
15
1 file changed, 7 insertions(+), 3 deletions(-)
11
16
12
diff --git a/hw/arm/armv7m.c b/hw/arm/armv7m.c
17
diff --git a/hw/display/omap_lcdc.c b/hw/display/omap_lcdc.c
13
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/arm/armv7m.c
19
--- a/hw/display/omap_lcdc.c
15
+++ b/hw/arm/armv7m.c
20
+++ b/hw/display/omap_lcdc.c
16
@@ -XXX,XX +XXX,XX @@ static void armv7m_instance_init(Object *obj)
21
@@ -XXX,XX +XXX,XX @@ static void omap_lcd_interrupts(struct omap_lcd_panel_s *s)
17
22
static void omap_update_display(void *opaque)
18
/* Can't init the cpu here, we don't yet know which model to use */
23
{
19
24
struct omap_lcd_panel_s *omap_lcd = (struct omap_lcd_panel_s *) opaque;
20
- object_property_add_link(obj, "memory",
25
- DisplaySurface *surface = qemu_console_surface(omap_lcd->con);
21
- TYPE_MEMORY_REGION,
26
+ DisplaySurface *surface;
22
- (Object **)&s->board_memory,
27
draw_line_func draw_line;
23
- qdev_prop_allow_set_link_before_realize,
28
int size, height, first, last;
24
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
29
int width, linesize, step, bpp, frame_offset;
25
- &error_abort);
30
hwaddr frame_base;
26
memory_region_init(&s->container, obj, "armv7m-container", UINT64_MAX);
31
27
32
- if (!omap_lcd || omap_lcd->plm == 1 || !omap_lcd->enable ||
28
object_initialize(&s->nvic, sizeof(s->nvic), TYPE_NVIC);
33
- !surface_bits_per_pixel(surface)) {
29
@@ -XXX,XX +XXX,XX @@ static void armv7m_realize(DeviceState *dev, Error **errp)
34
+ if (!omap_lcd || omap_lcd->plm == 1 || !omap_lcd->enable) {
30
35
+ return;
31
static Property armv7m_properties[] = {
36
+ }
32
DEFINE_PROP_STRING("cpu-model", ARMv7MState, cpu_model),
37
+
33
+ DEFINE_PROP_LINK("memory", ARMv7MState, board_memory, TYPE_MEMORY_REGION,
38
+ surface = qemu_console_surface(omap_lcd->con);
34
+ MemoryRegion *),
39
+ if (!surface_bits_per_pixel(surface)) {
35
DEFINE_PROP_END_OF_LIST(),
40
return;
36
};
41
}
37
42
38
--
43
--
39
2.7.4
44
2.20.1
40
45
41
46
diff view generated by jsdifflib
1
From: Fam Zheng <famz@redhat.com>
1
From: AlexChen <alex.chen@huawei.com>
2
2
3
Signed-off-by: Fam Zheng <famz@redhat.com>
3
In exynos4210_fimd_update(), the pointer s is dereferinced before
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
4
being check if it is valid, which may lead to NULL pointer dereference.
5
Message-id: 20170905131149.10669-7-famz@redhat.com
5
So move the assignment to global_width after checking that the s is valid.
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
7
Reported-by: Euler Robot <euler.robot@huawei.com>
8
Signed-off-by: Alex Chen <alex.chen@huawei.com>
9
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
10
Message-id: 5F9F8D88.9030102@huawei.com
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
---
12
---
9
hw/dma/xilinx_axidma.c | 16 ++++------------
13
hw/display/exynos4210_fimd.c | 4 +++-
10
1 file changed, 4 insertions(+), 12 deletions(-)
14
1 file changed, 3 insertions(+), 1 deletion(-)
11
15
12
diff --git a/hw/dma/xilinx_axidma.c b/hw/dma/xilinx_axidma.c
16
diff --git a/hw/display/exynos4210_fimd.c b/hw/display/exynos4210_fimd.c
13
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
14
--- a/hw/dma/xilinx_axidma.c
18
--- a/hw/display/exynos4210_fimd.c
15
+++ b/hw/dma/xilinx_axidma.c
19
+++ b/hw/display/exynos4210_fimd.c
16
@@ -XXX,XX +XXX,XX @@ static void xilinx_axidma_init(Object *obj)
20
@@ -XXX,XX +XXX,XX @@ static void exynos4210_fimd_update(void *opaque)
17
XilinxAXIDMA *s = XILINX_AXI_DMA(obj);
21
bool blend = false;
18
SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
22
uint8_t *host_fb_addr;
19
23
bool is_dirty = false;
20
- object_property_add_link(obj, "axistream-connected", TYPE_STREAM_SLAVE,
24
- const int global_width = (s->vidtcon[2] & FIMD_VIDTCON2_SIZE_MASK) + 1;
21
- (Object **)&s->tx_data_dev,
25
+ int global_width;
22
- qdev_prop_allow_set_link_before_realize,
26
23
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
27
if (!s || !s->console || !s->enabled ||
24
- &error_abort);
28
surface_bits_per_pixel(qemu_console_surface(s->console)) == 0) {
25
- object_property_add_link(obj, "axistream-control-connected",
29
return;
26
- TYPE_STREAM_SLAVE,
30
}
27
- (Object **)&s->tx_control_dev,
31
+
28
- qdev_prop_allow_set_link_before_realize,
32
+ global_width = (s->vidtcon[2] & FIMD_VIDTCON2_SIZE_MASK) + 1;
29
- OBJ_PROP_LINK_UNREF_ON_RELEASE,
33
exynos4210_update_resolution(s);
30
- &error_abort);
34
surface = qemu_console_surface(s->console);
31
-
32
object_initialize(&s->rx_data_dev, sizeof(s->rx_data_dev),
33
TYPE_XILINX_AXI_DMA_DATA_STREAM);
34
object_initialize(&s->rx_control_dev, sizeof(s->rx_control_dev),
35
@@ -XXX,XX +XXX,XX @@ static void xilinx_axidma_init(Object *obj)
36
37
static Property axidma_properties[] = {
38
DEFINE_PROP_UINT32("freqhz", XilinxAXIDMA, freqhz, 50000000),
39
+ DEFINE_PROP_LINK("axistream-connected", XilinxAXIDMA,
40
+ tx_data_dev, TYPE_STREAM_SLAVE, StreamSlave *),
41
+ DEFINE_PROP_LINK("axistream-control-connected", XilinxAXIDMA,
42
+ tx_control_dev, TYPE_STREAM_SLAVE, StreamSlave *),
43
DEFINE_PROP_END_OF_LIST(),
44
};
45
35
46
--
36
--
47
2.7.4
37
2.20.1
48
38
49
39
diff view generated by jsdifflib
Deleted patch
1
As part of ARMv8M, we need to add support for the PMSAv8 MPU
2
architecture.
3
1
4
PMSAv8 differs from PMSAv7 both in register/data layout (for instance
5
using base and limit registers rather than base and size) and also in
6
behaviour (for example it does not have subregions); rather than
7
trying to wedge it into the existing PMSAv7 code and data structures,
8
we define separate ones.
9
10
This commit adds the data structures which hold the state for a
11
PMSAv8 MPU and the register interface to it. The implementation of
12
the MPU behaviour will be added in a subsequent commit.
13
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
16
Message-id: 1503414539-28762-2-git-send-email-peter.maydell@linaro.org
17
---
18
target/arm/cpu.h | 13 ++++++
19
hw/intc/armv7m_nvic.c | 122 ++++++++++++++++++++++++++++++++++++++++++++++----
20
target/arm/cpu.c | 36 ++++++++++-----
21
target/arm/machine.c | 29 +++++++++++-
22
4 files changed, 180 insertions(+), 20 deletions(-)
23
24
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
25
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/cpu.h
27
+++ b/target/arm/cpu.h
28
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
29
uint32_t rnr;
30
} pmsav7;
31
32
+ /* PMSAv8 MPU */
33
+ struct {
34
+ /* The PMSAv8 implementation also shares some PMSAv7 config
35
+ * and state:
36
+ * pmsav7.rnr (region number register)
37
+ * pmsav7_dregion (number of configured regions)
38
+ */
39
+ uint32_t *rbar;
40
+ uint32_t *rlar;
41
+ uint32_t mair0;
42
+ uint32_t mair1;
43
+ } pmsav8;
44
+
45
void *nvic;
46
const struct arm_boot_info *boot_info;
47
/* Store GICv3CPUState to access from this struct */
48
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/hw/intc/armv7m_nvic.c
51
+++ b/hw/intc/armv7m_nvic.c
52
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset)
53
{
54
int region = cpu->env.pmsav7.rnr;
55
56
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
57
+ /* PMSAv8M handling of the aliases is different from v7M:
58
+ * aliases A1, A2, A3 override the low two bits of the region
59
+ * number in MPU_RNR, and there is no 'region' field in the
60
+ * RBAR register.
61
+ */
62
+ int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
63
+ if (aliasno) {
64
+ region = deposit32(region, 0, 2, aliasno);
65
+ }
66
+ if (region >= cpu->pmsav7_dregion) {
67
+ return 0;
68
+ }
69
+ return cpu->env.pmsav8.rbar[region];
70
+ }
71
+
72
if (region >= cpu->pmsav7_dregion) {
73
return 0;
74
}
75
return (cpu->env.pmsav7.drbar[region] & 0x1f) | (region & 0xf);
76
}
77
- case 0xda0: /* MPU_RASR */
78
- case 0xda8: /* MPU_RASR_A1 */
79
- case 0xdb0: /* MPU_RASR_A2 */
80
- case 0xdb8: /* MPU_RASR_A3 */
81
+ case 0xda0: /* MPU_RASR (v7M), MPU_RLAR (v8M) */
82
+ case 0xda8: /* MPU_RASR_A1 (v7M), MPU_RLAR_A1 (v8M) */
83
+ case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
84
+ case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
85
{
86
int region = cpu->env.pmsav7.rnr;
87
88
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
89
+ /* PMSAv8M handling of the aliases is different from v7M:
90
+ * aliases A1, A2, A3 override the low two bits of the region
91
+ * number in MPU_RNR.
92
+ */
93
+ int aliasno = (offset - 0xda0) / 8; /* 0..3 */
94
+ if (aliasno) {
95
+ region = deposit32(region, 0, 2, aliasno);
96
+ }
97
+ if (region >= cpu->pmsav7_dregion) {
98
+ return 0;
99
+ }
100
+ return cpu->env.pmsav8.rlar[region];
101
+ }
102
+
103
if (region >= cpu->pmsav7_dregion) {
104
return 0;
105
}
106
return ((cpu->env.pmsav7.dracr[region] & 0xffff) << 16) |
107
(cpu->env.pmsav7.drsr[region] & 0xffff);
108
}
109
+ case 0xdc0: /* MPU_MAIR0 */
110
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
111
+ goto bad_offset;
112
+ }
113
+ return cpu->env.pmsav8.mair0;
114
+ case 0xdc4: /* MPU_MAIR1 */
115
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
116
+ goto bad_offset;
117
+ }
118
+ return cpu->env.pmsav8.mair1;
119
default:
120
+ bad_offset:
121
qemu_log_mask(LOG_GUEST_ERROR, "NVIC: Bad read offset 0x%x\n", offset);
122
return 0;
123
}
124
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
125
{
126
int region;
127
128
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
129
+ /* PMSAv8M handling of the aliases is different from v7M:
130
+ * aliases A1, A2, A3 override the low two bits of the region
131
+ * number in MPU_RNR, and there is no 'region' field in the
132
+ * RBAR register.
133
+ */
134
+ int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
135
+
136
+ region = cpu->env.pmsav7.rnr;
137
+ if (aliasno) {
138
+ region = deposit32(region, 0, 2, aliasno);
139
+ }
140
+ if (region >= cpu->pmsav7_dregion) {
141
+ return;
142
+ }
143
+ cpu->env.pmsav8.rbar[region] = value;
144
+ tlb_flush(CPU(cpu));
145
+ return;
146
+ }
147
+
148
if (value & (1 << 4)) {
149
/* VALID bit means use the region number specified in this
150
* value and also update MPU_RNR.REGION with that value.
151
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
152
tlb_flush(CPU(cpu));
153
break;
154
}
155
- case 0xda0: /* MPU_RASR */
156
- case 0xda8: /* MPU_RASR_A1 */
157
- case 0xdb0: /* MPU_RASR_A2 */
158
- case 0xdb8: /* MPU_RASR_A3 */
159
+ case 0xda0: /* MPU_RASR (v7M), MPU_RLAR (v8M) */
160
+ case 0xda8: /* MPU_RASR_A1 (v7M), MPU_RLAR_A1 (v8M) */
161
+ case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
162
+ case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
163
{
164
int region = cpu->env.pmsav7.rnr;
165
166
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
167
+ /* PMSAv8M handling of the aliases is different from v7M:
168
+ * aliases A1, A2, A3 override the low two bits of the region
169
+ * number in MPU_RNR.
170
+ */
171
+ int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
172
+
173
+ region = cpu->env.pmsav7.rnr;
174
+ if (aliasno) {
175
+ region = deposit32(region, 0, 2, aliasno);
176
+ }
177
+ if (region >= cpu->pmsav7_dregion) {
178
+ return;
179
+ }
180
+ cpu->env.pmsav8.rlar[region] = value;
181
+ tlb_flush(CPU(cpu));
182
+ return;
183
+ }
184
+
185
if (region >= cpu->pmsav7_dregion) {
186
return;
187
}
188
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
189
tlb_flush(CPU(cpu));
190
break;
191
}
192
+ case 0xdc0: /* MPU_MAIR0 */
193
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
194
+ goto bad_offset;
195
+ }
196
+ if (cpu->pmsav7_dregion) {
197
+ /* Register is RES0 if no MPU regions are implemented */
198
+ cpu->env.pmsav8.mair0 = value;
199
+ }
200
+ /* We don't need to do anything else because memory attributes
201
+ * only affect cacheability, and we don't implement caching.
202
+ */
203
+ break;
204
+ case 0xdc4: /* MPU_MAIR1 */
205
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
206
+ goto bad_offset;
207
+ }
208
+ if (cpu->pmsav7_dregion) {
209
+ /* Register is RES0 if no MPU regions are implemented */
210
+ cpu->env.pmsav8.mair1 = value;
211
+ }
212
+ /* We don't need to do anything else because memory attributes
213
+ * only affect cacheability, and we don't implement caching.
214
+ */
215
+ break;
216
case 0xf00: /* Software Triggered Interrupt Register */
217
{
218
int excnum = (value & 0x1ff) + NVIC_FIRST_IRQ;
219
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
220
break;
221
}
222
default:
223
+ bad_offset:
224
qemu_log_mask(LOG_GUEST_ERROR,
225
"NVIC: Bad write offset 0x%x\n", offset);
226
}
227
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
228
index XXXXXXX..XXXXXXX 100644
229
--- a/target/arm/cpu.c
230
+++ b/target/arm/cpu.c
231
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
232
env->vfp.xregs[ARM_VFP_FPEXC] = 0;
233
#endif
234
235
- if (arm_feature(env, ARM_FEATURE_PMSA) &&
236
- arm_feature(env, ARM_FEATURE_V7)) {
237
+ if (arm_feature(env, ARM_FEATURE_PMSA)) {
238
if (cpu->pmsav7_dregion > 0) {
239
- memset(env->pmsav7.drbar, 0,
240
- sizeof(*env->pmsav7.drbar) * cpu->pmsav7_dregion);
241
- memset(env->pmsav7.drsr, 0,
242
- sizeof(*env->pmsav7.drsr) * cpu->pmsav7_dregion);
243
- memset(env->pmsav7.dracr, 0,
244
- sizeof(*env->pmsav7.dracr) * cpu->pmsav7_dregion);
245
+ if (arm_feature(env, ARM_FEATURE_V8)) {
246
+ memset(env->pmsav8.rbar, 0,
247
+ sizeof(*env->pmsav8.rbar) * cpu->pmsav7_dregion);
248
+ memset(env->pmsav8.rlar, 0,
249
+ sizeof(*env->pmsav8.rlar) * cpu->pmsav7_dregion);
250
+ } else if (arm_feature(env, ARM_FEATURE_V7)) {
251
+ memset(env->pmsav7.drbar, 0,
252
+ sizeof(*env->pmsav7.drbar) * cpu->pmsav7_dregion);
253
+ memset(env->pmsav7.drsr, 0,
254
+ sizeof(*env->pmsav7.drsr) * cpu->pmsav7_dregion);
255
+ memset(env->pmsav7.dracr, 0,
256
+ sizeof(*env->pmsav7.dracr) * cpu->pmsav7_dregion);
257
+ }
258
}
259
env->pmsav7.rnr = 0;
260
+ env->pmsav8.mair0 = 0;
261
+ env->pmsav8.mair1 = 0;
262
}
263
264
set_flush_to_zero(1, &env->vfp.standard_fp_status);
265
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
266
}
267
268
if (nr) {
269
- env->pmsav7.drbar = g_new0(uint32_t, nr);
270
- env->pmsav7.drsr = g_new0(uint32_t, nr);
271
- env->pmsav7.dracr = g_new0(uint32_t, nr);
272
+ if (arm_feature(env, ARM_FEATURE_V8)) {
273
+ /* PMSAv8 */
274
+ env->pmsav8.rbar = g_new0(uint32_t, nr);
275
+ env->pmsav8.rlar = g_new0(uint32_t, nr);
276
+ } else {
277
+ env->pmsav7.drbar = g_new0(uint32_t, nr);
278
+ env->pmsav7.drsr = g_new0(uint32_t, nr);
279
+ env->pmsav7.dracr = g_new0(uint32_t, nr);
280
+ }
281
}
282
}
283
284
diff --git a/target/arm/machine.c b/target/arm/machine.c
285
index XXXXXXX..XXXXXXX 100644
286
--- a/target/arm/machine.c
287
+++ b/target/arm/machine.c
288
@@ -XXX,XX +XXX,XX @@ static bool pmsav7_needed(void *opaque)
289
CPUARMState *env = &cpu->env;
290
291
return arm_feature(env, ARM_FEATURE_PMSA) &&
292
- arm_feature(env, ARM_FEATURE_V7);
293
+ arm_feature(env, ARM_FEATURE_V7) &&
294
+ !arm_feature(env, ARM_FEATURE_V8);
295
}
296
297
static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
298
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav7_rnr = {
299
}
300
};
301
302
+static bool pmsav8_needed(void *opaque)
303
+{
304
+ ARMCPU *cpu = opaque;
305
+ CPUARMState *env = &cpu->env;
306
+
307
+ return arm_feature(env, ARM_FEATURE_PMSA) &&
308
+ arm_feature(env, ARM_FEATURE_V8);
309
+}
310
+
311
+static const VMStateDescription vmstate_pmsav8 = {
312
+ .name = "cpu/pmsav8",
313
+ .version_id = 1,
314
+ .minimum_version_id = 1,
315
+ .needed = pmsav8_needed,
316
+ .fields = (VMStateField[]) {
317
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rbar, ARMCPU, pmsav7_dregion, 0,
318
+ vmstate_info_uint32, uint32_t),
319
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rlar, ARMCPU, pmsav7_dregion, 0,
320
+ vmstate_info_uint32, uint32_t),
321
+ VMSTATE_UINT32(env.pmsav8.mair0, ARMCPU),
322
+ VMSTATE_UINT32(env.pmsav8.mair1, ARMCPU),
323
+ VMSTATE_END_OF_LIST()
324
+ }
325
+};
326
+
327
static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
328
VMStateField *field)
329
{
330
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_arm_cpu = {
331
*/
332
&vmstate_pmsav7_rnr,
333
&vmstate_pmsav7,
334
+ &vmstate_pmsav8,
335
NULL
336
}
337
};
338
--
339
2.7.4
340
341
diff view generated by jsdifflib
Deleted patch
1
If a v8M CPU supports the security extension then we need to
2
give it two AddressSpaces, the same way we do already for
3
an A profile core with EL3.
4
1
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 1503414539-28762-5-git-send-email-peter.maydell@linaro.org
8
---
9
target/arm/cpu.c | 13 ++++++-------
10
1 file changed, 6 insertions(+), 7 deletions(-)
11
12
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/target/arm/cpu.c
15
+++ b/target/arm/cpu.c
16
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
17
init_cpreg_list(cpu);
18
19
#ifndef CONFIG_USER_ONLY
20
- if (cpu->has_el3) {
21
- cs->num_ases = 2;
22
- } else {
23
- cs->num_ases = 1;
24
- }
25
-
26
- if (cpu->has_el3) {
27
+ if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
28
AddressSpace *as;
29
30
+ cs->num_ases = 2;
31
+
32
if (!cpu->secure_memory) {
33
cpu->secure_memory = cs->memory;
34
}
35
as = address_space_init_shareable(cpu->secure_memory,
36
"cpu-secure-memory");
37
cpu_address_space_init(cs, as, ARMASIdx_S);
38
+ } else {
39
+ cs->num_ases = 1;
40
}
41
+
42
cpu_address_space_init(cs,
43
address_space_init_shareable(cs->memory,
44
"cpu-memory"),
45
--
46
2.7.4
47
48
diff view generated by jsdifflib
Deleted patch
1
Now that MPU lookups can return different results for v8M
2
when the CPU is in secure vs non-secure state, we need to
3
have separate MMU indexes; add the secure counterparts
4
to the existing three M profile MMU indexes.
5
1
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 1503414539-28762-6-git-send-email-peter.maydell@linaro.org
9
---
10
target/arm/cpu.h | 19 +++++++++++++++++--
11
target/arm/helper.c | 9 ++++++++-
12
2 files changed, 25 insertions(+), 3 deletions(-)
13
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.h
17
+++ b/target/arm/cpu.h
18
@@ -XXX,XX +XXX,XX @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx,
19
* Execution priority negative (this is like privileged, but the
20
* MPU HFNMIENA bit means that it may have different access permission
21
* check results to normal privileged code, so can't share a TLB).
22
+ * If the CPU supports the v8M Security Extension then there are also:
23
+ * Secure User
24
+ * Secure Privileged
25
+ * Secure, execution priority negative
26
*
27
* The ARMMMUIdx and the mmu index value used by the core QEMU TLB code
28
* are not quite the same -- different CPU types (most notably M profile
29
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdx {
30
ARMMMUIdx_MUser = 0 | ARM_MMU_IDX_M,
31
ARMMMUIdx_MPriv = 1 | ARM_MMU_IDX_M,
32
ARMMMUIdx_MNegPri = 2 | ARM_MMU_IDX_M,
33
+ ARMMMUIdx_MSUser = 3 | ARM_MMU_IDX_M,
34
+ ARMMMUIdx_MSPriv = 4 | ARM_MMU_IDX_M,
35
+ ARMMMUIdx_MSNegPri = 5 | ARM_MMU_IDX_M,
36
/* Indexes below here don't have TLBs and are used only for AT system
37
* instructions or for the first stage of an S12 page table walk.
38
*/
39
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
40
ARMMMUIdxBit_MUser = 1 << 0,
41
ARMMMUIdxBit_MPriv = 1 << 1,
42
ARMMMUIdxBit_MNegPri = 1 << 2,
43
+ ARMMMUIdxBit_MSUser = 1 << 3,
44
+ ARMMMUIdxBit_MSPriv = 1 << 4,
45
+ ARMMMUIdxBit_MSNegPri = 1 << 5,
46
} ARMMMUIdxBit;
47
48
#define MMU_USER_IDX 0
49
@@ -XXX,XX +XXX,XX @@ static inline int arm_mmu_idx_to_el(ARMMMUIdx mmu_idx)
50
case ARM_MMU_IDX_A:
51
return mmu_idx & 3;
52
case ARM_MMU_IDX_M:
53
- return mmu_idx == ARMMMUIdx_MUser ? 0 : 1;
54
+ return (mmu_idx == ARMMMUIdx_MUser || mmu_idx == ARMMMUIdx_MSUser)
55
+ ? 0 : 1;
56
default:
57
g_assert_not_reached();
58
}
59
@@ -XXX,XX +XXX,XX @@ static inline int cpu_mmu_index(CPUARMState *env, bool ifetch)
60
*/
61
if ((env->v7m.exception > 0 && env->v7m.exception <= 3)
62
|| env->v7m.faultmask) {
63
- return arm_to_core_mmu_idx(ARMMMUIdx_MNegPri);
64
+ mmu_idx = ARMMMUIdx_MNegPri;
65
+ }
66
+
67
+ if (env->v7m.secure) {
68
+ mmu_idx += ARMMMUIdx_MSUser;
69
}
70
71
return arm_to_core_mmu_idx(mmu_idx);
72
diff --git a/target/arm/helper.c b/target/arm/helper.c
73
index XXXXXXX..XXXXXXX 100644
74
--- a/target/arm/helper.c
75
+++ b/target/arm/helper.c
76
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
77
case ARMMMUIdx_MPriv:
78
case ARMMMUIdx_MNegPri:
79
case ARMMMUIdx_MUser:
80
+ case ARMMMUIdx_MSPriv:
81
+ case ARMMMUIdx_MSNegPri:
82
+ case ARMMMUIdx_MSUser:
83
return 1;
84
default:
85
g_assert_not_reached();
86
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_secure(CPUARMState *env, ARMMMUIdx mmu_idx)
87
case ARMMMUIdx_S1E3:
88
case ARMMMUIdx_S1SE0:
89
case ARMMMUIdx_S1SE1:
90
+ case ARMMMUIdx_MSPriv:
91
+ case ARMMMUIdx_MSNegPri:
92
+ case ARMMMUIdx_MSUser:
93
return true;
94
default:
95
g_assert_not_reached();
96
@@ -XXX,XX +XXX,XX @@ static inline bool regime_translation_disabled(CPUARMState *env,
97
(R_V7M_MPU_CTRL_ENABLE_MASK | R_V7M_MPU_CTRL_HFNMIENA_MASK)) {
98
case R_V7M_MPU_CTRL_ENABLE_MASK:
99
/* Enabled, but not for HardFault and NMI */
100
- return mmu_idx == ARMMMUIdx_MNegPri;
101
+ return mmu_idx == ARMMMUIdx_MNegPri ||
102
+ mmu_idx == ARMMMUIdx_MSNegPri;
103
case R_V7M_MPU_CTRL_ENABLE_MASK | R_V7M_MPU_CTRL_HFNMIENA_MASK:
104
/* Enabled for all cases */
105
return false;
106
--
107
2.7.4
108
109
diff view generated by jsdifflib
Deleted patch
1
Make the BASEPRI register banked if v8M security extensions are enabled.
2
1
3
Note that we do not yet implement the functionality of the new
4
AIRCR.PRIS bit (which allows the effect of the NS copy of BASEPRI to
5
be restricted).
6
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 1503414539-28762-7-git-send-email-peter.maydell@linaro.org
10
---
11
target/arm/cpu.h | 14 +++++++++++++-
12
hw/intc/armv7m_nvic.c | 4 ++--
13
target/arm/helper.c | 10 ++++++----
14
target/arm/machine.c | 3 ++-
15
4 files changed, 23 insertions(+), 8 deletions(-)
16
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@
22
#define ARMV7M_EXCP_PENDSV 14
23
#define ARMV7M_EXCP_SYSTICK 15
24
25
+/* For M profile, some registers are banked secure vs non-secure;
26
+ * these are represented as a 2-element array where the first element
27
+ * is the non-secure copy and the second is the secure copy.
28
+ * When the CPU does not have implement the security extension then
29
+ * only the first element is used.
30
+ * This means that the copy for the current security state can be
31
+ * accessed via env->registerfield[env->v7m.secure] (whether the security
32
+ * extension is implemented or not).
33
+ */
34
+#define M_REG_NS 0
35
+#define M_REG_S 1
36
+
37
/* ARM-specific interrupt pending bits. */
38
#define CPU_INTERRUPT_FIQ CPU_INTERRUPT_TGT_EXT_1
39
#define CPU_INTERRUPT_VIRQ CPU_INTERRUPT_TGT_EXT_2
40
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
41
struct {
42
uint32_t other_sp;
43
uint32_t vecbase;
44
- uint32_t basepri;
45
+ uint32_t basepri[2];
46
uint32_t control;
47
uint32_t ccr; /* Configuration and Control */
48
uint32_t cfsr; /* Configurable Fault Status */
49
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
50
index XXXXXXX..XXXXXXX 100644
51
--- a/hw/intc/armv7m_nvic.c
52
+++ b/hw/intc/armv7m_nvic.c
53
@@ -XXX,XX +XXX,XX @@ static inline int nvic_exec_prio(NVICState *s)
54
running = -1;
55
} else if (env->v7m.primask) {
56
running = 0;
57
- } else if (env->v7m.basepri > 0) {
58
- running = env->v7m.basepri & nvic_gprio_mask(s);
59
+ } else if (env->v7m.basepri[env->v7m.secure] > 0) {
60
+ running = env->v7m.basepri[env->v7m.secure] & nvic_gprio_mask(s);
61
} else {
62
running = NVIC_NOEXC_PRIO; /* lower than any possible priority */
63
}
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/helper.c
67
+++ b/target/arm/helper.c
68
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
69
return env->v7m.primask;
70
case 17: /* BASEPRI */
71
case 18: /* BASEPRI_MAX */
72
- return env->v7m.basepri;
73
+ return env->v7m.basepri[env->v7m.secure];
74
case 19: /* FAULTMASK */
75
return env->v7m.faultmask;
76
default:
77
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
78
env->v7m.primask = val & 1;
79
break;
80
case 17: /* BASEPRI */
81
- env->v7m.basepri = val & 0xff;
82
+ env->v7m.basepri[env->v7m.secure] = val & 0xff;
83
break;
84
case 18: /* BASEPRI_MAX */
85
val &= 0xff;
86
- if (val != 0 && (val < env->v7m.basepri || env->v7m.basepri == 0))
87
- env->v7m.basepri = val;
88
+ if (val != 0 && (val < env->v7m.basepri[env->v7m.secure]
89
+ || env->v7m.basepri[env->v7m.secure] == 0)) {
90
+ env->v7m.basepri[env->v7m.secure] = val;
91
+ }
92
break;
93
case 19: /* FAULTMASK */
94
env->v7m.faultmask = val & 1;
95
diff --git a/target/arm/machine.c b/target/arm/machine.c
96
index XXXXXXX..XXXXXXX 100644
97
--- a/target/arm/machine.c
98
+++ b/target/arm/machine.c
99
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
100
.needed = m_needed,
101
.fields = (VMStateField[]) {
102
VMSTATE_UINT32(env.v7m.vecbase, ARMCPU),
103
- VMSTATE_UINT32(env.v7m.basepri, ARMCPU),
104
+ VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
105
VMSTATE_UINT32(env.v7m.control, ARMCPU),
106
VMSTATE_UINT32(env.v7m.ccr, ARMCPU),
107
VMSTATE_UINT32(env.v7m.cfsr, ARMCPU),
108
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
109
.needed = m_security_needed,
110
.fields = (VMStateField[]) {
111
VMSTATE_UINT32(env.v7m.secure, ARMCPU),
112
+ VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
113
VMSTATE_END_OF_LIST()
114
}
115
};
116
--
117
2.7.4
118
119
diff view generated by jsdifflib
Deleted patch
1
Make the PRIMASK register banked if v8M security extensions are enabled.
2
1
3
Note that we do not yet implement the functionality of the new
4
AIRCR.PRIS bit (which allows the effect of the NS copy of PRIMASK to
5
be restricted).
6
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 1503414539-28762-8-git-send-email-peter.maydell@linaro.org
10
---
11
target/arm/cpu.h | 2 +-
12
hw/intc/armv7m_nvic.c | 2 +-
13
target/arm/helper.c | 4 ++--
14
target/arm/machine.c | 9 +++++++--
15
4 files changed, 11 insertions(+), 6 deletions(-)
16
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
uint32_t bfar; /* BusFault Address */
23
unsigned mpu_ctrl; /* MPU_CTRL */
24
int exception;
25
- uint32_t primask;
26
+ uint32_t primask[2];
27
uint32_t faultmask;
28
uint32_t secure; /* Is CPU in Secure state? (not guest visible) */
29
} v7m;
30
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/intc/armv7m_nvic.c
33
+++ b/hw/intc/armv7m_nvic.c
34
@@ -XXX,XX +XXX,XX @@ static inline int nvic_exec_prio(NVICState *s)
35
36
if (env->v7m.faultmask) {
37
running = -1;
38
- } else if (env->v7m.primask) {
39
+ } else if (env->v7m.primask[env->v7m.secure]) {
40
running = 0;
41
} else if (env->v7m.basepri[env->v7m.secure] > 0) {
42
running = env->v7m.basepri[env->v7m.secure] & nvic_gprio_mask(s);
43
diff --git a/target/arm/helper.c b/target/arm/helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/helper.c
46
+++ b/target/arm/helper.c
47
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
48
return (env->v7m.control & R_V7M_CONTROL_SPSEL_MASK) ?
49
env->regs[13] : env->v7m.other_sp;
50
case 16: /* PRIMASK */
51
- return env->v7m.primask;
52
+ return env->v7m.primask[env->v7m.secure];
53
case 17: /* BASEPRI */
54
case 18: /* BASEPRI_MAX */
55
return env->v7m.basepri[env->v7m.secure];
56
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
57
}
58
break;
59
case 16: /* PRIMASK */
60
- env->v7m.primask = val & 1;
61
+ env->v7m.primask[env->v7m.secure] = val & 1;
62
break;
63
case 17: /* BASEPRI */
64
env->v7m.basepri[env->v7m.secure] = val & 0xff;
65
diff --git a/target/arm/machine.c b/target/arm/machine.c
66
index XXXXXXX..XXXXXXX 100644
67
--- a/target/arm/machine.c
68
+++ b/target/arm/machine.c
69
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_faultmask_primask = {
70
.minimum_version_id = 1,
71
.fields = (VMStateField[]) {
72
VMSTATE_UINT32(env.v7m.faultmask, ARMCPU),
73
- VMSTATE_UINT32(env.v7m.primask, ARMCPU),
74
+ VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
75
VMSTATE_END_OF_LIST()
76
}
77
};
78
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
79
.fields = (VMStateField[]) {
80
VMSTATE_UINT32(env.v7m.secure, ARMCPU),
81
VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
82
+ VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
83
VMSTATE_END_OF_LIST()
84
}
85
};
86
@@ -XXX,XX +XXX,XX @@ static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
87
* differences are that the T bit is not in the same place, the
88
* primask/faultmask info may be in the CPSR I and F bits, and
89
* we do not want the mode bits.
90
+ * We know that this cleanup happened before v8M, so there
91
+ * is no complication with banked primask/faultmask.
92
*/
93
uint32_t newval = val;
94
95
+ assert(!arm_feature(env, ARM_FEATURE_M_SECURITY));
96
+
97
newval &= (CPSR_NZCV | CPSR_Q | CPSR_IT | CPSR_GE);
98
if (val & CPSR_T) {
99
newval |= XPSR_T;
100
@@ -XXX,XX +XXX,XX @@ static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
101
env->v7m.faultmask = 1;
102
}
103
if (val & CPSR_I) {
104
- env->v7m.primask = 1;
105
+ env->v7m.primask[M_REG_NS] = 1;
106
}
107
val = newval;
108
}
109
--
110
2.7.4
111
112
diff view generated by jsdifflib
1
Make the MPU registers MPU_MAIR0 and MPU_MAIR1 banked if v8M security
1
In arm_v7m_mmu_idx_for_secstate() we get the 'priv' level to pass to
2
extensions are enabled.
2
armv7m_mmu_idx_for_secstate_and_priv() by calling arm_current_el().
3
This is incorrect when the security state being queried is not the
4
current one, because arm_current_el() uses the current security state
5
to determine which of the banked CONTROL.nPRIV bits to look at.
6
The effect was that if (for instance) Secure state was in privileged
7
mode but Non-Secure was not then we would return the wrong MMU index.
3
8
4
We can freely add more items to vmstate_m_security without
9
The only places where we are using this function in a way that could
5
breaking migration compatibility, because no CPU currently
10
trigger this bug are for the stack loads during a v8M function-return
6
has the ARM_FEATURE_M_SECURITY bit enabled and so this
11
and for the instruction fetch of a v8M SG insn.
7
subsection is not yet used by anything.
12
13
Fix the bug by expanding out the M-profile version of the
14
arm_current_el() logic inline so it can use the passed in secstate
15
rather than env->v7m.secure.
8
16
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 1503414539-28762-14-git-send-email-peter.maydell@linaro.org
19
Message-id: 20201022164408.13214-1-peter.maydell@linaro.org
12
---
20
---
13
target/arm/cpu.h | 4 ++--
21
target/arm/m_helper.c | 3 ++-
14
hw/intc/armv7m_nvic.c | 8 ++++----
22
1 file changed, 2 insertions(+), 1 deletion(-)
15
target/arm/cpu.c | 26 ++++++++++++++++++++------
16
target/arm/helper.c | 11 ++++++-----
17
target/arm/machine.c | 12 ++++++++----
18
5 files changed, 40 insertions(+), 21 deletions(-)
19
23
20
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
24
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
21
index XXXXXXX..XXXXXXX 100644
25
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/cpu.h
26
--- a/target/arm/m_helper.c
23
+++ b/target/arm/cpu.h
27
+++ b/target/arm/m_helper.c
24
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
28
@@ -XXX,XX +XXX,XX @@ ARMMMUIdx arm_v7m_mmu_idx_for_secstate_and_priv(CPUARMState *env,
25
* pmsav7.rnr (region number register)
29
/* Return the MMU index for a v7M CPU in the specified security state */
26
* pmsav7_dregion (number of configured regions)
30
ARMMMUIdx arm_v7m_mmu_idx_for_secstate(CPUARMState *env, bool secstate)
27
*/
28
- uint32_t *rbar;
29
- uint32_t *rlar;
30
+ uint32_t *rbar[2];
31
+ uint32_t *rlar[2];
32
uint32_t mair0[2];
33
uint32_t mair1[2];
34
} pmsav8;
35
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/hw/intc/armv7m_nvic.c
38
+++ b/hw/intc/armv7m_nvic.c
39
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
40
if (region >= cpu->pmsav7_dregion) {
41
return 0;
42
}
43
- return cpu->env.pmsav8.rbar[region];
44
+ return cpu->env.pmsav8.rbar[attrs.secure][region];
45
}
46
47
if (region >= cpu->pmsav7_dregion) {
48
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
49
if (region >= cpu->pmsav7_dregion) {
50
return 0;
51
}
52
- return cpu->env.pmsav8.rlar[region];
53
+ return cpu->env.pmsav8.rlar[attrs.secure][region];
54
}
55
56
if (region >= cpu->pmsav7_dregion) {
57
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
58
if (region >= cpu->pmsav7_dregion) {
59
return;
60
}
61
- cpu->env.pmsav8.rbar[region] = value;
62
+ cpu->env.pmsav8.rbar[attrs.secure][region] = value;
63
tlb_flush(CPU(cpu));
64
return;
65
}
66
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
67
if (region >= cpu->pmsav7_dregion) {
68
return;
69
}
70
- cpu->env.pmsav8.rlar[region] = value;
71
+ cpu->env.pmsav8.rlar[attrs.secure][region] = value;
72
tlb_flush(CPU(cpu));
73
return;
74
}
75
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
76
index XXXXXXX..XXXXXXX 100644
77
--- a/target/arm/cpu.c
78
+++ b/target/arm/cpu.c
79
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
80
if (arm_feature(env, ARM_FEATURE_PMSA)) {
81
if (cpu->pmsav7_dregion > 0) {
82
if (arm_feature(env, ARM_FEATURE_V8)) {
83
- memset(env->pmsav8.rbar, 0,
84
- sizeof(*env->pmsav8.rbar) * cpu->pmsav7_dregion);
85
- memset(env->pmsav8.rlar, 0,
86
- sizeof(*env->pmsav8.rlar) * cpu->pmsav7_dregion);
87
+ memset(env->pmsav8.rbar[M_REG_NS], 0,
88
+ sizeof(*env->pmsav8.rbar[M_REG_NS])
89
+ * cpu->pmsav7_dregion);
90
+ memset(env->pmsav8.rlar[M_REG_NS], 0,
91
+ sizeof(*env->pmsav8.rlar[M_REG_NS])
92
+ * cpu->pmsav7_dregion);
93
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
94
+ memset(env->pmsav8.rbar[M_REG_S], 0,
95
+ sizeof(*env->pmsav8.rbar[M_REG_S])
96
+ * cpu->pmsav7_dregion);
97
+ memset(env->pmsav8.rlar[M_REG_S], 0,
98
+ sizeof(*env->pmsav8.rlar[M_REG_S])
99
+ * cpu->pmsav7_dregion);
100
+ }
101
} else if (arm_feature(env, ARM_FEATURE_V7)) {
102
memset(env->pmsav7.drbar, 0,
103
sizeof(*env->pmsav7.drbar) * cpu->pmsav7_dregion);
104
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
105
if (nr) {
106
if (arm_feature(env, ARM_FEATURE_V8)) {
107
/* PMSAv8 */
108
- env->pmsav8.rbar = g_new0(uint32_t, nr);
109
- env->pmsav8.rlar = g_new0(uint32_t, nr);
110
+ env->pmsav8.rbar[M_REG_NS] = g_new0(uint32_t, nr);
111
+ env->pmsav8.rlar[M_REG_NS] = g_new0(uint32_t, nr);
112
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
113
+ env->pmsav8.rbar[M_REG_S] = g_new0(uint32_t, nr);
114
+ env->pmsav8.rlar[M_REG_S] = g_new0(uint32_t, nr);
115
+ }
116
} else {
117
env->pmsav7.drbar = g_new0(uint32_t, nr);
118
env->pmsav7.drsr = g_new0(uint32_t, nr);
119
diff --git a/target/arm/helper.c b/target/arm/helper.c
120
index XXXXXXX..XXXXXXX 100644
121
--- a/target/arm/helper.c
122
+++ b/target/arm/helper.c
123
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_pmsav8(CPUARMState *env, uint32_t address,
124
{
31
{
125
ARMCPU *cpu = arm_env_get_cpu(env);
32
- bool priv = arm_current_el(env) != 0;
126
bool is_user = regime_is_user(env, mmu_idx);
33
+ bool priv = arm_v7m_is_handler_mode(env) ||
127
+ uint32_t secure = regime_is_secure(env, mmu_idx);
34
+ !(env->v7m.control[secstate] & 1);
128
int n;
35
129
int matchregion = -1;
36
return arm_v7m_mmu_idx_for_secstate_and_priv(env, secstate, priv);
130
bool hit = false;
37
}
131
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_pmsav8(CPUARMState *env, uint32_t address,
132
* with bits [4:0] all zeroes, but the limit address is bits
133
* [31:5] from the register with bits [4:0] all ones.
134
*/
135
- uint32_t base = env->pmsav8.rbar[n] & ~0x1f;
136
- uint32_t limit = env->pmsav8.rlar[n] | 0x1f;
137
+ uint32_t base = env->pmsav8.rbar[secure][n] & ~0x1f;
138
+ uint32_t limit = env->pmsav8.rlar[secure][n] | 0x1f;
139
140
- if (!(env->pmsav8.rlar[n] & 0x1)) {
141
+ if (!(env->pmsav8.rlar[secure][n] & 0x1)) {
142
/* Region disabled */
143
continue;
144
}
145
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_pmsav8(CPUARMState *env, uint32_t address,
146
/* hit using the background region */
147
get_phys_addr_pmsav7_default(env, mmu_idx, address, prot);
148
} else {
149
- uint32_t ap = extract32(env->pmsav8.rbar[matchregion], 1, 2);
150
- uint32_t xn = extract32(env->pmsav8.rbar[matchregion], 0, 1);
151
+ uint32_t ap = extract32(env->pmsav8.rbar[secure][matchregion], 1, 2);
152
+ uint32_t xn = extract32(env->pmsav8.rbar[secure][matchregion], 0, 1);
153
154
if (m_is_system_region(env, address)) {
155
/* System space is always execute never */
156
diff --git a/target/arm/machine.c b/target/arm/machine.c
157
index XXXXXXX..XXXXXXX 100644
158
--- a/target/arm/machine.c
159
+++ b/target/arm/machine.c
160
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav8 = {
161
.minimum_version_id = 1,
162
.needed = pmsav8_needed,
163
.fields = (VMStateField[]) {
164
- VMSTATE_VARRAY_UINT32(env.pmsav8.rbar, ARMCPU, pmsav7_dregion, 0,
165
- vmstate_info_uint32, uint32_t),
166
- VMSTATE_VARRAY_UINT32(env.pmsav8.rlar, ARMCPU, pmsav7_dregion, 0,
167
- vmstate_info_uint32, uint32_t),
168
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_NS], ARMCPU, pmsav7_dregion,
169
+ 0, vmstate_info_uint32, uint32_t),
170
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_NS], ARMCPU, pmsav7_dregion,
171
+ 0, vmstate_info_uint32, uint32_t),
172
VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
173
VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
174
VMSTATE_END_OF_LIST()
175
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
176
VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
177
VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
178
VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
179
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_S], ARMCPU, pmsav7_dregion,
180
+ 0, vmstate_info_uint32, uint32_t),
181
+ VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
182
+ 0, vmstate_info_uint32, uint32_t),
183
VMSTATE_END_OF_LIST()
184
}
185
};
186
--
38
--
187
2.7.4
39
2.20.1
188
40
189
41
diff view generated by jsdifflib
1
Make the CCR register banked if v8M security extensions are enabled.
1
On some hosts (eg Ubuntu Bionic) pkg-config returns a set of
2
libraries for gio-2.0 which don't actually work when compiling
3
statically. (Specifically, the returned library string includes
4
-lmount, but not -lblkid which -lmount depends upon, so linking
5
fails due to missing symbols.)
2
6
3
This is slightly more complicated than the other "add banking"
7
Check that the libraries work, and don't enable gio if they don't,
4
patches because there is one bit in the register which is not
8
in the same way we do for gnutls.
5
banked. We keep the live data in the NS copy of the register,
6
and adjust it on register reads and writes. (Since we don't
7
currently implement the behaviour that the bit controls, there
8
is nowhere else that needs to care.)
9
10
This patch includes the enforcement of the bits which are newly
11
RES1 in ARMv8M.
12
9
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Message-id: 1503414539-28762-17-git-send-email-peter.maydell@linaro.org
11
Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
12
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
13
Message-id: 20200928160402.7961-1-peter.maydell@linaro.org
15
---
14
---
16
target/arm/cpu.h | 2 +-
15
configure | 10 +++++++++-
17
hw/intc/armv7m_nvic.c | 33 +++++++++++++++++++++++++++------
16
1 file changed, 9 insertions(+), 1 deletion(-)
18
target/arm/cpu.c | 12 +++++++++---
19
target/arm/helper.c | 5 +++--
20
target/arm/machine.c | 3 ++-
21
5 files changed, 42 insertions(+), 13 deletions(-)
22
17
23
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
diff --git a/configure b/configure
24
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100755
25
--- a/target/arm/cpu.h
20
--- a/configure
26
+++ b/target/arm/cpu.h
21
+++ b/configure
27
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
22
@@ -XXX,XX +XXX,XX @@ if test "$static" = yes && test "$mingw32" = yes; then
28
uint32_t vecbase[2];
23
fi
29
uint32_t basepri[2];
24
30
uint32_t control[2];
25
if $pkg_config --atleast-version=$glib_req_ver gio-2.0; then
31
- uint32_t ccr; /* Configuration and Control */
26
- gio=yes
32
+ uint32_t ccr[2]; /* Configuration and Control */
27
gio_cflags=$($pkg_config --cflags gio-2.0)
33
uint32_t cfsr; /* Configurable Fault Status */
28
gio_libs=$($pkg_config --libs gio-2.0)
34
uint32_t hfsr; /* HardFault Status */
29
gdbus_codegen=$($pkg_config --variable=gdbus_codegen gio-2.0)
35
uint32_t dfsr; /* Debug Fault Status Register */
30
if [ ! -x "$gdbus_codegen" ]; then
36
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
31
gdbus_codegen=
37
index XXXXXXX..XXXXXXX 100644
32
fi
38
--- a/hw/intc/armv7m_nvic.c
33
+ # Check that the libraries actually work -- Ubuntu 18.04 ships
39
+++ b/hw/intc/armv7m_nvic.c
34
+ # with pkg-config --static --libs data for gio-2.0 that is missing
40
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
35
+ # -lblkid and will give a link error.
41
/* TODO: Implement SLEEPONEXIT. */
36
+ write_c_skeleton
42
return 0;
37
+ if compile_prog "" "gio_libs" ; then
43
case 0xd14: /* Configuration Control. */
38
+ gio=yes
44
- return cpu->env.v7m.ccr;
39
+ else
45
+ /* The BFHFNMIGN bit is the only non-banked bit; we
40
+ gio=no
46
+ * keep it in the non-secure copy of the register.
41
+ fi
47
+ */
42
else
48
+ val = cpu->env.v7m.ccr[attrs.secure];
43
gio=no
49
+ val |= cpu->env.v7m.ccr[M_REG_NS] & R_V7M_CCR_BFHFNMIGN_MASK;
44
fi
50
+ return val;
51
case 0xd24: /* System Handler Status. */
52
val = 0;
53
if (s->vectors[ARMV7M_EXCP_MEM].active) {
54
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
55
R_V7M_CCR_USERSETMPEND_MASK |
56
R_V7M_CCR_NONBASETHRDENA_MASK);
57
58
- cpu->env.v7m.ccr = value;
59
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
60
+ /* v8M makes NONBASETHRDENA and STKALIGN be RES1 */
61
+ value |= R_V7M_CCR_NONBASETHRDENA_MASK
62
+ | R_V7M_CCR_STKALIGN_MASK;
63
+ }
64
+ if (attrs.secure) {
65
+ /* the BFHFNMIGN bit is not banked; keep that in the NS copy */
66
+ cpu->env.v7m.ccr[M_REG_NS] =
67
+ (cpu->env.v7m.ccr[M_REG_NS] & ~R_V7M_CCR_BFHFNMIGN_MASK)
68
+ | (value & R_V7M_CCR_BFHFNMIGN_MASK);
69
+ value &= ~R_V7M_CCR_BFHFNMIGN_MASK;
70
+ }
71
+
72
+ cpu->env.v7m.ccr[attrs.secure] = value;
73
break;
74
case 0xd24: /* System Handler Control. */
75
s->vectors[ARMV7M_EXCP_MEM].active = (value & (1 << 0)) != 0;
76
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
77
}
78
}
79
80
-static bool nvic_user_access_ok(NVICState *s, hwaddr offset)
81
+static bool nvic_user_access_ok(NVICState *s, hwaddr offset, MemTxAttrs attrs)
82
{
83
/* Return true if unprivileged access to this register is permitted. */
84
switch (offset) {
85
case 0xf00: /* STIR: accessible only if CCR.USERSETMPEND permits */
86
- return s->cpu->env.v7m.ccr & R_V7M_CCR_USERSETMPEND_MASK;
87
+ /* For access via STIR_NS it is the NS CCR.USERSETMPEND that
88
+ * controls access even though the CPU is in Secure state (I_QDKX).
89
+ */
90
+ return s->cpu->env.v7m.ccr[attrs.secure] & R_V7M_CCR_USERSETMPEND_MASK;
91
default:
92
/* All other user accesses cause a BusFault unconditionally */
93
return false;
94
@@ -XXX,XX +XXX,XX @@ static MemTxResult nvic_sysreg_read(void *opaque, hwaddr addr,
95
unsigned i, startvec, end;
96
uint32_t val;
97
98
- if (attrs.user && !nvic_user_access_ok(s, addr)) {
99
+ if (attrs.user && !nvic_user_access_ok(s, addr, attrs)) {
100
/* Generate BusFault for unprivileged accesses */
101
return MEMTX_ERROR;
102
}
103
@@ -XXX,XX +XXX,XX @@ static MemTxResult nvic_sysreg_write(void *opaque, hwaddr addr,
104
105
trace_nvic_sysreg_write(addr, value, size);
106
107
- if (attrs.user && !nvic_user_access_ok(s, addr)) {
108
+ if (attrs.user && !nvic_user_access_ok(s, addr, attrs)) {
109
/* Generate BusFault for unprivileged accesses */
110
return MEMTX_ERROR;
111
}
112
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
113
index XXXXXXX..XXXXXXX 100644
114
--- a/target/arm/cpu.c
115
+++ b/target/arm/cpu.c
116
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
117
env->v7m.secure = true;
118
}
119
120
- /* The reset value of this bit is IMPDEF, but ARM recommends
121
+ /* In v7M the reset value of this bit is IMPDEF, but ARM recommends
122
* that it resets to 1, so QEMU always does that rather than making
123
- * it dependent on CPU model.
124
+ * it dependent on CPU model. In v8M it is RES1.
125
*/
126
- env->v7m.ccr = R_V7M_CCR_STKALIGN_MASK;
127
+ env->v7m.ccr[M_REG_NS] = R_V7M_CCR_STKALIGN_MASK;
128
+ env->v7m.ccr[M_REG_S] = R_V7M_CCR_STKALIGN_MASK;
129
+ if (arm_feature(env, ARM_FEATURE_V8)) {
130
+ /* in v8M the NONBASETHRDENA bit [0] is RES1 */
131
+ env->v7m.ccr[M_REG_NS] |= R_V7M_CCR_NONBASETHRDENA_MASK;
132
+ env->v7m.ccr[M_REG_S] |= R_V7M_CCR_NONBASETHRDENA_MASK;
133
+ }
134
135
/* Unlike A/R profile, M profile defines the reset LR value */
136
env->regs[14] = 0xffffffff;
137
diff --git a/target/arm/helper.c b/target/arm/helper.c
138
index XXXXXXX..XXXXXXX 100644
139
--- a/target/arm/helper.c
140
+++ b/target/arm/helper.c
141
@@ -XXX,XX +XXX,XX @@ static void v7m_push_stack(ARMCPU *cpu)
142
uint32_t xpsr = xpsr_read(env);
143
144
/* Align stack pointer if the guest wants that */
145
- if ((env->regs[13] & 4) && (env->v7m.ccr & R_V7M_CCR_STKALIGN_MASK)) {
146
+ if ((env->regs[13] & 4) &&
147
+ (env->v7m.ccr[env->v7m.secure] & R_V7M_CCR_STKALIGN_MASK)) {
148
env->regs[13] -= 4;
149
xpsr |= XPSR_SPREALIGN;
150
}
151
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
152
/* fall through */
153
case 9: /* Return to Thread using Main stack */
154
if (!rettobase &&
155
- !(env->v7m.ccr & R_V7M_CCR_NONBASETHRDENA_MASK)) {
156
+ !(env->v7m.ccr[env->v7m.secure] & R_V7M_CCR_NONBASETHRDENA_MASK)) {
157
ufault = true;
158
}
159
break;
160
diff --git a/target/arm/machine.c b/target/arm/machine.c
161
index XXXXXXX..XXXXXXX 100644
162
--- a/target/arm/machine.c
163
+++ b/target/arm/machine.c
164
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
165
VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
166
VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
167
VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
168
- VMSTATE_UINT32(env.v7m.ccr, ARMCPU),
169
+ VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
170
VMSTATE_UINT32(env.v7m.cfsr, ARMCPU),
171
VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
172
VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
173
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
174
VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
175
VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
176
VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
177
+ VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
178
VMSTATE_END_OF_LIST()
179
}
180
};
181
--
45
--
182
2.7.4
46
2.20.1
183
47
184
48
diff view generated by jsdifflib
1
Make the VTOR register banked if v8M security extensions are enabled.
1
In gicv3_init_cpuif() we copy the ARMCPU gicv3_maintenance_interrupt
2
into the GICv3CPUState struct's maintenance_irq field. This will
3
only work if the board happens to have already wired up the CPU
4
maintenance IRQ before the GIC was realized. Unfortunately this is
5
not the case for the 'virt' board, and so the value that gets copied
6
is NULL (since a qemu_irq is really a pointer to an IRQState struct
7
under the hood). The effect is that the CPU interface code never
8
actually raises the maintenance interrupt line.
2
9
10
Instead, since the GICv3CPUState has a pointer to the CPUState, make
11
the dereference at the point where we want to raise the interrupt, to
12
avoid an implicit requirement on board code to wire things up in a
13
particular order.
14
15
Reported-by: Jose Martins <josemartins90@gmail.com>
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
17
Message-id: 20201009153904.28529-1-peter.maydell@linaro.org
5
Message-id: 1503414539-28762-12-git-send-email-peter.maydell@linaro.org
18
Reviewed-by: Luc Michel <luc@lmichel.fr>
6
---
19
---
7
target/arm/cpu.h | 2 +-
20
include/hw/intc/arm_gicv3_common.h | 1 -
8
hw/intc/armv7m_nvic.c | 13 +++++++------
21
hw/intc/arm_gicv3_cpuif.c | 5 ++---
9
target/arm/helper.c | 2 +-
22
2 files changed, 2 insertions(+), 4 deletions(-)
10
target/arm/machine.c | 3 ++-
11
4 files changed, 11 insertions(+), 9 deletions(-)
12
23
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
24
diff --git a/include/hw/intc/arm_gicv3_common.h b/include/hw/intc/arm_gicv3_common.h
14
index XXXXXXX..XXXXXXX 100644
25
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.h
26
--- a/include/hw/intc/arm_gicv3_common.h
16
+++ b/target/arm/cpu.h
27
+++ b/include/hw/intc/arm_gicv3_common.h
17
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
28
@@ -XXX,XX +XXX,XX @@ struct GICv3CPUState {
18
29
qemu_irq parent_fiq;
19
struct {
30
qemu_irq parent_virq;
20
uint32_t other_sp;
31
qemu_irq parent_vfiq;
21
- uint32_t vecbase;
32
- qemu_irq maintenance_irq;
22
+ uint32_t vecbase[2];
33
23
uint32_t basepri[2];
34
/* Redistributor */
24
uint32_t control[2];
35
uint32_t level; /* Current IRQ level */
25
uint32_t ccr; /* Configuration and Control */
36
diff --git a/hw/intc/arm_gicv3_cpuif.c b/hw/intc/arm_gicv3_cpuif.c
26
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
27
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
28
--- a/hw/intc/armv7m_nvic.c
38
--- a/hw/intc/arm_gicv3_cpuif.c
29
+++ b/hw/intc/armv7m_nvic.c
39
+++ b/hw/intc/arm_gicv3_cpuif.c
30
@@ -XXX,XX +XXX,XX @@ static void set_irq_level(void *opaque, int n, int level)
40
@@ -XXX,XX +XXX,XX @@ static void gicv3_cpuif_virt_update(GICv3CPUState *cs)
31
}
41
int irqlevel = 0;
42
int fiqlevel = 0;
43
int maintlevel = 0;
44
+ ARMCPU *cpu = ARM_CPU(cs->cpu);
45
46
idx = hppvi_index(cs);
47
trace_gicv3_cpuif_virt_update(gicv3_redist_affid(cs), idx);
48
@@ -XXX,XX +XXX,XX @@ static void gicv3_cpuif_virt_update(GICv3CPUState *cs)
49
50
qemu_set_irq(cs->parent_vfiq, fiqlevel);
51
qemu_set_irq(cs->parent_virq, irqlevel);
52
- qemu_set_irq(cs->maintenance_irq, maintlevel);
53
+ qemu_set_irq(cpu->gicv3_maintenance_interrupt, maintlevel);
32
}
54
}
33
55
34
-static uint32_t nvic_readl(NVICState *s, uint32_t offset)
56
static uint64_t icv_ap_read(CPUARMState *env, const ARMCPRegInfo *ri)
35
+static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
57
@@ -XXX,XX +XXX,XX @@ void gicv3_init_cpuif(GICv3State *s)
36
{
58
&& cpu->gic_num_lrs) {
37
ARMCPU *cpu = s->cpu;
59
int j;
38
uint32_t val;
60
39
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset)
61
- cs->maintenance_irq = cpu->gicv3_maintenance_interrupt;
40
/* ISRPREEMPT not implemented */
62
-
41
return val;
63
cs->num_list_regs = cpu->gic_num_lrs;
42
case 0xd08: /* Vector Table Offset. */
64
cs->vpribits = cpu->gic_vpribits;
43
- return cpu->env.v7m.vecbase;
65
cs->vprebits = cpu->gic_vprebits;
44
+ return cpu->env.v7m.vecbase[attrs.secure];
45
case 0xd0c: /* Application Interrupt/Reset Control. */
46
return 0xfa050000 | (s->prigroup << 8);
47
case 0xd10: /* System Control. */
48
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset)
49
}
50
}
51
52
-static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
53
+static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
54
+ MemTxAttrs attrs)
55
{
56
ARMCPU *cpu = s->cpu;
57
58
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
59
}
60
break;
61
case 0xd08: /* Vector Table Offset. */
62
- cpu->env.v7m.vecbase = value & 0xffffff80;
63
+ cpu->env.v7m.vecbase[attrs.secure] = value & 0xffffff80;
64
break;
65
case 0xd0c: /* Application Interrupt/Reset Control. */
66
if ((value >> 16) == 0x05fa) {
67
@@ -XXX,XX +XXX,XX @@ static MemTxResult nvic_sysreg_read(void *opaque, hwaddr addr,
68
break;
69
default:
70
if (size == 4) {
71
- val = nvic_readl(s, offset);
72
+ val = nvic_readl(s, offset, attrs);
73
} else {
74
qemu_log_mask(LOG_GUEST_ERROR,
75
"NVIC: Bad read of size %d at offset 0x%x\n",
76
@@ -XXX,XX +XXX,XX @@ static MemTxResult nvic_sysreg_write(void *opaque, hwaddr addr,
77
return MEMTX_OK;
78
}
79
if (size == 4) {
80
- nvic_writel(s, offset, value);
81
+ nvic_writel(s, offset, value, attrs);
82
return MEMTX_OK;
83
}
84
qemu_log_mask(LOG_GUEST_ERROR,
85
diff --git a/target/arm/helper.c b/target/arm/helper.c
86
index XXXXXXX..XXXXXXX 100644
87
--- a/target/arm/helper.c
88
+++ b/target/arm/helper.c
89
@@ -XXX,XX +XXX,XX @@ static uint32_t arm_v7m_load_vector(ARMCPU *cpu)
90
CPUState *cs = CPU(cpu);
91
CPUARMState *env = &cpu->env;
92
MemTxResult result;
93
- hwaddr vec = env->v7m.vecbase + env->v7m.exception * 4;
94
+ hwaddr vec = env->v7m.vecbase[env->v7m.secure] + env->v7m.exception * 4;
95
uint32_t addr;
96
97
addr = address_space_ldl(cs->as, vec,
98
diff --git a/target/arm/machine.c b/target/arm/machine.c
99
index XXXXXXX..XXXXXXX 100644
100
--- a/target/arm/machine.c
101
+++ b/target/arm/machine.c
102
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m = {
103
.minimum_version_id = 4,
104
.needed = m_needed,
105
.fields = (VMStateField[]) {
106
- VMSTATE_UINT32(env.v7m.vecbase, ARMCPU),
107
+ VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
108
VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
109
VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
110
VMSTATE_UINT32(env.v7m.ccr, ARMCPU),
111
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
112
VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
113
VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
114
VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
115
+ VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
116
VMSTATE_END_OF_LIST()
117
}
118
};
119
--
66
--
120
2.7.4
67
2.20.1
121
68
122
69
diff view generated by jsdifflib
1
Make the FAULTMASK register banked if v8M security extensions are enabled.
1
The kerneldoc script currently emits Sphinx markup for a macro with
2
arguments that uses the c:function directive. This is correct for
3
Sphinx versions earlier than Sphinx 3, where c:macro doesn't allow
4
documentation of macros with arguments and c:function is not picky
5
about the syntax of what it is passed. However, in Sphinx 3 the
6
c:macro directive was enhanced to support macros with arguments,
7
and c:function was made more picky about what syntax it accepted.
2
8
3
Note that we do not yet implement the functionality of the new
9
When kerneldoc is told that it needs to produce output for Sphinx
4
AIRCR.PRIS bit (which allows the effect of the NS copy of FAULTMASK to
10
3 or later, make it emit c:function only for functions and c:macro
5
be restricted).
11
for macros with arguments. We assume that anything with a return
12
type is a function and anything without is a macro.
6
13
7
This patch includes the code to determine for v8M which copy
14
This fixes the Sphinx error:
8
of FAULTMASK should be updated on exception exit; further
9
changes will be required to the exception exit code in general
10
to support v8M, so this is just a small piece of that.
11
15
12
The v8M ARM ARM introduces a notation where individual paragraphs
16
/home/petmay01/linaro/qemu-from-laptop/qemu/docs/../include/qom/object.h:155:Error in declarator
13
are labelled with R (for rule) or I (for information) followed
17
If declarator-id with parameters (e.g., 'void f(int arg)'):
14
by a random group of subscript letters. In comments where we want
18
Invalid C declaration: Expected identifier in nested name. [error at 25]
15
to refer to a particular part of the manual we use this convention,
19
DECLARE_INSTANCE_CHECKER ( InstanceType, OBJ_NAME, TYPENAME)
16
which should be more stable across document revisions than using
20
-------------------------^
17
section or page numbers.
21
If parenthesis in noptr-declarator (e.g., 'void (*f(int arg))(double)'):
22
Error in declarator or parameters
23
Invalid C declaration: Expecting "(" in parameters. [error at 39]
24
DECLARE_INSTANCE_CHECKER ( InstanceType, OBJ_NAME, TYPENAME)
25
---------------------------------------^
18
26
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
27
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
28
Reviewed-by: Daniel P. Berrangé <berrange@redhat.com>
21
Message-id: 1503414539-28762-9-git-send-email-peter.maydell@linaro.org
29
Tested-by: Stefan Hajnoczi <stefanha@redhat.com>
30
Message-id: 20201030174700.7204-2-peter.maydell@linaro.org
22
---
31
---
23
target/arm/cpu.h | 14 ++++++++++++--
32
scripts/kernel-doc | 18 +++++++++++++++++-
24
hw/intc/armv7m_nvic.c | 9 ++++++++-
33
1 file changed, 17 insertions(+), 1 deletion(-)
25
target/arm/helper.c | 20 ++++++++++++++++----
26
target/arm/machine.c | 5 +++--
27
4 files changed, 39 insertions(+), 9 deletions(-)
28
34
29
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
35
diff --git a/scripts/kernel-doc b/scripts/kernel-doc
30
index XXXXXXX..XXXXXXX 100644
36
index XXXXXXX..XXXXXXX 100755
31
--- a/target/arm/cpu.h
37
--- a/scripts/kernel-doc
32
+++ b/target/arm/cpu.h
38
+++ b/scripts/kernel-doc
33
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
39
@@ -XXX,XX +XXX,XX @@ sub output_function_rst(%) {
34
unsigned mpu_ctrl; /* MPU_CTRL */
40
    output_highlight_rst($args{'purpose'});
35
int exception;
41
    $start = "\n\n**Syntax**\n\n ``";
36
uint32_t primask[2];
42
} else {
37
- uint32_t faultmask;
43
-    print ".. c:function:: ";
38
+ uint32_t faultmask[2];
44
+ if ((split(/\./, $sphinx_version))[0] >= 3) {
39
uint32_t secure; /* Is CPU in Secure state? (not guest visible) */
45
+ # Sphinx 3 and later distinguish macros and functions and
40
} v7m;
46
+ # complain if you use c:function with something that's not
41
47
+ # syntactically valid as a function declaration.
42
@@ -XXX,XX +XXX,XX @@ void armv7m_nvic_acknowledge_irq(void *opaque);
48
+ # We assume that anything with a return type is a function
43
* (Ignoring -1, this is the same as the RETTOBASE value before completion.)
49
+ # and anything without is a macro.
44
*/
50
+ if ($args{'functiontype'} ne "") {
45
int armv7m_nvic_complete_irq(void *opaque, int irq);
51
+ print ".. c:function:: ";
46
+/**
52
+ } else {
47
+ * armv7m_nvic_raw_execution_priority: return the raw execution priority
53
+ print ".. c:macro:: ";
48
+ * @opaque: the NVIC
49
+ *
50
+ * Returns: the raw execution priority as defined by the v8M architecture.
51
+ * This is the execution priority minus the effects of AIRCR.PRIS,
52
+ * and minus any PRIMASK/FAULTMASK/BASEPRI priority boosting.
53
+ * (v8M ARM ARM I_PKLD.)
54
+ */
55
+int armv7m_nvic_raw_execution_priority(void *opaque);
56
57
/* Interface for defining coprocessor registers.
58
* Registers are defined in tables of arm_cp_reginfo structs
59
@@ -XXX,XX +XXX,XX @@ static inline int cpu_mmu_index(CPUARMState *env, bool ifetch)
60
* we're in a HardFault or NMI handler.
61
*/
62
if ((env->v7m.exception > 0 && env->v7m.exception <= 3)
63
- || env->v7m.faultmask) {
64
+ || env->v7m.faultmask[env->v7m.secure]) {
65
mmu_idx = ARMMMUIdx_MNegPri;
66
}
67
68
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
69
index XXXXXXX..XXXXXXX 100644
70
--- a/hw/intc/armv7m_nvic.c
71
+++ b/hw/intc/armv7m_nvic.c
72
@@ -XXX,XX +XXX,XX @@ static inline int nvic_exec_prio(NVICState *s)
73
CPUARMState *env = &s->cpu->env;
74
int running;
75
76
- if (env->v7m.faultmask) {
77
+ if (env->v7m.faultmask[env->v7m.secure]) {
78
running = -1;
79
} else if (env->v7m.primask[env->v7m.secure]) {
80
running = 0;
81
@@ -XXX,XX +XXX,XX @@ bool armv7m_nvic_can_take_pending_exception(void *opaque)
82
return nvic_exec_prio(s) > nvic_pending_prio(s);
83
}
84
85
+int armv7m_nvic_raw_execution_priority(void *opaque)
86
+{
87
+ NVICState *s = opaque;
88
+
89
+ return s->exception_prio;
90
+}
91
+
92
/* caller must call nvic_irq_update() after this */
93
static void set_prio(NVICState *s, unsigned irq, uint8_t prio)
94
{
95
diff --git a/target/arm/helper.c b/target/arm/helper.c
96
index XXXXXXX..XXXXXXX 100644
97
--- a/target/arm/helper.c
98
+++ b/target/arm/helper.c
99
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
100
}
101
102
if (env->v7m.exception != ARMV7M_EXCP_NMI) {
103
- /* Auto-clear FAULTMASK on return from other than NMI */
104
- env->v7m.faultmask = 0;
105
+ /* Auto-clear FAULTMASK on return from other than NMI.
106
+ * If the security extension is implemented then this only
107
+ * happens if the raw execution priority is >= 0; the
108
+ * value of the ES bit in the exception return value indicates
109
+ * which security state's faultmask to clear. (v8M ARM ARM R_KBNF.)
110
+ */
111
+ if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
112
+ int es = type & 1;
113
+ if (armv7m_nvic_raw_execution_priority(env->nvic) >= 0) {
114
+ env->v7m.faultmask[es] = 0;
115
+ }
54
+ }
116
+ } else {
55
+ } else {
117
+ env->v7m.faultmask[M_REG_NS] = 0;
56
+ # Older Sphinx don't support documenting macros that take
57
+ # arguments with c:macro, and don't complain about the use
58
+ # of c:function for this.
59
+ print ".. c:function:: ";
118
+ }
60
+ }
119
}
61
}
120
62
if ($args{'functiontype'} ne "") {
121
switch (armv7m_nvic_complete_irq(env->nvic, env->v7m.exception)) {
63
    $start .= $args{'functiontype'} . " " . $args{'function'} . " (";
122
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(v7m_mrs)(CPUARMState *env, uint32_t reg)
123
case 18: /* BASEPRI_MAX */
124
return env->v7m.basepri[env->v7m.secure];
125
case 19: /* FAULTMASK */
126
- return env->v7m.faultmask;
127
+ return env->v7m.faultmask[env->v7m.secure];
128
default:
129
qemu_log_mask(LOG_GUEST_ERROR, "Attempt to read unknown special"
130
" register %d\n", reg);
131
@@ -XXX,XX +XXX,XX @@ void HELPER(v7m_msr)(CPUARMState *env, uint32_t maskreg, uint32_t val)
132
}
133
break;
134
case 19: /* FAULTMASK */
135
- env->v7m.faultmask = val & 1;
136
+ env->v7m.faultmask[env->v7m.secure] = val & 1;
137
break;
138
case 20: /* CONTROL */
139
/* Writing to the SPSEL bit only has an effect if we are in
140
diff --git a/target/arm/machine.c b/target/arm/machine.c
141
index XXXXXXX..XXXXXXX 100644
142
--- a/target/arm/machine.c
143
+++ b/target/arm/machine.c
144
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_faultmask_primask = {
145
.version_id = 1,
146
.minimum_version_id = 1,
147
.fields = (VMStateField[]) {
148
- VMSTATE_UINT32(env.v7m.faultmask, ARMCPU),
149
+ VMSTATE_UINT32(env.v7m.faultmask[M_REG_NS], ARMCPU),
150
VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
151
VMSTATE_END_OF_LIST()
152
}
153
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
154
VMSTATE_UINT32(env.v7m.secure, ARMCPU),
155
VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
156
VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
157
+ VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
158
VMSTATE_END_OF_LIST()
159
}
160
};
161
@@ -XXX,XX +XXX,XX @@ static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
162
* transferred using the vmstate_m_faultmask_primask subsection.
163
*/
164
if (val & CPSR_F) {
165
- env->v7m.faultmask = 1;
166
+ env->v7m.faultmask[M_REG_NS] = 1;
167
}
168
if (val & CPSR_I) {
169
env->v7m.primask[M_REG_NS] = 1;
170
--
64
--
171
2.7.4
65
2.20.1
172
66
173
67
diff view generated by jsdifflib
1
Make the MPU registers MPU_MAIR0 and MPU_MAIR1 banked if v8M security
1
Sphinx 3.2 is pickier than earlier versions about the option:: markup,
2
extensions are enabled.
2
and complains about our usage in qemu-option-trace.rst:
3
4
../../docs/qemu-option-trace.rst.inc:4:Malformed option description
5
'[enable=]PATTERN', should look like "opt", "-opt args", "--opt args",
6
"/opt args" or "+opt args"
7
8
In this file, we're really trying to document the different parts of
9
the top-level --trace option, which qemu-nbd.rst and qemu-img.rst
10
have already introduced with an option:: markup. So it's not right
11
to use option:: here anyway. Switch to a different markup
12
(definition lists) which gives about the same formatted output.
13
14
(Unlike option::, this markup doesn't produce index entries; but
15
at the moment we don't do anything much with indexes anyway, and
16
in any case I think it doesn't make much sense to have individual
17
index entries for the sub-parts of the --trace option.)
3
18
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
20
Reviewed-by: Daniel P. Berrangé <berrange@redhat.com>
6
Message-id: 1503414539-28762-13-git-send-email-peter.maydell@linaro.org
21
Tested-by: Stefan Hajnoczi <stefanha@redhat.com>
22
Message-id: 20201030174700.7204-3-peter.maydell@linaro.org
7
---
23
---
8
target/arm/cpu.h | 4 ++--
24
docs/qemu-option-trace.rst.inc | 6 +++---
9
hw/intc/armv7m_nvic.c | 8 ++++----
25
1 file changed, 3 insertions(+), 3 deletions(-)
10
target/arm/cpu.c | 6 ++++--
11
target/arm/machine.c | 6 ++++--
12
4 files changed, 14 insertions(+), 10 deletions(-)
13
26
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
27
diff --git a/docs/qemu-option-trace.rst.inc b/docs/qemu-option-trace.rst.inc
15
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/cpu.h
29
--- a/docs/qemu-option-trace.rst.inc
17
+++ b/target/arm/cpu.h
30
+++ b/docs/qemu-option-trace.rst.inc
18
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
31
@@ -XXX,XX +XXX,XX @@
19
*/
32
20
uint32_t *rbar;
33
Specify tracing options.
21
uint32_t *rlar;
34
22
- uint32_t mair0;
35
-.. option:: [enable=]PATTERN
23
- uint32_t mair1;
36
+``[enable=]PATTERN``
24
+ uint32_t mair0[2];
37
25
+ uint32_t mair1[2];
38
Immediately enable events matching *PATTERN*
26
} pmsav8;
39
(either event name or a globbing pattern). This option is only
27
40
@@ -XXX,XX +XXX,XX @@ Specify tracing options.
28
void *nvic;
41
29
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
42
Use :option:`-trace help` to print a list of names of trace points.
30
index XXXXXXX..XXXXXXX 100644
43
31
--- a/hw/intc/armv7m_nvic.c
44
-.. option:: events=FILE
32
+++ b/hw/intc/armv7m_nvic.c
45
+``events=FILE``
33
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
46
34
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
47
Immediately enable events listed in *FILE*.
35
goto bad_offset;
48
The file must contain one event name (as listed in the ``trace-events-all``
36
}
49
@@ -XXX,XX +XXX,XX @@ Specify tracing options.
37
- return cpu->env.pmsav8.mair0;
50
available if QEMU has been compiled with the ``simple``, ``log`` or
38
+ return cpu->env.pmsav8.mair0[attrs.secure];
51
``ftrace`` tracing backend.
39
case 0xdc4: /* MPU_MAIR1 */
52
40
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
53
-.. option:: file=FILE
41
goto bad_offset;
54
+``file=FILE``
42
}
55
43
- return cpu->env.pmsav8.mair1;
56
Log output traces to *FILE*.
44
+ return cpu->env.pmsav8.mair1[attrs.secure];
57
This option is only available if QEMU has been compiled with
45
default:
46
bad_offset:
47
qemu_log_mask(LOG_GUEST_ERROR, "NVIC: Bad read offset 0x%x\n", offset);
48
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
49
}
50
if (cpu->pmsav7_dregion) {
51
/* Register is RES0 if no MPU regions are implemented */
52
- cpu->env.pmsav8.mair0 = value;
53
+ cpu->env.pmsav8.mair0[attrs.secure] = value;
54
}
55
/* We don't need to do anything else because memory attributes
56
* only affect cacheability, and we don't implement caching.
57
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
58
}
59
if (cpu->pmsav7_dregion) {
60
/* Register is RES0 if no MPU regions are implemented */
61
- cpu->env.pmsav8.mair1 = value;
62
+ cpu->env.pmsav8.mair1[attrs.secure] = value;
63
}
64
/* We don't need to do anything else because memory attributes
65
* only affect cacheability, and we don't implement caching.
66
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
67
index XXXXXXX..XXXXXXX 100644
68
--- a/target/arm/cpu.c
69
+++ b/target/arm/cpu.c
70
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(CPUState *s)
71
}
72
}
73
env->pmsav7.rnr = 0;
74
- env->pmsav8.mair0 = 0;
75
- env->pmsav8.mair1 = 0;
76
+ env->pmsav8.mair0[M_REG_NS] = 0;
77
+ env->pmsav8.mair0[M_REG_S] = 0;
78
+ env->pmsav8.mair1[M_REG_NS] = 0;
79
+ env->pmsav8.mair1[M_REG_S] = 0;
80
}
81
82
set_flush_to_zero(1, &env->vfp.standard_fp_status);
83
diff --git a/target/arm/machine.c b/target/arm/machine.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/target/arm/machine.c
86
+++ b/target/arm/machine.c
87
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_pmsav8 = {
88
vmstate_info_uint32, uint32_t),
89
VMSTATE_VARRAY_UINT32(env.pmsav8.rlar, ARMCPU, pmsav7_dregion, 0,
90
vmstate_info_uint32, uint32_t),
91
- VMSTATE_UINT32(env.pmsav8.mair0, ARMCPU),
92
- VMSTATE_UINT32(env.pmsav8.mair1, ARMCPU),
93
+ VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
94
+ VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
95
VMSTATE_END_OF_LIST()
96
}
97
};
98
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_m_security = {
99
VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
100
VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
101
VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
102
+ VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
103
+ VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
104
VMSTATE_END_OF_LIST()
105
}
106
};
107
--
58
--
108
2.7.4
59
2.20.1
109
60
110
61
diff view generated by jsdifflib
1
Define a new MachineClass field ignore_memory_transaction_failures.
1
The randomness tests in the NPCM7xx RNG test fail intermittently
2
If this is flag is true then the CPU will ignore memory transaction
2
but fairly frequently. On my machine running the test in a loop:
3
failures which should cause the CPU to take an exception due to an
3
while QTEST_QEMU_BINARY=./qemu-system-aarch64 ./tests/qtest/npcm7xx_rng-test; do true; done
4
access to an unassigned physical address; the transaction will
5
instead return zero (for a read) or be ignored (for a write). This
6
should be set only by legacy board models which rely on the old
7
RAZ/WI behaviour for handling devices that QEMU does not yet model.
8
New board models should instead use "unimplemented-device" for all
9
memory ranges where the guest will attempt to probe for a device that
10
QEMU doesn't implement and a stub device is required.
11
4
12
We need this for ARM boards, where we're about to implement support for
5
will fail in less than a minute with an error like:
13
generating external aborts on memory transaction failures. Too many
6
ERROR:../../tests/qtest/npcm7xx_rng-test.c:256:test_first_byte_runs:
14
of our legacy board models rely on the RAZ/WI behaviour and we
7
assertion failed (calc_runs_p(buf.l, sizeof(buf) * BITS_PER_BYTE) > 0.01): (0.00286205989 > 0.01)
15
would break currently working guests when their "probe for device"
8
16
code provoked an external abort rather than a RAZ.
9
(Failures have been observed on all 4 of the randomness tests,
10
not just first_byte_runs.)
11
12
It's not clear why these tests are failing like this, but intermittent
13
failures make CI and merge testing awkward, so disable running them
14
unless a developer specifically sets QEMU_TEST_FLAKY_RNG_TESTS when
15
running the test suite, until we work out the cause.
17
16
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
20
Reviewed-by: Alistair Francis <alistair.francis@xilinx.com>
19
Message-id: 20201102152454.8287-1-peter.maydell@linaro.org
21
Message-id: 1504626814-23124-2-git-send-email-peter.maydell@linaro.org
20
Reviewed-by: Havard Skinnemoen <hskinnemoen@google.com>
22
---
21
---
23
include/hw/boards.h | 11 +++++++++++
22
tests/qtest/npcm7xx_rng-test.c | 14 ++++++++++----
24
include/qom/cpu.h | 7 ++++++-
23
1 file changed, 10 insertions(+), 4 deletions(-)
25
qom/cpu.c | 16 ++++++++++++++++
26
3 files changed, 33 insertions(+), 1 deletion(-)
27
24
28
diff --git a/include/hw/boards.h b/include/hw/boards.h
25
diff --git a/tests/qtest/npcm7xx_rng-test.c b/tests/qtest/npcm7xx_rng-test.c
29
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
30
--- a/include/hw/boards.h
27
--- a/tests/qtest/npcm7xx_rng-test.c
31
+++ b/include/hw/boards.h
28
+++ b/tests/qtest/npcm7xx_rng-test.c
32
@@ -XXX,XX +XXX,XX @@ typedef struct {
29
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
33
* size than the target architecture's minimum. (Attempting to create
30
34
* such a CPU will fail.) Note that changing this is a migration
31
qtest_add_func("npcm7xx_rng/enable_disable", test_enable_disable);
35
* compatibility break for the machine.
32
qtest_add_func("npcm7xx_rng/rosel", test_rosel);
36
+ * @ignore_memory_transaction_failures:
33
- qtest_add_func("npcm7xx_rng/continuous/monobit", test_continuous_monobit);
37
+ * If this is flag is true then the CPU will ignore memory transaction
34
- qtest_add_func("npcm7xx_rng/continuous/runs", test_continuous_runs);
38
+ * failures which should cause the CPU to take an exception due to an
35
- qtest_add_func("npcm7xx_rng/first_byte/monobit", test_first_byte_monobit);
39
+ * access to an unassigned physical address; the transaction will instead
36
- qtest_add_func("npcm7xx_rng/first_byte/runs", test_first_byte_runs);
40
+ * return zero (for a read) or be ignored (for a write). This should be
37
+ /*
41
+ * set only by legacy board models which rely on the old RAZ/WI behaviour
38
+ * These tests fail intermittently; only run them on explicit
42
+ * for handling devices that QEMU does not yet model. New board models
39
+ * request until we figure out why.
43
+ * should instead use "unimplemented-device" for all memory ranges where
44
+ * the guest will attempt to probe for a device that QEMU doesn't
45
+ * implement and a stub device is required.
46
*/
47
struct MachineClass {
48
/*< private >*/
49
@@ -XXX,XX +XXX,XX @@ struct MachineClass {
50
bool rom_file_has_mr;
51
int minimum_page_bits;
52
bool has_hotpluggable_cpus;
53
+ bool ignore_memory_transaction_failures;
54
int numa_mem_align_shift;
55
void (*numa_auto_assign_ram)(MachineClass *mc, NodeInfo *nodes,
56
int nb_nodes, ram_addr_t size);
57
diff --git a/include/qom/cpu.h b/include/qom/cpu.h
58
index XXXXXXX..XXXXXXX 100644
59
--- a/include/qom/cpu.h
60
+++ b/include/qom/cpu.h
61
@@ -XXX,XX +XXX,XX @@ struct qemu_work_item;
62
* @trace_dstate_delayed: Delayed changes to trace_dstate (includes all changes
63
* to @trace_dstate).
64
* @trace_dstate: Dynamic tracing state of events for this vCPU (bitmask).
65
+ * @ignore_memory_transaction_failures: Cached copy of the MachineState
66
+ * flag of the same name: allows the board to suppress calling of the
67
+ * CPU do_transaction_failed hook function.
68
*
69
* State of one CPU core or thread.
70
*/
71
@@ -XXX,XX +XXX,XX @@ struct CPUState {
72
*/
73
bool throttle_thread_scheduled;
74
75
+ bool ignore_memory_transaction_failures;
76
+
77
/* Note that this is accessed at the start of every TB via a negative
78
offset from AREG0. Leave this field at the end so as to make the
79
(absolute value) offset as small as possible. This reduces code
80
@@ -XXX,XX +XXX,XX @@ static inline void cpu_transaction_failed(CPUState *cpu, hwaddr physaddr,
81
{
82
CPUClass *cc = CPU_GET_CLASS(cpu);
83
84
- if (cc->do_transaction_failed) {
85
+ if (!cpu->ignore_memory_transaction_failures && cc->do_transaction_failed) {
86
cc->do_transaction_failed(cpu, physaddr, addr, size, access_type,
87
mmu_idx, attrs, response, retaddr);
88
}
89
diff --git a/qom/cpu.c b/qom/cpu.c
90
index XXXXXXX..XXXXXXX 100644
91
--- a/qom/cpu.c
92
+++ b/qom/cpu.c
93
@@ -XXX,XX +XXX,XX @@
94
#include "exec/cpu-common.h"
95
#include "qemu/error-report.h"
96
#include "sysemu/sysemu.h"
97
+#include "hw/boards.h"
98
#include "hw/qdev-properties.h"
99
#include "trace-root.h"
100
101
@@ -XXX,XX +XXX,XX @@ static void cpu_common_parse_features(const char *typename, char *features,
102
static void cpu_common_realizefn(DeviceState *dev, Error **errp)
103
{
104
CPUState *cpu = CPU(dev);
105
+ Object *machine = qdev_get_machine();
106
+
107
+ /* qdev_get_machine() can return something that's not TYPE_MACHINE
108
+ * if this is one of the user-only emulators; in that case there's
109
+ * no need to check the ignore_memory_transaction_failures board flag.
110
+ */
40
+ */
111
+ if (object_dynamic_cast(machine, TYPE_MACHINE)) {
41
+ if (getenv("QEMU_TEST_FLAKY_RNG_TESTS")) {
112
+ ObjectClass *oc = object_get_class(machine);
42
+ qtest_add_func("npcm7xx_rng/continuous/monobit", test_continuous_monobit);
113
+ MachineClass *mc = MACHINE_CLASS(oc);
43
+ qtest_add_func("npcm7xx_rng/continuous/runs", test_continuous_runs);
114
+
44
+ qtest_add_func("npcm7xx_rng/first_byte/monobit", test_first_byte_monobit);
115
+ if (mc) {
45
+ qtest_add_func("npcm7xx_rng/first_byte/runs", test_first_byte_runs);
116
+ cpu->ignore_memory_transaction_failures =
117
+ mc->ignore_memory_transaction_failures;
118
+ }
119
+ }
46
+ }
120
47
121
if (dev->hotplugged) {
48
qtest_start("-machine npcm750-evb");
122
cpu_synchronize_post_init(cpu);
49
ret = g_test_run();
123
--
50
--
124
2.7.4
51
2.20.1
125
52
126
53
diff view generated by jsdifflib