1
Another go at the v8.5-MemTag linux-user support, plus a
1
I don't have anything else queued up at the moment, so this is just
2
couple more npcm7xx devices.
2
Richard's SME patches.
3
3
4
-- PMM
4
-- PMM
5
5
6
The following changes since commit 8ba4bca570ace1e60614a0808631a517cf5df67a:
6
The following changes since commit 63b38f6c85acd312c2cab68554abf33adf4ee2b3:
7
7
8
Merge remote-tracking branch 'remotes/kevin/tags/for-upstream' into staging (2021-02-15 17:13:57 +0000)
8
Merge tag 'pull-target-arm-20220707' of https://git.linaro.org/people/pmaydell/qemu-arm into staging (2022-07-08 06:17:11 +0530)
9
9
10
are available in the Git repository at:
10
are available in the Git repository at:
11
11
12
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20210216
12
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20220711
13
13
14
for you to fetch changes up to 64fd5bddf3b71d1b92b55382ab39768bd87ecfbd:
14
for you to fetch changes up to f9982ceaf26df27d15547a3a7990a95019e9e3a8:
15
15
16
tests/qtests: Add npcm7xx emc model test (2021-02-16 14:27:05 +0000)
16
linux-user/aarch64: Add SME related hwcap entries (2022-07-11 13:43:52 +0100)
17
17
18
----------------------------------------------------------------
18
----------------------------------------------------------------
19
target-arm queue:
19
target-arm:
20
* Support ARMv8.5-MemTag for linux-user
20
* Implement SME emulation, for both system and linux-user
21
* ncpm7xx: Support SMBus, EMC ethernet devices
22
* MAINTAINERS: add section for Clock framework
23
21
24
----------------------------------------------------------------
22
----------------------------------------------------------------
25
Doug Evans (3):
23
Richard Henderson (45):
26
hw/net: Add npcm7xx emc model
24
target/arm: Handle SME in aarch64_cpu_dump_state
27
hw/arm: Add npcm7xx emc model
25
target/arm: Add infrastructure for disas_sme
28
tests/qtests: Add npcm7xx emc model test
26
target/arm: Trap non-streaming usage when Streaming SVE is active
27
target/arm: Mark ADR as non-streaming
28
target/arm: Mark RDFFR, WRFFR, SETFFR as non-streaming
29
target/arm: Mark BDEP, BEXT, BGRP, COMPACT, FEXPA, FTSSEL as non-streaming
30
target/arm: Mark PMULL, FMMLA as non-streaming
31
target/arm: Mark FTSMUL, FTMAD, FADDA as non-streaming
32
target/arm: Mark SMMLA, UMMLA, USMMLA as non-streaming
33
target/arm: Mark string/histo/crypto as non-streaming
34
target/arm: Mark gather/scatter load/store as non-streaming
35
target/arm: Mark gather prefetch as non-streaming
36
target/arm: Mark LDFF1 and LDNF1 as non-streaming
37
target/arm: Mark LD1RO as non-streaming
38
target/arm: Add SME enablement checks
39
target/arm: Handle SME in sve_access_check
40
target/arm: Implement SME RDSVL, ADDSVL, ADDSPL
41
target/arm: Implement SME ZERO
42
target/arm: Implement SME MOVA
43
target/arm: Implement SME LD1, ST1
44
target/arm: Export unpredicated ld/st from translate-sve.c
45
target/arm: Implement SME LDR, STR
46
target/arm: Implement SME ADDHA, ADDVA
47
target/arm: Implement FMOPA, FMOPS (non-widening)
48
target/arm: Implement BFMOPA, BFMOPS
49
target/arm: Implement FMOPA, FMOPS (widening)
50
target/arm: Implement SME integer outer product
51
target/arm: Implement PSEL
52
target/arm: Implement REVD
53
target/arm: Implement SCLAMP, UCLAMP
54
target/arm: Reset streaming sve state on exception boundaries
55
target/arm: Enable SME for -cpu max
56
linux-user/aarch64: Clear tpidr2_el0 if CLONE_SETTLS
57
linux-user/aarch64: Reset PSTATE.SM on syscalls
58
linux-user/aarch64: Add SM bit to SVE signal context
59
linux-user/aarch64: Tidy target_restore_sigframe error return
60
linux-user/aarch64: Do not allow duplicate or short sve records
61
linux-user/aarch64: Verify extra record lock succeeded
62
linux-user/aarch64: Move sve record checks into restore
63
linux-user/aarch64: Implement SME signal handling
64
linux-user: Rename sve prctls
65
linux-user/aarch64: Implement PR_SME_GET_VL, PR_SME_SET_VL
66
target/arm: Only set ZEN in reset if SVE present
67
target/arm: Enable SME for user-only
68
linux-user/aarch64: Add SME related hwcap entries
29
69
30
Hao Wu (5):
70
docs/system/arm/emulation.rst | 4 +
31
hw/i2c: Implement NPCM7XX SMBus Module Single Mode
71
linux-user/aarch64/target_cpu.h | 5 +-
32
hw/arm: Add I2C sensors for NPCM750 eval board
72
linux-user/aarch64/target_prctl.h | 62 +-
33
hw/arm: Add I2C sensors and EEPROM for GSJ machine
73
target/arm/cpu.h | 7 +
34
hw/i2c: Add a QTest for NPCM7XX SMBus Device
74
target/arm/helper-sme.h | 126 ++++
35
hw/i2c: Implement NPCM7XX SMBus Module FIFO Mode
75
target/arm/helper-sve.h | 4 +
36
76
target/arm/helper.h | 18 +
37
Luc Michel (1):
77
target/arm/translate-a64.h | 45 ++
38
MAINTAINERS: add myself maintainer for the clock framework
78
target/arm/translate.h | 16 +
39
79
target/arm/sme-fa64.decode | 60 ++
40
Richard Henderson (31):
80
target/arm/sme.decode | 88 +++
41
tcg: Introduce target-specific page data for user-only
81
target/arm/sve.decode | 41 +-
42
linux-user: Introduce PAGE_ANON
82
linux-user/aarch64/cpu_loop.c | 9 +
43
exec: Use uintptr_t for guest_base
83
linux-user/aarch64/signal.c | 243 ++++++--
44
exec: Use uintptr_t in cpu_ldst.h
84
linux-user/elfload.c | 20 +
45
exec: Improve types for guest_addr_valid
85
linux-user/syscall.c | 28 +-
46
linux-user: Check for overflow in access_ok
86
target/arm/cpu.c | 35 +-
47
linux-user: Tidy VERIFY_READ/VERIFY_WRITE
87
target/arm/cpu64.c | 11 +
48
bsd-user: Tidy VERIFY_READ/VERIFY_WRITE
88
target/arm/helper.c | 56 +-
49
linux-user: Do not use guest_addr_valid for h2g_valid
89
target/arm/sme_helper.c | 1140 +++++++++++++++++++++++++++++++++++++
50
linux-user: Fix guest_addr_valid vs reserved_va
90
target/arm/sve_helper.c | 28 +
51
exec: Introduce cpu_untagged_addr
91
target/arm/translate-a64.c | 103 +++-
52
exec: Use cpu_untagged_addr in g2h; split out g2h_untagged
92
target/arm/translate-sme.c | 373 ++++++++++++
53
linux-user: Explicitly untag memory management syscalls
93
target/arm/translate-sve.c | 393 ++++++++++---
54
linux-user: Use guest_range_valid in access_ok
94
target/arm/translate-vfp.c | 12 +
55
exec: Rename guest_{addr,range}_valid to *_untagged
95
target/arm/translate.c | 2 +
56
linux-user: Use cpu_untagged_addr in access_ok; split out *_untagged
96
target/arm/vec_helper.c | 24 +
57
linux-user: Move lock_user et al out of line
97
target/arm/meson.build | 3 +
58
linux-user: Fix types in uaccess.c
98
28 files changed, 2821 insertions(+), 135 deletions(-)
59
linux-user: Handle tags in lock_user/unlock_user
99
create mode 100644 target/arm/sme-fa64.decode
60
linux-user/aarch64: Implement PR_TAGGED_ADDR_ENABLE
100
create mode 100644 target/arm/sme.decode
61
target/arm: Improve gen_top_byte_ignore
101
create mode 100644 target/arm/translate-sme.c
62
target/arm: Use the proper TBI settings for linux-user
63
linux-user/aarch64: Implement PR_MTE_TCF and PR_MTE_TAG
64
linux-user/aarch64: Implement PROT_MTE
65
target/arm: Split out syndrome.h from internals.h
66
linux-user/aarch64: Pass syndrome to EXC_*_ABORT
67
linux-user/aarch64: Signal SEGV_MTESERR for sync tag check fault
68
linux-user/aarch64: Signal SEGV_MTEAERR for async tag check error
69
target/arm: Add allocation tag storage for user mode
70
target/arm: Enable MTE for user-only
71
tests/tcg/aarch64: Add mte smoke tests
72
73
docs/system/arm/nuvoton.rst | 5 +-
74
bsd-user/qemu.h | 17 +-
75
include/exec/cpu-all.h | 47 +-
76
include/exec/cpu_ldst.h | 39 +-
77
include/exec/exec-all.h | 2 +-
78
include/hw/arm/npcm7xx.h | 4 +
79
include/hw/i2c/npcm7xx_smbus.h | 113 ++++
80
include/hw/net/npcm7xx_emc.h | 286 +++++++++
81
linux-user/aarch64/target_signal.h | 3 +
82
linux-user/aarch64/target_syscall.h | 13 +
83
linux-user/qemu.h | 76 +--
84
linux-user/syscall_defs.h | 1 +
85
target/arm/cpu-param.h | 3 +
86
target/arm/cpu.h | 32 +
87
target/arm/internals.h | 249 +-------
88
target/arm/syndrome.h | 273 +++++++++
89
tests/tcg/aarch64/mte.h | 60 ++
90
accel/tcg/translate-all.c | 32 +-
91
accel/tcg/user-exec.c | 51 +-
92
bsd-user/elfload.c | 2 +-
93
bsd-user/main.c | 8 +-
94
bsd-user/mmap.c | 23 +-
95
hw/arm/npcm7xx.c | 118 +++-
96
hw/arm/npcm7xx_boards.c | 46 ++
97
hw/i2c/npcm7xx_smbus.c | 1099 +++++++++++++++++++++++++++++++++++
98
hw/net/npcm7xx_emc.c | 857 +++++++++++++++++++++++++++
99
linux-user/aarch64/cpu_loop.c | 38 +-
100
linux-user/elfload.c | 18 +-
101
linux-user/flatload.c | 2 +-
102
linux-user/hppa/cpu_loop.c | 39 +-
103
linux-user/i386/cpu_loop.c | 6 +-
104
linux-user/i386/signal.c | 5 +-
105
linux-user/main.c | 4 +-
106
linux-user/mmap.c | 88 +--
107
linux-user/ppc/signal.c | 4 +-
108
linux-user/syscall.c | 165 ++++--
109
linux-user/uaccess.c | 82 ++-
110
target/arm/cpu.c | 25 +-
111
target/arm/helper-a64.c | 4 +-
112
target/arm/mte_helper.c | 39 +-
113
target/arm/tlb_helper.c | 15 +-
114
target/arm/translate-a64.c | 25 +-
115
target/hppa/op_helper.c | 2 +-
116
target/i386/tcg/mem_helper.c | 2 +-
117
target/s390x/mem_helper.c | 4 +-
118
tests/qtest/npcm7xx_emc-test.c | 862 +++++++++++++++++++++++++++
119
tests/qtest/npcm7xx_smbus-test.c | 495 ++++++++++++++++
120
tests/tcg/aarch64/mte-1.c | 28 +
121
tests/tcg/aarch64/mte-2.c | 45 ++
122
tests/tcg/aarch64/mte-3.c | 51 ++
123
tests/tcg/aarch64/mte-4.c | 45 ++
124
tests/tcg/aarch64/pauth-2.c | 1 -
125
MAINTAINERS | 11 +
126
hw/arm/Kconfig | 1 +
127
hw/i2c/meson.build | 1 +
128
hw/i2c/trace-events | 12 +
129
hw/net/meson.build | 1 +
130
hw/net/trace-events | 17 +
131
tests/qtest/meson.build | 2 +
132
tests/tcg/aarch64/Makefile.target | 6 +
133
tests/tcg/configure.sh | 4 +
134
61 files changed, 5052 insertions(+), 556 deletions(-)
135
create mode 100644 include/hw/i2c/npcm7xx_smbus.h
136
create mode 100644 include/hw/net/npcm7xx_emc.h
137
create mode 100644 target/arm/syndrome.h
138
create mode 100644 tests/tcg/aarch64/mte.h
139
create mode 100644 hw/i2c/npcm7xx_smbus.c
140
create mode 100644 hw/net/npcm7xx_emc.c
141
create mode 100644 tests/qtest/npcm7xx_emc-test.c
142
create mode 100644 tests/qtest/npcm7xx_smbus-test.c
143
create mode 100644 tests/tcg/aarch64/mte-1.c
144
create mode 100644 tests/tcg/aarch64/mte-2.c
145
create mode 100644 tests/tcg/aarch64/mte-3.c
146
create mode 100644 tests/tcg/aarch64/mte-4.c
147
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Dump SVCR, plus use the correct access check for Streaming Mode.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20220708151540.18136-2-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/cpu.c | 17 ++++++++++++++++-
11
1 file changed, 16 insertions(+), 1 deletion(-)
12
13
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
14
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.c
16
+++ b/target/arm/cpu.c
17
@@ -XXX,XX +XXX,XX @@ static void aarch64_cpu_dump_state(CPUState *cs, FILE *f, int flags)
18
int i;
19
int el = arm_current_el(env);
20
const char *ns_status;
21
+ bool sve;
22
23
qemu_fprintf(f, " PC=%016" PRIx64 " ", env->pc);
24
for (i = 0; i < 32; i++) {
25
@@ -XXX,XX +XXX,XX @@ static void aarch64_cpu_dump_state(CPUState *cs, FILE *f, int flags)
26
el,
27
psr & PSTATE_SP ? 'h' : 't');
28
29
+ if (cpu_isar_feature(aa64_sme, cpu)) {
30
+ qemu_fprintf(f, " SVCR=%08" PRIx64 " %c%c",
31
+ env->svcr,
32
+ (FIELD_EX64(env->svcr, SVCR, ZA) ? 'Z' : '-'),
33
+ (FIELD_EX64(env->svcr, SVCR, SM) ? 'S' : '-'));
34
+ }
35
if (cpu_isar_feature(aa64_bti, cpu)) {
36
qemu_fprintf(f, " BTYPE=%d", (psr & PSTATE_BTYPE) >> 10);
37
}
38
@@ -XXX,XX +XXX,XX @@ static void aarch64_cpu_dump_state(CPUState *cs, FILE *f, int flags)
39
qemu_fprintf(f, " FPCR=%08x FPSR=%08x\n",
40
vfp_get_fpcr(env), vfp_get_fpsr(env));
41
42
- if (cpu_isar_feature(aa64_sve, cpu) && sve_exception_el(env, el) == 0) {
43
+ if (cpu_isar_feature(aa64_sme, cpu) && FIELD_EX64(env->svcr, SVCR, SM)) {
44
+ sve = sme_exception_el(env, el) == 0;
45
+ } else if (cpu_isar_feature(aa64_sve, cpu)) {
46
+ sve = sve_exception_el(env, el) == 0;
47
+ } else {
48
+ sve = false;
49
+ }
50
+
51
+ if (sve) {
52
int j, zcr_len = sve_vqm1_for_el(env, el);
53
54
for (i = 0; i <= FFR_PRED_NUM; i++) {
55
--
56
2.25.1
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
This includes the build rules for the decoder, and the
4
new file for translation, but excludes any instructions.
2
5
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20210212184902.1251044-32-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-3-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
10
---
8
tests/tcg/aarch64/mte.h | 60 +++++++++++++++++++++++++++++++
11
target/arm/translate-a64.h | 1 +
9
tests/tcg/aarch64/mte-1.c | 28 +++++++++++++++
12
target/arm/sme.decode | 20 ++++++++++++++++++++
10
tests/tcg/aarch64/mte-2.c | 45 +++++++++++++++++++++++
13
target/arm/translate-a64.c | 7 ++++++-
11
tests/tcg/aarch64/mte-3.c | 51 ++++++++++++++++++++++++++
14
target/arm/translate-sme.c | 35 +++++++++++++++++++++++++++++++++++
12
tests/tcg/aarch64/mte-4.c | 45 +++++++++++++++++++++++
15
target/arm/meson.build | 2 ++
13
tests/tcg/aarch64/Makefile.target | 6 ++++
16
5 files changed, 64 insertions(+), 1 deletion(-)
14
tests/tcg/configure.sh | 4 +++
17
create mode 100644 target/arm/sme.decode
15
7 files changed, 239 insertions(+)
18
create mode 100644 target/arm/translate-sme.c
16
create mode 100644 tests/tcg/aarch64/mte.h
17
create mode 100644 tests/tcg/aarch64/mte-1.c
18
create mode 100644 tests/tcg/aarch64/mte-2.c
19
create mode 100644 tests/tcg/aarch64/mte-3.c
20
create mode 100644 tests/tcg/aarch64/mte-4.c
21
19
22
diff --git a/tests/tcg/aarch64/mte.h b/tests/tcg/aarch64/mte.h
20
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
21
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/translate-a64.h
23
+++ b/target/arm/translate-a64.h
24
@@ -XXX,XX +XXX,XX @@ static inline int pred_gvec_reg_size(DisasContext *s)
25
}
26
27
bool disas_sve(DisasContext *, uint32_t);
28
+bool disas_sme(DisasContext *, uint32_t);
29
30
void gen_gvec_rax1(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
31
uint32_t rm_ofs, uint32_t opr_sz, uint32_t max_sz);
32
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
23
new file mode 100644
33
new file mode 100644
24
index XXXXXXX..XXXXXXX
34
index XXXXXXX..XXXXXXX
25
--- /dev/null
35
--- /dev/null
26
+++ b/tests/tcg/aarch64/mte.h
36
+++ b/target/arm/sme.decode
27
@@ -XXX,XX +XXX,XX @@
37
@@ -XXX,XX +XXX,XX @@
28
+/*
38
+# AArch64 SME instruction descriptions
29
+ * Linux kernel fallback API definitions for MTE and test helpers.
39
+#
30
+ *
40
+# Copyright (c) 2022 Linaro, Ltd
31
+ * Copyright (c) 2021 Linaro Ltd
41
+#
32
+ * SPDX-License-Identifier: GPL-2.0-or-later
42
+# This library is free software; you can redistribute it and/or
33
+ */
43
+# modify it under the terms of the GNU Lesser General Public
44
+# License as published by the Free Software Foundation; either
45
+# version 2.1 of the License, or (at your option) any later version.
46
+#
47
+# This library is distributed in the hope that it will be useful,
48
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
49
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
50
+# Lesser General Public License for more details.
51
+#
52
+# You should have received a copy of the GNU Lesser General Public
53
+# License along with this library; if not, see <http://www.gnu.org/licenses/>.
34
+
54
+
35
+#include <assert.h>
55
+#
36
+#include <string.h>
56
+# This file is processed by scripts/decodetree.py
37
+#include <stdlib.h>
57
+#
38
+#include <stdio.h>
58
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
39
+#include <unistd.h>
59
index XXXXXXX..XXXXXXX 100644
40
+#include <signal.h>
60
--- a/target/arm/translate-a64.c
41
+#include <sys/mman.h>
61
+++ b/target/arm/translate-a64.c
42
+#include <sys/prctl.h>
62
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
43
+
63
}
44
+#ifndef PR_SET_TAGGED_ADDR_CTRL
64
45
+# define PR_SET_TAGGED_ADDR_CTRL 55
65
switch (extract32(insn, 25, 4)) {
46
+#endif
66
- case 0x0: case 0x1: case 0x3: /* UNALLOCATED */
47
+#ifndef PR_TAGGED_ADDR_ENABLE
67
+ case 0x0:
48
+# define PR_TAGGED_ADDR_ENABLE (1UL << 0)
68
+ if (!extract32(insn, 31, 1) || !disas_sme(s, insn)) {
49
+#endif
69
+ unallocated_encoding(s);
50
+#ifndef PR_MTE_TCF_SHIFT
70
+ }
51
+# define PR_MTE_TCF_SHIFT 1
71
+ break;
52
+# define PR_MTE_TCF_NONE (0UL << PR_MTE_TCF_SHIFT)
72
+ case 0x1: case 0x3: /* UNALLOCATED */
53
+# define PR_MTE_TCF_SYNC (1UL << PR_MTE_TCF_SHIFT)
73
unallocated_encoding(s);
54
+# define PR_MTE_TCF_ASYNC (2UL << PR_MTE_TCF_SHIFT)
74
break;
55
+# define PR_MTE_TAG_SHIFT 3
75
case 0x2:
56
+#endif
76
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
57
+
58
+#ifndef PROT_MTE
59
+# define PROT_MTE 0x20
60
+#endif
61
+
62
+#ifndef SEGV_MTEAERR
63
+# define SEGV_MTEAERR 8
64
+# define SEGV_MTESERR 9
65
+#endif
66
+
67
+static void enable_mte(int tcf)
68
+{
69
+ int r = prctl(PR_SET_TAGGED_ADDR_CTRL,
70
+ PR_TAGGED_ADDR_ENABLE | tcf | (0xfffe << PR_MTE_TAG_SHIFT),
71
+ 0, 0, 0);
72
+ if (r < 0) {
73
+ perror("PR_SET_TAGGED_ADDR_CTRL");
74
+ exit(2);
75
+ }
76
+}
77
+
78
+static void *alloc_mte_mem(size_t size)
79
+{
80
+ void *p = mmap(NULL, size, PROT_READ | PROT_WRITE | PROT_MTE,
81
+ MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
82
+ if (p == MAP_FAILED) {
83
+ perror("mmap PROT_MTE");
84
+ exit(2);
85
+ }
86
+ return p;
87
+}
88
diff --git a/tests/tcg/aarch64/mte-1.c b/tests/tcg/aarch64/mte-1.c
89
new file mode 100644
77
new file mode 100644
90
index XXXXXXX..XXXXXXX
78
index XXXXXXX..XXXXXXX
91
--- /dev/null
79
--- /dev/null
92
+++ b/tests/tcg/aarch64/mte-1.c
80
+++ b/target/arm/translate-sme.c
93
@@ -XXX,XX +XXX,XX @@
81
@@ -XXX,XX +XXX,XX @@
94
+/*
82
+/*
95
+ * Memory tagging, basic pass cases.
83
+ * AArch64 SME translation
96
+ *
84
+ *
97
+ * Copyright (c) 2021 Linaro Ltd
85
+ * Copyright (c) 2022 Linaro, Ltd
98
+ * SPDX-License-Identifier: GPL-2.0-or-later
86
+ *
87
+ * This library is free software; you can redistribute it and/or
88
+ * modify it under the terms of the GNU Lesser General Public
89
+ * License as published by the Free Software Foundation; either
90
+ * version 2.1 of the License, or (at your option) any later version.
91
+ *
92
+ * This library is distributed in the hope that it will be useful,
93
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
94
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
95
+ * Lesser General Public License for more details.
96
+ *
97
+ * You should have received a copy of the GNU Lesser General Public
98
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
99
+ */
99
+ */
100
+
100
+
101
+#include "mte.h"
101
+#include "qemu/osdep.h"
102
+#include "cpu.h"
103
+#include "tcg/tcg-op.h"
104
+#include "tcg/tcg-op-gvec.h"
105
+#include "tcg/tcg-gvec-desc.h"
106
+#include "translate.h"
107
+#include "exec/helper-gen.h"
108
+#include "translate-a64.h"
109
+#include "fpu/softfloat.h"
102
+
110
+
103
+int main(int ac, char **av)
104
+{
105
+ int *p0, *p1, *p2;
106
+ long c;
107
+
111
+
108
+ enable_mte(PR_MTE_TCF_NONE);
109
+ p0 = alloc_mte_mem(sizeof(*p0));
110
+
111
+ asm("irg %0,%1,%2" : "=r"(p1) : "r"(p0), "r"(1));
112
+ assert(p1 != p0);
113
+ asm("subp %0,%1,%2" : "=r"(c) : "r"(p0), "r"(p1));
114
+ assert(c == 0);
115
+
116
+ asm("stg %0, [%0]" : : "r"(p1));
117
+ asm("ldg %0, [%1]" : "=r"(p2) : "r"(p0), "0"(p0));
118
+ assert(p1 == p2);
119
+
120
+ return 0;
121
+}
122
diff --git a/tests/tcg/aarch64/mte-2.c b/tests/tcg/aarch64/mte-2.c
123
new file mode 100644
124
index XXXXXXX..XXXXXXX
125
--- /dev/null
126
+++ b/tests/tcg/aarch64/mte-2.c
127
@@ -XXX,XX +XXX,XX @@
128
+/*
112
+/*
129
+ * Memory tagging, basic fail cases, synchronous signals.
113
+ * Include the generated decoder.
130
+ *
131
+ * Copyright (c) 2021 Linaro Ltd
132
+ * SPDX-License-Identifier: GPL-2.0-or-later
133
+ */
114
+ */
134
+
115
+
135
+#include "mte.h"
116
+#include "decode-sme.c.inc"
136
+
117
diff --git a/target/arm/meson.build b/target/arm/meson.build
137
+void pass(int sig, siginfo_t *info, void *uc)
118
index XXXXXXX..XXXXXXX 100644
138
+{
119
--- a/target/arm/meson.build
139
+ assert(info->si_code == SEGV_MTESERR);
120
+++ b/target/arm/meson.build
140
+ exit(0);
141
+}
142
+
143
+int main(int ac, char **av)
144
+{
145
+ struct sigaction sa;
146
+ int *p0, *p1, *p2;
147
+ long excl = 1;
148
+
149
+ enable_mte(PR_MTE_TCF_SYNC);
150
+ p0 = alloc_mte_mem(sizeof(*p0));
151
+
152
+ /* Create two differently tagged pointers. */
153
+ asm("irg %0,%1,%2" : "=r"(p1) : "r"(p0), "r"(excl));
154
+ asm("gmi %0,%1,%0" : "+r"(excl) : "r" (p1));
155
+ assert(excl != 1);
156
+ asm("irg %0,%1,%2" : "=r"(p2) : "r"(p0), "r"(excl));
157
+ assert(p1 != p2);
158
+
159
+ /* Store the tag from the first pointer. */
160
+ asm("stg %0, [%0]" : : "r"(p1));
161
+
162
+ *p1 = 0;
163
+
164
+ memset(&sa, 0, sizeof(sa));
165
+ sa.sa_sigaction = pass;
166
+ sa.sa_flags = SA_SIGINFO;
167
+ sigaction(SIGSEGV, &sa, NULL);
168
+
169
+ *p2 = 0;
170
+
171
+ abort();
172
+}
173
diff --git a/tests/tcg/aarch64/mte-3.c b/tests/tcg/aarch64/mte-3.c
174
new file mode 100644
175
index XXXXXXX..XXXXXXX
176
--- /dev/null
177
+++ b/tests/tcg/aarch64/mte-3.c
178
@@ -XXX,XX +XXX,XX @@
121
@@ -XXX,XX +XXX,XX @@
179
+/*
122
gen = [
180
+ * Memory tagging, basic fail cases, asynchronous signals.
123
decodetree.process('sve.decode', extra_args: '--decode=disas_sve'),
181
+ *
124
+ decodetree.process('sme.decode', extra_args: '--decode=disas_sme'),
182
+ * Copyright (c) 2021 Linaro Ltd
125
decodetree.process('neon-shared.decode', extra_args: '--decode=disas_neon_shared'),
183
+ * SPDX-License-Identifier: GPL-2.0-or-later
126
decodetree.process('neon-dp.decode', extra_args: '--decode=disas_neon_dp'),
184
+ */
127
decodetree.process('neon-ls.decode', extra_args: '--decode=disas_neon_ls'),
185
+
128
@@ -XXX,XX +XXX,XX @@ arm_ss.add(when: 'TARGET_AARCH64', if_true: files(
186
+#include "mte.h"
129
'sme_helper.c',
187
+
130
'translate-a64.c',
188
+void pass(int sig, siginfo_t *info, void *uc)
131
'translate-sve.c',
189
+{
132
+ 'translate-sme.c',
190
+ assert(info->si_code == SEGV_MTEAERR);
133
))
191
+ exit(0);
134
192
+}
135
arm_softmmu_ss = ss.source_set()
193
+
194
+int main(int ac, char **av)
195
+{
196
+ struct sigaction sa;
197
+ long *p0, *p1, *p2;
198
+ long excl = 1;
199
+
200
+ enable_mte(PR_MTE_TCF_ASYNC);
201
+ p0 = alloc_mte_mem(sizeof(*p0));
202
+
203
+ /* Create two differently tagged pointers. */
204
+ asm("irg %0,%1,%2" : "=r"(p1) : "r"(p0), "r"(excl));
205
+ asm("gmi %0,%1,%0" : "+r"(excl) : "r" (p1));
206
+ assert(excl != 1);
207
+ asm("irg %0,%1,%2" : "=r"(p2) : "r"(p0), "r"(excl));
208
+ assert(p1 != p2);
209
+
210
+ /* Store the tag from the first pointer. */
211
+ asm("stg %0, [%0]" : : "r"(p1));
212
+
213
+ *p1 = 0;
214
+
215
+ memset(&sa, 0, sizeof(sa));
216
+ sa.sa_sigaction = pass;
217
+ sa.sa_flags = SA_SIGINFO;
218
+ sigaction(SIGSEGV, &sa, NULL);
219
+
220
+ /*
221
+ * Signal for async error will happen eventually.
222
+ * For a real kernel this should be after the next IRQ (e.g. timer).
223
+ * For qemu linux-user, we kick the cpu and exit at the next TB.
224
+ * In either case, loop until this happens (or killed by timeout).
225
+ * For extra sauce, yield, producing EXCP_YIELD to cpu_loop().
226
+ */
227
+ asm("str %0, [%0]; yield" : : "r"(p2));
228
+ while (1);
229
+}
230
diff --git a/tests/tcg/aarch64/mte-4.c b/tests/tcg/aarch64/mte-4.c
231
new file mode 100644
232
index XXXXXXX..XXXXXXX
233
--- /dev/null
234
+++ b/tests/tcg/aarch64/mte-4.c
235
@@ -XXX,XX +XXX,XX @@
236
+/*
237
+ * Memory tagging, re-reading tag checks.
238
+ *
239
+ * Copyright (c) 2021 Linaro Ltd
240
+ * SPDX-License-Identifier: GPL-2.0-or-later
241
+ */
242
+
243
+#include "mte.h"
244
+
245
+void __attribute__((noinline)) tagset(void *p, size_t size)
246
+{
247
+ size_t i;
248
+ for (i = 0; i < size; i += 16) {
249
+ asm("stg %0, [%0]" : : "r"(p + i));
250
+ }
251
+}
252
+
253
+void __attribute__((noinline)) tagcheck(void *p, size_t size)
254
+{
255
+ size_t i;
256
+ void *c;
257
+
258
+ for (i = 0; i < size; i += 16) {
259
+ asm("ldg %0, [%1]" : "=r"(c) : "r"(p + i), "0"(p));
260
+ assert(c == p);
261
+ }
262
+}
263
+
264
+int main(int ac, char **av)
265
+{
266
+ size_t size = getpagesize() * 4;
267
+ long excl = 1;
268
+ int *p0, *p1;
269
+
270
+ enable_mte(PR_MTE_TCF_ASYNC);
271
+ p0 = alloc_mte_mem(size);
272
+
273
+ /* Tag the pointer. */
274
+ asm("irg %0,%1,%2" : "=r"(p1) : "r"(p0), "r"(excl));
275
+
276
+ tagset(p1, size);
277
+ tagcheck(p1, size);
278
+
279
+ return 0;
280
+}
281
diff --git a/tests/tcg/aarch64/Makefile.target b/tests/tcg/aarch64/Makefile.target
282
index XXXXXXX..XXXXXXX 100644
283
--- a/tests/tcg/aarch64/Makefile.target
284
+++ b/tests/tcg/aarch64/Makefile.target
285
@@ -XXX,XX +XXX,XX @@ endif
286
# bti-2 tests PROT_BTI, so no special compiler support required.
287
AARCH64_TESTS += bti-2
288
289
+# MTE Tests
290
+ifneq ($(DOCKER_IMAGE)$(CROSS_CC_HAS_ARMV8_MTE),)
291
+AARCH64_TESTS += mte-1 mte-2 mte-3 mte-4
292
+mte-%: CFLAGS += -march=armv8.5-a+memtag
293
+endif
294
+
295
# Semihosting smoke test for linux-user
296
AARCH64_TESTS += semihosting
297
run-semihosting: semihosting
298
diff --git a/tests/tcg/configure.sh b/tests/tcg/configure.sh
299
index XXXXXXX..XXXXXXX 100755
300
--- a/tests/tcg/configure.sh
301
+++ b/tests/tcg/configure.sh
302
@@ -XXX,XX +XXX,XX @@ for target in $target_list; do
303
-mbranch-protection=standard -o $TMPE $TMPC; then
304
echo "CROSS_CC_HAS_ARMV8_BTI=y" >> $config_target_mak
305
fi
306
+ if do_compiler "$target_compiler" $target_compiler_cflags \
307
+ -march=armv8.5-a+memtag -o $TMPE $TMPC; then
308
+ echo "CROSS_CC_HAS_ARMV8_MTE=y" >> $config_target_mak
309
+ fi
310
;;
311
esac
312
313
--
136
--
314
2.20.1
137
2.25.1
315
316
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Remember the PROT_MTE bit as PAGE_MTE/PAGE_TARGET_2.
3
This new behaviour is in the ARM pseudocode function
4
Otherwise this does not yet have effect.
4
AArch64.CheckFPAdvSIMDEnabled, which applies to AArch32
5
via AArch32.CheckAdvSIMDOrFPEnabled when the EL to which
6
the trap would be delivered is in AArch64 mode.
7
8
Given that ARMv9 drops support for AArch32 outside EL0, the trap EL
9
detection ought to be trivially true, but the pseudocode still contains
10
a number of conditions, and QEMU has not yet committed to dropping A32
11
support for EL[12] when v9 features are present.
12
13
Since the computation of SME_TRAP_NONSTREAMING is necessarily different
14
for the two modes, we might as well preserve bits within TBFLAG_ANY and
15
allocate separate bits within TBFLAG_A32 and TBFLAG_A64 instead.
16
17
Note that DDI0616A.a has typos for bits [22:21] of LD1RO in the table
18
of instructions illegal in streaming mode.
5
19
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
20
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
21
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-25-richard.henderson@linaro.org
22
Message-id: 20220708151540.18136-4-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
23
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
24
---
11
include/exec/cpu-all.h | 1 +
25
target/arm/cpu.h | 7 +++
12
linux-user/syscall_defs.h | 1 +
26
target/arm/translate.h | 4 ++
13
target/arm/cpu.h | 1 +
27
target/arm/sme-fa64.decode | 90 ++++++++++++++++++++++++++++++++++++++
14
linux-user/mmap.c | 22 ++++++++++++++--------
28
target/arm/helper.c | 41 +++++++++++++++++
15
4 files changed, 17 insertions(+), 8 deletions(-)
29
target/arm/translate-a64.c | 40 ++++++++++++++++-
16
30
target/arm/translate-vfp.c | 12 +++++
17
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
31
target/arm/translate.c | 2 +
18
index XXXXXXX..XXXXXXX 100644
32
target/arm/meson.build | 1 +
19
--- a/include/exec/cpu-all.h
33
8 files changed, 195 insertions(+), 2 deletions(-)
20
+++ b/include/exec/cpu-all.h
34
create mode 100644 target/arm/sme-fa64.decode
21
@@ -XXX,XX +XXX,XX @@ extern intptr_t qemu_host_page_mask;
35
22
#endif
23
/* Target-specific bits that will be used via page_get_flags(). */
24
#define PAGE_TARGET_1 0x0080
25
+#define PAGE_TARGET_2 0x0200
26
27
#if defined(CONFIG_USER_ONLY)
28
void page_dump(FILE *f);
29
diff --git a/linux-user/syscall_defs.h b/linux-user/syscall_defs.h
30
index XXXXXXX..XXXXXXX 100644
31
--- a/linux-user/syscall_defs.h
32
+++ b/linux-user/syscall_defs.h
33
@@ -XXX,XX +XXX,XX @@ struct target_winsize {
34
35
#ifdef TARGET_AARCH64
36
#define TARGET_PROT_BTI 0x10
37
+#define TARGET_PROT_MTE 0x20
38
#endif
39
40
/* Common */
41
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
36
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
42
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
43
--- a/target/arm/cpu.h
38
--- a/target/arm/cpu.h
44
+++ b/target/arm/cpu.h
39
+++ b/target/arm/cpu.h
45
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A32, HSTR_ACTIVE, 9, 1)
46
* AArch64 usage of the PAGE_TARGET_* bits for linux-user.
41
* the same thing as the current security state of the processor!
47
*/
42
*/
48
#define PAGE_BTI PAGE_TARGET_1
43
FIELD(TBFLAG_A32, NS, 10, 1)
49
+#define PAGE_MTE PAGE_TARGET_2
44
+/*
50
45
+ * Indicates that SME Streaming mode is active, and SMCR_ELx.FA64 is not.
51
#ifdef TARGET_TAGGED_ADDRESSES
46
+ * This requires an SME trap from AArch32 mode when using NEON.
47
+ */
48
+FIELD(TBFLAG_A32, SME_TRAP_NONSTREAMING, 11, 1)
49
50
/*
51
* Bit usage when in AArch32 state, for M-profile only.
52
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, SMEEXC_EL, 20, 2)
53
FIELD(TBFLAG_A64, PSTATE_SM, 22, 1)
54
FIELD(TBFLAG_A64, PSTATE_ZA, 23, 1)
55
FIELD(TBFLAG_A64, SVL, 24, 4)
56
+/* Indicates that SME Streaming mode is active, and SMCR_ELx.FA64 is not. */
57
+FIELD(TBFLAG_A64, SME_TRAP_NONSTREAMING, 28, 1)
58
59
/*
60
* Helpers for using the above.
61
diff --git a/target/arm/translate.h b/target/arm/translate.h
62
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate.h
64
+++ b/target/arm/translate.h
65
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
66
bool pstate_sm;
67
/* True if PSTATE.ZA is set. */
68
bool pstate_za;
69
+ /* True if non-streaming insns should raise an SME Streaming exception. */
70
+ bool sme_trap_nonstreaming;
71
+ /* True if the current instruction is non-streaming. */
72
+ bool is_nonstreaming;
73
/* True if MVE insns are definitely not predicated by VPR or LTPSIZE */
74
bool mve_no_pred;
75
/*
76
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
77
new file mode 100644
78
index XXXXXXX..XXXXXXX
79
--- /dev/null
80
+++ b/target/arm/sme-fa64.decode
81
@@ -XXX,XX +XXX,XX @@
82
+# AArch64 SME allowed instruction decoding
83
+#
84
+# Copyright (c) 2022 Linaro, Ltd
85
+#
86
+# This library is free software; you can redistribute it and/or
87
+# modify it under the terms of the GNU Lesser General Public
88
+# License as published by the Free Software Foundation; either
89
+# version 2.1 of the License, or (at your option) any later version.
90
+#
91
+# This library is distributed in the hope that it will be useful,
92
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
93
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
94
+# Lesser General Public License for more details.
95
+#
96
+# You should have received a copy of the GNU Lesser General Public
97
+# License along with this library; if not, see <http://www.gnu.org/licenses/>.
98
+
99
+#
100
+# This file is processed by scripts/decodetree.py
101
+#
102
+
103
+# These patterns are taken from Appendix E1.1 of DDI0616 A.a,
104
+# Arm Architecture Reference Manual Supplement,
105
+# The Scalable Matrix Extension (SME), for Armv9-A
106
+
107
+{
108
+ [
109
+ OK 0-00 1110 0000 0001 0010 11-- ---- ---- # SMOV W|Xd,Vn.B[0]
110
+ OK 0-00 1110 0000 0010 0010 11-- ---- ---- # SMOV W|Xd,Vn.H[0]
111
+ OK 0100 1110 0000 0100 0010 11-- ---- ---- # SMOV Xd,Vn.S[0]
112
+ OK 0000 1110 0000 0001 0011 11-- ---- ---- # UMOV Wd,Vn.B[0]
113
+ OK 0000 1110 0000 0010 0011 11-- ---- ---- # UMOV Wd,Vn.H[0]
114
+ OK 0000 1110 0000 0100 0011 11-- ---- ---- # UMOV Wd,Vn.S[0]
115
+ OK 0100 1110 0000 1000 0011 11-- ---- ---- # UMOV Xd,Vn.D[0]
116
+ ]
117
+ FAIL 0--0 111- ---- ---- ---- ---- ---- ---- # Advanced SIMD vector operations
118
+}
119
+
120
+{
121
+ [
122
+ OK 0101 1110 --1- ---- 11-1 11-- ---- ---- # FMULX/FRECPS/FRSQRTS (scalar)
123
+ OK 0101 1110 -10- ---- 00-1 11-- ---- ---- # FMULX/FRECPS/FRSQRTS (scalar, FP16)
124
+ OK 01-1 1110 1-10 0001 11-1 10-- ---- ---- # FRECPE/FRSQRTE/FRECPX (scalar)
125
+ OK 01-1 1110 1111 1001 11-1 10-- ---- ---- # FRECPE/FRSQRTE/FRECPX (scalar, FP16)
126
+ ]
127
+ FAIL 01-1 111- ---- ---- ---- ---- ---- ---- # Advanced SIMD single-element operations
128
+}
129
+
130
+FAIL 0-00 110- ---- ---- ---- ---- ---- ---- # Advanced SIMD structure load/store
131
+FAIL 1100 1110 ---- ---- ---- ---- ---- ---- # Advanced SIMD cryptography extensions
132
+FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
133
+
134
+# These are the "avoidance of doubt" final table of Illegal Advanced SIMD instructions
135
+# We don't actually need to include these, as the default is OK.
136
+# -001 111- ---- ---- ---- ---- ---- ---- # Scalar floating-point operations
137
+# --10 110- ---- ---- ---- ---- ---- ---- # Load/store pair of FP registers
138
+# --01 1100 ---- ---- ---- ---- ---- ---- # Load FP register (PC-relative literal)
139
+# --11 1100 --0- ---- ---- ---- ---- ---- # Load/store FP register (unscaled imm)
140
+# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
141
+# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
142
+
143
+FAIL 0000 0100 --1- ---- 1010 ---- ---- ---- # ADR
144
+FAIL 0000 0100 --1- ---- 1011 -0-- ---- ---- # FTSSEL, FEXPA
145
+FAIL 0000 0101 --10 0001 100- ---- ---- ---- # COMPACT
146
+FAIL 0010 0101 --01 100- 1111 000- ---0 ---- # RDFFR, RDFFRS
147
+FAIL 0010 0101 --10 1--- 1001 ---- ---- ---- # WRFFR, SETFFR
148
+FAIL 0100 0101 --0- ---- 1011 ---- ---- ---- # BDEP, BEXT, BGRP
149
+FAIL 0100 0101 000- ---- 0110 1--- ---- ---- # PMULLB, PMULLT (128b result)
150
+FAIL 0110 0100 --1- ---- 1110 01-- ---- ---- # FMMLA, BFMMLA
151
+FAIL 0110 0101 --0- ---- 0000 11-- ---- ---- # FTSMUL
152
+FAIL 0110 0101 --01 0--- 100- ---- ---- ---- # FTMAD
153
+FAIL 0110 0101 --01 1--- 001- ---- ---- ---- # FADDA
154
+FAIL 0100 0101 --0- ---- 1001 10-- ---- ---- # SMMLA, UMMLA, USMMLA
155
+FAIL 0100 0101 --1- ---- 1--- ---- ---- ---- # SVE2 string/histo/crypto instructions
156
+FAIL 1000 010- -00- ---- 10-- ---- ---- ---- # SVE2 32-bit gather NT load (vector+scalar)
157
+FAIL 1000 010- -00- ---- 111- ---- ---- ---- # SVE 32-bit gather prefetch (vector+imm)
158
+FAIL 1000 0100 0-1- ---- 0--- ---- ---- ---- # SVE 32-bit gather prefetch (scalar+vector)
159
+FAIL 1000 010- -01- ---- 1--- ---- ---- ---- # SVE 32-bit gather load (vector+imm)
160
+FAIL 1000 0100 0-0- ---- 0--- ---- ---- ---- # SVE 32-bit gather load byte (scalar+vector)
161
+FAIL 1000 0100 1--- ---- 0--- ---- ---- ---- # SVE 32-bit gather load half (scalar+vector)
162
+FAIL 1000 0101 0--- ---- 0--- ---- ---- ---- # SVE 32-bit gather load word (scalar+vector)
163
+FAIL 1010 010- ---- ---- 011- ---- ---- ---- # SVE contiguous FF load (scalar+scalar)
164
+FAIL 1010 010- ---1 ---- 101- ---- ---- ---- # SVE contiguous NF load (scalar+imm)
165
+FAIL 1010 010- -01- ---- 000- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+scalar)
166
+FAIL 1010 010- -010 ---- 001- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+imm)
167
+FAIL 1100 010- ---- ---- ---- ---- ---- ---- # SVE 64-bit gather load/prefetch
168
+FAIL 1110 010- -00- ---- 001- ---- ---- ---- # SVE2 64-bit scatter NT store (vector+scalar)
169
+FAIL 1110 010- -10- ---- 001- ---- ---- ---- # SVE2 32-bit scatter NT store (vector+scalar)
170
+FAIL 1110 010- ---- ---- 1-0- ---- ---- ---- # SVE scatter store (scalar+32-bit vector)
171
+FAIL 1110 010- ---- ---- 101- ---- ---- ---- # SVE scatter store (misc)
172
diff --git a/target/arm/helper.c b/target/arm/helper.c
173
index XXXXXXX..XXXXXXX 100644
174
--- a/target/arm/helper.c
175
+++ b/target/arm/helper.c
176
@@ -XXX,XX +XXX,XX @@ int sme_exception_el(CPUARMState *env, int el)
177
return 0;
178
}
179
180
+/* This corresponds to the ARM pseudocode function IsFullA64Enabled(). */
181
+static bool sme_fa64(CPUARMState *env, int el)
182
+{
183
+ if (!cpu_isar_feature(aa64_sme_fa64, env_archcpu(env))) {
184
+ return false;
185
+ }
186
+
187
+ if (el <= 1 && !el_is_in_host(env, el)) {
188
+ if (!FIELD_EX64(env->vfp.smcr_el[1], SMCR, FA64)) {
189
+ return false;
190
+ }
191
+ }
192
+ if (el <= 2 && arm_is_el2_enabled(env)) {
193
+ if (!FIELD_EX64(env->vfp.smcr_el[2], SMCR, FA64)) {
194
+ return false;
195
+ }
196
+ }
197
+ if (arm_feature(env, ARM_FEATURE_EL3)) {
198
+ if (!FIELD_EX64(env->vfp.smcr_el[3], SMCR, FA64)) {
199
+ return false;
200
+ }
201
+ }
202
+
203
+ return true;
204
+}
205
+
206
/*
207
* Given that SVE is enabled, return the vector length for EL.
208
*/
209
@@ -XXX,XX +XXX,XX @@ static CPUARMTBFlags rebuild_hflags_a32(CPUARMState *env, int fp_el,
210
DP_TBFLAG_ANY(flags, PSTATE__IL, 1);
211
}
212
213
+ /*
214
+ * The SME exception we are testing for is raised via
215
+ * AArch64.CheckFPAdvSIMDEnabled(), as called from
216
+ * AArch32.CheckAdvSIMDOrFPEnabled().
217
+ */
218
+ if (el == 0
219
+ && FIELD_EX64(env->svcr, SVCR, SM)
220
+ && (!arm_is_el2_enabled(env)
221
+ || (arm_el_is_aa64(env, 2) && !(env->cp15.hcr_el2 & HCR_TGE)))
222
+ && arm_el_is_aa64(env, 1)
223
+ && !sme_fa64(env, el)) {
224
+ DP_TBFLAG_A32(flags, SME_TRAP_NONSTREAMING, 1);
225
+ }
226
+
227
return rebuild_hflags_common_32(env, fp_el, mmu_idx, flags);
228
}
229
230
@@ -XXX,XX +XXX,XX @@ static CPUARMTBFlags rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
231
}
232
if (FIELD_EX64(env->svcr, SVCR, SM)) {
233
DP_TBFLAG_A64(flags, PSTATE_SM, 1);
234
+ DP_TBFLAG_A64(flags, SME_TRAP_NONSTREAMING, !sme_fa64(env, el));
235
}
236
DP_TBFLAG_A64(flags, PSTATE_ZA, FIELD_EX64(env->svcr, SVCR, ZA));
237
}
238
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
239
index XXXXXXX..XXXXXXX 100644
240
--- a/target/arm/translate-a64.c
241
+++ b/target/arm/translate-a64.c
242
@@ -XXX,XX +XXX,XX @@ static void do_vec_ld(DisasContext *s, int destidx, int element,
243
* unallocated-encoding checks (otherwise the syndrome information
244
* for the resulting exception will be incorrect).
245
*/
246
-static bool fp_access_check(DisasContext *s)
247
+static bool fp_access_check_only(DisasContext *s)
248
{
249
if (s->fp_excp_el) {
250
assert(!s->fp_access_checked);
251
@@ -XXX,XX +XXX,XX @@ static bool fp_access_check(DisasContext *s)
252
return true;
253
}
254
255
+static bool fp_access_check(DisasContext *s)
256
+{
257
+ if (!fp_access_check_only(s)) {
258
+ return false;
259
+ }
260
+ if (s->sme_trap_nonstreaming && s->is_nonstreaming) {
261
+ gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
262
+ syn_smetrap(SME_ET_Streaming, false));
263
+ return false;
264
+ }
265
+ return true;
266
+}
267
+
268
/* Check that SVE access is enabled. If it is, return true.
269
* If not, emit code to generate an appropriate exception and return false.
270
*/
271
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
272
default:
273
g_assert_not_reached();
274
}
275
- if ((ri->type & ARM_CP_FPU) && !fp_access_check(s)) {
276
+ if ((ri->type & ARM_CP_FPU) && !fp_access_check_only(s)) {
277
return;
278
} else if ((ri->type & ARM_CP_SVE) && !sve_access_check(s)) {
279
return;
280
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_simd_fp(DisasContext *s, uint32_t insn)
281
}
282
}
283
284
+/*
285
+ * Include the generated SME FA64 decoder.
286
+ */
287
+
288
+#include "decode-sme-fa64.c.inc"
289
+
290
+static bool trans_OK(DisasContext *s, arg_OK *a)
291
+{
292
+ return true;
293
+}
294
+
295
+static bool trans_FAIL(DisasContext *s, arg_OK *a)
296
+{
297
+ s->is_nonstreaming = true;
298
+ return true;
299
+}
300
+
52
/**
301
/**
53
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
302
* is_guarded_page:
54
index XXXXXXX..XXXXXXX 100644
303
* @env: The cpu environment
55
--- a/linux-user/mmap.c
304
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
56
+++ b/linux-user/mmap.c
305
dc->mte_active[1] = EX_TBFLAG_A64(tb_flags, MTE0_ACTIVE);
57
@@ -XXX,XX +XXX,XX @@ static int validate_prot_to_pageflags(int *host_prot, int prot)
306
dc->pstate_sm = EX_TBFLAG_A64(tb_flags, PSTATE_SM);
58
| (prot & PROT_EXEC ? PROT_READ : 0);
307
dc->pstate_za = EX_TBFLAG_A64(tb_flags, PSTATE_ZA);
59
308
+ dc->sme_trap_nonstreaming = EX_TBFLAG_A64(tb_flags, SME_TRAP_NONSTREAMING);
60
#ifdef TARGET_AARCH64
309
dc->vec_len = 0;
61
- /*
310
dc->vec_stride = 0;
62
- * The PROT_BTI bit is only accepted if the cpu supports the feature.
311
dc->cp_regs = arm_cpu->cp_regs;
63
- * Since this is the unusual case, don't bother checking unless
312
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
64
- * the bit has been requested. If set and valid, record the bit
65
- * within QEMU's page_flags.
66
- */
67
- if (prot & TARGET_PROT_BTI) {
68
+ {
69
ARMCPU *cpu = ARM_CPU(thread_cpu);
70
- if (cpu_isar_feature(aa64_bti, cpu)) {
71
+
72
+ /*
73
+ * The PROT_BTI bit is only accepted if the cpu supports the feature.
74
+ * Since this is the unusual case, don't bother checking unless
75
+ * the bit has been requested. If set and valid, record the bit
76
+ * within QEMU's page_flags.
77
+ */
78
+ if ((prot & TARGET_PROT_BTI) && cpu_isar_feature(aa64_bti, cpu)) {
79
valid |= TARGET_PROT_BTI;
80
page_flags |= PAGE_BTI;
81
}
313
}
82
+ /* Similarly for the PROT_MTE bit. */
314
}
83
+ if ((prot & TARGET_PROT_MTE) && cpu_isar_feature(aa64_mte, cpu)) {
315
84
+ valid |= TARGET_PROT_MTE;
316
+ s->is_nonstreaming = false;
85
+ page_flags |= PAGE_MTE;
317
+ if (s->sme_trap_nonstreaming) {
86
+ }
318
+ disas_sme_fa64(s, insn);
87
}
319
+ }
88
#endif
320
+
89
321
switch (extract32(insn, 25, 4)) {
322
case 0x0:
323
if (!extract32(insn, 31, 1) || !disas_sme(s, insn)) {
324
diff --git a/target/arm/translate-vfp.c b/target/arm/translate-vfp.c
325
index XXXXXXX..XXXXXXX 100644
326
--- a/target/arm/translate-vfp.c
327
+++ b/target/arm/translate-vfp.c
328
@@ -XXX,XX +XXX,XX @@ static bool vfp_access_check_a(DisasContext *s, bool ignore_vfp_enabled)
329
return false;
330
}
331
332
+ /*
333
+ * Note that rebuild_hflags_a32 has already accounted for being in EL0
334
+ * and the higher EL in A64 mode, etc. Unlike A64 mode, there do not
335
+ * appear to be any insns which touch VFP which are allowed.
336
+ */
337
+ if (s->sme_trap_nonstreaming) {
338
+ gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
339
+ syn_smetrap(SME_ET_Streaming,
340
+ s->base.pc_next - s->pc_curr == 2));
341
+ return false;
342
+ }
343
+
344
if (!s->vfp_enabled && !ignore_vfp_enabled) {
345
assert(!arm_dc_feature(s, ARM_FEATURE_M));
346
unallocated_encoding(s);
347
diff --git a/target/arm/translate.c b/target/arm/translate.c
348
index XXXXXXX..XXXXXXX 100644
349
--- a/target/arm/translate.c
350
+++ b/target/arm/translate.c
351
@@ -XXX,XX +XXX,XX @@ static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
352
dc->vec_len = EX_TBFLAG_A32(tb_flags, VECLEN);
353
dc->vec_stride = EX_TBFLAG_A32(tb_flags, VECSTRIDE);
354
}
355
+ dc->sme_trap_nonstreaming =
356
+ EX_TBFLAG_A32(tb_flags, SME_TRAP_NONSTREAMING);
357
}
358
dc->cp_regs = cpu->cp_regs;
359
dc->features = env->features;
360
diff --git a/target/arm/meson.build b/target/arm/meson.build
361
index XXXXXXX..XXXXXXX 100644
362
--- a/target/arm/meson.build
363
+++ b/target/arm/meson.build
364
@@ -XXX,XX +XXX,XX @@
365
gen = [
366
decodetree.process('sve.decode', extra_args: '--decode=disas_sve'),
367
decodetree.process('sme.decode', extra_args: '--decode=disas_sme'),
368
+ decodetree.process('sme-fa64.decode', extra_args: '--static-decode=disas_sme_fa64'),
369
decodetree.process('neon-shared.decode', extra_args: '--decode=disas_neon_shared'),
370
decodetree.process('neon-dp.decode', extra_args: '--decode=disas_neon_dp'),
371
decodetree.process('neon-ls.decode', extra_args: '--decode=disas_neon_ls'),
90
--
372
--
91
2.20.1
373
2.25.1
92
93
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Mark ADR as a non-streaming instruction, which should trap
4
if full a64 support is not enabled in streaming mode.
5
6
Removing entries from sme-fa64.decode is an easy way to see
7
what remains to be done.
8
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20220708151540.18136-5-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
target/arm/translate.h | 7 +++++++
15
target/arm/sme-fa64.decode | 1 -
16
target/arm/translate-sve.c | 8 ++++----
17
3 files changed, 11 insertions(+), 5 deletions(-)
18
19
diff --git a/target/arm/translate.h b/target/arm/translate.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/translate.h
22
+++ b/target/arm/translate.h
23
@@ -XXX,XX +XXX,XX @@ uint64_t asimd_imm_const(uint32_t imm, int cmode, int op);
24
static bool trans_##NAME(DisasContext *s, arg_##NAME *a) \
25
{ return dc_isar_feature(FEAT, s) && FUNC(s, __VA_ARGS__); }
26
27
+#define TRANS_FEAT_NONSTREAMING(NAME, FEAT, FUNC, ...) \
28
+ static bool trans_##NAME(DisasContext *s, arg_##NAME *a) \
29
+ { \
30
+ s->is_nonstreaming = true; \
31
+ return dc_isar_feature(FEAT, s) && FUNC(s, __VA_ARGS__); \
32
+ }
33
+
34
#endif /* TARGET_ARM_TRANSLATE_H */
35
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
36
index XXXXXXX..XXXXXXX 100644
37
--- a/target/arm/sme-fa64.decode
38
+++ b/target/arm/sme-fa64.decode
39
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
40
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
41
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
42
43
-FAIL 0000 0100 --1- ---- 1010 ---- ---- ---- # ADR
44
FAIL 0000 0100 --1- ---- 1011 -0-- ---- ---- # FTSSEL, FEXPA
45
FAIL 0000 0101 --10 0001 100- ---- ---- ---- # COMPACT
46
FAIL 0010 0101 --01 100- 1111 000- ---0 ---- # RDFFR, RDFFRS
47
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/translate-sve.c
50
+++ b/target/arm/translate-sve.c
51
@@ -XXX,XX +XXX,XX @@ static bool do_adr(DisasContext *s, arg_rrri *a, gen_helper_gvec_3 *fn)
52
return gen_gvec_ool_zzz(s, fn, a->rd, a->rn, a->rm, a->imm);
53
}
54
55
-TRANS_FEAT(ADR_p32, aa64_sve, do_adr, a, gen_helper_sve_adr_p32)
56
-TRANS_FEAT(ADR_p64, aa64_sve, do_adr, a, gen_helper_sve_adr_p64)
57
-TRANS_FEAT(ADR_s32, aa64_sve, do_adr, a, gen_helper_sve_adr_s32)
58
-TRANS_FEAT(ADR_u32, aa64_sve, do_adr, a, gen_helper_sve_adr_u32)
59
+TRANS_FEAT_NONSTREAMING(ADR_p32, aa64_sve, do_adr, a, gen_helper_sve_adr_p32)
60
+TRANS_FEAT_NONSTREAMING(ADR_p64, aa64_sve, do_adr, a, gen_helper_sve_adr_p64)
61
+TRANS_FEAT_NONSTREAMING(ADR_s32, aa64_sve, do_adr, a, gen_helper_sve_adr_s32)
62
+TRANS_FEAT_NONSTREAMING(ADR_u32, aa64_sve, do_adr, a, gen_helper_sve_adr_u32)
63
64
/*
65
*** SVE Integer Misc - Unpredicated Group
66
--
67
2.25.1
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Mark these as a non-streaming instructions, which should trap
4
if full a64 support is not enabled in streaming mode.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220708151540.18136-6-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/sme-fa64.decode | 2 --
12
target/arm/translate-sve.c | 9 ++++++---
13
2 files changed, 6 insertions(+), 5 deletions(-)
14
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/sme-fa64.decode
18
+++ b/target/arm/sme-fa64.decode
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
20
21
FAIL 0000 0100 --1- ---- 1011 -0-- ---- ---- # FTSSEL, FEXPA
22
FAIL 0000 0101 --10 0001 100- ---- ---- ---- # COMPACT
23
-FAIL 0010 0101 --01 100- 1111 000- ---0 ---- # RDFFR, RDFFRS
24
-FAIL 0010 0101 --10 1--- 1001 ---- ---- ---- # WRFFR, SETFFR
25
FAIL 0100 0101 --0- ---- 1011 ---- ---- ---- # BDEP, BEXT, BGRP
26
FAIL 0100 0101 000- ---- 0110 1--- ---- ---- # PMULLB, PMULLT (128b result)
27
FAIL 0110 0100 --1- ---- 1110 01-- ---- ---- # FMMLA, BFMMLA
28
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
29
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/translate-sve.c
31
+++ b/target/arm/translate-sve.c
32
@@ -XXX,XX +XXX,XX @@ static bool do_predset(DisasContext *s, int esz, int rd, int pat, bool setflag)
33
TRANS_FEAT(PTRUE, aa64_sve, do_predset, a->esz, a->rd, a->pat, a->s)
34
35
/* Note pat == 31 is #all, to set all elements. */
36
-TRANS_FEAT(SETFFR, aa64_sve, do_predset, 0, FFR_PRED_NUM, 31, false)
37
+TRANS_FEAT_NONSTREAMING(SETFFR, aa64_sve,
38
+ do_predset, 0, FFR_PRED_NUM, 31, false)
39
40
/* Note pat == 32 is #unimp, to set no elements. */
41
TRANS_FEAT(PFALSE, aa64_sve, do_predset, 0, a->rd, 32, false)
42
@@ -XXX,XX +XXX,XX @@ static bool trans_RDFFR_p(DisasContext *s, arg_RDFFR_p *a)
43
.rd = a->rd, .pg = a->pg, .s = a->s,
44
.rn = FFR_PRED_NUM, .rm = FFR_PRED_NUM,
45
};
46
+
47
+ s->is_nonstreaming = true;
48
return trans_AND_pppp(s, &alt_a);
49
}
50
51
-TRANS_FEAT(RDFFR, aa64_sve, do_mov_p, a->rd, FFR_PRED_NUM)
52
-TRANS_FEAT(WRFFR, aa64_sve, do_mov_p, FFR_PRED_NUM, a->rn)
53
+TRANS_FEAT_NONSTREAMING(RDFFR, aa64_sve, do_mov_p, a->rd, FFR_PRED_NUM)
54
+TRANS_FEAT_NONSTREAMING(WRFFR, aa64_sve, do_mov_p, FFR_PRED_NUM, a->rn)
55
56
static bool do_pfirst_pnext(DisasContext *s, arg_rr_esz *a,
57
void (*gen_fn)(TCGv_i32, TCGv_ptr,
58
--
59
2.25.1
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is more descriptive than 'unsigned long'.
3
Mark these as a non-streaming instructions, which should trap
4
No functional change, since these match on all linux+bsd hosts.
4
if full a64 support is not enabled in streaming mode.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20220708151540.18136-7-richard.henderson@linaro.org
9
Message-id: 20210212184902.1251044-4-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
include/exec/cpu-all.h | 2 +-
11
target/arm/sme-fa64.decode | 3 ---
13
bsd-user/main.c | 4 ++--
12
target/arm/translate-sve.c | 22 ++++++++++++----------
14
linux-user/elfload.c | 4 ++--
13
2 files changed, 12 insertions(+), 13 deletions(-)
15
linux-user/main.c | 4 ++--
16
4 files changed, 7 insertions(+), 7 deletions(-)
17
14
18
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
19
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
20
--- a/include/exec/cpu-all.h
17
--- a/target/arm/sme-fa64.decode
21
+++ b/include/exec/cpu-all.h
18
+++ b/target/arm/sme-fa64.decode
22
@@ -XXX,XX +XXX,XX @@ static inline void tswap64s(uint64_t *s)
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
23
/* On some host systems the guest address space is reserved on the host.
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
24
* This allows the guest address space to be offset to a convenient location.
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
25
*/
22
26
-extern unsigned long guest_base;
23
-FAIL 0000 0100 --1- ---- 1011 -0-- ---- ---- # FTSSEL, FEXPA
27
+extern uintptr_t guest_base;
24
-FAIL 0000 0101 --10 0001 100- ---- ---- ---- # COMPACT
28
extern bool have_guest_base;
25
-FAIL 0100 0101 --0- ---- 1011 ---- ---- ---- # BDEP, BEXT, BGRP
29
extern unsigned long reserved_va;
26
FAIL 0100 0101 000- ---- 0110 1--- ---- ---- # PMULLB, PMULLT (128b result)
30
27
FAIL 0110 0100 --1- ---- 1110 01-- ---- ---- # FMMLA, BFMMLA
31
diff --git a/bsd-user/main.c b/bsd-user/main.c
28
FAIL 0110 0101 --0- ---- 0000 11-- ---- ---- # FTSMUL
29
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
32
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
33
--- a/bsd-user/main.c
31
--- a/target/arm/translate-sve.c
34
+++ b/bsd-user/main.c
32
+++ b/target/arm/translate-sve.c
35
@@ -XXX,XX +XXX,XX @@
33
@@ -XXX,XX +XXX,XX @@ static gen_helper_gvec_2 * const fexpa_fns[4] = {
36
34
NULL, gen_helper_sve_fexpa_h,
37
int singlestep;
35
gen_helper_sve_fexpa_s, gen_helper_sve_fexpa_d,
38
unsigned long mmap_min_addr;
36
};
39
-unsigned long guest_base;
37
-TRANS_FEAT(FEXPA, aa64_sve, gen_gvec_ool_zz,
40
+uintptr_t guest_base;
38
- fexpa_fns[a->esz], a->rd, a->rn, 0)
41
bool have_guest_base;
39
+TRANS_FEAT_NONSTREAMING(FEXPA, aa64_sve, gen_gvec_ool_zz,
42
unsigned long reserved_va;
40
+ fexpa_fns[a->esz], a->rd, a->rn, 0)
43
41
44
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
42
static gen_helper_gvec_3 * const ftssel_fns[4] = {
45
g_free(target_environ);
43
NULL, gen_helper_sve_ftssel_h,
46
44
gen_helper_sve_ftssel_s, gen_helper_sve_ftssel_d,
47
if (qemu_loglevel_mask(CPU_LOG_PAGE)) {
45
};
48
- qemu_log("guest_base 0x%lx\n", guest_base);
46
-TRANS_FEAT(FTSSEL, aa64_sve, gen_gvec_ool_arg_zzz, ftssel_fns[a->esz], a, 0)
49
+ qemu_log("guest_base %p\n", (void *)guest_base);
47
+TRANS_FEAT_NONSTREAMING(FTSSEL, aa64_sve, gen_gvec_ool_arg_zzz,
50
log_page_dump("binary load");
48
+ ftssel_fns[a->esz], a, 0)
51
52
qemu_log("start_brk 0x" TARGET_ABI_FMT_lx "\n", info->start_brk);
53
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
54
index XXXXXXX..XXXXXXX 100644
55
--- a/linux-user/elfload.c
56
+++ b/linux-user/elfload.c
57
@@ -XXX,XX +XXX,XX @@ static void pgb_have_guest_base(const char *image_name, abi_ulong guest_loaddr,
58
void *addr, *test;
59
60
if (!QEMU_IS_ALIGNED(guest_base, align)) {
61
- fprintf(stderr, "Requested guest base 0x%lx does not satisfy "
62
+ fprintf(stderr, "Requested guest base %p does not satisfy "
63
"host minimum alignment (0x%lx)\n",
64
- guest_base, align);
65
+ (void *)guest_base, align);
66
exit(EXIT_FAILURE);
67
}
68
69
diff --git a/linux-user/main.c b/linux-user/main.c
70
index XXXXXXX..XXXXXXX 100644
71
--- a/linux-user/main.c
72
+++ b/linux-user/main.c
73
@@ -XXX,XX +XXX,XX @@ static const char *cpu_model;
74
static const char *cpu_type;
75
static const char *seed_optarg;
76
unsigned long mmap_min_addr;
77
-unsigned long guest_base;
78
+uintptr_t guest_base;
79
bool have_guest_base;
80
49
81
/*
50
/*
82
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv, char **envp)
51
*** SVE Predicate Logical Operations Group
83
g_free(target_environ);
52
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(TRN2_q, aa64_sve_f64mm, gen_gvec_ool_arg_zzz,
84
53
static gen_helper_gvec_3 * const compact_fns[4] = {
85
if (qemu_loglevel_mask(CPU_LOG_PAGE)) {
54
NULL, NULL, gen_helper_sve_compact_s, gen_helper_sve_compact_d
86
- qemu_log("guest_base 0x%lx\n", guest_base);
55
};
87
+ qemu_log("guest_base %p\n", (void *)guest_base);
56
-TRANS_FEAT(COMPACT, aa64_sve, gen_gvec_ool_arg_zpz, compact_fns[a->esz], a, 0)
88
log_page_dump("binary load");
57
+TRANS_FEAT_NONSTREAMING(COMPACT, aa64_sve, gen_gvec_ool_arg_zpz,
89
58
+ compact_fns[a->esz], a, 0)
90
qemu_log("start_brk 0x" TARGET_ABI_FMT_lx "\n", info->start_brk);
59
60
/* Call the helper that computes the ARM LastActiveElement pseudocode
61
* function, scaled by the element size. This includes the not found
62
@@ -XXX,XX +XXX,XX @@ static gen_helper_gvec_3 * const bext_fns[4] = {
63
gen_helper_sve2_bext_b, gen_helper_sve2_bext_h,
64
gen_helper_sve2_bext_s, gen_helper_sve2_bext_d,
65
};
66
-TRANS_FEAT(BEXT, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
67
- bext_fns[a->esz], a, 0)
68
+TRANS_FEAT_NONSTREAMING(BEXT, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
69
+ bext_fns[a->esz], a, 0)
70
71
static gen_helper_gvec_3 * const bdep_fns[4] = {
72
gen_helper_sve2_bdep_b, gen_helper_sve2_bdep_h,
73
gen_helper_sve2_bdep_s, gen_helper_sve2_bdep_d,
74
};
75
-TRANS_FEAT(BDEP, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
76
- bdep_fns[a->esz], a, 0)
77
+TRANS_FEAT_NONSTREAMING(BDEP, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
78
+ bdep_fns[a->esz], a, 0)
79
80
static gen_helper_gvec_3 * const bgrp_fns[4] = {
81
gen_helper_sve2_bgrp_b, gen_helper_sve2_bgrp_h,
82
gen_helper_sve2_bgrp_s, gen_helper_sve2_bgrp_d,
83
};
84
-TRANS_FEAT(BGRP, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
85
- bgrp_fns[a->esz], a, 0)
86
+TRANS_FEAT_NONSTREAMING(BGRP, aa64_sve2_bitperm, gen_gvec_ool_arg_zzz,
87
+ bgrp_fns[a->esz], a, 0)
88
89
static gen_helper_gvec_3 * const cadd_fns[4] = {
90
gen_helper_sve2_cadd_b, gen_helper_sve2_cadd_h,
91
--
91
--
92
2.20.1
92
2.25.1
93
94
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We must always use GUEST_ADDR_MAX, because even 32-bit hosts can
3
Mark these as a non-streaming instructions, which should trap
4
use -R <reserved_va> to restrict the memory address of the guest.
4
if full a64 support is not enabled in streaming mode.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-11-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-8-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
include/exec/cpu_ldst.h | 9 ++++-----
11
target/arm/sme-fa64.decode | 2 --
12
1 file changed, 4 insertions(+), 5 deletions(-)
12
target/arm/translate-sve.c | 24 +++++++++++++++---------
13
2 files changed, 15 insertions(+), 11 deletions(-)
13
14
14
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/include/exec/cpu_ldst.h
17
--- a/target/arm/sme-fa64.decode
17
+++ b/include/exec/cpu_ldst.h
18
+++ b/target/arm/sme-fa64.decode
18
@@ -XXX,XX +XXX,XX @@ typedef uint64_t abi_ptr;
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
19
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
20
#define g2h(x) ((void *)((uintptr_t)(abi_ptr)(x) + guest_base))
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
21
22
22
-#if HOST_LONG_BITS <= TARGET_VIRT_ADDR_SPACE_BITS
23
-FAIL 0100 0101 000- ---- 0110 1--- ---- ---- # PMULLB, PMULLT (128b result)
23
-#define guest_addr_valid(x) (1)
24
-FAIL 0110 0100 --1- ---- 1110 01-- ---- ---- # FMMLA, BFMMLA
24
-#else
25
FAIL 0110 0101 --0- ---- 0000 11-- ---- ---- # FTSMUL
25
-#define guest_addr_valid(x) ((x) <= GUEST_ADDR_MAX)
26
FAIL 0110 0101 --01 0--- 100- ---- ---- ---- # FTMAD
26
-#endif
27
FAIL 0110 0101 --01 1--- 001- ---- ---- ---- # FADDA
27
+static inline bool guest_addr_valid(abi_ulong x)
28
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
28
+{
29
index XXXXXXX..XXXXXXX 100644
29
+ return x <= GUEST_ADDR_MAX;
30
--- a/target/arm/translate-sve.c
30
+}
31
+++ b/target/arm/translate-sve.c
31
32
@@ -XXX,XX +XXX,XX @@ static bool do_trans_pmull(DisasContext *s, arg_rrr_esz *a, bool sel)
32
static inline bool guest_range_valid(abi_ulong start, abi_ulong len)
33
gen_helper_gvec_pmull_q, gen_helper_sve2_pmull_h,
34
NULL, gen_helper_sve2_pmull_d,
35
};
36
- if (a->esz == 0
37
- ? !dc_isar_feature(aa64_sve2_pmull128, s)
38
- : !dc_isar_feature(aa64_sve, s)) {
39
+
40
+ if (a->esz == 0) {
41
+ if (!dc_isar_feature(aa64_sve2_pmull128, s)) {
42
+ return false;
43
+ }
44
+ s->is_nonstreaming = true;
45
+ } else if (!dc_isar_feature(aa64_sve, s)) {
46
return false;
47
}
48
return gen_gvec_ool_arg_zzz(s, fns[a->esz], a, sel);
49
@@ -XXX,XX +XXX,XX @@ DO_ZPZZ_FP(FMINP, aa64_sve2, sve2_fminp_zpzz)
50
* SVE Integer Multiply-Add (unpredicated)
51
*/
52
53
-TRANS_FEAT(FMMLA_s, aa64_sve_f32mm, gen_gvec_fpst_zzzz, gen_helper_fmmla_s,
54
- a->rd, a->rn, a->rm, a->ra, 0, FPST_FPCR)
55
-TRANS_FEAT(FMMLA_d, aa64_sve_f64mm, gen_gvec_fpst_zzzz, gen_helper_fmmla_d,
56
- a->rd, a->rn, a->rm, a->ra, 0, FPST_FPCR)
57
+TRANS_FEAT_NONSTREAMING(FMMLA_s, aa64_sve_f32mm, gen_gvec_fpst_zzzz,
58
+ gen_helper_fmmla_s, a->rd, a->rn, a->rm, a->ra,
59
+ 0, FPST_FPCR)
60
+TRANS_FEAT_NONSTREAMING(FMMLA_d, aa64_sve_f64mm, gen_gvec_fpst_zzzz,
61
+ gen_helper_fmmla_d, a->rd, a->rn, a->rm, a->ra,
62
+ 0, FPST_FPCR)
63
64
static gen_helper_gvec_4 * const sqdmlal_zzzw_fns[] = {
65
NULL, gen_helper_sve2_sqdmlal_zzzw_h,
66
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(BFDOT_zzzz, aa64_sve_bf16, gen_gvec_ool_arg_zzzz,
67
TRANS_FEAT(BFDOT_zzxz, aa64_sve_bf16, gen_gvec_ool_arg_zzxz,
68
gen_helper_gvec_bfdot_idx, a)
69
70
-TRANS_FEAT(BFMMLA, aa64_sve_bf16, gen_gvec_ool_arg_zzzz,
71
- gen_helper_gvec_bfmmla, a, 0)
72
+TRANS_FEAT_NONSTREAMING(BFMMLA, aa64_sve_bf16, gen_gvec_ool_arg_zzzz,
73
+ gen_helper_gvec_bfmmla, a, 0)
74
75
static bool do_BFMLAL_zzzw(DisasContext *s, arg_rrrr_esz *a, bool sel)
33
{
76
{
34
--
77
--
35
2.20.1
78
2.25.1
36
37
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is the prctl bit that controls whether syscalls accept tagged
3
Mark these as a non-streaming instructions, which should trap
4
addresses. See Documentation/arm64/tagged-address-abi.rst in the
4
if full a64 support is not enabled in streaming mode.
5
linux kernel.
6
5
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20210212184902.1251044-21-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-9-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
linux-user/aarch64/target_syscall.h | 4 ++++
11
target/arm/sme-fa64.decode | 3 ---
13
target/arm/cpu-param.h | 3 +++
12
target/arm/translate-sve.c | 15 +++++++++++----
14
target/arm/cpu.h | 31 +++++++++++++++++++++++++++++
13
2 files changed, 11 insertions(+), 7 deletions(-)
15
linux-user/syscall.c | 24 ++++++++++++++++++++++
16
4 files changed, 62 insertions(+)
17
14
18
diff --git a/linux-user/aarch64/target_syscall.h b/linux-user/aarch64/target_syscall.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
19
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
20
--- a/linux-user/aarch64/target_syscall.h
17
--- a/target/arm/sme-fa64.decode
21
+++ b/linux-user/aarch64/target_syscall.h
18
+++ b/target/arm/sme-fa64.decode
22
@@ -XXX,XX +XXX,XX @@ struct target_pt_regs {
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
23
# define TARGET_PR_PAC_APDBKEY (1 << 3)
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
24
# define TARGET_PR_PAC_APGAKEY (1 << 4)
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
25
22
26
+#define TARGET_PR_SET_TAGGED_ADDR_CTRL 55
23
-FAIL 0110 0101 --0- ---- 0000 11-- ---- ---- # FTSMUL
27
+#define TARGET_PR_GET_TAGGED_ADDR_CTRL 56
24
-FAIL 0110 0101 --01 0--- 100- ---- ---- ---- # FTMAD
28
+# define TARGET_PR_TAGGED_ADDR_ENABLE (1UL << 0)
25
-FAIL 0110 0101 --01 1--- 001- ---- ---- ---- # FADDA
29
+
26
FAIL 0100 0101 --0- ---- 1001 10-- ---- ---- # SMMLA, UMMLA, USMMLA
30
#endif /* AARCH64_TARGET_SYSCALL_H */
27
FAIL 0100 0101 --1- ---- 1--- ---- ---- ---- # SVE2 string/histo/crypto instructions
31
diff --git a/target/arm/cpu-param.h b/target/arm/cpu-param.h
28
FAIL 1000 010- -00- ---- 10-- ---- ---- ---- # SVE2 32-bit gather NT load (vector+scalar)
29
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
32
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/cpu-param.h
31
--- a/target/arm/translate-sve.c
34
+++ b/target/arm/cpu-param.h
32
+++ b/target/arm/translate-sve.c
35
@@ -XXX,XX +XXX,XX @@
33
@@ -XXX,XX +XXX,XX @@ static gen_helper_gvec_3_ptr * const ftmad_fns[4] = {
36
34
NULL, gen_helper_sve_ftmad_h,
37
#ifdef CONFIG_USER_ONLY
35
gen_helper_sve_ftmad_s, gen_helper_sve_ftmad_d,
38
#define TARGET_PAGE_BITS 12
36
};
39
+# ifdef TARGET_AARCH64
37
-TRANS_FEAT(FTMAD, aa64_sve, gen_gvec_fpst_zzz,
40
+# define TARGET_TAGGED_ADDRESSES
38
- ftmad_fns[a->esz], a->rd, a->rn, a->rm, a->imm,
41
+# endif
39
- a->esz == MO_16 ? FPST_FPCR_F16 : FPST_FPCR)
42
#else
40
+TRANS_FEAT_NONSTREAMING(FTMAD, aa64_sve, gen_gvec_fpst_zzz,
41
+ ftmad_fns[a->esz], a->rd, a->rn, a->rm, a->imm,
42
+ a->esz == MO_16 ? FPST_FPCR_F16 : FPST_FPCR)
43
43
/*
44
/*
44
* ARMv7 and later CPUs have 4K pages minimum, but ARMv5 and v6
45
*** SVE Floating Point Accumulating Reduction Group
45
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
46
@@ -XXX,XX +XXX,XX @@ static bool trans_FADDA(DisasContext *s, arg_rprr_esz *a)
46
index XXXXXXX..XXXXXXX 100644
47
if (a->esz == 0 || !dc_isar_feature(aa64_sve, s)) {
47
--- a/target/arm/cpu.h
48
return false;
48
+++ b/target/arm/cpu.h
49
}
49
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
50
+ s->is_nonstreaming = true;
50
const struct arm_boot_info *boot_info;
51
if (!sve_access_check(s)) {
51
/* Store GICv3CPUState to access from this struct */
52
return true;
52
void *gicv3state;
53
}
53
+
54
@@ -XXX,XX +XXX,XX @@ static bool trans_FADDA(DisasContext *s, arg_rprr_esz *a)
54
+#ifdef TARGET_TAGGED_ADDRESSES
55
DO_FP3(FADD_zzz, fadd)
55
+ /* Linux syscall tagged address support */
56
DO_FP3(FSUB_zzz, fsub)
56
+ bool tagged_addr_enable;
57
DO_FP3(FMUL_zzz, fmul)
57
+#endif
58
-DO_FP3(FTSMUL, ftsmul)
58
} CPUARMState;
59
DO_FP3(FRECPS, recps)
59
60
DO_FP3(FRSQRTS, rsqrts)
60
static inline void set_feature(CPUARMState *env, int feature)
61
61
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
62
#undef DO_FP3
62
*/
63
63
#define PAGE_BTI PAGE_TARGET_1
64
+static gen_helper_gvec_3_ptr * const ftsmul_fns[4] = {
64
65
+ NULL, gen_helper_gvec_ftsmul_h,
65
+#ifdef TARGET_TAGGED_ADDRESSES
66
+ gen_helper_gvec_ftsmul_s, gen_helper_gvec_ftsmul_d
66
+/**
67
+};
67
+ * cpu_untagged_addr:
68
+TRANS_FEAT_NONSTREAMING(FTSMUL, aa64_sve, gen_gvec_fpst_arg_zzz,
68
+ * @cs: CPU context
69
+ ftsmul_fns[a->esz], a, 0)
69
+ * @x: tagged address
70
+ *
71
+ * Remove any address tag from @x. This is explicitly related to the
72
+ * linux syscall TIF_TAGGED_ADDR setting, not TBI in general.
73
+ *
74
+ * There should be a better place to put this, but we need this in
75
+ * include/exec/cpu_ldst.h, and not some place linux-user specific.
76
+ */
77
+static inline target_ulong cpu_untagged_addr(CPUState *cs, target_ulong x)
78
+{
79
+ ARMCPU *cpu = ARM_CPU(cs);
80
+ if (cpu->env.tagged_addr_enable) {
81
+ /*
82
+ * TBI is enabled for userspace but not kernelspace addresses.
83
+ * Only clear the tag if bit 55 is clear.
84
+ */
85
+ x &= sextract64(x, 0, 56);
86
+ }
87
+ return x;
88
+}
89
+#endif
90
+
70
+
91
/*
71
/*
92
* Naming convention for isar_feature functions:
72
*** SVE Floating Point Arithmetic - Predicated Group
93
* Functions which test 32-bit ID registers should have _aa32_ in
73
*/
94
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
95
index XXXXXXX..XXXXXXX 100644
96
--- a/linux-user/syscall.c
97
+++ b/linux-user/syscall.c
98
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
99
}
100
}
101
return -TARGET_EINVAL;
102
+ case TARGET_PR_SET_TAGGED_ADDR_CTRL:
103
+ {
104
+ abi_ulong valid_mask = TARGET_PR_TAGGED_ADDR_ENABLE;
105
+ CPUARMState *env = cpu_env;
106
+
107
+ if ((arg2 & ~valid_mask) || arg3 || arg4 || arg5) {
108
+ return -TARGET_EINVAL;
109
+ }
110
+ env->tagged_addr_enable = arg2 & TARGET_PR_TAGGED_ADDR_ENABLE;
111
+ return 0;
112
+ }
113
+ case TARGET_PR_GET_TAGGED_ADDR_CTRL:
114
+ {
115
+ abi_long ret = 0;
116
+ CPUARMState *env = cpu_env;
117
+
118
+ if (arg2 || arg3 || arg4 || arg5) {
119
+ return -TARGET_EINVAL;
120
+ }
121
+ if (env->tagged_addr_enable) {
122
+ ret |= TARGET_PR_TAGGED_ADDR_ENABLE;
123
+ }
124
+ return ret;
125
+ }
126
#endif /* AARCH64 */
127
case PR_GET_SECCOMP:
128
case PR_SET_SECCOMP:
129
--
74
--
130
2.20.1
75
2.25.1
131
132
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Mark these as a non-streaming instructions, which should trap
4
if full a64 support is not enabled in streaming mode.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220708151540.18136-10-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/sme-fa64.decode | 1 -
12
target/arm/translate-sve.c | 12 ++++++------
13
2 files changed, 6 insertions(+), 7 deletions(-)
14
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/sme-fa64.decode
18
+++ b/target/arm/sme-fa64.decode
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
22
23
-FAIL 0100 0101 --0- ---- 1001 10-- ---- ---- # SMMLA, UMMLA, USMMLA
24
FAIL 0100 0101 --1- ---- 1--- ---- ---- ---- # SVE2 string/histo/crypto instructions
25
FAIL 1000 010- -00- ---- 10-- ---- ---- ---- # SVE2 32-bit gather NT load (vector+scalar)
26
FAIL 1000 010- -00- ---- 111- ---- ---- ---- # SVE 32-bit gather prefetch (vector+imm)
27
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-sve.c
30
+++ b/target/arm/translate-sve.c
31
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(FMLALT_zzxw, aa64_sve2, do_FMLAL_zzxw, a, false, true)
32
TRANS_FEAT(FMLSLB_zzxw, aa64_sve2, do_FMLAL_zzxw, a, true, false)
33
TRANS_FEAT(FMLSLT_zzxw, aa64_sve2, do_FMLAL_zzxw, a, true, true)
34
35
-TRANS_FEAT(SMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
36
- gen_helper_gvec_smmla_b, a, 0)
37
-TRANS_FEAT(USMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
38
- gen_helper_gvec_usmmla_b, a, 0)
39
-TRANS_FEAT(UMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
40
- gen_helper_gvec_ummla_b, a, 0)
41
+TRANS_FEAT_NONSTREAMING(SMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
42
+ gen_helper_gvec_smmla_b, a, 0)
43
+TRANS_FEAT_NONSTREAMING(USMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
44
+ gen_helper_gvec_usmmla_b, a, 0)
45
+TRANS_FEAT_NONSTREAMING(UMMLA, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
46
+ gen_helper_gvec_ummla_b, a, 0)
47
48
TRANS_FEAT(BFDOT_zzzz, aa64_sve_bf16, gen_gvec_ool_arg_zzzz,
49
gen_helper_gvec_bfdot, a, 0)
50
--
51
2.25.1
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
The real kernel collects _TIF_MTE_ASYNC_FAULT into the current thread's
3
Mark these as non-streaming instructions, which should trap
4
state on any kernel entry (interrupt, exception etc), and then delivers
4
if full a64 support is not enabled in streaming mode.
5
the signal in advance of resuming the thread.
6
7
This means that while the signal won't be delivered immediately, it will
8
not be delayed forever -- at minimum it will be delivered after the next
9
clock interrupt.
10
11
We don't have a clock interrupt in linux-user, so we issue a cpu_kick
12
to signal a return to the main loop at the end of the current TB.
13
5
14
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
15
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
16
Message-id: 20210212184902.1251044-29-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-11-richard.henderson@linaro.org
17
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
18
---
10
---
19
linux-user/aarch64/target_signal.h | 1 +
11
target/arm/sme-fa64.decode | 1 -
20
linux-user/aarch64/cpu_loop.c | 11 +++++++++++
12
target/arm/translate-sve.c | 35 ++++++++++++++++++-----------------
21
target/arm/mte_helper.c | 10 ++++++++++
13
2 files changed, 18 insertions(+), 18 deletions(-)
22
3 files changed, 22 insertions(+)
23
14
24
diff --git a/linux-user/aarch64/target_signal.h b/linux-user/aarch64/target_signal.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
25
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
26
--- a/linux-user/aarch64/target_signal.h
17
--- a/target/arm/sme-fa64.decode
27
+++ b/linux-user/aarch64/target_signal.h
18
+++ b/target/arm/sme-fa64.decode
28
@@ -XXX,XX +XXX,XX @@ typedef struct target_sigaltstack {
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
29
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
30
#include "../generic/signal.h"
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
31
22
32
+#define TARGET_SEGV_MTEAERR 8 /* Asynchronous ARM MTE error */
23
-FAIL 0100 0101 --1- ---- 1--- ---- ---- ---- # SVE2 string/histo/crypto instructions
33
#define TARGET_SEGV_MTESERR 9 /* Synchronous ARM MTE exception */
24
FAIL 1000 010- -00- ---- 10-- ---- ---- ---- # SVE2 32-bit gather NT load (vector+scalar)
34
25
FAIL 1000 010- -00- ---- 111- ---- ---- ---- # SVE 32-bit gather prefetch (vector+imm)
35
#define TARGET_ARCH_HAS_SETUP_FRAME
26
FAIL 1000 0100 0-1- ---- 0--- ---- ---- ---- # SVE 32-bit gather prefetch (scalar+vector)
36
diff --git a/linux-user/aarch64/cpu_loop.c b/linux-user/aarch64/cpu_loop.c
27
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
37
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
38
--- a/linux-user/aarch64/cpu_loop.c
29
--- a/target/arm/translate-sve.c
39
+++ b/linux-user/aarch64/cpu_loop.c
30
+++ b/target/arm/translate-sve.c
40
@@ -XXX,XX +XXX,XX @@ void cpu_loop(CPUARMState *env)
31
@@ -XXX,XX +XXX,XX @@ DO_SVE2_ZZZ_NARROW(RSUBHNT, rsubhnt)
41
EXCP_DUMP(env, "qemu: unhandled CPU exception 0x%x - aborting\n", trapnr);
32
static gen_helper_gvec_flags_4 * const match_fns[4] = {
42
abort();
33
gen_helper_sve2_match_ppzz_b, gen_helper_sve2_match_ppzz_h, NULL, NULL
43
}
34
};
44
+
35
-TRANS_FEAT(MATCH, aa64_sve2, do_ppzz_flags, a, match_fns[a->esz])
45
+ /* Check for MTE asynchronous faults */
36
+TRANS_FEAT_NONSTREAMING(MATCH, aa64_sve2, do_ppzz_flags, a, match_fns[a->esz])
46
+ if (unlikely(env->cp15.tfsr_el[0])) {
37
47
+ env->cp15.tfsr_el[0] = 0;
38
static gen_helper_gvec_flags_4 * const nmatch_fns[4] = {
48
+ info.si_signo = TARGET_SIGSEGV;
39
gen_helper_sve2_nmatch_ppzz_b, gen_helper_sve2_nmatch_ppzz_h, NULL, NULL
49
+ info.si_errno = 0;
40
};
50
+ info._sifields._sigfault._addr = 0;
41
-TRANS_FEAT(NMATCH, aa64_sve2, do_ppzz_flags, a, nmatch_fns[a->esz])
51
+ info.si_code = TARGET_SEGV_MTEAERR;
42
+TRANS_FEAT_NONSTREAMING(NMATCH, aa64_sve2, do_ppzz_flags, a, nmatch_fns[a->esz])
52
+ queue_signal(env, info.si_signo, QEMU_SI_FAULT, &info);
43
53
+ }
44
static gen_helper_gvec_4 * const histcnt_fns[4] = {
54
+
45
NULL, NULL, gen_helper_sve2_histcnt_s, gen_helper_sve2_histcnt_d
55
process_pending_signals(env);
46
};
56
/* Exception return on AArch64 always clears the exclusive monitor,
47
-TRANS_FEAT(HISTCNT, aa64_sve2, gen_gvec_ool_arg_zpzz,
57
* so any return to running guest code implies this.
48
- histcnt_fns[a->esz], a, 0)
58
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
49
+TRANS_FEAT_NONSTREAMING(HISTCNT, aa64_sve2, gen_gvec_ool_arg_zpzz,
59
index XXXXXXX..XXXXXXX 100644
50
+ histcnt_fns[a->esz], a, 0)
60
--- a/target/arm/mte_helper.c
51
61
+++ b/target/arm/mte_helper.c
52
-TRANS_FEAT(HISTSEG, aa64_sve2, gen_gvec_ool_arg_zzz,
62
@@ -XXX,XX +XXX,XX @@ static void mte_check_fail(CPUARMState *env, uint32_t desc,
53
- a->esz == 0 ? gen_helper_sve2_histseg : NULL, a, 0)
63
select = 0;
54
+TRANS_FEAT_NONSTREAMING(HISTSEG, aa64_sve2, gen_gvec_ool_arg_zzz,
64
}
55
+ a->esz == 0 ? gen_helper_sve2_histseg : NULL, a, 0)
65
env->cp15.tfsr_el[el] |= 1 << select;
56
66
+#ifdef CONFIG_USER_ONLY
57
DO_ZPZZ_FP(FADDP, aa64_sve2, sve2_faddp_zpzz)
67
+ /*
58
DO_ZPZZ_FP(FMAXNMP, aa64_sve2, sve2_fmaxnmp_zpzz)
68
+ * Stand in for a timer irq, setting _TIF_MTE_ASYNC_FAULT,
59
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(SQRDCMLAH_zzzz, aa64_sve2, gen_gvec_ool_zzzz,
69
+ * which then sends a SIGSEGV when the thread is next scheduled.
60
TRANS_FEAT(USDOT_zzzz, aa64_sve_i8mm, gen_gvec_ool_arg_zzzz,
70
+ * This cpu will return to the main loop at the end of the TB,
61
a->esz == 2 ? gen_helper_gvec_usdot_b : NULL, a, 0)
71
+ * which is rather sooner than "normal". But the alternative
62
72
+ * is waiting until the next syscall.
63
-TRANS_FEAT(AESMC, aa64_sve2_aes, gen_gvec_ool_zz,
73
+ */
64
- gen_helper_crypto_aesmc, a->rd, a->rd, a->decrypt)
74
+ qemu_cpu_kick(env_cpu(env));
65
+TRANS_FEAT_NONSTREAMING(AESMC, aa64_sve2_aes, gen_gvec_ool_zz,
75
+#endif
66
+ gen_helper_crypto_aesmc, a->rd, a->rd, a->decrypt)
76
break;
67
77
68
-TRANS_FEAT(AESE, aa64_sve2_aes, gen_gvec_ool_arg_zzz,
78
default:
69
- gen_helper_crypto_aese, a, false)
70
-TRANS_FEAT(AESD, aa64_sve2_aes, gen_gvec_ool_arg_zzz,
71
- gen_helper_crypto_aese, a, true)
72
+TRANS_FEAT_NONSTREAMING(AESE, aa64_sve2_aes, gen_gvec_ool_arg_zzz,
73
+ gen_helper_crypto_aese, a, false)
74
+TRANS_FEAT_NONSTREAMING(AESD, aa64_sve2_aes, gen_gvec_ool_arg_zzz,
75
+ gen_helper_crypto_aese, a, true)
76
77
-TRANS_FEAT(SM4E, aa64_sve2_sm4, gen_gvec_ool_arg_zzz,
78
- gen_helper_crypto_sm4e, a, 0)
79
-TRANS_FEAT(SM4EKEY, aa64_sve2_sm4, gen_gvec_ool_arg_zzz,
80
- gen_helper_crypto_sm4ekey, a, 0)
81
+TRANS_FEAT_NONSTREAMING(SM4E, aa64_sve2_sm4, gen_gvec_ool_arg_zzz,
82
+ gen_helper_crypto_sm4e, a, 0)
83
+TRANS_FEAT_NONSTREAMING(SM4EKEY, aa64_sve2_sm4, gen_gvec_ool_arg_zzz,
84
+ gen_helper_crypto_sm4ekey, a, 0)
85
86
-TRANS_FEAT(RAX1, aa64_sve2_sha3, gen_gvec_fn_arg_zzz, gen_gvec_rax1, a)
87
+TRANS_FEAT_NONSTREAMING(RAX1, aa64_sve2_sha3, gen_gvec_fn_arg_zzz,
88
+ gen_gvec_rax1, a)
89
90
TRANS_FEAT(FCVTNT_sh, aa64_sve2, gen_gvec_fpst_arg_zpz,
91
gen_helper_sve2_fcvtnt_sh, a, 0, FPST_FPCR)
79
--
92
--
80
2.20.1
93
2.25.1
81
82
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
A proper syndrome is required to fill in the proper si_code.
3
Mark these as a non-streaming instructions, which should trap
4
Use page_get_flags to determine permission vs translation for user-only.
4
if full a64 support is not enabled in streaming mode.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-27-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-12-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
linux-user/aarch64/cpu_loop.c | 24 +++++++++++++++++++++---
11
target/arm/sme-fa64.decode | 9 ---------
12
target/arm/tlb_helper.c | 15 +++++++++------
12
target/arm/translate-sve.c | 6 ++++++
13
2 files changed, 30 insertions(+), 9 deletions(-)
13
2 files changed, 6 insertions(+), 9 deletions(-)
14
14
15
diff --git a/linux-user/aarch64/cpu_loop.c b/linux-user/aarch64/cpu_loop.c
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/linux-user/aarch64/cpu_loop.c
17
--- a/target/arm/sme-fa64.decode
18
+++ b/linux-user/aarch64/cpu_loop.c
18
+++ b/target/arm/sme-fa64.decode
19
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
20
#include "cpu_loop-common.h"
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
21
#include "qemu/guest-random.h"
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
22
#include "hw/semihosting/common-semi.h"
22
23
+#include "target/arm/syndrome.h"
23
-FAIL 1000 010- -00- ---- 10-- ---- ---- ---- # SVE2 32-bit gather NT load (vector+scalar)
24
24
FAIL 1000 010- -00- ---- 111- ---- ---- ---- # SVE 32-bit gather prefetch (vector+imm)
25
#define get_user_code_u32(x, gaddr, env) \
25
FAIL 1000 0100 0-1- ---- 0--- ---- ---- ---- # SVE 32-bit gather prefetch (scalar+vector)
26
({ abi_long __r = get_user_u32((x), (gaddr)); \
26
-FAIL 1000 010- -01- ---- 1--- ---- ---- ---- # SVE 32-bit gather load (vector+imm)
27
@@ -XXX,XX +XXX,XX @@
27
-FAIL 1000 0100 0-0- ---- 0--- ---- ---- ---- # SVE 32-bit gather load byte (scalar+vector)
28
void cpu_loop(CPUARMState *env)
28
-FAIL 1000 0100 1--- ---- 0--- ---- ---- ---- # SVE 32-bit gather load half (scalar+vector)
29
{
29
-FAIL 1000 0101 0--- ---- 0--- ---- ---- ---- # SVE 32-bit gather load word (scalar+vector)
30
CPUState *cs = env_cpu(env);
30
FAIL 1010 010- ---- ---- 011- ---- ---- ---- # SVE contiguous FF load (scalar+scalar)
31
- int trapnr;
31
FAIL 1010 010- ---1 ---- 101- ---- ---- ---- # SVE contiguous NF load (scalar+imm)
32
+ int trapnr, ec, fsc;
32
FAIL 1010 010- -01- ---- 000- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+scalar)
33
abi_long ret;
33
FAIL 1010 010- -010 ---- 001- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+imm)
34
target_siginfo_t info;
34
FAIL 1100 010- ---- ---- ---- ---- ---- ---- # SVE 64-bit gather load/prefetch
35
35
-FAIL 1110 010- -00- ---- 001- ---- ---- ---- # SVE2 64-bit scatter NT store (vector+scalar)
36
@@ -XXX,XX +XXX,XX @@ void cpu_loop(CPUARMState *env)
36
-FAIL 1110 010- -10- ---- 001- ---- ---- ---- # SVE2 32-bit scatter NT store (vector+scalar)
37
case EXCP_DATA_ABORT:
37
-FAIL 1110 010- ---- ---- 1-0- ---- ---- ---- # SVE scatter store (scalar+32-bit vector)
38
info.si_signo = TARGET_SIGSEGV;
38
-FAIL 1110 010- ---- ---- 101- ---- ---- ---- # SVE scatter store (misc)
39
info.si_errno = 0;
39
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
40
- /* XXX: check env->error_code */
41
- info.si_code = TARGET_SEGV_MAPERR;
42
info._sifields._sigfault._addr = env->exception.vaddress;
43
+
44
+ /* We should only arrive here with EC in {DATAABORT, INSNABORT}. */
45
+ ec = syn_get_ec(env->exception.syndrome);
46
+ assert(ec == EC_DATAABORT || ec == EC_INSNABORT);
47
+
48
+ /* Both EC have the same format for FSC, or close enough. */
49
+ fsc = extract32(env->exception.syndrome, 0, 6);
50
+ switch (fsc) {
51
+ case 0x04 ... 0x07: /* Translation fault, level {0-3} */
52
+ info.si_code = TARGET_SEGV_MAPERR;
53
+ break;
54
+ case 0x09 ... 0x0b: /* Access flag fault, level {1-3} */
55
+ case 0x0d ... 0x0f: /* Permission fault, level {1-3} */
56
+ info.si_code = TARGET_SEGV_ACCERR;
57
+ break;
58
+ default:
59
+ g_assert_not_reached();
60
+ }
61
+
62
queue_signal(env, info.si_signo, QEMU_SI_FAULT, &info);
63
break;
64
case EXCP_DEBUG:
65
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
66
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
67
--- a/target/arm/tlb_helper.c
41
--- a/target/arm/translate-sve.c
68
+++ b/target/arm/tlb_helper.c
42
+++ b/target/arm/translate-sve.c
69
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
43
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
70
bool probe, uintptr_t retaddr)
44
if (!dc_isar_feature(aa64_sve, s)) {
71
{
45
return false;
72
ARMCPU *cpu = ARM_CPU(cs);
73
+ ARMMMUFaultInfo fi = {};
74
75
#ifdef CONFIG_USER_ONLY
76
- cpu->env.exception.vaddress = address;
77
- if (access_type == MMU_INST_FETCH) {
78
- cs->exception_index = EXCP_PREFETCH_ABORT;
79
+ int flags = page_get_flags(useronly_clean_ptr(address));
80
+ if (flags & PAGE_VALID) {
81
+ fi.type = ARMFault_Permission;
82
} else {
83
- cs->exception_index = EXCP_DATA_ABORT;
84
+ fi.type = ARMFault_Translation;
85
}
46
}
86
- cpu_loop_exit_restore(cs, retaddr);
47
+ s->is_nonstreaming = true;
87
+
48
if (!sve_access_check(s)) {
88
+ /* now we have a real cpu fault */
49
return true;
89
+ cpu_restore_state(cs, retaddr, true);
50
}
90
+ arm_deliver_fault(cpu, address, access_type, mmu_idx, &fi);
51
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
91
#else
52
if (!dc_isar_feature(aa64_sve, s)) {
92
hwaddr phys_addr;
53
return false;
93
target_ulong page_size;
54
}
94
int prot, ret;
55
+ s->is_nonstreaming = true;
95
MemTxAttrs attrs = {};
56
if (!sve_access_check(s)) {
96
- ARMMMUFaultInfo fi = {};
57
return true;
97
ARMCacheAttrs cacheattrs = {};
58
}
98
59
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNT1_zprz(DisasContext *s, arg_LD1_zprz *a)
99
/*
60
if (!dc_isar_feature(aa64_sve2, s)) {
61
return false;
62
}
63
+ s->is_nonstreaming = true;
64
if (!sve_access_check(s)) {
65
return true;
66
}
67
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
68
if (!dc_isar_feature(aa64_sve, s)) {
69
return false;
70
}
71
+ s->is_nonstreaming = true;
72
if (!sve_access_check(s)) {
73
return true;
74
}
75
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
76
if (!dc_isar_feature(aa64_sve, s)) {
77
return false;
78
}
79
+ s->is_nonstreaming = true;
80
if (!sve_access_check(s)) {
81
return true;
82
}
83
@@ -XXX,XX +XXX,XX @@ static bool trans_STNT1_zprz(DisasContext *s, arg_ST1_zprz *a)
84
if (!dc_isar_feature(aa64_sve2, s)) {
85
return false;
86
}
87
+ s->is_nonstreaming = true;
88
if (!sve_access_check(s)) {
89
return true;
90
}
100
--
91
--
101
2.20.1
92
2.25.1
102
103
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This data can be allocated by page_alloc_target_data() and
3
Mark these as a non-streaming instructions, which should trap if full
4
released by page_set_flags(start, end, prot | PAGE_RESET).
4
a64 support is not enabled in streaming mode. In this case, introduce
5
5
PRF_ns (prefetch non-streaming) to handle the checks.
6
This data will be used to hold tag memory for AArch64 MTE.
7
6
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20210212184902.1251044-2-richard.henderson@linaro.org
9
Message-id: 20220708151540.18136-13-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
11
---
13
include/exec/cpu-all.h | 42 +++++++++++++++++++++++++++++++++------
12
target/arm/sme-fa64.decode | 3 ---
14
accel/tcg/translate-all.c | 28 ++++++++++++++++++++++++++
13
target/arm/sve.decode | 10 +++++-----
15
linux-user/mmap.c | 4 +++-
14
target/arm/translate-sve.c | 11 +++++++++++
16
linux-user/syscall.c | 4 ++--
15
3 files changed, 16 insertions(+), 8 deletions(-)
17
4 files changed, 69 insertions(+), 9 deletions(-)
18
16
19
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
17
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
20
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
21
--- a/include/exec/cpu-all.h
19
--- a/target/arm/sme-fa64.decode
22
+++ b/include/exec/cpu-all.h
20
+++ b/target/arm/sme-fa64.decode
23
@@ -XXX,XX +XXX,XX @@ extern intptr_t qemu_host_page_mask;
21
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
24
#define PAGE_EXEC 0x0004
22
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
25
#define PAGE_BITS (PAGE_READ | PAGE_WRITE | PAGE_EXEC)
23
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
26
#define PAGE_VALID 0x0008
24
27
-/* original state of the write flag (used when tracking self-modifying
25
-FAIL 1000 010- -00- ---- 111- ---- ---- ---- # SVE 32-bit gather prefetch (vector+imm)
28
- code */
26
-FAIL 1000 0100 0-1- ---- 0--- ---- ---- ---- # SVE 32-bit gather prefetch (scalar+vector)
29
+/*
27
FAIL 1010 010- ---- ---- 011- ---- ---- ---- # SVE contiguous FF load (scalar+scalar)
30
+ * Original state of the write flag (used when tracking self-modifying code)
28
FAIL 1010 010- ---1 ---- 101- ---- ---- ---- # SVE contiguous NF load (scalar+imm)
31
+ */
29
FAIL 1010 010- -01- ---- 000- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+scalar)
32
#define PAGE_WRITE_ORG 0x0010
30
FAIL 1010 010- -010 ---- 001- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+imm)
33
-/* Invalidate the TLB entry immediately, helpful for s390x
31
-FAIL 1100 010- ---- ---- ---- ---- ---- ---- # SVE 64-bit gather load/prefetch
34
- * Low-Address-Protection. Used with PAGE_WRITE in tlb_set_page_with_attrs() */
32
diff --git a/target/arm/sve.decode b/target/arm/sve.decode
35
-#define PAGE_WRITE_INV 0x0040
36
+/*
37
+ * Invalidate the TLB entry immediately, helpful for s390x
38
+ * Low-Address-Protection. Used with PAGE_WRITE in tlb_set_page_with_attrs()
39
+ */
40
+#define PAGE_WRITE_INV 0x0020
41
+/* For use with page_set_flags: page is being replaced; target_data cleared. */
42
+#define PAGE_RESET 0x0040
43
+
44
#if defined(CONFIG_BSD) && defined(CONFIG_USER_ONLY)
45
/* FIXME: Code that sets/uses this is broken and needs to go away. */
46
-#define PAGE_RESERVED 0x0020
47
+#define PAGE_RESERVED 0x0100
48
#endif
49
/* Target-specific bits that will be used via page_get_flags(). */
50
#define PAGE_TARGET_1 0x0080
51
@@ -XXX,XX +XXX,XX @@ int walk_memory_regions(void *, walk_memory_regions_fn);
52
int page_get_flags(target_ulong address);
53
void page_set_flags(target_ulong start, target_ulong end, int flags);
54
int page_check_range(target_ulong start, target_ulong len, int flags);
55
+
56
+/**
57
+ * page_alloc_target_data(address, size)
58
+ * @address: guest virtual address
59
+ * @size: size of data to allocate
60
+ *
61
+ * Allocate @size bytes of out-of-band data to associate with the
62
+ * guest page at @address. If the page is not mapped, NULL will
63
+ * be returned. If there is existing data associated with @address,
64
+ * no new memory will be allocated.
65
+ *
66
+ * The memory will be freed when the guest page is deallocated,
67
+ * e.g. with the munmap system call.
68
+ */
69
+void *page_alloc_target_data(target_ulong address, size_t size);
70
+
71
+/**
72
+ * page_get_target_data(address)
73
+ * @address: guest virtual address
74
+ *
75
+ * Return any out-of-bound memory assocated with the guest page
76
+ * at @address, as per page_alloc_target_data.
77
+ */
78
+void *page_get_target_data(target_ulong address);
79
#endif
80
81
CPUArchState *cpu_copy(CPUArchState *env);
82
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
83
index XXXXXXX..XXXXXXX 100644
33
index XXXXXXX..XXXXXXX 100644
84
--- a/accel/tcg/translate-all.c
34
--- a/target/arm/sve.decode
85
+++ b/accel/tcg/translate-all.c
35
+++ b/target/arm/sve.decode
86
@@ -XXX,XX +XXX,XX @@ typedef struct PageDesc {
36
@@ -XXX,XX +XXX,XX @@ LD1RO_zpri 1010010 .. 01 0.... 001 ... ..... ..... \
87
unsigned int code_write_count;
37
@rpri_load_msz nreg=0
88
#else
38
89
unsigned long flags;
39
# SVE 32-bit gather prefetch (scalar plus 32-bit scaled offsets)
90
+ void *target_data;
40
-PRF 1000010 00 -1 ----- 0-- --- ----- 0 ----
91
#endif
41
+PRF_ns 1000010 00 -1 ----- 0-- --- ----- 0 ----
92
#ifndef CONFIG_USER_ONLY
42
93
QemuSpin lock;
43
# SVE 32-bit gather prefetch (vector plus immediate)
94
@@ -XXX,XX +XXX,XX @@ int page_get_flags(target_ulong address)
44
-PRF 1000010 -- 00 ----- 111 --- ----- 0 ----
95
void page_set_flags(target_ulong start, target_ulong end, int flags)
45
+PRF_ns 1000010 -- 00 ----- 111 --- ----- 0 ----
96
{
46
97
target_ulong addr, len;
47
# SVE contiguous prefetch (scalar plus immediate)
98
+ bool reset_target_data;
48
PRF 1000010 11 1- ----- 0-- --- ----- 0 ----
99
49
@@ -XXX,XX +XXX,XX @@ LD1_zpiz 1100010 .. 01 ..... 1.. ... ..... ..... \
100
/* This function should never be called with addresses outside the
50
@rpri_g_load esz=3
101
guest address space. If this assert fires, it probably indicates
51
102
@@ -XXX,XX +XXX,XX @@ void page_set_flags(target_ulong start, target_ulong end, int flags)
52
# SVE 64-bit gather prefetch (scalar plus 64-bit scaled offsets)
103
if (flags & PAGE_WRITE) {
53
-PRF 1100010 00 11 ----- 1-- --- ----- 0 ----
104
flags |= PAGE_WRITE_ORG;
54
+PRF_ns 1100010 00 11 ----- 1-- --- ----- 0 ----
105
}
55
106
+ reset_target_data = !(flags & PAGE_VALID) || (flags & PAGE_RESET);
56
# SVE 64-bit gather prefetch (scalar plus unpacked 32-bit scaled offsets)
107
+ flags &= ~PAGE_RESET;
57
-PRF 1100010 00 -1 ----- 0-- --- ----- 0 ----
108
58
+PRF_ns 1100010 00 -1 ----- 0-- --- ----- 0 ----
109
for (addr = start, len = end - start;
59
110
len != 0;
60
# SVE 64-bit gather prefetch (vector plus immediate)
111
@@ -XXX,XX +XXX,XX @@ void page_set_flags(target_ulong start, target_ulong end, int flags)
61
-PRF 1100010 -- 00 ----- 111 --- ----- 0 ----
112
p->first_tb) {
62
+PRF_ns 1100010 -- 00 ----- 111 --- ----- 0 ----
113
tb_invalidate_phys_page(addr, 0);
63
114
}
64
### SVE Memory Store Group
115
+ if (reset_target_data && p->target_data) {
65
116
+ g_free(p->target_data);
66
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
117
+ p->target_data = NULL;
67
index XXXXXXX..XXXXXXX 100644
118
+ }
68
--- a/target/arm/translate-sve.c
119
p->flags = flags;
69
+++ b/target/arm/translate-sve.c
120
}
70
@@ -XXX,XX +XXX,XX @@ static bool trans_PRF_rr(DisasContext *s, arg_PRF_rr *a)
71
return true;
121
}
72
}
122
73
123
+void *page_get_target_data(target_ulong address)
74
+static bool trans_PRF_ns(DisasContext *s, arg_PRF_ns *a)
124
+{
75
+{
125
+ PageDesc *p = page_find(address >> TARGET_PAGE_BITS);
76
+ if (!dc_isar_feature(aa64_sve, s)) {
126
+ return p ? p->target_data : NULL;
77
+ return false;
78
+ }
79
+ /* Prefetch is a nop within QEMU. */
80
+ s->is_nonstreaming = true;
81
+ (void)sve_access_check(s);
82
+ return true;
127
+}
83
+}
128
+
84
+
129
+void *page_alloc_target_data(target_ulong address, size_t size)
85
/*
130
+{
86
* Move Prefix
131
+ PageDesc *p = page_find(address >> TARGET_PAGE_BITS);
87
*
132
+ void *ret = NULL;
133
+
134
+ if (p->flags & PAGE_VALID) {
135
+ ret = p->target_data;
136
+ if (!ret) {
137
+ p->target_data = ret = g_malloc0(size);
138
+ }
139
+ }
140
+ return ret;
141
+}
142
+
143
int page_check_range(target_ulong start, target_ulong len, int flags)
144
{
145
PageDesc *p;
146
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
147
index XXXXXXX..XXXXXXX 100644
148
--- a/linux-user/mmap.c
149
+++ b/linux-user/mmap.c
150
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
151
}
152
}
153
the_end1:
154
+ page_flags |= PAGE_RESET;
155
page_set_flags(start, start + len, page_flags);
156
the_end:
157
trace_target_mmap_complete(start);
158
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
159
new_addr = h2g(host_addr);
160
prot = page_get_flags(old_addr);
161
page_set_flags(old_addr, old_addr + old_size, 0);
162
- page_set_flags(new_addr, new_addr + new_size, prot | PAGE_VALID);
163
+ page_set_flags(new_addr, new_addr + new_size,
164
+ prot | PAGE_VALID | PAGE_RESET);
165
}
166
tb_invalidate_phys_range(new_addr, new_addr + new_size);
167
mmap_unlock();
168
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
169
index XXXXXXX..XXXXXXX 100644
170
--- a/linux-user/syscall.c
171
+++ b/linux-user/syscall.c
172
@@ -XXX,XX +XXX,XX @@ static inline abi_ulong do_shmat(CPUArchState *cpu_env,
173
raddr=h2g((unsigned long)host_raddr);
174
175
page_set_flags(raddr, raddr + shm_info.shm_segsz,
176
- PAGE_VALID | PAGE_READ |
177
- ((shmflg & SHM_RDONLY)? 0 : PAGE_WRITE));
178
+ PAGE_VALID | PAGE_RESET | PAGE_READ |
179
+ (shmflg & SHM_RDONLY ? 0 : PAGE_WRITE));
180
181
for (i = 0; i < N_SHM_REGIONS; i++) {
182
if (!shm_regions[i].in_use) {
183
--
88
--
184
2.20.1
89
2.25.1
185
186
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We're currently open-coding the range check in access_ok;
3
Mark these as a non-streaming instructions, which should trap
4
use guest_range_valid when size != 0.
4
if full a64 support is not enabled in streaming mode.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-15-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-14-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
linux-user/qemu.h | 9 +++------
11
target/arm/sme-fa64.decode | 2 --
12
1 file changed, 3 insertions(+), 6 deletions(-)
12
target/arm/translate-sve.c | 2 ++
13
2 files changed, 2 insertions(+), 2 deletions(-)
13
14
14
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/qemu.h
17
--- a/target/arm/sme-fa64.decode
17
+++ b/linux-user/qemu.h
18
+++ b/target/arm/sme-fa64.decode
18
@@ -XXX,XX +XXX,XX @@ extern unsigned long guest_stack_size;
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
19
20
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
20
static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
21
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
21
{
22
22
- if (!guest_addr_valid(addr)) {
23
-FAIL 1010 010- ---- ---- 011- ---- ---- ---- # SVE contiguous FF load (scalar+scalar)
23
- return false;
24
-FAIL 1010 010- ---1 ---- 101- ---- ---- ---- # SVE contiguous NF load (scalar+imm)
24
- }
25
FAIL 1010 010- -01- ---- 000- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+scalar)
25
- if (size != 0 &&
26
FAIL 1010 010- -010 ---- 001- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+imm)
26
- (addr + size - 1 < addr ||
27
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
27
- !guest_addr_valid(addr + size - 1))) {
28
index XXXXXXX..XXXXXXX 100644
28
+ if (size == 0
29
--- a/target/arm/translate-sve.c
29
+ ? !guest_addr_valid(addr)
30
+++ b/target/arm/translate-sve.c
30
+ : !guest_range_valid(addr, size)) {
31
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
32
if (!dc_isar_feature(aa64_sve, s)) {
31
return false;
33
return false;
32
}
34
}
33
return page_check_range((target_ulong)addr, size, type) == 0;
35
+ s->is_nonstreaming = true;
36
if (sve_access_check(s)) {
37
TCGv_i64 addr = new_tmp_a64(s);
38
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
39
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
40
if (!dc_isar_feature(aa64_sve, s)) {
41
return false;
42
}
43
+ s->is_nonstreaming = true;
44
if (sve_access_check(s)) {
45
int vsz = vec_full_reg_size(s);
46
int elements = vsz >> dtype_esz[a->dtype];
34
--
47
--
35
2.20.1
48
2.25.1
36
37
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
These constants are only ever used with access_ok, and friends.
3
Mark these as a non-streaming instructions, which should trap
4
Rather than translating them to PAGE_* bits, let them equal
4
if full a64 support is not enabled in streaming mode.
5
the PAGE_* bits to begin.
6
5
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20210212184902.1251044-8-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-15-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
linux-user/qemu.h | 8 +++-----
11
target/arm/sme-fa64.decode | 3 ---
13
1 file changed, 3 insertions(+), 5 deletions(-)
12
target/arm/translate-sve.c | 2 ++
13
2 files changed, 2 insertions(+), 3 deletions(-)
14
14
15
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
15
diff --git a/target/arm/sme-fa64.decode b/target/arm/sme-fa64.decode
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/linux-user/qemu.h
17
--- a/target/arm/sme-fa64.decode
18
+++ b/linux-user/qemu.h
18
+++ b/target/arm/sme-fa64.decode
19
@@ -XXX,XX +XXX,XX @@ extern unsigned long guest_stack_size;
19
@@ -XXX,XX +XXX,XX @@ FAIL 0001 1110 0111 1110 0000 00-- ---- ---- # FJCVTZS
20
20
# --11 1100 --0- ---- ---- ---- ---- ---- # Load/store FP register (unscaled imm)
21
/* user access */
21
# --11 1100 --1- ---- ---- ---- ---- --10 # Load/store FP register (register offset)
22
22
# --11 1101 ---- ---- ---- ---- ---- ---- # Load/store FP register (scaled imm)
23
-#define VERIFY_READ 0
23
-
24
-#define VERIFY_WRITE 1 /* implies read access */
24
-FAIL 1010 010- -01- ---- 000- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+scalar)
25
+#define VERIFY_READ PAGE_READ
25
-FAIL 1010 010- -010 ---- 001- ---- ---- ---- # SVE load & replicate 32 bytes (scalar+imm)
26
+#define VERIFY_WRITE (PAGE_READ | PAGE_WRITE)
26
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
27
27
index XXXXXXX..XXXXXXX 100644
28
static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
28
--- a/target/arm/translate-sve.c
29
{
29
+++ b/target/arm/translate-sve.c
30
@@ -XXX,XX +XXX,XX @@ static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
30
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RO_zprr(DisasContext *s, arg_rprr_load *a)
31
!guest_addr_valid(addr + size - 1))) {
31
if (a->rm == 31) {
32
return false;
32
return false;
33
}
33
}
34
- return page_check_range((target_ulong)addr, size,
34
+ s->is_nonstreaming = true;
35
- (type == VERIFY_READ) ? PAGE_READ :
35
if (sve_access_check(s)) {
36
- (PAGE_READ | PAGE_WRITE)) == 0;
36
TCGv_i64 addr = new_tmp_a64(s);
37
+ return page_check_range((target_ulong)addr, size, type) == 0;
37
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
38
}
38
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RO_zpri(DisasContext *s, arg_rpri_load *a)
39
39
if (!dc_isar_feature(aa64_sve_f64mm, s)) {
40
/* NOTE __get_user and __put_user use host pointers and don't check access.
40
return false;
41
}
42
+ s->is_nonstreaming = true;
43
if (sve_access_check(s)) {
44
TCGv_i64 addr = new_tmp_a64(s);
45
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), a->imm * 32);
41
--
46
--
42
2.20.1
47
2.25.1
43
44
diff view generated by jsdifflib
1
From: Doug Evans <dje@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is a 10/100 ethernet device that has several features.
3
These functions will be used to verify that the cpu
4
Only the ones needed by the Linux driver have been implemented.
4
is in the correct state for a given instruction.
5
See npcm7xx_emc.c for a list of unimplemented features.
6
5
7
Reviewed-by: Hao Wu <wuhaotsh@google.com>
8
Reviewed-by: Avi Fishman <avi.fishman@nuvoton.com>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Doug Evans <dje@google.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20210213002520.1374134-2-dje@google.com
8
Message-id: 20220708151540.18136-16-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
10
---
14
include/hw/net/npcm7xx_emc.h | 286 ++++++++++++
11
target/arm/translate-a64.h | 21 +++++++++++++++++++++
15
hw/net/npcm7xx_emc.c | 857 +++++++++++++++++++++++++++++++++++
12
target/arm/translate-a64.c | 34 ++++++++++++++++++++++++++++++++++
16
hw/net/meson.build | 1 +
13
2 files changed, 55 insertions(+)
17
hw/net/trace-events | 17 +
18
4 files changed, 1161 insertions(+)
19
create mode 100644 include/hw/net/npcm7xx_emc.h
20
create mode 100644 hw/net/npcm7xx_emc.c
21
14
22
diff --git a/include/hw/net/npcm7xx_emc.h b/include/hw/net/npcm7xx_emc.h
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
23
new file mode 100644
16
index XXXXXXX..XXXXXXX 100644
24
index XXXXXXX..XXXXXXX
17
--- a/target/arm/translate-a64.h
25
--- /dev/null
18
+++ b/target/arm/translate-a64.h
26
+++ b/include/hw/net/npcm7xx_emc.h
19
@@ -XXX,XX +XXX,XX @@ void write_fp_dreg(DisasContext *s, int reg, TCGv_i64 v);
27
@@ -XXX,XX +XXX,XX @@
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
28
+/*
21
unsigned int imms, unsigned int immr);
29
+ * Nuvoton NPCM7xx EMC Module
22
bool sve_access_check(DisasContext *s);
30
+ *
23
+bool sme_enabled_check(DisasContext *s);
31
+ * Copyright 2020 Google LLC
24
+bool sme_enabled_check_with_svcr(DisasContext *s, unsigned);
32
+ *
33
+ * This program is free software; you can redistribute it and/or modify it
34
+ * under the terms of the GNU General Public License as published by the
35
+ * Free Software Foundation; either version 2 of the License, or
36
+ * (at your option) any later version.
37
+ *
38
+ * This program is distributed in the hope that it will be useful, but WITHOUT
39
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
40
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
41
+ * for more details.
42
+ */
43
+
25
+
44
+#ifndef NPCM7XX_EMC_H
26
+/* This function corresponds to CheckStreamingSVEEnabled. */
45
+#define NPCM7XX_EMC_H
27
+static inline bool sme_sm_enabled_check(DisasContext *s)
46
+
47
+#include "hw/irq.h"
48
+#include "hw/sysbus.h"
49
+#include "net/net.h"
50
+
51
+/* 32-bit register indices. */
52
+enum NPCM7xxPWMRegister {
53
+ /* Control registers. */
54
+ REG_CAMCMR,
55
+ REG_CAMEN,
56
+
57
+ /* There are 16 CAMn[ML] registers. */
58
+ REG_CAMM_BASE,
59
+ REG_CAML_BASE,
60
+ REG_CAMML_LAST = 0x21,
61
+
62
+ REG_TXDLSA = 0x22,
63
+ REG_RXDLSA,
64
+ REG_MCMDR,
65
+ REG_MIID,
66
+ REG_MIIDA,
67
+ REG_FFTCR,
68
+ REG_TSDR,
69
+ REG_RSDR,
70
+ REG_DMARFC,
71
+ REG_MIEN,
72
+
73
+ /* Status registers. */
74
+ REG_MISTA,
75
+ REG_MGSTA,
76
+ REG_MPCNT,
77
+ REG_MRPC,
78
+ REG_MRPCC,
79
+ REG_MREPC,
80
+ REG_DMARFS,
81
+ REG_CTXDSA,
82
+ REG_CTXBSA,
83
+ REG_CRXDSA,
84
+ REG_CRXBSA,
85
+
86
+ NPCM7XX_NUM_EMC_REGS,
87
+};
88
+
89
+/* REG_CAMCMR fields */
90
+/* Enable CAM Compare */
91
+#define REG_CAMCMR_ECMP (1 << 4)
92
+/* Complement CAM Compare */
93
+#define REG_CAMCMR_CCAM (1 << 3)
94
+/* Accept Broadcast Packet */
95
+#define REG_CAMCMR_ABP (1 << 2)
96
+/* Accept Multicast Packet */
97
+#define REG_CAMCMR_AMP (1 << 1)
98
+/* Accept Unicast Packet */
99
+#define REG_CAMCMR_AUP (1 << 0)
100
+
101
+/* REG_MCMDR fields */
102
+/* Software Reset */
103
+#define REG_MCMDR_SWR (1 << 24)
104
+/* Internal Loopback Select */
105
+#define REG_MCMDR_LBK (1 << 21)
106
+/* Operation Mode Select */
107
+#define REG_MCMDR_OPMOD (1 << 20)
108
+/* Enable MDC Clock Generation */
109
+#define REG_MCMDR_ENMDC (1 << 19)
110
+/* Full-Duplex Mode Select */
111
+#define REG_MCMDR_FDUP (1 << 18)
112
+/* Enable SQE Checking */
113
+#define REG_MCMDR_ENSEQ (1 << 17)
114
+/* Send PAUSE Frame */
115
+#define REG_MCMDR_SDPZ (1 << 16)
116
+/* No Defer */
117
+#define REG_MCMDR_NDEF (1 << 9)
118
+/* Frame Transmission On */
119
+#define REG_MCMDR_TXON (1 << 8)
120
+/* Strip CRC Checksum */
121
+#define REG_MCMDR_SPCRC (1 << 5)
122
+/* Accept CRC Error Packet */
123
+#define REG_MCMDR_AEP (1 << 4)
124
+/* Accept Control Packet */
125
+#define REG_MCMDR_ACP (1 << 3)
126
+/* Accept Runt Packet */
127
+#define REG_MCMDR_ARP (1 << 2)
128
+/* Accept Long Packet */
129
+#define REG_MCMDR_ALP (1 << 1)
130
+/* Frame Reception On */
131
+#define REG_MCMDR_RXON (1 << 0)
132
+
133
+/* REG_MIEN fields */
134
+/* Enable Transmit Descriptor Unavailable Interrupt */
135
+#define REG_MIEN_ENTDU (1 << 23)
136
+/* Enable Transmit Completion Interrupt */
137
+#define REG_MIEN_ENTXCP (1 << 18)
138
+/* Enable Transmit Interrupt */
139
+#define REG_MIEN_ENTXINTR (1 << 16)
140
+/* Enable Receive Descriptor Unavailable Interrupt */
141
+#define REG_MIEN_ENRDU (1 << 10)
142
+/* Enable Receive Good Interrupt */
143
+#define REG_MIEN_ENRXGD (1 << 4)
144
+/* Enable Receive Interrupt */
145
+#define REG_MIEN_ENRXINTR (1 << 0)
146
+
147
+/* REG_MISTA fields */
148
+/* TODO: Add error fields and support simulated errors? */
149
+/* Transmit Bus Error Interrupt */
150
+#define REG_MISTA_TXBERR (1 << 24)
151
+/* Transmit Descriptor Unavailable Interrupt */
152
+#define REG_MISTA_TDU (1 << 23)
153
+/* Transmit Completion Interrupt */
154
+#define REG_MISTA_TXCP (1 << 18)
155
+/* Transmit Interrupt */
156
+#define REG_MISTA_TXINTR (1 << 16)
157
+/* Receive Bus Error Interrupt */
158
+#define REG_MISTA_RXBERR (1 << 11)
159
+/* Receive Descriptor Unavailable Interrupt */
160
+#define REG_MISTA_RDU (1 << 10)
161
+/* DMA Early Notification Interrupt */
162
+#define REG_MISTA_DENI (1 << 9)
163
+/* Maximum Frame Length Interrupt */
164
+#define REG_MISTA_DFOI (1 << 8)
165
+/* Receive Good Interrupt */
166
+#define REG_MISTA_RXGD (1 << 4)
167
+/* Packet Too Long Interrupt */
168
+#define REG_MISTA_PTLE (1 << 3)
169
+/* Receive Interrupt */
170
+#define REG_MISTA_RXINTR (1 << 0)
171
+
172
+/* REG_MGSTA fields */
173
+/* Transmission Halted */
174
+#define REG_MGSTA_TXHA (1 << 11)
175
+/* Receive Halted */
176
+#define REG_MGSTA_RXHA (1 << 11)
177
+
178
+/* REG_DMARFC fields */
179
+/* Maximum Receive Frame Length */
180
+#define REG_DMARFC_RXMS(word) extract32((word), 0, 16)
181
+
182
+/* REG MIIDA fields */
183
+/* Busy Bit */
184
+#define REG_MIIDA_BUSY (1 << 17)
185
+
186
+/* Transmit and receive descriptors */
187
+typedef struct NPCM7xxEMCTxDesc NPCM7xxEMCTxDesc;
188
+typedef struct NPCM7xxEMCRxDesc NPCM7xxEMCRxDesc;
189
+
190
+struct NPCM7xxEMCTxDesc {
191
+ uint32_t flags;
192
+ uint32_t txbsa;
193
+ uint32_t status_and_length;
194
+ uint32_t ntxdsa;
195
+};
196
+
197
+struct NPCM7xxEMCRxDesc {
198
+ uint32_t status_and_length;
199
+ uint32_t rxbsa;
200
+ uint32_t reserved;
201
+ uint32_t nrxdsa;
202
+};
203
+
204
+/* NPCM7xxEMCTxDesc.flags values */
205
+/* Owner: 0 = cpu, 1 = emc */
206
+#define TX_DESC_FLAG_OWNER_MASK (1 << 31)
207
+/* Transmit interrupt enable */
208
+#define TX_DESC_FLAG_INTEN (1 << 2)
209
+/* CRC append */
210
+#define TX_DESC_FLAG_CRCAPP (1 << 1)
211
+/* Padding enable */
212
+#define TX_DESC_FLAG_PADEN (1 << 0)
213
+
214
+/* NPCM7xxEMCTxDesc.status_and_length values */
215
+/* Collision count */
216
+#define TX_DESC_STATUS_CCNT_SHIFT 28
217
+#define TX_DESC_STATUS_CCNT_BITSIZE 4
218
+/* SQE error */
219
+#define TX_DESC_STATUS_SQE (1 << 26)
220
+/* Transmission paused */
221
+#define TX_DESC_STATUS_PAU (1 << 25)
222
+/* P transmission halted */
223
+#define TX_DESC_STATUS_TXHA (1 << 24)
224
+/* Late collision */
225
+#define TX_DESC_STATUS_LC (1 << 23)
226
+/* Transmission abort */
227
+#define TX_DESC_STATUS_TXABT (1 << 22)
228
+/* No carrier sense */
229
+#define TX_DESC_STATUS_NCS (1 << 21)
230
+/* Defer exceed */
231
+#define TX_DESC_STATUS_EXDEF (1 << 20)
232
+/* Transmission complete */
233
+#define TX_DESC_STATUS_TXCP (1 << 19)
234
+/* Transmission deferred */
235
+#define TX_DESC_STATUS_DEF (1 << 17)
236
+/* Transmit interrupt */
237
+#define TX_DESC_STATUS_TXINTR (1 << 16)
238
+
239
+#define TX_DESC_PKT_LEN(word) extract32((word), 0, 16)
240
+
241
+/* Transmit buffer start address */
242
+#define TX_DESC_TXBSA(word) ((uint32_t) (word) & ~3u)
243
+
244
+/* Next transmit descriptor start address */
245
+#define TX_DESC_NTXDSA(word) ((uint32_t) (word) & ~3u)
246
+
247
+/* NPCM7xxEMCRxDesc.status_and_length values */
248
+/* Owner: 0b00 = cpu, 0b01 = undefined, 0b10 = emc, 0b11 = undefined */
249
+#define RX_DESC_STATUS_OWNER_SHIFT 30
250
+#define RX_DESC_STATUS_OWNER_BITSIZE 2
251
+#define RX_DESC_STATUS_OWNER_MASK (3 << RX_DESC_STATUS_OWNER_SHIFT)
252
+/* Runt packet */
253
+#define RX_DESC_STATUS_RP (1 << 22)
254
+/* Alignment error */
255
+#define RX_DESC_STATUS_ALIE (1 << 21)
256
+/* Frame reception complete */
257
+#define RX_DESC_STATUS_RXGD (1 << 20)
258
+/* Packet too long */
259
+#define RX_DESC_STATUS_PTLE (1 << 19)
260
+/* CRC error */
261
+#define RX_DESC_STATUS_CRCE (1 << 17)
262
+/* Receive interrupt */
263
+#define RX_DESC_STATUS_RXINTR (1 << 16)
264
+
265
+#define RX_DESC_PKT_LEN(word) extract32((word), 0, 16)
266
+
267
+/* Receive buffer start address */
268
+#define RX_DESC_RXBSA(word) ((uint32_t) (word) & ~3u)
269
+
270
+/* Next receive descriptor start address */
271
+#define RX_DESC_NRXDSA(word) ((uint32_t) (word) & ~3u)
272
+
273
+/* Minimum packet length, when TX_DESC_FLAG_PADEN is set. */
274
+#define MIN_PACKET_LENGTH 64
275
+
276
+struct NPCM7xxEMCState {
277
+ /*< private >*/
278
+ SysBusDevice parent;
279
+ /*< public >*/
280
+
281
+ MemoryRegion iomem;
282
+
283
+ qemu_irq tx_irq;
284
+ qemu_irq rx_irq;
285
+
286
+ NICState *nic;
287
+ NICConf conf;
288
+
289
+ /* 0 or 1, for log messages */
290
+ uint8_t emc_num;
291
+
292
+ uint32_t regs[NPCM7XX_NUM_EMC_REGS];
293
+
294
+ /*
295
+ * tx is active. Set to true by TSDR and then switches off when out of
296
+ * descriptors. If the TXON bit in REG_MCMDR is off then this is off.
297
+ */
298
+ bool tx_active;
299
+
300
+ /*
301
+ * rx is active. Set to true by RSDR and then switches off when out of
302
+ * descriptors. If the RXON bit in REG_MCMDR is off then this is off.
303
+ */
304
+ bool rx_active;
305
+};
306
+
307
+typedef struct NPCM7xxEMCState NPCM7xxEMCState;
308
+
309
+#define TYPE_NPCM7XX_EMC "npcm7xx-emc"
310
+#define NPCM7XX_EMC(obj) \
311
+ OBJECT_CHECK(NPCM7xxEMCState, (obj), TYPE_NPCM7XX_EMC)
312
+
313
+#endif /* NPCM7XX_EMC_H */
314
diff --git a/hw/net/npcm7xx_emc.c b/hw/net/npcm7xx_emc.c
315
new file mode 100644
316
index XXXXXXX..XXXXXXX
317
--- /dev/null
318
+++ b/hw/net/npcm7xx_emc.c
319
@@ -XXX,XX +XXX,XX @@
320
+/*
321
+ * Nuvoton NPCM7xx EMC Module
322
+ *
323
+ * Copyright 2020 Google LLC
324
+ *
325
+ * This program is free software; you can redistribute it and/or modify it
326
+ * under the terms of the GNU General Public License as published by the
327
+ * Free Software Foundation; either version 2 of the License, or
328
+ * (at your option) any later version.
329
+ *
330
+ * This program is distributed in the hope that it will be useful, but WITHOUT
331
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
332
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
333
+ * for more details.
334
+ *
335
+ * Unsupported/unimplemented features:
336
+ * - MCMDR.FDUP (full duplex) is ignored, half duplex is not supported
337
+ * - Only CAM0 is supported, CAM[1-15] are not
338
+ * - writes to CAMEN.[1-15] are ignored, these bits always read as zeroes
339
+ * - MII is not implemented, MIIDA.BUSY and MIID always return zero
340
+ * - MCMDR.LBK is not implemented
341
+ * - MCMDR.{OPMOD,ENSQE,AEP,ARP} are not supported
342
+ * - H/W FIFOs are not supported, MCMDR.FFTCR is ignored
343
+ * - MGSTA.SQE is not supported
344
+ * - pause and control frames are not implemented
345
+ * - MGSTA.CCNT is not supported
346
+ * - MPCNT, DMARFS are not implemented
347
+ */
348
+
349
+#include "qemu/osdep.h"
350
+
351
+/* For crc32 */
352
+#include <zlib.h>
353
+
354
+#include "qemu-common.h"
355
+#include "hw/irq.h"
356
+#include "hw/qdev-clock.h"
357
+#include "hw/qdev-properties.h"
358
+#include "hw/net/npcm7xx_emc.h"
359
+#include "net/eth.h"
360
+#include "migration/vmstate.h"
361
+#include "qemu/bitops.h"
362
+#include "qemu/error-report.h"
363
+#include "qemu/log.h"
364
+#include "qemu/module.h"
365
+#include "qemu/units.h"
366
+#include "sysemu/dma.h"
367
+#include "trace.h"
368
+
369
+#define CRC_LENGTH 4
370
+
371
+/*
372
+ * The maximum size of a (layer 2) ethernet frame as defined by 802.3.
373
+ * 1518 = 6(dest macaddr) + 6(src macaddr) + 2(proto) + 4(crc) + 1500(payload)
374
+ * This does not include an additional 4 for the vlan field (802.1q).
375
+ */
376
+#define MAX_ETH_FRAME_SIZE 1518
377
+
378
+static const char *emc_reg_name(int regno)
379
+{
28
+{
380
+#define REG(name) case REG_ ## name: return #name;
29
+ return sme_enabled_check_with_svcr(s, R_SVCR_SM_MASK);
381
+ switch (regno) {
382
+ REG(CAMCMR)
383
+ REG(CAMEN)
384
+ REG(TXDLSA)
385
+ REG(RXDLSA)
386
+ REG(MCMDR)
387
+ REG(MIID)
388
+ REG(MIIDA)
389
+ REG(FFTCR)
390
+ REG(TSDR)
391
+ REG(RSDR)
392
+ REG(DMARFC)
393
+ REG(MIEN)
394
+ REG(MISTA)
395
+ REG(MGSTA)
396
+ REG(MPCNT)
397
+ REG(MRPC)
398
+ REG(MRPCC)
399
+ REG(MREPC)
400
+ REG(DMARFS)
401
+ REG(CTXDSA)
402
+ REG(CTXBSA)
403
+ REG(CRXDSA)
404
+ REG(CRXBSA)
405
+ case REG_CAMM_BASE + 0: return "CAM0M";
406
+ case REG_CAML_BASE + 0: return "CAM0L";
407
+ case REG_CAMM_BASE + 2 ... REG_CAMML_LAST:
408
+ /* Only CAM0 is supported, fold the others into something simple. */
409
+ if (regno & 1) {
410
+ return "CAM<n>L";
411
+ } else {
412
+ return "CAM<n>M";
413
+ }
414
+ default: return "UNKNOWN";
415
+ }
416
+#undef REG
417
+}
30
+}
418
+
31
+
419
+static void emc_reset(NPCM7xxEMCState *emc)
32
+/* This function corresponds to CheckSMEAndZAEnabled. */
33
+static inline bool sme_za_enabled_check(DisasContext *s)
420
+{
34
+{
421
+ trace_npcm7xx_emc_reset(emc->emc_num);
35
+ return sme_enabled_check_with_svcr(s, R_SVCR_ZA_MASK);
422
+
423
+ memset(&emc->regs[0], 0, sizeof(emc->regs));
424
+
425
+ /* These regs have non-zero reset values. */
426
+ emc->regs[REG_TXDLSA] = 0xfffffffc;
427
+ emc->regs[REG_RXDLSA] = 0xfffffffc;
428
+ emc->regs[REG_MIIDA] = 0x00900000;
429
+ emc->regs[REG_FFTCR] = 0x0101;
430
+ emc->regs[REG_DMARFC] = 0x0800;
431
+ emc->regs[REG_MPCNT] = 0x7fff;
432
+
433
+ emc->tx_active = false;
434
+ emc->rx_active = false;
435
+}
36
+}
436
+
37
+
437
+static void npcm7xx_emc_reset(DeviceState *dev)
38
+/* Note that this function corresponds to CheckStreamingSVEAndZAEnabled. */
39
+static inline bool sme_smza_enabled_check(DisasContext *s)
438
+{
40
+{
439
+ NPCM7xxEMCState *emc = NPCM7XX_EMC(dev);
41
+ return sme_enabled_check_with_svcr(s, R_SVCR_SM_MASK | R_SVCR_ZA_MASK);
440
+ emc_reset(emc);
441
+}
42
+}
442
+
43
+
443
+static void emc_soft_reset(NPCM7xxEMCState *emc)
44
TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
45
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
46
bool tag_checked, int log2_size);
47
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/target/arm/translate-a64.c
50
+++ b/target/arm/translate-a64.c
51
@@ -XXX,XX +XXX,XX @@ static bool sme_access_check(DisasContext *s)
52
return true;
53
}
54
55
+/* This function corresponds to CheckSMEEnabled. */
56
+bool sme_enabled_check(DisasContext *s)
444
+{
57
+{
445
+ /*
58
+ /*
446
+ * The docs say at least MCMDR.{LBK,OPMOD} bits are not changed during a
59
+ * Note that unlike sve_excp_el, we have not constrained sme_excp_el
447
+ * soft reset, but does not go into further detail. For now, KISS.
60
+ * to be zero when fp_excp_el has priority. This is because we need
61
+ * sme_excp_el by itself for cpregs access checks.
448
+ */
62
+ */
449
+ uint32_t mcmdr = emc->regs[REG_MCMDR];
63
+ if (!s->fp_excp_el || s->sme_excp_el < s->fp_excp_el) {
450
+ emc_reset(emc);
64
+ s->fp_access_checked = true;
451
+ emc->regs[REG_MCMDR] = mcmdr & (REG_MCMDR_LBK | REG_MCMDR_OPMOD);
65
+ return sme_access_check(s);
452
+
66
+ }
453
+ qemu_set_irq(emc->tx_irq, 0);
67
+ return fp_access_check_only(s);
454
+ qemu_set_irq(emc->rx_irq, 0);
455
+}
68
+}
456
+
69
+
457
+static void emc_set_link(NetClientState *nc)
70
+/* Common subroutine for CheckSMEAnd*Enabled. */
71
+bool sme_enabled_check_with_svcr(DisasContext *s, unsigned req)
458
+{
72
+{
459
+ /* Nothing to do yet. */
73
+ if (!sme_enabled_check(s)) {
74
+ return false;
75
+ }
76
+ if (FIELD_EX64(req, SVCR, SM) && !s->pstate_sm) {
77
+ gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
78
+ syn_smetrap(SME_ET_NotStreaming, false));
79
+ return false;
80
+ }
81
+ if (FIELD_EX64(req, SVCR, ZA) && !s->pstate_za) {
82
+ gen_exception_insn(s, s->pc_curr, EXCP_UDEF,
83
+ syn_smetrap(SME_ET_InactiveZA, false));
84
+ return false;
85
+ }
86
+ return true;
460
+}
87
+}
461
+
88
+
462
+/* MISTA.TXINTR is the union of the individual bits with their enables. */
89
/*
463
+static void emc_update_mista_txintr(NPCM7xxEMCState *emc)
90
* This utility function is for doing register extension with an
464
+{
91
* optional shift. You will likely want to pass a temporary for the
465
+ /* Only look at the bits we support. */
466
+ uint32_t mask = (REG_MISTA_TXBERR |
467
+ REG_MISTA_TDU |
468
+ REG_MISTA_TXCP);
469
+ if (emc->regs[REG_MISTA] & emc->regs[REG_MIEN] & mask) {
470
+ emc->regs[REG_MISTA] |= REG_MISTA_TXINTR;
471
+ } else {
472
+ emc->regs[REG_MISTA] &= ~REG_MISTA_TXINTR;
473
+ }
474
+}
475
+
476
+/* MISTA.RXINTR is the union of the individual bits with their enables. */
477
+static void emc_update_mista_rxintr(NPCM7xxEMCState *emc)
478
+{
479
+ /* Only look at the bits we support. */
480
+ uint32_t mask = (REG_MISTA_RXBERR |
481
+ REG_MISTA_RDU |
482
+ REG_MISTA_RXGD);
483
+ if (emc->regs[REG_MISTA] & emc->regs[REG_MIEN] & mask) {
484
+ emc->regs[REG_MISTA] |= REG_MISTA_RXINTR;
485
+ } else {
486
+ emc->regs[REG_MISTA] &= ~REG_MISTA_RXINTR;
487
+ }
488
+}
489
+
490
+/* N.B. emc_update_mista_txintr must have already been called. */
491
+static void emc_update_tx_irq(NPCM7xxEMCState *emc)
492
+{
493
+ int level = !!(emc->regs[REG_MISTA] &
494
+ emc->regs[REG_MIEN] &
495
+ REG_MISTA_TXINTR);
496
+ trace_npcm7xx_emc_update_tx_irq(level);
497
+ qemu_set_irq(emc->tx_irq, level);
498
+}
499
+
500
+/* N.B. emc_update_mista_rxintr must have already been called. */
501
+static void emc_update_rx_irq(NPCM7xxEMCState *emc)
502
+{
503
+ int level = !!(emc->regs[REG_MISTA] &
504
+ emc->regs[REG_MIEN] &
505
+ REG_MISTA_RXINTR);
506
+ trace_npcm7xx_emc_update_rx_irq(level);
507
+ qemu_set_irq(emc->rx_irq, level);
508
+}
509
+
510
+/* Update IRQ states due to changes in MIEN,MISTA. */
511
+static void emc_update_irq_from_reg_change(NPCM7xxEMCState *emc)
512
+{
513
+ emc_update_mista_txintr(emc);
514
+ emc_update_tx_irq(emc);
515
+
516
+ emc_update_mista_rxintr(emc);
517
+ emc_update_rx_irq(emc);
518
+}
519
+
520
+static int emc_read_tx_desc(dma_addr_t addr, NPCM7xxEMCTxDesc *desc)
521
+{
522
+ if (dma_memory_read(&address_space_memory, addr, desc, sizeof(*desc))) {
523
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Failed to read descriptor @ 0x%"
524
+ HWADDR_PRIx "\n", __func__, addr);
525
+ return -1;
526
+ }
527
+ desc->flags = le32_to_cpu(desc->flags);
528
+ desc->txbsa = le32_to_cpu(desc->txbsa);
529
+ desc->status_and_length = le32_to_cpu(desc->status_and_length);
530
+ desc->ntxdsa = le32_to_cpu(desc->ntxdsa);
531
+ return 0;
532
+}
533
+
534
+static int emc_write_tx_desc(const NPCM7xxEMCTxDesc *desc, dma_addr_t addr)
535
+{
536
+ NPCM7xxEMCTxDesc le_desc;
537
+
538
+ le_desc.flags = cpu_to_le32(desc->flags);
539
+ le_desc.txbsa = cpu_to_le32(desc->txbsa);
540
+ le_desc.status_and_length = cpu_to_le32(desc->status_and_length);
541
+ le_desc.ntxdsa = cpu_to_le32(desc->ntxdsa);
542
+ if (dma_memory_write(&address_space_memory, addr, &le_desc,
543
+ sizeof(le_desc))) {
544
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Failed to write descriptor @ 0x%"
545
+ HWADDR_PRIx "\n", __func__, addr);
546
+ return -1;
547
+ }
548
+ return 0;
549
+}
550
+
551
+static int emc_read_rx_desc(dma_addr_t addr, NPCM7xxEMCRxDesc *desc)
552
+{
553
+ if (dma_memory_read(&address_space_memory, addr, desc, sizeof(*desc))) {
554
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Failed to read descriptor @ 0x%"
555
+ HWADDR_PRIx "\n", __func__, addr);
556
+ return -1;
557
+ }
558
+ desc->status_and_length = le32_to_cpu(desc->status_and_length);
559
+ desc->rxbsa = le32_to_cpu(desc->rxbsa);
560
+ desc->reserved = le32_to_cpu(desc->reserved);
561
+ desc->nrxdsa = le32_to_cpu(desc->nrxdsa);
562
+ return 0;
563
+}
564
+
565
+static int emc_write_rx_desc(const NPCM7xxEMCRxDesc *desc, dma_addr_t addr)
566
+{
567
+ NPCM7xxEMCRxDesc le_desc;
568
+
569
+ le_desc.status_and_length = cpu_to_le32(desc->status_and_length);
570
+ le_desc.rxbsa = cpu_to_le32(desc->rxbsa);
571
+ le_desc.reserved = cpu_to_le32(desc->reserved);
572
+ le_desc.nrxdsa = cpu_to_le32(desc->nrxdsa);
573
+ if (dma_memory_write(&address_space_memory, addr, &le_desc,
574
+ sizeof(le_desc))) {
575
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Failed to write descriptor @ 0x%"
576
+ HWADDR_PRIx "\n", __func__, addr);
577
+ return -1;
578
+ }
579
+ return 0;
580
+}
581
+
582
+static void emc_set_mista(NPCM7xxEMCState *emc, uint32_t flags)
583
+{
584
+ trace_npcm7xx_emc_set_mista(flags);
585
+ emc->regs[REG_MISTA] |= flags;
586
+ if (extract32(flags, 16, 16)) {
587
+ emc_update_mista_txintr(emc);
588
+ }
589
+ if (extract32(flags, 0, 16)) {
590
+ emc_update_mista_rxintr(emc);
591
+ }
592
+}
593
+
594
+static void emc_halt_tx(NPCM7xxEMCState *emc, uint32_t mista_flag)
595
+{
596
+ emc->tx_active = false;
597
+ emc_set_mista(emc, mista_flag);
598
+}
599
+
600
+static void emc_halt_rx(NPCM7xxEMCState *emc, uint32_t mista_flag)
601
+{
602
+ emc->rx_active = false;
603
+ emc_set_mista(emc, mista_flag);
604
+}
605
+
606
+static void emc_set_next_tx_descriptor(NPCM7xxEMCState *emc,
607
+ const NPCM7xxEMCTxDesc *tx_desc,
608
+ uint32_t desc_addr)
609
+{
610
+ /* Update the current descriptor, if only to reset the owner flag. */
611
+ if (emc_write_tx_desc(tx_desc, desc_addr)) {
612
+ /*
613
+ * We just read it so this shouldn't generally happen.
614
+ * Error already reported.
615
+ */
616
+ emc_set_mista(emc, REG_MISTA_TXBERR);
617
+ }
618
+ emc->regs[REG_CTXDSA] = TX_DESC_NTXDSA(tx_desc->ntxdsa);
619
+}
620
+
621
+static void emc_set_next_rx_descriptor(NPCM7xxEMCState *emc,
622
+ const NPCM7xxEMCRxDesc *rx_desc,
623
+ uint32_t desc_addr)
624
+{
625
+ /* Update the current descriptor, if only to reset the owner flag. */
626
+ if (emc_write_rx_desc(rx_desc, desc_addr)) {
627
+ /*
628
+ * We just read it so this shouldn't generally happen.
629
+ * Error already reported.
630
+ */
631
+ emc_set_mista(emc, REG_MISTA_RXBERR);
632
+ }
633
+ emc->regs[REG_CRXDSA] = RX_DESC_NRXDSA(rx_desc->nrxdsa);
634
+}
635
+
636
+static void emc_try_send_next_packet(NPCM7xxEMCState *emc)
637
+{
638
+ /* Working buffer for sending out packets. Most packets fit in this. */
639
+#define TX_BUFFER_SIZE 2048
640
+ uint8_t tx_send_buffer[TX_BUFFER_SIZE];
641
+ uint32_t desc_addr = TX_DESC_NTXDSA(emc->regs[REG_CTXDSA]);
642
+ NPCM7xxEMCTxDesc tx_desc;
643
+ uint32_t next_buf_addr, length;
644
+ uint8_t *buf;
645
+ g_autofree uint8_t *malloced_buf = NULL;
646
+
647
+ if (emc_read_tx_desc(desc_addr, &tx_desc)) {
648
+ /* Error reading descriptor, already reported. */
649
+ emc_halt_tx(emc, REG_MISTA_TXBERR);
650
+ emc_update_tx_irq(emc);
651
+ return;
652
+ }
653
+
654
+ /* Nothing we can do if we don't own the descriptor. */
655
+ if (!(tx_desc.flags & TX_DESC_FLAG_OWNER_MASK)) {
656
+ trace_npcm7xx_emc_cpu_owned_desc(desc_addr);
657
+ emc_halt_tx(emc, REG_MISTA_TDU);
658
+ emc_update_tx_irq(emc);
659
+ return;
660
+ }
661
+
662
+ /* Give the descriptor back regardless of what happens. */
663
+ tx_desc.flags &= ~TX_DESC_FLAG_OWNER_MASK;
664
+ tx_desc.status_and_length &= 0xffff;
665
+
666
+ /*
667
+ * Despite the h/w documentation saying the tx buffer is word aligned,
668
+ * the linux driver does not word align the buffer. There is value in not
669
+ * aligning the buffer: See the description of NET_IP_ALIGN in linux
670
+ * kernel sources.
671
+ */
672
+ next_buf_addr = tx_desc.txbsa;
673
+ emc->regs[REG_CTXBSA] = next_buf_addr;
674
+ length = TX_DESC_PKT_LEN(tx_desc.status_and_length);
675
+ buf = &tx_send_buffer[0];
676
+
677
+ if (length > sizeof(tx_send_buffer)) {
678
+ malloced_buf = g_malloc(length);
679
+ buf = malloced_buf;
680
+ }
681
+
682
+ if (dma_memory_read(&address_space_memory, next_buf_addr, buf, length)) {
683
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Failed to read packet @ 0x%x\n",
684
+ __func__, next_buf_addr);
685
+ emc_set_mista(emc, REG_MISTA_TXBERR);
686
+ emc_set_next_tx_descriptor(emc, &tx_desc, desc_addr);
687
+ emc_update_tx_irq(emc);
688
+ trace_npcm7xx_emc_tx_done(emc->regs[REG_CTXDSA]);
689
+ return;
690
+ }
691
+
692
+ if ((tx_desc.flags & TX_DESC_FLAG_PADEN) && (length < MIN_PACKET_LENGTH)) {
693
+ memset(buf + length, 0, MIN_PACKET_LENGTH - length);
694
+ length = MIN_PACKET_LENGTH;
695
+ }
696
+
697
+ /* N.B. emc_receive can get called here. */
698
+ qemu_send_packet(qemu_get_queue(emc->nic), buf, length);
699
+ trace_npcm7xx_emc_sent_packet(length);
700
+
701
+ tx_desc.status_and_length |= TX_DESC_STATUS_TXCP;
702
+ if (tx_desc.flags & TX_DESC_FLAG_INTEN) {
703
+ emc_set_mista(emc, REG_MISTA_TXCP);
704
+ }
705
+ if (emc->regs[REG_MISTA] & emc->regs[REG_MIEN] & REG_MISTA_TXINTR) {
706
+ tx_desc.status_and_length |= TX_DESC_STATUS_TXINTR;
707
+ }
708
+
709
+ emc_set_next_tx_descriptor(emc, &tx_desc, desc_addr);
710
+ emc_update_tx_irq(emc);
711
+ trace_npcm7xx_emc_tx_done(emc->regs[REG_CTXDSA]);
712
+}
713
+
714
+static bool emc_can_receive(NetClientState *nc)
715
+{
716
+ NPCM7xxEMCState *emc = NPCM7XX_EMC(qemu_get_nic_opaque(nc));
717
+
718
+ bool can_receive = emc->rx_active;
719
+ trace_npcm7xx_emc_can_receive(can_receive);
720
+ return can_receive;
721
+}
722
+
723
+/* If result is false then *fail_reason contains the reason. */
724
+static bool emc_receive_filter1(NPCM7xxEMCState *emc, const uint8_t *buf,
725
+ size_t len, const char **fail_reason)
726
+{
727
+ eth_pkt_types_e pkt_type = get_eth_packet_type(PKT_GET_ETH_HDR(buf));
728
+
729
+ switch (pkt_type) {
730
+ case ETH_PKT_BCAST:
731
+ if (emc->regs[REG_CAMCMR] & REG_CAMCMR_CCAM) {
732
+ return true;
733
+ } else {
734
+ *fail_reason = "Broadcast packet disabled";
735
+ return !!(emc->regs[REG_CAMCMR] & REG_CAMCMR_ABP);
736
+ }
737
+ case ETH_PKT_MCAST:
738
+ if (emc->regs[REG_CAMCMR] & REG_CAMCMR_CCAM) {
739
+ return true;
740
+ } else {
741
+ *fail_reason = "Multicast packet disabled";
742
+ return !!(emc->regs[REG_CAMCMR] & REG_CAMCMR_AMP);
743
+ }
744
+ case ETH_PKT_UCAST: {
745
+ bool matches;
746
+ if (emc->regs[REG_CAMCMR] & REG_CAMCMR_AUP) {
747
+ return true;
748
+ }
749
+ matches = ((emc->regs[REG_CAMCMR] & REG_CAMCMR_ECMP) &&
750
+ /* We only support one CAM register, CAM0. */
751
+ (emc->regs[REG_CAMEN] & (1 << 0)) &&
752
+ memcmp(buf, emc->conf.macaddr.a, ETH_ALEN) == 0);
753
+ if (emc->regs[REG_CAMCMR] & REG_CAMCMR_CCAM) {
754
+ *fail_reason = "MACADDR matched, comparison complemented";
755
+ return !matches;
756
+ } else {
757
+ *fail_reason = "MACADDR didn't match";
758
+ return matches;
759
+ }
760
+ }
761
+ default:
762
+ g_assert_not_reached();
763
+ }
764
+}
765
+
766
+static bool emc_receive_filter(NPCM7xxEMCState *emc, const uint8_t *buf,
767
+ size_t len)
768
+{
769
+ const char *fail_reason = NULL;
770
+ bool ok = emc_receive_filter1(emc, buf, len, &fail_reason);
771
+ if (!ok) {
772
+ trace_npcm7xx_emc_packet_filtered_out(fail_reason);
773
+ }
774
+ return ok;
775
+}
776
+
777
+static ssize_t emc_receive(NetClientState *nc, const uint8_t *buf, size_t len1)
778
+{
779
+ NPCM7xxEMCState *emc = NPCM7XX_EMC(qemu_get_nic_opaque(nc));
780
+ const uint32_t len = len1;
781
+ size_t max_frame_len;
782
+ bool long_frame;
783
+ uint32_t desc_addr;
784
+ NPCM7xxEMCRxDesc rx_desc;
785
+ uint32_t crc;
786
+ uint8_t *crc_ptr;
787
+ uint32_t buf_addr;
788
+
789
+ trace_npcm7xx_emc_receiving_packet(len);
790
+
791
+ if (!emc_can_receive(nc)) {
792
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Unexpected packet\n", __func__);
793
+ return -1;
794
+ }
795
+
796
+ if (len < ETH_HLEN ||
797
+ /* Defensive programming: drop unsupportable large packets. */
798
+ len > 0xffff - CRC_LENGTH) {
799
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Dropped frame of %u bytes\n",
800
+ __func__, len);
801
+ return len;
802
+ }
803
+
804
+ /*
805
+ * DENI is set if EMC received the Length/Type field of the incoming
806
+ * packet, so it will be set regardless of what happens next.
807
+ */
808
+ emc_set_mista(emc, REG_MISTA_DENI);
809
+
810
+ if (!emc_receive_filter(emc, buf, len)) {
811
+ emc_update_rx_irq(emc);
812
+ return len;
813
+ }
814
+
815
+ /* Huge frames (> DMARFC) are dropped. */
816
+ max_frame_len = REG_DMARFC_RXMS(emc->regs[REG_DMARFC]);
817
+ if (len + CRC_LENGTH > max_frame_len) {
818
+ trace_npcm7xx_emc_packet_dropped(len);
819
+ emc_set_mista(emc, REG_MISTA_DFOI);
820
+ emc_update_rx_irq(emc);
821
+ return len;
822
+ }
823
+
824
+ /*
825
+ * Long Frames (> MAX_ETH_FRAME_SIZE) are also dropped, unless MCMDR.ALP
826
+ * is set.
827
+ */
828
+ long_frame = false;
829
+ if (len + CRC_LENGTH > MAX_ETH_FRAME_SIZE) {
830
+ if (emc->regs[REG_MCMDR] & REG_MCMDR_ALP) {
831
+ long_frame = true;
832
+ } else {
833
+ trace_npcm7xx_emc_packet_dropped(len);
834
+ emc_set_mista(emc, REG_MISTA_PTLE);
835
+ emc_update_rx_irq(emc);
836
+ return len;
837
+ }
838
+ }
839
+
840
+ desc_addr = RX_DESC_NRXDSA(emc->regs[REG_CRXDSA]);
841
+ if (emc_read_rx_desc(desc_addr, &rx_desc)) {
842
+ /* Error reading descriptor, already reported. */
843
+ emc_halt_rx(emc, REG_MISTA_RXBERR);
844
+ emc_update_rx_irq(emc);
845
+ return len;
846
+ }
847
+
848
+ /* Nothing we can do if we don't own the descriptor. */
849
+ if (!(rx_desc.status_and_length & RX_DESC_STATUS_OWNER_MASK)) {
850
+ trace_npcm7xx_emc_cpu_owned_desc(desc_addr);
851
+ emc_halt_rx(emc, REG_MISTA_RDU);
852
+ emc_update_rx_irq(emc);
853
+ return len;
854
+ }
855
+
856
+ crc = 0;
857
+ crc_ptr = (uint8_t *) &crc;
858
+ if (!(emc->regs[REG_MCMDR] & REG_MCMDR_SPCRC)) {
859
+ crc = cpu_to_be32(crc32(~0, buf, len));
860
+ }
861
+
862
+ /* Give the descriptor back regardless of what happens. */
863
+ rx_desc.status_and_length &= ~RX_DESC_STATUS_OWNER_MASK;
864
+
865
+ buf_addr = rx_desc.rxbsa;
866
+ emc->regs[REG_CRXBSA] = buf_addr;
867
+ if (dma_memory_write(&address_space_memory, buf_addr, buf, len) ||
868
+ (!(emc->regs[REG_MCMDR] & REG_MCMDR_SPCRC) &&
869
+ dma_memory_write(&address_space_memory, buf_addr + len, crc_ptr,
870
+ 4))) {
871
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Bus error writing packet\n",
872
+ __func__);
873
+ emc_set_mista(emc, REG_MISTA_RXBERR);
874
+ emc_set_next_rx_descriptor(emc, &rx_desc, desc_addr);
875
+ emc_update_rx_irq(emc);
876
+ trace_npcm7xx_emc_rx_done(emc->regs[REG_CRXDSA]);
877
+ return len;
878
+ }
879
+
880
+ trace_npcm7xx_emc_received_packet(len);
881
+
882
+ /* Note: We've already verified len+4 <= 0xffff. */
883
+ rx_desc.status_and_length = len;
884
+ if (!(emc->regs[REG_MCMDR] & REG_MCMDR_SPCRC)) {
885
+ rx_desc.status_and_length += 4;
886
+ }
887
+ rx_desc.status_and_length |= RX_DESC_STATUS_RXGD;
888
+ emc_set_mista(emc, REG_MISTA_RXGD);
889
+
890
+ if (emc->regs[REG_MISTA] & emc->regs[REG_MIEN] & REG_MISTA_RXINTR) {
891
+ rx_desc.status_and_length |= RX_DESC_STATUS_RXINTR;
892
+ }
893
+ if (long_frame) {
894
+ rx_desc.status_and_length |= RX_DESC_STATUS_PTLE;
895
+ }
896
+
897
+ emc_set_next_rx_descriptor(emc, &rx_desc, desc_addr);
898
+ emc_update_rx_irq(emc);
899
+ trace_npcm7xx_emc_rx_done(emc->regs[REG_CRXDSA]);
900
+ return len;
901
+}
902
+
903
+static void emc_try_receive_next_packet(NPCM7xxEMCState *emc)
904
+{
905
+ if (emc_can_receive(qemu_get_queue(emc->nic))) {
906
+ qemu_flush_queued_packets(qemu_get_queue(emc->nic));
907
+ }
908
+}
909
+
910
+static uint64_t npcm7xx_emc_read(void *opaque, hwaddr offset, unsigned size)
911
+{
912
+ NPCM7xxEMCState *emc = opaque;
913
+ uint32_t reg = offset / sizeof(uint32_t);
914
+ uint32_t result;
915
+
916
+ if (reg >= NPCM7XX_NUM_EMC_REGS) {
917
+ qemu_log_mask(LOG_GUEST_ERROR,
918
+ "%s: Invalid offset 0x%04" HWADDR_PRIx "\n",
919
+ __func__, offset);
920
+ return 0;
921
+ }
922
+
923
+ switch (reg) {
924
+ case REG_MIID:
925
+ /*
926
+ * We don't implement MII. For determinism, always return zero as
927
+ * writes record the last value written for debugging purposes.
928
+ */
929
+ qemu_log_mask(LOG_UNIMP, "%s: Read of MIID, returning 0\n", __func__);
930
+ result = 0;
931
+ break;
932
+ case REG_TSDR:
933
+ case REG_RSDR:
934
+ qemu_log_mask(LOG_GUEST_ERROR,
935
+ "%s: Read of write-only reg, %s/%d\n",
936
+ __func__, emc_reg_name(reg), reg);
937
+ return 0;
938
+ default:
939
+ result = emc->regs[reg];
940
+ break;
941
+ }
942
+
943
+ trace_npcm7xx_emc_reg_read(emc->emc_num, result, emc_reg_name(reg), reg);
944
+ return result;
945
+}
946
+
947
+static void npcm7xx_emc_write(void *opaque, hwaddr offset,
948
+ uint64_t v, unsigned size)
949
+{
950
+ NPCM7xxEMCState *emc = opaque;
951
+ uint32_t reg = offset / sizeof(uint32_t);
952
+ uint32_t value = v;
953
+
954
+ g_assert(size == sizeof(uint32_t));
955
+
956
+ if (reg >= NPCM7XX_NUM_EMC_REGS) {
957
+ qemu_log_mask(LOG_GUEST_ERROR,
958
+ "%s: Invalid offset 0x%04" HWADDR_PRIx "\n",
959
+ __func__, offset);
960
+ return;
961
+ }
962
+
963
+ trace_npcm7xx_emc_reg_write(emc->emc_num, emc_reg_name(reg), reg, value);
964
+
965
+ switch (reg) {
966
+ case REG_CAMCMR:
967
+ emc->regs[reg] = value;
968
+ break;
969
+ case REG_CAMEN:
970
+ /* Only CAM0 is supported, don't pretend otherwise. */
971
+ if (value & ~1) {
972
+ qemu_log_mask(LOG_GUEST_ERROR,
973
+ "%s: Only CAM0 is supported, cannot enable others"
974
+ ": 0x%x\n",
975
+ __func__, value);
976
+ }
977
+ emc->regs[reg] = value & 1;
978
+ break;
979
+ case REG_CAMM_BASE + 0:
980
+ emc->regs[reg] = value;
981
+ emc->conf.macaddr.a[0] = value >> 24;
982
+ emc->conf.macaddr.a[1] = value >> 16;
983
+ emc->conf.macaddr.a[2] = value >> 8;
984
+ emc->conf.macaddr.a[3] = value >> 0;
985
+ break;
986
+ case REG_CAML_BASE + 0:
987
+ emc->regs[reg] = value;
988
+ emc->conf.macaddr.a[4] = value >> 24;
989
+ emc->conf.macaddr.a[5] = value >> 16;
990
+ break;
991
+ case REG_MCMDR: {
992
+ uint32_t prev;
993
+ if (value & REG_MCMDR_SWR) {
994
+ emc_soft_reset(emc);
995
+ /* On h/w the reset happens over multiple cycles. For now KISS. */
996
+ break;
997
+ }
998
+ prev = emc->regs[reg];
999
+ emc->regs[reg] = value;
1000
+ /* Update tx state. */
1001
+ if (!(prev & REG_MCMDR_TXON) &&
1002
+ (value & REG_MCMDR_TXON)) {
1003
+ emc->regs[REG_CTXDSA] = emc->regs[REG_TXDLSA];
1004
+ /*
1005
+ * Linux kernel turns TX on with CPU still holding descriptor,
1006
+ * which suggests we should wait for a write to TSDR before trying
1007
+ * to send a packet: so we don't send one here.
1008
+ */
1009
+ } else if ((prev & REG_MCMDR_TXON) &&
1010
+ !(value & REG_MCMDR_TXON)) {
1011
+ emc->regs[REG_MGSTA] |= REG_MGSTA_TXHA;
1012
+ }
1013
+ if (!(value & REG_MCMDR_TXON)) {
1014
+ emc_halt_tx(emc, 0);
1015
+ }
1016
+ /* Update rx state. */
1017
+ if (!(prev & REG_MCMDR_RXON) &&
1018
+ (value & REG_MCMDR_RXON)) {
1019
+ emc->regs[REG_CRXDSA] = emc->regs[REG_RXDLSA];
1020
+ } else if ((prev & REG_MCMDR_RXON) &&
1021
+ !(value & REG_MCMDR_RXON)) {
1022
+ emc->regs[REG_MGSTA] |= REG_MGSTA_RXHA;
1023
+ }
1024
+ if (!(value & REG_MCMDR_RXON)) {
1025
+ emc_halt_rx(emc, 0);
1026
+ }
1027
+ break;
1028
+ }
1029
+ case REG_TXDLSA:
1030
+ case REG_RXDLSA:
1031
+ case REG_DMARFC:
1032
+ case REG_MIID:
1033
+ emc->regs[reg] = value;
1034
+ break;
1035
+ case REG_MIEN:
1036
+ emc->regs[reg] = value;
1037
+ emc_update_irq_from_reg_change(emc);
1038
+ break;
1039
+ case REG_MISTA:
1040
+ /* Clear the bits that have 1 in "value". */
1041
+ emc->regs[reg] &= ~value;
1042
+ emc_update_irq_from_reg_change(emc);
1043
+ break;
1044
+ case REG_MGSTA:
1045
+ /* Clear the bits that have 1 in "value". */
1046
+ emc->regs[reg] &= ~value;
1047
+ break;
1048
+ case REG_TSDR:
1049
+ if (emc->regs[REG_MCMDR] & REG_MCMDR_TXON) {
1050
+ emc->tx_active = true;
1051
+ /* Keep trying to send packets until we run out. */
1052
+ while (emc->tx_active) {
1053
+ emc_try_send_next_packet(emc);
1054
+ }
1055
+ }
1056
+ break;
1057
+ case REG_RSDR:
1058
+ if (emc->regs[REG_MCMDR] & REG_MCMDR_RXON) {
1059
+ emc->rx_active = true;
1060
+ emc_try_receive_next_packet(emc);
1061
+ }
1062
+ break;
1063
+ case REG_MIIDA:
1064
+ emc->regs[reg] = value & ~REG_MIIDA_BUSY;
1065
+ break;
1066
+ case REG_MRPC:
1067
+ case REG_MRPCC:
1068
+ case REG_MREPC:
1069
+ case REG_CTXDSA:
1070
+ case REG_CTXBSA:
1071
+ case REG_CRXDSA:
1072
+ case REG_CRXBSA:
1073
+ qemu_log_mask(LOG_GUEST_ERROR,
1074
+ "%s: Write to read-only reg %s/%d\n",
1075
+ __func__, emc_reg_name(reg), reg);
1076
+ break;
1077
+ default:
1078
+ qemu_log_mask(LOG_UNIMP, "%s: Write to unimplemented reg %s/%d\n",
1079
+ __func__, emc_reg_name(reg), reg);
1080
+ break;
1081
+ }
1082
+}
1083
+
1084
+static const struct MemoryRegionOps npcm7xx_emc_ops = {
1085
+ .read = npcm7xx_emc_read,
1086
+ .write = npcm7xx_emc_write,
1087
+ .endianness = DEVICE_LITTLE_ENDIAN,
1088
+ .valid = {
1089
+ .min_access_size = 4,
1090
+ .max_access_size = 4,
1091
+ .unaligned = false,
1092
+ },
1093
+};
1094
+
1095
+static void emc_cleanup(NetClientState *nc)
1096
+{
1097
+ /* Nothing to do yet. */
1098
+}
1099
+
1100
+static NetClientInfo net_npcm7xx_emc_info = {
1101
+ .type = NET_CLIENT_DRIVER_NIC,
1102
+ .size = sizeof(NICState),
1103
+ .can_receive = emc_can_receive,
1104
+ .receive = emc_receive,
1105
+ .cleanup = emc_cleanup,
1106
+ .link_status_changed = emc_set_link,
1107
+};
1108
+
1109
+static void npcm7xx_emc_realize(DeviceState *dev, Error **errp)
1110
+{
1111
+ NPCM7xxEMCState *emc = NPCM7XX_EMC(dev);
1112
+ SysBusDevice *sbd = SYS_BUS_DEVICE(emc);
1113
+
1114
+ memory_region_init_io(&emc->iomem, OBJECT(emc), &npcm7xx_emc_ops, emc,
1115
+ TYPE_NPCM7XX_EMC, 4 * KiB);
1116
+ sysbus_init_mmio(sbd, &emc->iomem);
1117
+ sysbus_init_irq(sbd, &emc->tx_irq);
1118
+ sysbus_init_irq(sbd, &emc->rx_irq);
1119
+
1120
+ qemu_macaddr_default_if_unset(&emc->conf.macaddr);
1121
+ emc->nic = qemu_new_nic(&net_npcm7xx_emc_info, &emc->conf,
1122
+ object_get_typename(OBJECT(dev)), dev->id, emc);
1123
+ qemu_format_nic_info_str(qemu_get_queue(emc->nic), emc->conf.macaddr.a);
1124
+}
1125
+
1126
+static void npcm7xx_emc_unrealize(DeviceState *dev)
1127
+{
1128
+ NPCM7xxEMCState *emc = NPCM7XX_EMC(dev);
1129
+
1130
+ qemu_del_nic(emc->nic);
1131
+}
1132
+
1133
+static const VMStateDescription vmstate_npcm7xx_emc = {
1134
+ .name = TYPE_NPCM7XX_EMC,
1135
+ .version_id = 0,
1136
+ .minimum_version_id = 0,
1137
+ .fields = (VMStateField[]) {
1138
+ VMSTATE_UINT8(emc_num, NPCM7xxEMCState),
1139
+ VMSTATE_UINT32_ARRAY(regs, NPCM7xxEMCState, NPCM7XX_NUM_EMC_REGS),
1140
+ VMSTATE_BOOL(tx_active, NPCM7xxEMCState),
1141
+ VMSTATE_BOOL(rx_active, NPCM7xxEMCState),
1142
+ VMSTATE_END_OF_LIST(),
1143
+ },
1144
+};
1145
+
1146
+static Property npcm7xx_emc_properties[] = {
1147
+ DEFINE_NIC_PROPERTIES(NPCM7xxEMCState, conf),
1148
+ DEFINE_PROP_END_OF_LIST(),
1149
+};
1150
+
1151
+static void npcm7xx_emc_class_init(ObjectClass *klass, void *data)
1152
+{
1153
+ DeviceClass *dc = DEVICE_CLASS(klass);
1154
+
1155
+ set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
1156
+ dc->desc = "NPCM7xx EMC Controller";
1157
+ dc->realize = npcm7xx_emc_realize;
1158
+ dc->unrealize = npcm7xx_emc_unrealize;
1159
+ dc->reset = npcm7xx_emc_reset;
1160
+ dc->vmsd = &vmstate_npcm7xx_emc;
1161
+ device_class_set_props(dc, npcm7xx_emc_properties);
1162
+}
1163
+
1164
+static const TypeInfo npcm7xx_emc_info = {
1165
+ .name = TYPE_NPCM7XX_EMC,
1166
+ .parent = TYPE_SYS_BUS_DEVICE,
1167
+ .instance_size = sizeof(NPCM7xxEMCState),
1168
+ .class_init = npcm7xx_emc_class_init,
1169
+};
1170
+
1171
+static void npcm7xx_emc_register_type(void)
1172
+{
1173
+ type_register_static(&npcm7xx_emc_info);
1174
+}
1175
+
1176
+type_init(npcm7xx_emc_register_type)
1177
diff --git a/hw/net/meson.build b/hw/net/meson.build
1178
index XXXXXXX..XXXXXXX 100644
1179
--- a/hw/net/meson.build
1180
+++ b/hw/net/meson.build
1181
@@ -XXX,XX +XXX,XX @@ softmmu_ss.add(when: 'CONFIG_I82596_COMMON', if_true: files('i82596.c'))
1182
softmmu_ss.add(when: 'CONFIG_SUNHME', if_true: files('sunhme.c'))
1183
softmmu_ss.add(when: 'CONFIG_FTGMAC100', if_true: files('ftgmac100.c'))
1184
softmmu_ss.add(when: 'CONFIG_SUNGEM', if_true: files('sungem.c'))
1185
+softmmu_ss.add(when: 'CONFIG_NPCM7XX', if_true: files('npcm7xx_emc.c'))
1186
1187
softmmu_ss.add(when: 'CONFIG_ETRAXFS', if_true: files('etraxfs_eth.c'))
1188
softmmu_ss.add(when: 'CONFIG_COLDFIRE', if_true: files('mcf_fec.c'))
1189
diff --git a/hw/net/trace-events b/hw/net/trace-events
1190
index XXXXXXX..XXXXXXX 100644
1191
--- a/hw/net/trace-events
1192
+++ b/hw/net/trace-events
1193
@@ -XXX,XX +XXX,XX @@ imx_fec_receive_last(int last) "rx frame flags 0x%04x"
1194
imx_enet_receive(size_t size) "len %zu"
1195
imx_enet_receive_len(uint64_t addr, int len) "rx_bd 0x%"PRIx64" length %d"
1196
imx_enet_receive_last(int last) "rx frame flags 0x%04x"
1197
+
1198
+# npcm7xx_emc.c
1199
+npcm7xx_emc_reset(int emc_num) "Resetting emc%d"
1200
+npcm7xx_emc_update_tx_irq(int level) "Setting tx irq to %d"
1201
+npcm7xx_emc_update_rx_irq(int level) "Setting rx irq to %d"
1202
+npcm7xx_emc_set_mista(uint32_t flags) "ORing 0x%x into MISTA"
1203
+npcm7xx_emc_cpu_owned_desc(uint32_t addr) "Can't process cpu-owned descriptor @0x%x"
1204
+npcm7xx_emc_sent_packet(uint32_t len) "Sent %u byte packet"
1205
+npcm7xx_emc_tx_done(uint32_t ctxdsa) "TX done, CTXDSA=0x%x"
1206
+npcm7xx_emc_can_receive(int can_receive) "Can receive: %d"
1207
+npcm7xx_emc_packet_filtered_out(const char* fail_reason) "Packet filtered out: %s"
1208
+npcm7xx_emc_packet_dropped(uint32_t len) "%u byte packet dropped"
1209
+npcm7xx_emc_receiving_packet(uint32_t len) "Receiving %u byte packet"
1210
+npcm7xx_emc_received_packet(uint32_t len) "Received %u byte packet"
1211
+npcm7xx_emc_rx_done(uint32_t crxdsa) "RX done, CRXDSA=0x%x"
1212
+npcm7xx_emc_reg_read(int emc_num, uint32_t result, const char *name, int regno) "emc%d: 0x%x = reg[%s/%d]"
1213
+npcm7xx_emc_reg_write(int emc_num, const char *name, int regno, uint32_t value) "emc%d: reg[%s/%d] = 0x%x"
1214
--
92
--
1215
2.20.1
93
2.25.1
1216
1217
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Use simple arithmetic instead of a conditional
3
The pseudocode for CheckSVEEnabled gains a check for Streaming
4
move when tbi0 != tbi1.
4
SVE mode, and for SME present but SVE absent.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-22-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-17-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
target/arm/translate-a64.c | 25 ++++++++++++++-----------
11
target/arm/translate-a64.c | 22 ++++++++++++++++------
12
1 file changed, 14 insertions(+), 11 deletions(-)
12
1 file changed, 16 insertions(+), 6 deletions(-)
13
13
14
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
14
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
15
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate-a64.c
16
--- a/target/arm/translate-a64.c
17
+++ b/target/arm/translate-a64.c
17
+++ b/target/arm/translate-a64.c
18
@@ -XXX,XX +XXX,XX @@ static void gen_top_byte_ignore(DisasContext *s, TCGv_i64 dst,
18
@@ -XXX,XX +XXX,XX @@ static bool fp_access_check(DisasContext *s)
19
/* Sign-extend from bit 55. */
19
return true;
20
tcg_gen_sextract_i64(dst, src, 0, 56);
20
}
21
21
22
- if (tbi != 3) {
22
-/* Check that SVE access is enabled. If it is, return true.
23
- TCGv_i64 tcg_zero = tcg_const_i64(0);
23
+/*
24
+ * Check that SVE access is enabled. If it is, return true.
25
* If not, emit code to generate an appropriate exception and return false.
26
+ * This function corresponds to CheckSVEEnabled().
27
*/
28
bool sve_access_check(DisasContext *s)
29
{
30
- if (s->sve_excp_el) {
31
- assert(!s->sve_access_checked);
32
- s->sve_access_checked = true;
24
-
33
-
25
- /*
34
+ if (s->pstate_sm || !dc_isar_feature(aa64_sve, s)) {
26
- * The two TBI bits differ.
35
+ assert(dc_isar_feature(aa64_sme, s));
27
- * If tbi0, then !tbi1: only use the extension if positive.
36
+ if (!sme_sm_enabled_check(s)) {
28
- * if !tbi0, then tbi1: only use the extension if negative.
37
+ goto fail_exit;
29
- */
38
+ }
30
- tcg_gen_movcond_i64(tbi == 1 ? TCG_COND_GE : TCG_COND_LT,
39
+ } else if (s->sve_excp_el) {
31
- dst, dst, tcg_zero, dst, src);
40
gen_exception_insn_el(s, s->pc_curr, EXCP_UDEF,
32
- tcg_temp_free_i64(tcg_zero);
41
syn_sve_access_trap(), s->sve_excp_el);
33
+ switch (tbi) {
42
- return false;
34
+ case 1:
43
+ goto fail_exit;
35
+ /* tbi0 but !tbi1: only use the extension if positive */
36
+ tcg_gen_and_i64(dst, dst, src);
37
+ break;
38
+ case 2:
39
+ /* !tbi0 but tbi1: only use the extension if negative */
40
+ tcg_gen_or_i64(dst, dst, src);
41
+ break;
42
+ case 3:
43
+ /* tbi0 and tbi1: always use the extension */
44
+ break;
45
+ default:
46
+ g_assert_not_reached();
47
}
48
}
44
}
45
s->sve_access_checked = true;
46
return fp_access_check(s);
47
+
48
+ fail_exit:
49
+ /* Assert that we only raise one exception per instruction. */
50
+ assert(!s->sve_access_checked);
51
+ s->sve_access_checked = true;
52
+ return false;
49
}
53
}
54
55
/*
50
--
56
--
51
2.20.1
57
2.25.1
52
53
diff view generated by jsdifflib
1
From: Hao Wu <wuhaotsh@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Add AT24 EEPROM and temperature sensors for GSJ machine.
3
These SME instructions are nominally within the SVE decode space,
4
so we add them to sve.decode and translate-sve.c.
4
5
5
Reviewed-by: Doug Evans<dje@google.com>
6
Reviewed-by: Tyrong Ting<kfting@nuvoton.com>
7
Signed-off-by: Hao Wu <wuhaotsh@google.com>
8
Message-id: 20210210220426.3577804-4-wuhaotsh@google.com
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20220708151540.18136-18-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
hw/arm/npcm7xx_boards.c | 27 +++++++++++++++++++++++++++
11
target/arm/translate-a64.h | 12 ++++++++++++
13
hw/arm/Kconfig | 1 +
12
target/arm/sve.decode | 5 ++++-
14
2 files changed, 28 insertions(+)
13
target/arm/translate-sve.c | 38 ++++++++++++++++++++++++++++++++++++++
14
3 files changed, 54 insertions(+), 1 deletion(-)
15
15
16
diff --git a/hw/arm/npcm7xx_boards.c b/hw/arm/npcm7xx_boards.c
16
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
17
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
18
--- a/hw/arm/npcm7xx_boards.c
18
--- a/target/arm/translate-a64.h
19
+++ b/hw/arm/npcm7xx_boards.c
19
+++ b/target/arm/translate-a64.h
20
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ static inline int vec_full_reg_size(DisasContext *s)
21
#include "exec/address-spaces.h"
21
return s->vl;
22
#include "hw/arm/npcm7xx.h"
23
#include "hw/core/cpu.h"
24
+#include "hw/i2c/smbus_eeprom.h"
25
#include "hw/loader.h"
26
#include "hw/qdev-properties.h"
27
#include "qapi/error.h"
28
@@ -XXX,XX +XXX,XX @@ static I2CBus *npcm7xx_i2c_get_bus(NPCM7xxState *soc, uint32_t num)
29
return I2C_BUS(qdev_get_child_bus(DEVICE(&soc->smbus[num]), "i2c-bus"));
30
}
22
}
31
23
32
+static void at24c_eeprom_init(NPCM7xxState *soc, int bus, uint8_t addr,
24
+/* Return the byte size of the vector register, SVL / 8. */
33
+ uint32_t rsize)
25
+static inline int streaming_vec_reg_size(DisasContext *s)
34
+{
26
+{
35
+ I2CBus *i2c_bus = npcm7xx_i2c_get_bus(soc, bus);
27
+ return s->svl;
36
+ I2CSlave *i2c_dev = i2c_slave_new("at24c-eeprom", addr);
37
+ DeviceState *dev = DEVICE(i2c_dev);
38
+
39
+ qdev_prop_set_uint32(dev, "rom-size", rsize);
40
+ i2c_slave_realize_and_unref(i2c_dev, i2c_bus, &error_abort);
41
+}
28
+}
42
+
29
+
43
static void npcm750_evb_i2c_init(NPCM7xxState *soc)
30
/*
44
{
31
* Return the offset info CPUARMState of the predicate vector register Pn.
45
/* lm75 temperature sensor on SVB, tmp105 is compatible */
32
* Note for this purpose, FFR is P16.
46
@@ -XXX,XX +XXX,XX @@ static void npcm750_evb_i2c_init(NPCM7xxState *soc)
33
@@ -XXX,XX +XXX,XX @@ static inline int pred_full_reg_size(DisasContext *s)
47
i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 6), "tmp105", 0x48);
34
return s->vl >> 3;
48
}
35
}
49
36
50
+static void quanta_gsj_i2c_init(NPCM7xxState *soc)
37
+/* Return the byte size of the predicate register, SVL / 64. */
38
+static inline int streaming_pred_reg_size(DisasContext *s)
51
+{
39
+{
52
+ /* GSJ machine have 4 max31725 temperature sensors, tmp105 is compatible. */
40
+ return s->svl >> 3;
53
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 1), "tmp105", 0x5c);
54
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 2), "tmp105", 0x5c);
55
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 3), "tmp105", 0x5c);
56
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 4), "tmp105", 0x5c);
57
+
58
+ at24c_eeprom_init(soc, 9, 0x55, 8192);
59
+ at24c_eeprom_init(soc, 10, 0x55, 8192);
60
+
61
+ /* TODO: Add additional i2c devices. */
62
+}
41
+}
63
+
42
+
64
static void npcm750_evb_init(MachineState *machine)
43
/*
44
* Round up the size of a register to a size allowed by
45
* the tcg vector infrastructure. Any operation which uses this
46
diff --git a/target/arm/sve.decode b/target/arm/sve.decode
47
index XXXXXXX..XXXXXXX 100644
48
--- a/target/arm/sve.decode
49
+++ b/target/arm/sve.decode
50
@@ -XXX,XX +XXX,XX @@ INDEX_ri 00000100 esz:2 1 imm:s5 010001 rn:5 rd:5
51
# SVE index generation (register start, register increment)
52
INDEX_rr 00000100 .. 1 ..... 010011 ..... ..... @rd_rn_rm
53
54
-### SVE Stack Allocation Group
55
+### SVE / Streaming SVE Stack Allocation Group
56
57
# SVE stack frame adjustment
58
ADDVL 00000100 001 ..... 01010 ...... ..... @rd_rn_i6
59
+ADDSVL 00000100 001 ..... 01011 ...... ..... @rd_rn_i6
60
ADDPL 00000100 011 ..... 01010 ...... ..... @rd_rn_i6
61
+ADDSPL 00000100 011 ..... 01011 ...... ..... @rd_rn_i6
62
63
# SVE stack frame size
64
RDVL 00000100 101 11111 01010 imm:s6 rd:5
65
+RDSVL 00000100 101 11111 01011 imm:s6 rd:5
66
67
### SVE Bitwise Shift - Unpredicated Group
68
69
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
70
index XXXXXXX..XXXXXXX 100644
71
--- a/target/arm/translate-sve.c
72
+++ b/target/arm/translate-sve.c
73
@@ -XXX,XX +XXX,XX @@ static bool trans_ADDVL(DisasContext *s, arg_ADDVL *a)
74
return true;
75
}
76
77
+static bool trans_ADDSVL(DisasContext *s, arg_ADDSVL *a)
78
+{
79
+ if (!dc_isar_feature(aa64_sme, s)) {
80
+ return false;
81
+ }
82
+ if (sme_enabled_check(s)) {
83
+ TCGv_i64 rd = cpu_reg_sp(s, a->rd);
84
+ TCGv_i64 rn = cpu_reg_sp(s, a->rn);
85
+ tcg_gen_addi_i64(rd, rn, a->imm * streaming_vec_reg_size(s));
86
+ }
87
+ return true;
88
+}
89
+
90
static bool trans_ADDPL(DisasContext *s, arg_ADDPL *a)
65
{
91
{
66
NPCM7xxState *soc;
92
if (!dc_isar_feature(aa64_sve, s)) {
67
@@ -XXX,XX +XXX,XX @@ static void quanta_gsj_init(MachineState *machine)
93
@@ -XXX,XX +XXX,XX @@ static bool trans_ADDPL(DisasContext *s, arg_ADDPL *a)
68
npcm7xx_load_bootrom(machine, soc);
94
return true;
69
npcm7xx_connect_flash(&soc->fiu[0], 0, "mx25l25635e",
70
drive_get(IF_MTD, 0, 0));
71
+ quanta_gsj_i2c_init(soc);
72
npcm7xx_load_kernel(machine, soc);
73
}
95
}
74
96
75
diff --git a/hw/arm/Kconfig b/hw/arm/Kconfig
97
+static bool trans_ADDSPL(DisasContext *s, arg_ADDSPL *a)
76
index XXXXXXX..XXXXXXX 100644
98
+{
77
--- a/hw/arm/Kconfig
99
+ if (!dc_isar_feature(aa64_sme, s)) {
78
+++ b/hw/arm/Kconfig
100
+ return false;
79
@@ -XXX,XX +XXX,XX @@ config NPCM7XX
101
+ }
80
bool
102
+ if (sme_enabled_check(s)) {
81
select A9MPCORE
103
+ TCGv_i64 rd = cpu_reg_sp(s, a->rd);
82
select ARM_GIC
104
+ TCGv_i64 rn = cpu_reg_sp(s, a->rn);
83
+ select AT24C # EEPROM
105
+ tcg_gen_addi_i64(rd, rn, a->imm * streaming_pred_reg_size(s));
84
select PL310 # cache controller
106
+ }
85
select SERIAL
107
+ return true;
86
select SSI
108
+}
109
+
110
static bool trans_RDVL(DisasContext *s, arg_RDVL *a)
111
{
112
if (!dc_isar_feature(aa64_sve, s)) {
113
@@ -XXX,XX +XXX,XX @@ static bool trans_RDVL(DisasContext *s, arg_RDVL *a)
114
return true;
115
}
116
117
+static bool trans_RDSVL(DisasContext *s, arg_RDSVL *a)
118
+{
119
+ if (!dc_isar_feature(aa64_sme, s)) {
120
+ return false;
121
+ }
122
+ if (sme_enabled_check(s)) {
123
+ TCGv_i64 reg = cpu_reg(s, a->rd);
124
+ tcg_gen_movi_i64(reg, a->imm * streaming_vec_reg_size(s));
125
+ }
126
+ return true;
127
+}
128
+
129
/*
130
*** SVE Compute Vector Address Group
131
*/
87
--
132
--
88
2.20.1
133
2.25.1
89
90
diff view generated by jsdifflib
1
From: Hao Wu <wuhaotsh@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This patch implements the FIFO mode of the SMBus module. In FIFO, the
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
user transmits or receives at most 16 bytes at a time. The FIFO mode
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
allows the module to transmit large amount of data faster than single
5
Message-id: 20220708151540.18136-19-richard.henderson@linaro.org
6
byte mode.
7
8
Since we only added the device in a patch that is only a few commits
9
away in the same patch set. We do not increase the VMstate version
10
number in this special case.
11
12
Reviewed-by: Doug Evans<dje@google.com>
13
Reviewed-by: Tyrong Ting<kfting@nuvoton.com>
14
Signed-off-by: Hao Wu <wuhaotsh@google.com>
15
Reviewed-by: Corey Minyard <cminyard@mvista.com>
16
Message-id: 20210210220426.3577804-6-wuhaotsh@google.com
17
Acked-by: Corey Minyard <cminyard@mvista.com>
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
---
7
---
20
include/hw/i2c/npcm7xx_smbus.h | 25 +++
8
target/arm/helper-sme.h | 2 ++
21
hw/i2c/npcm7xx_smbus.c | 342 +++++++++++++++++++++++++++++--
9
target/arm/sme.decode | 4 ++++
22
tests/qtest/npcm7xx_smbus-test.c | 149 +++++++++++++-
10
target/arm/sme_helper.c | 25 +++++++++++++++++++++++++
23
hw/i2c/trace-events | 1 +
11
target/arm/translate-sme.c | 13 +++++++++++++
24
4 files changed, 501 insertions(+), 16 deletions(-)
12
4 files changed, 44 insertions(+)
25
13
26
diff --git a/include/hw/i2c/npcm7xx_smbus.h b/include/hw/i2c/npcm7xx_smbus.h
14
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
27
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
28
--- a/include/hw/i2c/npcm7xx_smbus.h
16
--- a/target/arm/helper-sme.h
29
+++ b/include/hw/i2c/npcm7xx_smbus.h
17
+++ b/target/arm/helper-sme.h
30
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@
31
*/
19
32
#define NPCM7XX_SMBUS_NR_ADDRS 10
20
DEF_HELPER_FLAGS_2(set_pstate_sm, TCG_CALL_NO_RWG, void, env, i32)
33
21
DEF_HELPER_FLAGS_2(set_pstate_za, TCG_CALL_NO_RWG, void, env, i32)
34
+/* Size of the FIFO buffer. */
35
+#define NPCM7XX_SMBUS_FIFO_SIZE 16
36
+
22
+
37
typedef enum NPCM7xxSMBusStatus {
23
+DEF_HELPER_FLAGS_3(sme_zero, TCG_CALL_NO_RWG, void, env, i32, i32)
38
NPCM7XX_SMBUS_STATUS_IDLE,
24
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
39
NPCM7XX_SMBUS_STATUS_SENDING,
25
index XXXXXXX..XXXXXXX 100644
40
@@ -XXX,XX +XXX,XX @@ typedef enum NPCM7xxSMBusStatus {
26
--- a/target/arm/sme.decode
41
* @addr: The SMBus module's own addresses on the I2C bus.
27
+++ b/target/arm/sme.decode
42
* @scllt: The SCL low time register.
28
@@ -XXX,XX +XXX,XX @@
43
* @sclht: The SCL high time register.
29
#
44
+ * @fif_ctl: The FIFO control register.
30
# This file is processed by scripts/decodetree.py
45
+ * @fif_cts: The FIFO control status register.
31
#
46
+ * @fair_per: The fair preriod register.
47
+ * @txf_ctl: The transmit FIFO control register.
48
+ * @t_out: The SMBus timeout register.
49
+ * @txf_sts: The transmit FIFO status register.
50
+ * @rxf_sts: The receive FIFO status register.
51
+ * @rxf_ctl: The receive FIFO control register.
52
+ * @rx_fifo: The FIFO buffer for receiving in FIFO mode.
53
+ * @rx_cur: The current position of rx_fifo.
54
* @status: The current status of the SMBus.
55
*/
56
typedef struct NPCM7xxSMBusState {
57
@@ -XXX,XX +XXX,XX @@ typedef struct NPCM7xxSMBusState {
58
uint8_t scllt;
59
uint8_t sclht;
60
61
+ uint8_t fif_ctl;
62
+ uint8_t fif_cts;
63
+ uint8_t fair_per;
64
+ uint8_t txf_ctl;
65
+ uint8_t t_out;
66
+ uint8_t txf_sts;
67
+ uint8_t rxf_sts;
68
+ uint8_t rxf_ctl;
69
+
32
+
70
+ uint8_t rx_fifo[NPCM7XX_SMBUS_FIFO_SIZE];
33
+### SME Misc
71
+ uint8_t rx_cur;
72
+
34
+
73
NPCM7xxSMBusStatus status;
35
+ZERO 11000000 00 001 00000000000 imm:8
74
} NPCM7xxSMBusState;
36
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
75
76
diff --git a/hw/i2c/npcm7xx_smbus.c b/hw/i2c/npcm7xx_smbus.c
77
index XXXXXXX..XXXXXXX 100644
37
index XXXXXXX..XXXXXXX 100644
78
--- a/hw/i2c/npcm7xx_smbus.c
38
--- a/target/arm/sme_helper.c
79
+++ b/hw/i2c/npcm7xx_smbus.c
39
+++ b/target/arm/sme_helper.c
80
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxSMBusBank1Register {
40
@@ -XXX,XX +XXX,XX @@ void helper_set_pstate_za(CPUARMState *env, uint32_t i)
81
#define NPCM7XX_ADDR_EN BIT(7)
41
memset(env->zarray, 0, sizeof(env->zarray));
82
#define NPCM7XX_ADDR_A(rv) extract8((rv), 0, 6)
42
}
83
43
}
84
+/* FIFO Mode Register Fields */
85
+/* FIF_CTL fields */
86
+#define NPCM7XX_SMBFIF_CTL_FIFO_EN BIT(4)
87
+#define NPCM7XX_SMBFIF_CTL_FAIR_RDY_IE BIT(2)
88
+#define NPCM7XX_SMBFIF_CTL_FAIR_RDY BIT(1)
89
+#define NPCM7XX_SMBFIF_CTL_FAIR_BUSY BIT(0)
90
+/* FIF_CTS fields */
91
+#define NPCM7XX_SMBFIF_CTS_STR BIT(7)
92
+#define NPCM7XX_SMBFIF_CTS_CLR_FIFO BIT(6)
93
+#define NPCM7XX_SMBFIF_CTS_RFTE_IE BIT(3)
94
+#define NPCM7XX_SMBFIF_CTS_RXF_TXE BIT(1)
95
+/* TXF_CTL fields */
96
+#define NPCM7XX_SMBTXF_CTL_THR_TXIE BIT(6)
97
+#define NPCM7XX_SMBTXF_CTL_TX_THR(rv) extract8((rv), 0, 5)
98
+/* T_OUT fields */
99
+#define NPCM7XX_SMBT_OUT_ST BIT(7)
100
+#define NPCM7XX_SMBT_OUT_IE BIT(6)
101
+#define NPCM7XX_SMBT_OUT_CLKDIV(rv) extract8((rv), 0, 6)
102
+/* TXF_STS fields */
103
+#define NPCM7XX_SMBTXF_STS_TX_THST BIT(6)
104
+#define NPCM7XX_SMBTXF_STS_TX_BYTES(rv) extract8((rv), 0, 5)
105
+/* RXF_STS fields */
106
+#define NPCM7XX_SMBRXF_STS_RX_THST BIT(6)
107
+#define NPCM7XX_SMBRXF_STS_RX_BYTES(rv) extract8((rv), 0, 5)
108
+/* RXF_CTL fields */
109
+#define NPCM7XX_SMBRXF_CTL_THR_RXIE BIT(6)
110
+#define NPCM7XX_SMBRXF_CTL_LAST BIT(5)
111
+#define NPCM7XX_SMBRXF_CTL_RX_THR(rv) extract8((rv), 0, 5)
112
+
44
+
113
#define KEEP_OLD_BIT(o, n, b) (((n) & (~(b))) | ((o) & (b)))
45
+void helper_sme_zero(CPUARMState *env, uint32_t imm, uint32_t svl)
114
#define WRITE_ONE_CLEAR(o, n, b) ((n) & (b) ? (o) & (~(b)) : (o))
115
116
#define NPCM7XX_SMBUS_ENABLED(s) ((s)->ctl2 & NPCM7XX_SMBCTL2_ENABLE)
117
+#define NPCM7XX_SMBUS_FIFO_ENABLED(s) ((s)->fif_ctl & \
118
+ NPCM7XX_SMBFIF_CTL_FIFO_EN)
119
120
/* VERSION fields values, read-only. */
121
#define NPCM7XX_SMBUS_VERSION_NUMBER 1
122
-#define NPCM7XX_SMBUS_VERSION_FIFO_SUPPORTED 0
123
+#define NPCM7XX_SMBUS_VERSION_FIFO_SUPPORTED 1
124
125
/* Reset values */
126
#define NPCM7XX_SMB_ST_INIT_VAL 0x00
127
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxSMBusBank1Register {
128
#define NPCM7XX_SMB_ADDR_INIT_VAL 0x00
129
#define NPCM7XX_SMB_SCLLT_INIT_VAL 0x00
130
#define NPCM7XX_SMB_SCLHT_INIT_VAL 0x00
131
+#define NPCM7XX_SMB_FIF_CTL_INIT_VAL 0x00
132
+#define NPCM7XX_SMB_FIF_CTS_INIT_VAL 0x00
133
+#define NPCM7XX_SMB_FAIR_PER_INIT_VAL 0x00
134
+#define NPCM7XX_SMB_TXF_CTL_INIT_VAL 0x00
135
+#define NPCM7XX_SMB_T_OUT_INIT_VAL 0x3f
136
+#define NPCM7XX_SMB_TXF_STS_INIT_VAL 0x00
137
+#define NPCM7XX_SMB_RXF_STS_INIT_VAL 0x00
138
+#define NPCM7XX_SMB_RXF_CTL_INIT_VAL 0x01
139
140
static uint8_t npcm7xx_smbus_get_version(void)
141
{
142
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_update_irq(NPCM7xxSMBusState *s)
143
(s->ctl1 & NPCM7XX_SMBCTL1_STASTRE &&
144
s->st & NPCM7XX_SMBST_SDAST) ||
145
(s->ctl1 & NPCM7XX_SMBCTL1_EOBINTE &&
146
- s->cst3 & NPCM7XX_SMBCST3_EO_BUSY));
147
+ s->cst3 & NPCM7XX_SMBCST3_EO_BUSY) ||
148
+ (s->rxf_ctl & NPCM7XX_SMBRXF_CTL_THR_RXIE &&
149
+ s->rxf_sts & NPCM7XX_SMBRXF_STS_RX_THST) ||
150
+ (s->txf_ctl & NPCM7XX_SMBTXF_CTL_THR_TXIE &&
151
+ s->txf_sts & NPCM7XX_SMBTXF_STS_TX_THST) ||
152
+ (s->fif_cts & NPCM7XX_SMBFIF_CTS_RFTE_IE &&
153
+ s->fif_cts & NPCM7XX_SMBFIF_CTS_RXF_TXE));
154
155
if (level) {
156
s->cst2 |= NPCM7XX_SMBCST2_INTSTS;
157
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_nack(NPCM7xxSMBusState *s)
158
s->status = NPCM7XX_SMBUS_STATUS_NEGACK;
159
}
160
161
+static void npcm7xx_smbus_clear_buffer(NPCM7xxSMBusState *s)
162
+{
46
+{
163
+ s->fif_cts &= ~NPCM7XX_SMBFIF_CTS_RXF_TXE;
47
+ uint32_t i;
164
+ s->txf_sts = 0;
165
+ s->rxf_sts = 0;
166
+}
167
+
48
+
168
static void npcm7xx_smbus_send_byte(NPCM7xxSMBusState *s, uint8_t value)
49
+ /*
169
{
50
+ * Special case clearing the entire ZA space.
170
int rv = i2c_send(s->bus, value);
51
+ * This falls into the CONSTRAINED UNPREDICTABLE zeroing of any
171
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_send_byte(NPCM7xxSMBusState *s, uint8_t value)
52
+ * parts of the ZA storage outside of SVL.
172
npcm7xx_smbus_nack(s);
53
+ */
173
} else {
54
+ if (imm == 0xff) {
174
s->st |= NPCM7XX_SMBST_SDAST;
55
+ memset(env->zarray, 0, sizeof(env->zarray));
175
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
176
+ s->fif_cts |= NPCM7XX_SMBFIF_CTS_RXF_TXE;
177
+ if (NPCM7XX_SMBTXF_STS_TX_BYTES(s->txf_sts) ==
178
+ NPCM7XX_SMBTXF_CTL_TX_THR(s->txf_ctl)) {
179
+ s->txf_sts = NPCM7XX_SMBTXF_STS_TX_THST;
180
+ } else {
181
+ s->txf_sts = 0;
182
+ }
183
+ }
184
}
185
trace_npcm7xx_smbus_send_byte((DEVICE(s)->canonical_path), value, !rv);
186
npcm7xx_smbus_update_irq(s);
187
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_recv_byte(NPCM7xxSMBusState *s)
188
npcm7xx_smbus_update_irq(s);
189
}
190
191
+static void npcm7xx_smbus_recv_fifo(NPCM7xxSMBusState *s)
192
+{
193
+ uint8_t expected_bytes = NPCM7XX_SMBRXF_CTL_RX_THR(s->rxf_ctl);
194
+ uint8_t received_bytes = NPCM7XX_SMBRXF_STS_RX_BYTES(s->rxf_sts);
195
+ uint8_t pos;
196
+
197
+ if (received_bytes == expected_bytes) {
198
+ return;
56
+ return;
199
+ }
57
+ }
200
+
58
+
201
+ while (received_bytes < expected_bytes &&
59
+ /*
202
+ received_bytes < NPCM7XX_SMBUS_FIFO_SIZE) {
60
+ * Recall that ZAnH.D[m] is spread across ZA[n+8*m],
203
+ pos = (s->rx_cur + received_bytes) % NPCM7XX_SMBUS_FIFO_SIZE;
61
+ * so each row is discontiguous within ZA[].
204
+ s->rx_fifo[pos] = i2c_recv(s->bus);
62
+ */
205
+ trace_npcm7xx_smbus_recv_byte((DEVICE(s)->canonical_path),
63
+ for (i = 0; i < svl; i++) {
206
+ s->rx_fifo[pos]);
64
+ if (imm & (1 << (i % 8))) {
207
+ ++received_bytes;
65
+ memset(&env->zarray[i], 0, svl);
208
+ }
209
+
210
+ trace_npcm7xx_smbus_recv_fifo((DEVICE(s)->canonical_path),
211
+ received_bytes, expected_bytes);
212
+ s->rxf_sts = received_bytes;
213
+ if (unlikely(received_bytes < expected_bytes)) {
214
+ qemu_log_mask(LOG_GUEST_ERROR,
215
+ "%s: invalid rx_thr value: 0x%02x\n",
216
+ DEVICE(s)->canonical_path, expected_bytes);
217
+ return;
218
+ }
219
+
220
+ s->rxf_sts |= NPCM7XX_SMBRXF_STS_RX_THST;
221
+ if (s->rxf_ctl & NPCM7XX_SMBRXF_CTL_LAST) {
222
+ trace_npcm7xx_smbus_nack(DEVICE(s)->canonical_path);
223
+ i2c_nack(s->bus);
224
+ s->rxf_ctl &= ~NPCM7XX_SMBRXF_CTL_LAST;
225
+ }
226
+ if (received_bytes == NPCM7XX_SMBUS_FIFO_SIZE) {
227
+ s->st |= NPCM7XX_SMBST_SDAST;
228
+ s->fif_cts |= NPCM7XX_SMBFIF_CTS_RXF_TXE;
229
+ } else if (!(s->rxf_ctl & NPCM7XX_SMBRXF_CTL_THR_RXIE)) {
230
+ s->st |= NPCM7XX_SMBST_SDAST;
231
+ } else {
232
+ s->st &= ~NPCM7XX_SMBST_SDAST;
233
+ }
234
+ npcm7xx_smbus_update_irq(s);
235
+}
236
+
237
+static void npcm7xx_smbus_read_byte_fifo(NPCM7xxSMBusState *s)
238
+{
239
+ uint8_t received_bytes = NPCM7XX_SMBRXF_STS_RX_BYTES(s->rxf_sts);
240
+
241
+ if (received_bytes == 0) {
242
+ npcm7xx_smbus_recv_fifo(s);
243
+ return;
244
+ }
245
+
246
+ s->sda = s->rx_fifo[s->rx_cur];
247
+ s->rx_cur = (s->rx_cur + 1u) % NPCM7XX_SMBUS_FIFO_SIZE;
248
+ --s->rxf_sts;
249
+ npcm7xx_smbus_update_irq(s);
250
+}
251
+
252
static void npcm7xx_smbus_start(NPCM7xxSMBusState *s)
253
{
254
/*
255
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_start(NPCM7xxSMBusState *s)
256
if (available) {
257
s->st |= NPCM7XX_SMBST_MODE | NPCM7XX_SMBST_XMIT | NPCM7XX_SMBST_SDAST;
258
s->cst |= NPCM7XX_SMBCST_BUSY;
259
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
260
+ s->fif_cts |= NPCM7XX_SMBFIF_CTS_RXF_TXE;
261
+ }
262
} else {
263
s->st &= ~NPCM7XX_SMBST_MODE;
264
s->cst &= ~NPCM7XX_SMBCST_BUSY;
265
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_send_address(NPCM7xxSMBusState *s, uint8_t value)
266
s->st |= NPCM7XX_SMBST_SDAST;
267
}
268
} else if (recv) {
269
- npcm7xx_smbus_recv_byte(s);
270
+ s->st |= NPCM7XX_SMBST_SDAST;
271
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
272
+ npcm7xx_smbus_recv_fifo(s);
273
+ } else {
274
+ npcm7xx_smbus_recv_byte(s);
275
+ }
276
+ } else if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
277
+ s->st |= NPCM7XX_SMBST_SDAST;
278
+ s->fif_cts |= NPCM7XX_SMBFIF_CTS_RXF_TXE;
279
}
280
npcm7xx_smbus_update_irq(s);
281
}
282
@@ -XXX,XX +XXX,XX @@ static uint8_t npcm7xx_smbus_read_sda(NPCM7xxSMBusState *s)
283
284
switch (s->status) {
285
case NPCM7XX_SMBUS_STATUS_STOPPING_LAST_RECEIVE:
286
- npcm7xx_smbus_execute_stop(s);
287
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
288
+ if (NPCM7XX_SMBRXF_STS_RX_BYTES(s->rxf_sts) <= 1) {
289
+ npcm7xx_smbus_execute_stop(s);
290
+ }
291
+ if (NPCM7XX_SMBRXF_STS_RX_BYTES(s->rxf_sts) == 0) {
292
+ qemu_log_mask(LOG_GUEST_ERROR,
293
+ "%s: read to SDA with an empty rx-fifo buffer, "
294
+ "result undefined: %u\n",
295
+ DEVICE(s)->canonical_path, s->sda);
296
+ break;
297
+ }
298
+ npcm7xx_smbus_read_byte_fifo(s);
299
+ value = s->sda;
300
+ } else {
301
+ npcm7xx_smbus_execute_stop(s);
302
+ }
303
break;
304
305
case NPCM7XX_SMBUS_STATUS_RECEIVING:
306
- npcm7xx_smbus_recv_byte(s);
307
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
308
+ npcm7xx_smbus_read_byte_fifo(s);
309
+ value = s->sda;
310
+ } else {
311
+ npcm7xx_smbus_recv_byte(s);
312
+ }
313
break;
314
315
default:
316
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_write_st(NPCM7xxSMBusState *s, uint8_t value)
317
}
318
319
if (value & NPCM7XX_SMBST_STASTR &&
320
- s->status == NPCM7XX_SMBUS_STATUS_RECEIVING) {
321
- npcm7xx_smbus_recv_byte(s);
322
+ s->status == NPCM7XX_SMBUS_STATUS_RECEIVING) {
323
+ if (NPCM7XX_SMBUS_FIFO_ENABLED(s)) {
324
+ npcm7xx_smbus_recv_fifo(s);
325
+ } else {
326
+ npcm7xx_smbus_recv_byte(s);
327
+ }
328
}
329
330
npcm7xx_smbus_update_irq(s);
331
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_write_ctl2(NPCM7xxSMBusState *s, uint8_t value)
332
s->st = 0;
333
s->cst3 = s->cst3 & (~NPCM7XX_SMBCST3_EO_BUSY);
334
s->cst = 0;
335
+ npcm7xx_smbus_clear_buffer(s);
336
}
337
}
338
339
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_write_ctl3(NPCM7xxSMBusState *s, uint8_t value)
340
NPCM7XX_SMBCTL3_SCL_LVL | NPCM7XX_SMBCTL3_SDA_LVL);
341
}
342
343
+static void npcm7xx_smbus_write_fif_ctl(NPCM7xxSMBusState *s, uint8_t value)
344
+{
345
+ uint8_t new_ctl = value;
346
+
347
+ new_ctl = KEEP_OLD_BIT(s->fif_ctl, new_ctl, NPCM7XX_SMBFIF_CTL_FAIR_RDY);
348
+ new_ctl = WRITE_ONE_CLEAR(new_ctl, value, NPCM7XX_SMBFIF_CTL_FAIR_RDY);
349
+ new_ctl = KEEP_OLD_BIT(s->fif_ctl, new_ctl, NPCM7XX_SMBFIF_CTL_FAIR_BUSY);
350
+ s->fif_ctl = new_ctl;
351
+}
352
+
353
+static void npcm7xx_smbus_write_fif_cts(NPCM7xxSMBusState *s, uint8_t value)
354
+{
355
+ s->fif_cts = WRITE_ONE_CLEAR(s->fif_cts, value, NPCM7XX_SMBFIF_CTS_STR);
356
+ s->fif_cts = WRITE_ONE_CLEAR(s->fif_cts, value, NPCM7XX_SMBFIF_CTS_RXF_TXE);
357
+ s->fif_cts = KEEP_OLD_BIT(value, s->fif_cts, NPCM7XX_SMBFIF_CTS_RFTE_IE);
358
+
359
+ if (value & NPCM7XX_SMBFIF_CTS_CLR_FIFO) {
360
+ npcm7xx_smbus_clear_buffer(s);
361
+ }
362
+}
363
+
364
+static void npcm7xx_smbus_write_txf_ctl(NPCM7xxSMBusState *s, uint8_t value)
365
+{
366
+ s->txf_ctl = value;
367
+}
368
+
369
+static void npcm7xx_smbus_write_t_out(NPCM7xxSMBusState *s, uint8_t value)
370
+{
371
+ uint8_t new_t_out = value;
372
+
373
+ if ((value & NPCM7XX_SMBT_OUT_ST) || (!(s->t_out & NPCM7XX_SMBT_OUT_ST))) {
374
+ new_t_out &= ~NPCM7XX_SMBT_OUT_ST;
375
+ } else {
376
+ new_t_out |= NPCM7XX_SMBT_OUT_ST;
377
+ }
378
+
379
+ s->t_out = new_t_out;
380
+}
381
+
382
+static void npcm7xx_smbus_write_txf_sts(NPCM7xxSMBusState *s, uint8_t value)
383
+{
384
+ s->txf_sts = WRITE_ONE_CLEAR(s->txf_sts, value, NPCM7XX_SMBTXF_STS_TX_THST);
385
+}
386
+
387
+static void npcm7xx_smbus_write_rxf_sts(NPCM7xxSMBusState *s, uint8_t value)
388
+{
389
+ if (value & NPCM7XX_SMBRXF_STS_RX_THST) {
390
+ s->rxf_sts &= ~NPCM7XX_SMBRXF_STS_RX_THST;
391
+ if (s->status == NPCM7XX_SMBUS_STATUS_RECEIVING) {
392
+ npcm7xx_smbus_recv_fifo(s);
393
+ }
66
+ }
394
+ }
67
+ }
395
+}
68
+}
396
+
69
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
397
+static void npcm7xx_smbus_write_rxf_ctl(NPCM7xxSMBusState *s, uint8_t value)
398
+{
399
+ uint8_t new_ctl = value;
400
+
401
+ if (!(value & NPCM7XX_SMBRXF_CTL_LAST)) {
402
+ new_ctl = KEEP_OLD_BIT(s->rxf_ctl, new_ctl, NPCM7XX_SMBRXF_CTL_LAST);
403
+ }
404
+ s->rxf_ctl = new_ctl;
405
+}
406
+
407
static uint64_t npcm7xx_smbus_read(void *opaque, hwaddr offset, unsigned size)
408
{
409
NPCM7xxSMBusState *s = opaque;
410
@@ -XXX,XX +XXX,XX @@ static uint64_t npcm7xx_smbus_read(void *opaque, hwaddr offset, unsigned size)
411
default:
412
if (bank) {
413
/* Bank 1 */
414
- qemu_log_mask(LOG_GUEST_ERROR,
415
- "%s: read from invalid offset 0x%" HWADDR_PRIx "\n",
416
- DEVICE(s)->canonical_path, offset);
417
+ switch (offset) {
418
+ case NPCM7XX_SMB_FIF_CTS:
419
+ value = s->fif_cts;
420
+ break;
421
+
422
+ case NPCM7XX_SMB_FAIR_PER:
423
+ value = s->fair_per;
424
+ break;
425
+
426
+ case NPCM7XX_SMB_TXF_CTL:
427
+ value = s->txf_ctl;
428
+ break;
429
+
430
+ case NPCM7XX_SMB_T_OUT:
431
+ value = s->t_out;
432
+ break;
433
+
434
+ case NPCM7XX_SMB_TXF_STS:
435
+ value = s->txf_sts;
436
+ break;
437
+
438
+ case NPCM7XX_SMB_RXF_STS:
439
+ value = s->rxf_sts;
440
+ break;
441
+
442
+ case NPCM7XX_SMB_RXF_CTL:
443
+ value = s->rxf_ctl;
444
+ break;
445
+
446
+ default:
447
+ qemu_log_mask(LOG_GUEST_ERROR,
448
+ "%s: read from invalid offset 0x%" HWADDR_PRIx "\n",
449
+ DEVICE(s)->canonical_path, offset);
450
+ break;
451
+ }
452
} else {
453
/* Bank 0 */
454
switch (offset) {
455
@@ -XXX,XX +XXX,XX @@ static uint64_t npcm7xx_smbus_read(void *opaque, hwaddr offset, unsigned size)
456
value = s->scllt;
457
break;
458
459
+ case NPCM7XX_SMB_FIF_CTL:
460
+ value = s->fif_ctl;
461
+ break;
462
+
463
case NPCM7XX_SMB_SCLHT:
464
value = s->sclht;
465
break;
466
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_write(void *opaque, hwaddr offset, uint64_t value,
467
default:
468
if (bank) {
469
/* Bank 1 */
470
- qemu_log_mask(LOG_GUEST_ERROR,
471
- "%s: write to invalid offset 0x%" HWADDR_PRIx "\n",
472
- DEVICE(s)->canonical_path, offset);
473
+ switch (offset) {
474
+ case NPCM7XX_SMB_FIF_CTS:
475
+ npcm7xx_smbus_write_fif_cts(s, value);
476
+ break;
477
+
478
+ case NPCM7XX_SMB_FAIR_PER:
479
+ s->fair_per = value;
480
+ break;
481
+
482
+ case NPCM7XX_SMB_TXF_CTL:
483
+ npcm7xx_smbus_write_txf_ctl(s, value);
484
+ break;
485
+
486
+ case NPCM7XX_SMB_T_OUT:
487
+ npcm7xx_smbus_write_t_out(s, value);
488
+ break;
489
+
490
+ case NPCM7XX_SMB_TXF_STS:
491
+ npcm7xx_smbus_write_txf_sts(s, value);
492
+ break;
493
+
494
+ case NPCM7XX_SMB_RXF_STS:
495
+ npcm7xx_smbus_write_rxf_sts(s, value);
496
+ break;
497
+
498
+ case NPCM7XX_SMB_RXF_CTL:
499
+ npcm7xx_smbus_write_rxf_ctl(s, value);
500
+ break;
501
+
502
+ default:
503
+ qemu_log_mask(LOG_GUEST_ERROR,
504
+ "%s: write to invalid offset 0x%" HWADDR_PRIx "\n",
505
+ DEVICE(s)->canonical_path, offset);
506
+ break;
507
+ }
508
} else {
509
/* Bank 0 */
510
switch (offset) {
511
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_write(void *opaque, hwaddr offset, uint64_t value,
512
s->scllt = value;
513
break;
514
515
+ case NPCM7XX_SMB_FIF_CTL:
516
+ npcm7xx_smbus_write_fif_ctl(s, value);
517
+ break;
518
+
519
case NPCM7XX_SMB_SCLHT:
520
s->sclht = value;
521
break;
522
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_smbus_enter_reset(Object *obj, ResetType type)
523
s->scllt = NPCM7XX_SMB_SCLLT_INIT_VAL;
524
s->sclht = NPCM7XX_SMB_SCLHT_INIT_VAL;
525
526
+ s->fif_ctl = NPCM7XX_SMB_FIF_CTL_INIT_VAL;
527
+ s->fif_cts = NPCM7XX_SMB_FIF_CTS_INIT_VAL;
528
+ s->fair_per = NPCM7XX_SMB_FAIR_PER_INIT_VAL;
529
+ s->txf_ctl = NPCM7XX_SMB_TXF_CTL_INIT_VAL;
530
+ s->t_out = NPCM7XX_SMB_T_OUT_INIT_VAL;
531
+ s->txf_sts = NPCM7XX_SMB_TXF_STS_INIT_VAL;
532
+ s->rxf_sts = NPCM7XX_SMB_RXF_STS_INIT_VAL;
533
+ s->rxf_ctl = NPCM7XX_SMB_RXF_CTL_INIT_VAL;
534
+
535
+ npcm7xx_smbus_clear_buffer(s);
536
s->status = NPCM7XX_SMBUS_STATUS_IDLE;
537
+ s->rx_cur = 0;
538
}
539
540
static void npcm7xx_smbus_hold_reset(Object *obj)
541
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription vmstate_npcm7xx_smbus = {
542
VMSTATE_UINT8_ARRAY(addr, NPCM7xxSMBusState, NPCM7XX_SMBUS_NR_ADDRS),
543
VMSTATE_UINT8(scllt, NPCM7xxSMBusState),
544
VMSTATE_UINT8(sclht, NPCM7xxSMBusState),
545
+ VMSTATE_UINT8(fif_ctl, NPCM7xxSMBusState),
546
+ VMSTATE_UINT8(fif_cts, NPCM7xxSMBusState),
547
+ VMSTATE_UINT8(fair_per, NPCM7xxSMBusState),
548
+ VMSTATE_UINT8(txf_ctl, NPCM7xxSMBusState),
549
+ VMSTATE_UINT8(t_out, NPCM7xxSMBusState),
550
+ VMSTATE_UINT8(txf_sts, NPCM7xxSMBusState),
551
+ VMSTATE_UINT8(rxf_sts, NPCM7xxSMBusState),
552
+ VMSTATE_UINT8(rxf_ctl, NPCM7xxSMBusState),
553
+ VMSTATE_UINT8_ARRAY(rx_fifo, NPCM7xxSMBusState,
554
+ NPCM7XX_SMBUS_FIFO_SIZE),
555
+ VMSTATE_UINT8(rx_cur, NPCM7xxSMBusState),
556
VMSTATE_END_OF_LIST(),
557
},
558
};
559
diff --git a/tests/qtest/npcm7xx_smbus-test.c b/tests/qtest/npcm7xx_smbus-test.c
560
index XXXXXXX..XXXXXXX 100644
70
index XXXXXXX..XXXXXXX 100644
561
--- a/tests/qtest/npcm7xx_smbus-test.c
71
--- a/target/arm/translate-sme.c
562
+++ b/tests/qtest/npcm7xx_smbus-test.c
72
+++ b/target/arm/translate-sme.c
563
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxSMBusBank1Register {
73
@@ -XXX,XX +XXX,XX @@
564
#define ADDR_EN BIT(7)
74
*/
565
#define ADDR_A(rv) extract8((rv), 0, 6)
75
566
76
#include "decode-sme.c.inc"
567
+/* FIF_CTL fields */
568
+#define FIF_CTL_FIFO_EN BIT(4)
569
+
570
+/* FIF_CTS fields */
571
+#define FIF_CTS_CLR_FIFO BIT(6)
572
+#define FIF_CTS_RFTE_IE BIT(3)
573
+#define FIF_CTS_RXF_TXE BIT(1)
574
+
575
+/* TXF_CTL fields */
576
+#define TXF_CTL_THR_TXIE BIT(6)
577
+#define TXF_CTL_TX_THR(rv) extract8((rv), 0, 5)
578
+
579
+/* TXF_STS fields */
580
+#define TXF_STS_TX_THST BIT(6)
581
+#define TXF_STS_TX_BYTES(rv) extract8((rv), 0, 5)
582
+
583
+/* RXF_CTL fields */
584
+#define RXF_CTL_THR_RXIE BIT(6)
585
+#define RXF_CTL_LAST BIT(5)
586
+#define RXF_CTL_RX_THR(rv) extract8((rv), 0, 5)
587
+
588
+/* RXF_STS fields */
589
+#define RXF_STS_RX_THST BIT(6)
590
+#define RXF_STS_RX_BYTES(rv) extract8((rv), 0, 5)
591
+
77
+
592
+
78
+
593
+static void choose_bank(QTestState *qts, uint64_t base_addr, uint8_t bank)
79
+static bool trans_ZERO(DisasContext *s, arg_ZERO *a)
594
+{
80
+{
595
+ uint8_t ctl3 = qtest_readb(qts, base_addr + OFFSET_CTL3);
81
+ if (!dc_isar_feature(aa64_sme, s)) {
596
+
82
+ return false;
597
+ if (bank) {
598
+ ctl3 |= CTL3_BNK_SEL;
599
+ } else {
600
+ ctl3 &= ~CTL3_BNK_SEL;
601
+ }
83
+ }
602
+
84
+ if (sme_za_enabled_check(s)) {
603
+ qtest_writeb(qts, base_addr + OFFSET_CTL3, ctl3);
85
+ gen_helper_sme_zero(cpu_env, tcg_constant_i32(a->imm),
86
+ tcg_constant_i32(streaming_vec_reg_size(s)));
87
+ }
88
+ return true;
604
+}
89
+}
605
606
static void check_running(QTestState *qts, uint64_t base_addr)
607
{
608
@@ -XXX,XX +XXX,XX @@ static void send_byte(QTestState *qts, uint64_t base_addr, uint8_t byte)
609
qtest_writeb(qts, base_addr + OFFSET_SDA, byte);
610
}
611
612
+static bool check_recv(QTestState *qts, uint64_t base_addr)
613
+{
614
+ uint8_t st, fif_ctl, rxf_ctl, rxf_sts;
615
+ bool fifo;
616
+
617
+ st = qtest_readb(qts, base_addr + OFFSET_ST);
618
+ choose_bank(qts, base_addr, 0);
619
+ fif_ctl = qtest_readb(qts, base_addr + OFFSET_FIF_CTL);
620
+ fifo = fif_ctl & FIF_CTL_FIFO_EN;
621
+ if (!fifo) {
622
+ return st == (ST_MODE | ST_SDAST);
623
+ }
624
+
625
+ choose_bank(qts, base_addr, 1);
626
+ rxf_ctl = qtest_readb(qts, base_addr + OFFSET_RXF_CTL);
627
+ rxf_sts = qtest_readb(qts, base_addr + OFFSET_RXF_STS);
628
+
629
+ if ((rxf_ctl & RXF_CTL_THR_RXIE) && RXF_STS_RX_BYTES(rxf_sts) < 16) {
630
+ return st == ST_MODE;
631
+ } else {
632
+ return st == (ST_MODE | ST_SDAST);
633
+ }
634
+}
635
+
636
static uint8_t recv_byte(QTestState *qts, uint64_t base_addr)
637
{
638
- g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==,
639
- ST_MODE | ST_SDAST);
640
+ g_assert_true(check_recv(qts, base_addr));
641
return qtest_readb(qts, base_addr + OFFSET_SDA);
642
}
643
644
@@ -XXX,XX +XXX,XX @@ static void send_address(QTestState *qts, uint64_t base_addr, uint8_t addr,
645
qtest_writeb(qts, base_addr + OFFSET_ST, ST_STASTR);
646
st = qtest_readb(qts, base_addr + OFFSET_ST);
647
if (recv) {
648
- g_assert_cmphex(st, ==, ST_MODE | ST_SDAST);
649
+ g_assert_true(check_recv(qts, base_addr));
650
} else {
651
g_assert_cmphex(st, ==, ST_MODE | ST_XMIT | ST_SDAST);
652
}
653
@@ -XXX,XX +XXX,XX @@ static void send_nack(QTestState *qts, uint64_t base_addr)
654
qtest_writeb(qts, base_addr + OFFSET_CTL1, ctl1);
655
}
656
657
+static void start_fifo_mode(QTestState *qts, uint64_t base_addr)
658
+{
659
+ choose_bank(qts, base_addr, 0);
660
+ qtest_writeb(qts, base_addr + OFFSET_FIF_CTL, FIF_CTL_FIFO_EN);
661
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_FIF_CTL) &
662
+ FIF_CTL_FIFO_EN);
663
+ choose_bank(qts, base_addr, 1);
664
+ qtest_writeb(qts, base_addr + OFFSET_FIF_CTS,
665
+ FIF_CTS_CLR_FIFO | FIF_CTS_RFTE_IE);
666
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_FIF_CTS), ==,
667
+ FIF_CTS_RFTE_IE);
668
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_TXF_STS), ==, 0);
669
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_RXF_STS), ==, 0);
670
+}
671
+
672
+static void start_recv_fifo(QTestState *qts, uint64_t base_addr, uint8_t bytes)
673
+{
674
+ choose_bank(qts, base_addr, 1);
675
+ qtest_writeb(qts, base_addr + OFFSET_TXF_CTL, 0);
676
+ qtest_writeb(qts, base_addr + OFFSET_RXF_CTL,
677
+ RXF_CTL_THR_RXIE | RXF_CTL_LAST | bytes);
678
+}
679
+
680
/* Check the SMBus's status is set correctly when disabled. */
681
static void test_disable_bus(gconstpointer data)
682
{
683
@@ -XXX,XX +XXX,XX @@ static void test_single_mode(gconstpointer data)
684
qtest_quit(qts);
685
}
686
687
+/* Check the SMBus can send and receive bytes in FIFO mode. */
688
+static void test_fifo_mode(gconstpointer data)
689
+{
690
+ intptr_t index = (intptr_t)data;
691
+ uint64_t base_addr = SMBUS_ADDR(index);
692
+ int irq = SMBUS_IRQ(index);
693
+ uint8_t value = 0x60;
694
+ QTestState *qts = qtest_init("-machine npcm750-evb");
695
+
696
+ qtest_irq_intercept_in(qts, "/machine/soc/a9mpcore/gic");
697
+ enable_bus(qts, base_addr);
698
+ start_fifo_mode(qts, base_addr);
699
+ g_assert_false(qtest_get_irq(qts, irq));
700
+
701
+ /* Sending */
702
+ start_transfer(qts, base_addr);
703
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, false, true);
704
+ choose_bank(qts, base_addr, 1);
705
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_FIF_CTS) &
706
+ FIF_CTS_RXF_TXE);
707
+ qtest_writeb(qts, base_addr + OFFSET_TXF_CTL, TXF_CTL_THR_TXIE);
708
+ send_byte(qts, base_addr, TMP105_REG_CONFIG);
709
+ send_byte(qts, base_addr, value);
710
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_FIF_CTS) &
711
+ FIF_CTS_RXF_TXE);
712
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_TXF_STS) &
713
+ TXF_STS_TX_THST);
714
+ g_assert_cmpuint(TXF_STS_TX_BYTES(
715
+ qtest_readb(qts, base_addr + OFFSET_TXF_STS)), ==, 0);
716
+ g_assert_true(qtest_get_irq(qts, irq));
717
+ stop_transfer(qts, base_addr);
718
+ check_stopped(qts, base_addr);
719
+
720
+ /* Receiving */
721
+ start_fifo_mode(qts, base_addr);
722
+ start_transfer(qts, base_addr);
723
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, false, true);
724
+ send_byte(qts, base_addr, TMP105_REG_CONFIG);
725
+ start_transfer(qts, base_addr);
726
+ qtest_writeb(qts, base_addr + OFFSET_FIF_CTS, FIF_CTS_RXF_TXE);
727
+ start_recv_fifo(qts, base_addr, 1);
728
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, true, true);
729
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_FIF_CTS) &
730
+ FIF_CTS_RXF_TXE);
731
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_RXF_STS) &
732
+ RXF_STS_RX_THST);
733
+ g_assert_cmpuint(RXF_STS_RX_BYTES(
734
+ qtest_readb(qts, base_addr + OFFSET_RXF_STS)), ==, 1);
735
+ send_nack(qts, base_addr);
736
+ stop_transfer(qts, base_addr);
737
+ check_running(qts, base_addr);
738
+ g_assert_cmphex(recv_byte(qts, base_addr), ==, value);
739
+ g_assert_cmpuint(RXF_STS_RX_BYTES(
740
+ qtest_readb(qts, base_addr + OFFSET_RXF_STS)), ==, 0);
741
+ check_stopped(qts, base_addr);
742
+ qtest_quit(qts);
743
+}
744
+
745
static void smbus_add_test(const char *name, int index, GTestDataFunc fn)
746
{
747
g_autofree char *full_name = g_strdup_printf(
748
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
749
750
for (i = 0; i < ARRAY_SIZE(evb_bus_list); ++i) {
751
add_test(single_mode, evb_bus_list[i]);
752
+ add_test(fifo_mode, evb_bus_list[i]);
753
}
754
755
return g_test_run();
756
diff --git a/hw/i2c/trace-events b/hw/i2c/trace-events
757
index XXXXXXX..XXXXXXX 100644
758
--- a/hw/i2c/trace-events
759
+++ b/hw/i2c/trace-events
760
@@ -XXX,XX +XXX,XX @@ npcm7xx_smbus_send_byte(const char *id, uint8_t value, int success) "%s send byt
761
npcm7xx_smbus_recv_byte(const char *id, uint8_t value) "%s recv byte: 0x%02x"
762
npcm7xx_smbus_stop(const char *id) "%s stopping"
763
npcm7xx_smbus_nack(const char *id) "%s nacking"
764
+npcm7xx_smbus_recv_fifo(const char *id, uint8_t received, uint8_t expected) "%s recv fifo: received %u, expected %u"
765
--
90
--
766
2.20.1
91
2.25.1
767
768
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
The places that use these are better off using untagged
3
We can reuse the SVE functions for implementing moves to/from
4
addresses, so do not provide a tagged versions. Rename
4
horizontal tile slices, but we need new ones for moves to/from
5
to make it clear about the address type.
5
vertical tile slices.
6
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20210212184902.1251044-16-richard.henderson@linaro.org
9
Message-id: 20220708151540.18136-20-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
11
---
12
include/exec/cpu_ldst.h | 4 ++--
12
target/arm/helper-sme.h | 12 +++
13
linux-user/qemu.h | 4 ++--
13
target/arm/helper-sve.h | 2 +
14
accel/tcg/user-exec.c | 3 ++-
14
target/arm/translate-a64.h | 8 ++
15
linux-user/mmap.c | 14 +++++++-------
15
target/arm/translate.h | 5 ++
16
linux-user/syscall.c | 2 +-
16
target/arm/sme.decode | 15 ++++
17
5 files changed, 14 insertions(+), 13 deletions(-)
17
target/arm/sme_helper.c | 151 ++++++++++++++++++++++++++++++++++++-
18
target/arm/sve_helper.c | 12 +++
19
target/arm/translate-sme.c | 127 +++++++++++++++++++++++++++++++
20
8 files changed, 331 insertions(+), 1 deletion(-)
18
21
19
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
22
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
20
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
21
--- a/include/exec/cpu_ldst.h
24
--- a/target/arm/helper-sme.h
22
+++ b/include/exec/cpu_ldst.h
25
+++ b/target/arm/helper-sme.h
23
@@ -XXX,XX +XXX,XX @@ static inline void *g2h(CPUState *cs, abi_ptr x)
26
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(set_pstate_sm, TCG_CALL_NO_RWG, void, env, i32)
24
return g2h_untagged(cpu_untagged_addr(cs, x));
27
DEF_HELPER_FLAGS_2(set_pstate_za, TCG_CALL_NO_RWG, void, env, i32)
28
29
DEF_HELPER_FLAGS_3(sme_zero, TCG_CALL_NO_RWG, void, env, i32, i32)
30
+
31
+/* Move to/from vertical array slices, i.e. columns, so 'c'. */
32
+DEF_HELPER_FLAGS_4(sme_mova_cz_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
33
+DEF_HELPER_FLAGS_4(sme_mova_zc_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
34
+DEF_HELPER_FLAGS_4(sme_mova_cz_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
35
+DEF_HELPER_FLAGS_4(sme_mova_zc_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
36
+DEF_HELPER_FLAGS_4(sme_mova_cz_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
37
+DEF_HELPER_FLAGS_4(sme_mova_zc_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
38
+DEF_HELPER_FLAGS_4(sme_mova_cz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
39
+DEF_HELPER_FLAGS_4(sme_mova_zc_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
40
+DEF_HELPER_FLAGS_4(sme_mova_cz_q, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
41
+DEF_HELPER_FLAGS_4(sme_mova_zc_q, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
42
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
43
index XXXXXXX..XXXXXXX 100644
44
--- a/target/arm/helper-sve.h
45
+++ b/target/arm/helper-sve.h
46
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sve_sel_zpzz_s, TCG_CALL_NO_RWG,
47
void, ptr, ptr, ptr, ptr, i32)
48
DEF_HELPER_FLAGS_5(sve_sel_zpzz_d, TCG_CALL_NO_RWG,
49
void, ptr, ptr, ptr, ptr, i32)
50
+DEF_HELPER_FLAGS_5(sve_sel_zpzz_q, TCG_CALL_NO_RWG,
51
+ void, ptr, ptr, ptr, ptr, i32)
52
53
DEF_HELPER_FLAGS_5(sve2_addp_zpzz_b, TCG_CALL_NO_RWG,
54
void, ptr, ptr, ptr, ptr, i32)
55
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
56
index XXXXXXX..XXXXXXX 100644
57
--- a/target/arm/translate-a64.h
58
+++ b/target/arm/translate-a64.h
59
@@ -XXX,XX +XXX,XX @@ static inline int pred_gvec_reg_size(DisasContext *s)
60
return size_for_gvec(pred_full_reg_size(s));
25
}
61
}
26
62
27
-static inline bool guest_addr_valid(abi_ulong x)
63
+/* Return a newly allocated pointer to the predicate register. */
28
+static inline bool guest_addr_valid_untagged(abi_ulong x)
64
+static inline TCGv_ptr pred_full_reg_ptr(DisasContext *s, int regno)
65
+{
66
+ TCGv_ptr ret = tcg_temp_new_ptr();
67
+ tcg_gen_addi_ptr(ret, cpu_env, pred_full_reg_offset(s, regno));
68
+ return ret;
69
+}
70
+
71
bool disas_sve(DisasContext *, uint32_t);
72
bool disas_sme(DisasContext *, uint32_t);
73
74
diff --git a/target/arm/translate.h b/target/arm/translate.h
75
index XXXXXXX..XXXXXXX 100644
76
--- a/target/arm/translate.h
77
+++ b/target/arm/translate.h
78
@@ -XXX,XX +XXX,XX @@ static inline int plus_2(DisasContext *s, int x)
79
return x + 2;
80
}
81
82
+static inline int plus_12(DisasContext *s, int x)
83
+{
84
+ return x + 12;
85
+}
86
+
87
static inline int times_2(DisasContext *s, int x)
29
{
88
{
30
return x <= GUEST_ADDR_MAX;
89
return x * 2;
31
}
90
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
32
91
index XXXXXXX..XXXXXXX 100644
33
-static inline bool guest_range_valid(abi_ulong start, abi_ulong len)
92
--- a/target/arm/sme.decode
34
+static inline bool guest_range_valid_untagged(abi_ulong start, abi_ulong len)
93
+++ b/target/arm/sme.decode
35
{
94
@@ -XXX,XX +XXX,XX @@
36
return len - 1 <= GUEST_ADDR_MAX && start <= GUEST_ADDR_MAX - len + 1;
95
### SME Misc
37
}
96
38
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
97
ZERO 11000000 00 001 00000000000 imm:8
39
index XXXXXXX..XXXXXXX 100644
98
+
40
--- a/linux-user/qemu.h
99
+### SME Move into/from Array
41
+++ b/linux-user/qemu.h
100
+
42
@@ -XXX,XX +XXX,XX @@ extern unsigned long guest_stack_size;
101
+%mova_rs 13:2 !function=plus_12
43
static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
102
+&mova esz rs pg zr za_imm v:bool to_vec:bool
44
{
103
+
45
if (size == 0
104
+MOVA 11000000 esz:2 00000 0 v:1 .. pg:3 zr:5 0 za_imm:4 \
46
- ? !guest_addr_valid(addr)
105
+ &mova to_vec=0 rs=%mova_rs
47
- : !guest_range_valid(addr, size)) {
106
+MOVA 11000000 11 00000 1 v:1 .. pg:3 zr:5 0 za_imm:4 \
48
+ ? !guest_addr_valid_untagged(addr)
107
+ &mova to_vec=0 rs=%mova_rs esz=4
49
+ : !guest_range_valid_untagged(addr, size)) {
108
+
50
return false;
109
+MOVA 11000000 esz:2 00001 0 v:1 .. pg:3 0 za_imm:4 zr:5 \
51
}
110
+ &mova to_vec=1 rs=%mova_rs
52
return page_check_range((target_ulong)addr, size, type) == 0;
111
+MOVA 11000000 11 00001 1 v:1 .. pg:3 0 za_imm:4 zr:5 \
53
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
112
+ &mova to_vec=1 rs=%mova_rs esz=4
54
index XXXXXXX..XXXXXXX 100644
113
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
55
--- a/accel/tcg/user-exec.c
114
index XXXXXXX..XXXXXXX 100644
56
+++ b/accel/tcg/user-exec.c
115
--- a/target/arm/sme_helper.c
57
@@ -XXX,XX +XXX,XX @@ static int probe_access_internal(CPUArchState *env, target_ulong addr,
116
+++ b/target/arm/sme_helper.c
58
g_assert_not_reached();
117
@@ -XXX,XX +XXX,XX @@
59
}
118
60
119
#include "qemu/osdep.h"
61
- if (!guest_addr_valid(addr) || page_check_range(addr, 1, flags) < 0) {
120
#include "cpu.h"
62
+ if (!guest_addr_valid_untagged(addr) ||
121
-#include "internals.h"
63
+ page_check_range(addr, 1, flags) < 0) {
122
+#include "tcg/tcg-gvec-desc.h"
64
if (nonfault) {
123
#include "exec/helper-proto.h"
65
return TLB_INVALID_MASK;
124
+#include "qemu/int128.h"
66
} else {
125
+#include "vec_internal.h"
67
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
126
68
index XXXXXXX..XXXXXXX 100644
127
/* ResetSVEState */
69
--- a/linux-user/mmap.c
128
void arm_reset_sve_state(CPUARMState *env)
70
+++ b/linux-user/mmap.c
129
@@ -XXX,XX +XXX,XX @@ void helper_sme_zero(CPUARMState *env, uint32_t imm, uint32_t svl)
71
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
72
}
73
len = TARGET_PAGE_ALIGN(len);
74
end = start + len;
75
- if (!guest_range_valid(start, len)) {
76
+ if (!guest_range_valid_untagged(start, len)) {
77
return -TARGET_ENOMEM;
78
}
79
if (len == 0) {
80
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
81
* It can fail only on 64-bit host with 32-bit target.
82
* On any other target/host host mmap() handles this error correctly.
83
*/
84
- if (end < start || !guest_range_valid(start, len)) {
85
+ if (end < start || !guest_range_valid_untagged(start, len)) {
86
errno = ENOMEM;
87
goto fail;
88
}
89
@@ -XXX,XX +XXX,XX @@ int target_munmap(abi_ulong start, abi_ulong len)
90
if (start & ~TARGET_PAGE_MASK)
91
return -TARGET_EINVAL;
92
len = TARGET_PAGE_ALIGN(len);
93
- if (len == 0 || !guest_range_valid(start, len)) {
94
+ if (len == 0 || !guest_range_valid_untagged(start, len)) {
95
return -TARGET_EINVAL;
96
}
97
98
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
99
int prot;
100
void *host_addr;
101
102
- if (!guest_range_valid(old_addr, old_size) ||
103
+ if (!guest_range_valid_untagged(old_addr, old_size) ||
104
((flags & MREMAP_FIXED) &&
105
- !guest_range_valid(new_addr, new_size)) ||
106
+ !guest_range_valid_untagged(new_addr, new_size)) ||
107
((flags & MREMAP_MAYMOVE) == 0 &&
108
- !guest_range_valid(old_addr, new_size))) {
109
+ !guest_range_valid_untagged(old_addr, new_size))) {
110
errno = ENOMEM;
111
return -1;
112
}
113
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
114
115
if (host_addr != MAP_FAILED) {
116
/* Check if address fits target address space */
117
- if (!guest_range_valid(h2g(host_addr), new_size)) {
118
+ if (!guest_range_valid_untagged(h2g(host_addr), new_size)) {
119
/* Revert mremap() changes */
120
host_addr = mremap(g2h_untagged(old_addr),
121
new_size, old_size, flags);
122
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
123
index XXXXXXX..XXXXXXX 100644
124
--- a/linux-user/syscall.c
125
+++ b/linux-user/syscall.c
126
@@ -XXX,XX +XXX,XX @@ static inline abi_ulong do_shmat(CPUArchState *cpu_env,
127
return -TARGET_EINVAL;
128
}
130
}
129
}
131
}
130
- if (!guest_range_valid(shmaddr, shm_info.shm_segsz)) {
132
}
131
+ if (!guest_range_valid_untagged(shmaddr, shm_info.shm_segsz)) {
133
+
132
return -TARGET_EINVAL;
134
+
135
+/*
136
+ * When considering the ZA storage as an array of elements of
137
+ * type T, the index within that array of the Nth element of
138
+ * a vertical slice of a tile can be calculated like this,
139
+ * regardless of the size of type T. This is because the tiles
140
+ * are interleaved, so if type T is size N bytes then row 1 of
141
+ * the tile is N rows away from row 0. The division by N to
142
+ * convert a byte offset into an array index and the multiplication
143
+ * by N to convert from vslice-index-within-the-tile to
144
+ * the index within the ZA storage cancel out.
145
+ */
146
+#define tile_vslice_index(i) ((i) * sizeof(ARMVectorReg))
147
+
148
+/*
149
+ * When doing byte arithmetic on the ZA storage, the element
150
+ * byteoff bytes away in a tile vertical slice is always this
151
+ * many bytes away in the ZA storage, regardless of the
152
+ * size of the tile element, assuming that byteoff is a multiple
153
+ * of the element size. Again this is because of the interleaving
154
+ * of the tiles. For instance if we have 1 byte per element then
155
+ * each row of the ZA storage has one byte of the vslice data,
156
+ * and (counting from 0) byte 8 goes in row 8 of the storage
157
+ * at offset (8 * row-size-in-bytes).
158
+ * If we have 8 bytes per element then each row of the ZA storage
159
+ * has 8 bytes of the data, but there are 8 interleaved tiles and
160
+ * so byte 8 of the data goes into row 1 of the tile,
161
+ * which is again row 8 of the storage, so the offset is still
162
+ * (8 * row-size-in-bytes). Similarly for other element sizes.
163
+ */
164
+#define tile_vslice_offset(byteoff) ((byteoff) * sizeof(ARMVectorReg))
165
+
166
+
167
+/*
168
+ * Move Zreg vector to ZArray column.
169
+ */
170
+#define DO_MOVA_C(NAME, TYPE, H) \
171
+void HELPER(NAME)(void *za, void *vn, void *vg, uint32_t desc) \
172
+{ \
173
+ int i, oprsz = simd_oprsz(desc); \
174
+ for (i = 0; i < oprsz; ) { \
175
+ uint16_t pg = *(uint16_t *)(vg + H1_2(i >> 3)); \
176
+ do { \
177
+ if (pg & 1) { \
178
+ *(TYPE *)(za + tile_vslice_offset(i)) = *(TYPE *)(vn + H(i)); \
179
+ } \
180
+ i += sizeof(TYPE); \
181
+ pg >>= sizeof(TYPE); \
182
+ } while (i & 15); \
183
+ } \
184
+}
185
+
186
+DO_MOVA_C(sme_mova_cz_b, uint8_t, H1)
187
+DO_MOVA_C(sme_mova_cz_h, uint16_t, H1_2)
188
+DO_MOVA_C(sme_mova_cz_s, uint32_t, H1_4)
189
+
190
+void HELPER(sme_mova_cz_d)(void *za, void *vn, void *vg, uint32_t desc)
191
+{
192
+ int i, oprsz = simd_oprsz(desc) / 8;
193
+ uint8_t *pg = vg;
194
+ uint64_t *n = vn;
195
+ uint64_t *a = za;
196
+
197
+ for (i = 0; i < oprsz; i++) {
198
+ if (pg[H1(i)] & 1) {
199
+ a[tile_vslice_index(i)] = n[i];
200
+ }
201
+ }
202
+}
203
+
204
+void HELPER(sme_mova_cz_q)(void *za, void *vn, void *vg, uint32_t desc)
205
+{
206
+ int i, oprsz = simd_oprsz(desc) / 16;
207
+ uint16_t *pg = vg;
208
+ Int128 *n = vn;
209
+ Int128 *a = za;
210
+
211
+ /*
212
+ * Int128 is used here simply to copy 16 bytes, and to simplify
213
+ * the address arithmetic.
214
+ */
215
+ for (i = 0; i < oprsz; i++) {
216
+ if (pg[H2(i)] & 1) {
217
+ a[tile_vslice_index(i)] = n[i];
218
+ }
219
+ }
220
+}
221
+
222
+#undef DO_MOVA_C
223
+
224
+/*
225
+ * Move ZArray column to Zreg vector.
226
+ */
227
+#define DO_MOVA_Z(NAME, TYPE, H) \
228
+void HELPER(NAME)(void *vd, void *za, void *vg, uint32_t desc) \
229
+{ \
230
+ int i, oprsz = simd_oprsz(desc); \
231
+ for (i = 0; i < oprsz; ) { \
232
+ uint16_t pg = *(uint16_t *)(vg + H1_2(i >> 3)); \
233
+ do { \
234
+ if (pg & 1) { \
235
+ *(TYPE *)(vd + H(i)) = *(TYPE *)(za + tile_vslice_offset(i)); \
236
+ } \
237
+ i += sizeof(TYPE); \
238
+ pg >>= sizeof(TYPE); \
239
+ } while (i & 15); \
240
+ } \
241
+}
242
+
243
+DO_MOVA_Z(sme_mova_zc_b, uint8_t, H1)
244
+DO_MOVA_Z(sme_mova_zc_h, uint16_t, H1_2)
245
+DO_MOVA_Z(sme_mova_zc_s, uint32_t, H1_4)
246
+
247
+void HELPER(sme_mova_zc_d)(void *vd, void *za, void *vg, uint32_t desc)
248
+{
249
+ int i, oprsz = simd_oprsz(desc) / 8;
250
+ uint8_t *pg = vg;
251
+ uint64_t *d = vd;
252
+ uint64_t *a = za;
253
+
254
+ for (i = 0; i < oprsz; i++) {
255
+ if (pg[H1(i)] & 1) {
256
+ d[i] = a[tile_vslice_index(i)];
257
+ }
258
+ }
259
+}
260
+
261
+void HELPER(sme_mova_zc_q)(void *vd, void *za, void *vg, uint32_t desc)
262
+{
263
+ int i, oprsz = simd_oprsz(desc) / 16;
264
+ uint16_t *pg = vg;
265
+ Int128 *d = vd;
266
+ Int128 *a = za;
267
+
268
+ /*
269
+ * Int128 is used here simply to copy 16 bytes, and to simplify
270
+ * the address arithmetic.
271
+ */
272
+ for (i = 0; i < oprsz; i++, za += sizeof(ARMVectorReg)) {
273
+ if (pg[H2(i)] & 1) {
274
+ d[i] = a[tile_vslice_index(i)];
275
+ }
276
+ }
277
+}
278
+
279
+#undef DO_MOVA_Z
280
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
281
index XXXXXXX..XXXXXXX 100644
282
--- a/target/arm/sve_helper.c
283
+++ b/target/arm/sve_helper.c
284
@@ -XXX,XX +XXX,XX @@ void HELPER(sve_sel_zpzz_d)(void *vd, void *vn, void *vm,
133
}
285
}
134
286
}
287
288
+void HELPER(sve_sel_zpzz_q)(void *vd, void *vn, void *vm,
289
+ void *vg, uint32_t desc)
290
+{
291
+ intptr_t i, opr_sz = simd_oprsz(desc) / 16;
292
+ Int128 *d = vd, *n = vn, *m = vm;
293
+ uint16_t *pg = vg;
294
+
295
+ for (i = 0; i < opr_sz; i += 1) {
296
+ d[i] = (pg[H2(i)] & 1 ? n : m)[i];
297
+ }
298
+}
299
+
300
/* Two operand comparison controlled by a predicate.
301
* ??? It is very tempting to want to be able to expand this inline
302
* with x86 instructions, e.g.
303
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
304
index XXXXXXX..XXXXXXX 100644
305
--- a/target/arm/translate-sme.c
306
+++ b/target/arm/translate-sme.c
307
@@ -XXX,XX +XXX,XX @@
308
#include "decode-sme.c.inc"
309
310
311
+/*
312
+ * Resolve tile.size[index] to a host pointer, where tile and index
313
+ * are always decoded together, dependent on the element size.
314
+ */
315
+static TCGv_ptr get_tile_rowcol(DisasContext *s, int esz, int rs,
316
+ int tile_index, bool vertical)
317
+{
318
+ int tile = tile_index >> (4 - esz);
319
+ int index = esz == MO_128 ? 0 : extract32(tile_index, 0, 4 - esz);
320
+ int pos, len, offset;
321
+ TCGv_i32 tmp;
322
+ TCGv_ptr addr;
323
+
324
+ /* Compute the final index, which is Rs+imm. */
325
+ tmp = tcg_temp_new_i32();
326
+ tcg_gen_trunc_tl_i32(tmp, cpu_reg(s, rs));
327
+ tcg_gen_addi_i32(tmp, tmp, index);
328
+
329
+ /* Prepare a power-of-two modulo via extraction of @len bits. */
330
+ len = ctz32(streaming_vec_reg_size(s)) - esz;
331
+
332
+ if (vertical) {
333
+ /*
334
+ * Compute the byte offset of the index within the tile:
335
+ * (index % (svl / size)) * size
336
+ * = (index % (svl >> esz)) << esz
337
+ * Perform the power-of-two modulo via extraction of the low @len bits.
338
+ * Perform the multiply by shifting left by @pos bits.
339
+ * Perform these operations simultaneously via deposit into zero.
340
+ */
341
+ pos = esz;
342
+ tcg_gen_deposit_z_i32(tmp, tmp, pos, len);
343
+
344
+ /*
345
+ * For big-endian, adjust the indexed column byte offset within
346
+ * the uint64_t host words that make up env->zarray[].
347
+ */
348
+ if (HOST_BIG_ENDIAN && esz < MO_64) {
349
+ tcg_gen_xori_i32(tmp, tmp, 8 - (1 << esz));
350
+ }
351
+ } else {
352
+ /*
353
+ * Compute the byte offset of the index within the tile:
354
+ * (index % (svl / size)) * (size * sizeof(row))
355
+ * = (index % (svl >> esz)) << (esz + log2(sizeof(row)))
356
+ */
357
+ pos = esz + ctz32(sizeof(ARMVectorReg));
358
+ tcg_gen_deposit_z_i32(tmp, tmp, pos, len);
359
+
360
+ /* Row slices are always aligned and need no endian adjustment. */
361
+ }
362
+
363
+ /* The tile byte offset within env->zarray is the row. */
364
+ offset = tile * sizeof(ARMVectorReg);
365
+
366
+ /* Include the byte offset of zarray to make this relative to env. */
367
+ offset += offsetof(CPUARMState, zarray);
368
+ tcg_gen_addi_i32(tmp, tmp, offset);
369
+
370
+ /* Add the byte offset to env to produce the final pointer. */
371
+ addr = tcg_temp_new_ptr();
372
+ tcg_gen_ext_i32_ptr(addr, tmp);
373
+ tcg_temp_free_i32(tmp);
374
+ tcg_gen_add_ptr(addr, addr, cpu_env);
375
+
376
+ return addr;
377
+}
378
+
379
static bool trans_ZERO(DisasContext *s, arg_ZERO *a)
380
{
381
if (!dc_isar_feature(aa64_sme, s)) {
382
@@ -XXX,XX +XXX,XX @@ static bool trans_ZERO(DisasContext *s, arg_ZERO *a)
383
}
384
return true;
385
}
386
+
387
+static bool trans_MOVA(DisasContext *s, arg_MOVA *a)
388
+{
389
+ static gen_helper_gvec_4 * const h_fns[5] = {
390
+ gen_helper_sve_sel_zpzz_b, gen_helper_sve_sel_zpzz_h,
391
+ gen_helper_sve_sel_zpzz_s, gen_helper_sve_sel_zpzz_d,
392
+ gen_helper_sve_sel_zpzz_q
393
+ };
394
+ static gen_helper_gvec_3 * const cz_fns[5] = {
395
+ gen_helper_sme_mova_cz_b, gen_helper_sme_mova_cz_h,
396
+ gen_helper_sme_mova_cz_s, gen_helper_sme_mova_cz_d,
397
+ gen_helper_sme_mova_cz_q,
398
+ };
399
+ static gen_helper_gvec_3 * const zc_fns[5] = {
400
+ gen_helper_sme_mova_zc_b, gen_helper_sme_mova_zc_h,
401
+ gen_helper_sme_mova_zc_s, gen_helper_sme_mova_zc_d,
402
+ gen_helper_sme_mova_zc_q,
403
+ };
404
+
405
+ TCGv_ptr t_za, t_zr, t_pg;
406
+ TCGv_i32 t_desc;
407
+ int svl;
408
+
409
+ if (!dc_isar_feature(aa64_sme, s)) {
410
+ return false;
411
+ }
412
+ if (!sme_smza_enabled_check(s)) {
413
+ return true;
414
+ }
415
+
416
+ t_za = get_tile_rowcol(s, a->esz, a->rs, a->za_imm, a->v);
417
+ t_zr = vec_full_reg_ptr(s, a->zr);
418
+ t_pg = pred_full_reg_ptr(s, a->pg);
419
+
420
+ svl = streaming_vec_reg_size(s);
421
+ t_desc = tcg_constant_i32(simd_desc(svl, svl, 0));
422
+
423
+ if (a->v) {
424
+ /* Vertical slice -- use sme mova helpers. */
425
+ if (a->to_vec) {
426
+ zc_fns[a->esz](t_zr, t_za, t_pg, t_desc);
427
+ } else {
428
+ cz_fns[a->esz](t_za, t_zr, t_pg, t_desc);
429
+ }
430
+ } else {
431
+ /* Horizontal slice -- reuse sve sel helpers. */
432
+ if (a->to_vec) {
433
+ h_fns[a->esz](t_zr, t_za, t_zr, t_pg, t_desc);
434
+ } else {
435
+ h_fns[a->esz](t_za, t_zr, t_za, t_pg, t_desc);
436
+ }
437
+ }
438
+
439
+ tcg_temp_free_ptr(t_za);
440
+ tcg_temp_free_ptr(t_zr);
441
+ tcg_temp_free_ptr(t_pg);
442
+
443
+ return true;
444
+}
135
--
445
--
136
2.20.1
446
2.25.1
137
138
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Move everything related to syndromes to a new file,
3
We cannot reuse the SVE functions for LD[1-4] and ST[1-4],
4
which can be shared with linux-user.
4
because those functions accept only a Zreg register number.
5
For SME, we want to pass a pointer into ZA storage.
5
6
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Message-id: 20220708151540.18136-21-richard.henderson@linaro.org
9
Message-id: 20210212184902.1251044-26-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
11
---
12
target/arm/internals.h | 245 +-----------------------------------
12
target/arm/helper-sme.h | 82 +++++
13
target/arm/syndrome.h | 273 +++++++++++++++++++++++++++++++++++++++++
13
target/arm/sme.decode | 9 +
14
2 files changed, 274 insertions(+), 244 deletions(-)
14
target/arm/sme_helper.c | 595 +++++++++++++++++++++++++++++++++++++
15
create mode 100644 target/arm/syndrome.h
15
target/arm/translate-sme.c | 70 +++++
16
4 files changed, 756 insertions(+)
16
17
17
diff --git a/target/arm/internals.h b/target/arm/internals.h
18
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
18
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/internals.h
20
--- a/target/arm/helper-sme.h
20
+++ b/target/arm/internals.h
21
+++ b/target/arm/helper-sme.h
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sme_mova_cz_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
23
DEF_HELPER_FLAGS_4(sme_mova_zc_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
24
DEF_HELPER_FLAGS_4(sme_mova_cz_q, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
25
DEF_HELPER_FLAGS_4(sme_mova_zc_q, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
26
+
27
+DEF_HELPER_FLAGS_5(sme_ld1b_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
28
+DEF_HELPER_FLAGS_5(sme_ld1b_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_5(sme_ld1b_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_5(sme_ld1b_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
31
+
32
+DEF_HELPER_FLAGS_5(sme_ld1h_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
33
+DEF_HELPER_FLAGS_5(sme_ld1h_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
34
+DEF_HELPER_FLAGS_5(sme_ld1h_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_5(sme_ld1h_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_5(sme_ld1h_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_5(sme_ld1h_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
38
+DEF_HELPER_FLAGS_5(sme_ld1h_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_5(sme_ld1h_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
40
+
41
+DEF_HELPER_FLAGS_5(sme_ld1s_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
42
+DEF_HELPER_FLAGS_5(sme_ld1s_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
43
+DEF_HELPER_FLAGS_5(sme_ld1s_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_5(sme_ld1s_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_5(sme_ld1s_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_5(sme_ld1s_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_5(sme_ld1s_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_5(sme_ld1s_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
49
+
50
+DEF_HELPER_FLAGS_5(sme_ld1d_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_5(sme_ld1d_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_5(sme_ld1d_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
53
+DEF_HELPER_FLAGS_5(sme_ld1d_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_5(sme_ld1d_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_5(sme_ld1d_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_5(sme_ld1d_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_5(sme_ld1d_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_5(sme_ld1q_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_5(sme_ld1q_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_5(sme_ld1q_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
62
+DEF_HELPER_FLAGS_5(sme_ld1q_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_5(sme_ld1q_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_5(sme_ld1q_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_5(sme_ld1q_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_5(sme_ld1q_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_5(sme_st1b_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_5(sme_st1b_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_5(sme_st1b_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_5(sme_st1b_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_5(sme_st1h_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_5(sme_st1h_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_5(sme_st1h_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_5(sme_st1h_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_5(sme_st1h_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_5(sme_st1h_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_5(sme_st1h_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
80
+DEF_HELPER_FLAGS_5(sme_st1h_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
81
+
82
+DEF_HELPER_FLAGS_5(sme_st1s_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_5(sme_st1s_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_5(sme_st1s_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_5(sme_st1s_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_5(sme_st1s_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_5(sme_st1s_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_5(sme_st1s_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
89
+DEF_HELPER_FLAGS_5(sme_st1s_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
90
+
91
+DEF_HELPER_FLAGS_5(sme_st1d_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_5(sme_st1d_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
93
+DEF_HELPER_FLAGS_5(sme_st1d_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_5(sme_st1d_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
95
+DEF_HELPER_FLAGS_5(sme_st1d_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_5(sme_st1d_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
97
+DEF_HELPER_FLAGS_5(sme_st1d_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_5(sme_st1d_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
99
+
100
+DEF_HELPER_FLAGS_5(sme_st1q_be_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_5(sme_st1q_le_h, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_5(sme_st1q_be_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_5(sme_st1q_le_v, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_5(sme_st1q_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_5(sme_st1q_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_5(sme_st1q_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_5(sme_st1q_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
108
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
109
index XXXXXXX..XXXXXXX 100644
110
--- a/target/arm/sme.decode
111
+++ b/target/arm/sme.decode
112
@@ -XXX,XX +XXX,XX @@ MOVA 11000000 esz:2 00001 0 v:1 .. pg:3 0 za_imm:4 zr:5 \
113
&mova to_vec=1 rs=%mova_rs
114
MOVA 11000000 11 00001 1 v:1 .. pg:3 0 za_imm:4 zr:5 \
115
&mova to_vec=1 rs=%mova_rs esz=4
116
+
117
+### SME Memory
118
+
119
+&ldst esz rs pg rn rm za_imm v:bool st:bool
120
+
121
+LDST1 1110000 0 esz:2 st:1 rm:5 v:1 .. pg:3 rn:5 0 za_imm:4 \
122
+ &ldst rs=%mova_rs
123
+LDST1 1110000 111 st:1 rm:5 v:1 .. pg:3 rn:5 0 za_imm:4 \
124
+ &ldst esz=4 rs=%mova_rs
125
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
126
index XXXXXXX..XXXXXXX 100644
127
--- a/target/arm/sme_helper.c
128
+++ b/target/arm/sme_helper.c
21
@@ -XXX,XX +XXX,XX @@
129
@@ -XXX,XX +XXX,XX @@
22
#define TARGET_ARM_INTERNALS_H
130
23
131
#include "qemu/osdep.h"
24
#include "hw/registerfields.h"
132
#include "cpu.h"
25
+#include "syndrome.h"
133
+#include "internals.h"
26
134
#include "tcg/tcg-gvec-desc.h"
27
/* register banks for CPU modes */
135
#include "exec/helper-proto.h"
28
#define BANK_USRSYS 0
136
+#include "exec/cpu_ldst.h"
29
@@ -XXX,XX +XXX,XX @@ static inline bool extended_addresses_enabled(CPUARMState *env)
137
+#include "exec/exec-all.h"
30
(arm_feature(env, ARM_FEATURE_LPAE) && (tcr->raw_tcr & TTBCR_EAE));
138
#include "qemu/int128.h"
139
#include "vec_internal.h"
140
+#include "sve_ldst_internal.h"
141
142
/* ResetSVEState */
143
void arm_reset_sve_state(CPUARMState *env)
144
@@ -XXX,XX +XXX,XX @@ void HELPER(sme_mova_zc_q)(void *vd, void *za, void *vg, uint32_t desc)
31
}
145
}
32
146
33
-/* Valid Syndrome Register EC field values */
147
#undef DO_MOVA_Z
34
-enum arm_exception_class {
148
+
35
- EC_UNCATEGORIZED = 0x00,
36
- EC_WFX_TRAP = 0x01,
37
- EC_CP15RTTRAP = 0x03,
38
- EC_CP15RRTTRAP = 0x04,
39
- EC_CP14RTTRAP = 0x05,
40
- EC_CP14DTTRAP = 0x06,
41
- EC_ADVSIMDFPACCESSTRAP = 0x07,
42
- EC_FPIDTRAP = 0x08,
43
- EC_PACTRAP = 0x09,
44
- EC_CP14RRTTRAP = 0x0c,
45
- EC_BTITRAP = 0x0d,
46
- EC_ILLEGALSTATE = 0x0e,
47
- EC_AA32_SVC = 0x11,
48
- EC_AA32_HVC = 0x12,
49
- EC_AA32_SMC = 0x13,
50
- EC_AA64_SVC = 0x15,
51
- EC_AA64_HVC = 0x16,
52
- EC_AA64_SMC = 0x17,
53
- EC_SYSTEMREGISTERTRAP = 0x18,
54
- EC_SVEACCESSTRAP = 0x19,
55
- EC_INSNABORT = 0x20,
56
- EC_INSNABORT_SAME_EL = 0x21,
57
- EC_PCALIGNMENT = 0x22,
58
- EC_DATAABORT = 0x24,
59
- EC_DATAABORT_SAME_EL = 0x25,
60
- EC_SPALIGNMENT = 0x26,
61
- EC_AA32_FPTRAP = 0x28,
62
- EC_AA64_FPTRAP = 0x2c,
63
- EC_SERROR = 0x2f,
64
- EC_BREAKPOINT = 0x30,
65
- EC_BREAKPOINT_SAME_EL = 0x31,
66
- EC_SOFTWARESTEP = 0x32,
67
- EC_SOFTWARESTEP_SAME_EL = 0x33,
68
- EC_WATCHPOINT = 0x34,
69
- EC_WATCHPOINT_SAME_EL = 0x35,
70
- EC_AA32_BKPT = 0x38,
71
- EC_VECTORCATCH = 0x3a,
72
- EC_AA64_BKPT = 0x3c,
73
-};
74
-
75
-#define ARM_EL_EC_SHIFT 26
76
-#define ARM_EL_IL_SHIFT 25
77
-#define ARM_EL_ISV_SHIFT 24
78
-#define ARM_EL_IL (1 << ARM_EL_IL_SHIFT)
79
-#define ARM_EL_ISV (1 << ARM_EL_ISV_SHIFT)
80
-
81
-static inline uint32_t syn_get_ec(uint32_t syn)
82
-{
83
- return syn >> ARM_EL_EC_SHIFT;
84
-}
85
-
86
-/* Utility functions for constructing various kinds of syndrome value.
87
- * Note that in general we follow the AArch64 syndrome values; in a
88
- * few cases the value in HSR for exceptions taken to AArch32 Hyp
89
- * mode differs slightly, and we fix this up when populating HSR in
90
- * arm_cpu_do_interrupt_aarch32_hyp().
91
- * The exception is FP/SIMD access traps -- these report extra information
92
- * when taking an exception to AArch32. For those we include the extra coproc
93
- * and TA fields, and mask them out when taking the exception to AArch64.
94
- */
95
-static inline uint32_t syn_uncategorized(void)
96
-{
97
- return (EC_UNCATEGORIZED << ARM_EL_EC_SHIFT) | ARM_EL_IL;
98
-}
99
-
100
-static inline uint32_t syn_aa64_svc(uint32_t imm16)
101
-{
102
- return (EC_AA64_SVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
103
-}
104
-
105
-static inline uint32_t syn_aa64_hvc(uint32_t imm16)
106
-{
107
- return (EC_AA64_HVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
108
-}
109
-
110
-static inline uint32_t syn_aa64_smc(uint32_t imm16)
111
-{
112
- return (EC_AA64_SMC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
113
-}
114
-
115
-static inline uint32_t syn_aa32_svc(uint32_t imm16, bool is_16bit)
116
-{
117
- return (EC_AA32_SVC << ARM_EL_EC_SHIFT) | (imm16 & 0xffff)
118
- | (is_16bit ? 0 : ARM_EL_IL);
119
-}
120
-
121
-static inline uint32_t syn_aa32_hvc(uint32_t imm16)
122
-{
123
- return (EC_AA32_HVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
124
-}
125
-
126
-static inline uint32_t syn_aa32_smc(void)
127
-{
128
- return (EC_AA32_SMC << ARM_EL_EC_SHIFT) | ARM_EL_IL;
129
-}
130
-
131
-static inline uint32_t syn_aa64_bkpt(uint32_t imm16)
132
-{
133
- return (EC_AA64_BKPT << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
134
-}
135
-
136
-static inline uint32_t syn_aa32_bkpt(uint32_t imm16, bool is_16bit)
137
-{
138
- return (EC_AA32_BKPT << ARM_EL_EC_SHIFT) | (imm16 & 0xffff)
139
- | (is_16bit ? 0 : ARM_EL_IL);
140
-}
141
-
142
-static inline uint32_t syn_aa64_sysregtrap(int op0, int op1, int op2,
143
- int crn, int crm, int rt,
144
- int isread)
145
-{
146
- return (EC_SYSTEMREGISTERTRAP << ARM_EL_EC_SHIFT) | ARM_EL_IL
147
- | (op0 << 20) | (op2 << 17) | (op1 << 14) | (crn << 10) | (rt << 5)
148
- | (crm << 1) | isread;
149
-}
150
-
151
-static inline uint32_t syn_cp14_rt_trap(int cv, int cond, int opc1, int opc2,
152
- int crn, int crm, int rt, int isread,
153
- bool is_16bit)
154
-{
155
- return (EC_CP14RTTRAP << ARM_EL_EC_SHIFT)
156
- | (is_16bit ? 0 : ARM_EL_IL)
157
- | (cv << 24) | (cond << 20) | (opc2 << 17) | (opc1 << 14)
158
- | (crn << 10) | (rt << 5) | (crm << 1) | isread;
159
-}
160
-
161
-static inline uint32_t syn_cp15_rt_trap(int cv, int cond, int opc1, int opc2,
162
- int crn, int crm, int rt, int isread,
163
- bool is_16bit)
164
-{
165
- return (EC_CP15RTTRAP << ARM_EL_EC_SHIFT)
166
- | (is_16bit ? 0 : ARM_EL_IL)
167
- | (cv << 24) | (cond << 20) | (opc2 << 17) | (opc1 << 14)
168
- | (crn << 10) | (rt << 5) | (crm << 1) | isread;
169
-}
170
-
171
-static inline uint32_t syn_cp14_rrt_trap(int cv, int cond, int opc1, int crm,
172
- int rt, int rt2, int isread,
173
- bool is_16bit)
174
-{
175
- return (EC_CP14RRTTRAP << ARM_EL_EC_SHIFT)
176
- | (is_16bit ? 0 : ARM_EL_IL)
177
- | (cv << 24) | (cond << 20) | (opc1 << 16)
178
- | (rt2 << 10) | (rt << 5) | (crm << 1) | isread;
179
-}
180
-
181
-static inline uint32_t syn_cp15_rrt_trap(int cv, int cond, int opc1, int crm,
182
- int rt, int rt2, int isread,
183
- bool is_16bit)
184
-{
185
- return (EC_CP15RRTTRAP << ARM_EL_EC_SHIFT)
186
- | (is_16bit ? 0 : ARM_EL_IL)
187
- | (cv << 24) | (cond << 20) | (opc1 << 16)
188
- | (rt2 << 10) | (rt << 5) | (crm << 1) | isread;
189
-}
190
-
191
-static inline uint32_t syn_fp_access_trap(int cv, int cond, bool is_16bit)
192
-{
193
- /* AArch32 FP trap or any AArch64 FP/SIMD trap: TA == 0 coproc == 0xa */
194
- return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
195
- | (is_16bit ? 0 : ARM_EL_IL)
196
- | (cv << 24) | (cond << 20) | 0xa;
197
-}
198
-
199
-static inline uint32_t syn_simd_access_trap(int cv, int cond, bool is_16bit)
200
-{
201
- /* AArch32 SIMD trap: TA == 1 coproc == 0 */
202
- return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
203
- | (is_16bit ? 0 : ARM_EL_IL)
204
- | (cv << 24) | (cond << 20) | (1 << 5);
205
-}
206
-
207
-static inline uint32_t syn_sve_access_trap(void)
208
-{
209
- return EC_SVEACCESSTRAP << ARM_EL_EC_SHIFT;
210
-}
211
-
212
-static inline uint32_t syn_pactrap(void)
213
-{
214
- return EC_PACTRAP << ARM_EL_EC_SHIFT;
215
-}
216
-
217
-static inline uint32_t syn_btitrap(int btype)
218
-{
219
- return (EC_BTITRAP << ARM_EL_EC_SHIFT) | btype;
220
-}
221
-
222
-static inline uint32_t syn_insn_abort(int same_el, int ea, int s1ptw, int fsc)
223
-{
224
- return (EC_INSNABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
225
- | ARM_EL_IL | (ea << 9) | (s1ptw << 7) | fsc;
226
-}
227
-
228
-static inline uint32_t syn_data_abort_no_iss(int same_el, int fnv,
229
- int ea, int cm, int s1ptw,
230
- int wnr, int fsc)
231
-{
232
- return (EC_DATAABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
233
- | ARM_EL_IL
234
- | (fnv << 10) | (ea << 9) | (cm << 8) | (s1ptw << 7)
235
- | (wnr << 6) | fsc;
236
-}
237
-
238
-static inline uint32_t syn_data_abort_with_iss(int same_el,
239
- int sas, int sse, int srt,
240
- int sf, int ar,
241
- int ea, int cm, int s1ptw,
242
- int wnr, int fsc,
243
- bool is_16bit)
244
-{
245
- return (EC_DATAABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
246
- | (is_16bit ? 0 : ARM_EL_IL)
247
- | ARM_EL_ISV | (sas << 22) | (sse << 21) | (srt << 16)
248
- | (sf << 15) | (ar << 14)
249
- | (ea << 9) | (cm << 8) | (s1ptw << 7) | (wnr << 6) | fsc;
250
-}
251
-
252
-static inline uint32_t syn_swstep(int same_el, int isv, int ex)
253
-{
254
- return (EC_SOFTWARESTEP << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
255
- | ARM_EL_IL | (isv << 24) | (ex << 6) | 0x22;
256
-}
257
-
258
-static inline uint32_t syn_watchpoint(int same_el, int cm, int wnr)
259
-{
260
- return (EC_WATCHPOINT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
261
- | ARM_EL_IL | (cm << 8) | (wnr << 6) | 0x22;
262
-}
263
-
264
-static inline uint32_t syn_breakpoint(int same_el)
265
-{
266
- return (EC_BREAKPOINT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
267
- | ARM_EL_IL | 0x22;
268
-}
269
-
270
-static inline uint32_t syn_wfx(int cv, int cond, int ti, bool is_16bit)
271
-{
272
- return (EC_WFX_TRAP << ARM_EL_EC_SHIFT) |
273
- (is_16bit ? 0 : (1 << ARM_EL_IL_SHIFT)) |
274
- (cv << 24) | (cond << 20) | ti;
275
-}
276
-
277
/* Update a QEMU watchpoint based on the information the guest has set in the
278
* DBGWCR<n>_EL1 and DBGWVR<n>_EL1 registers.
279
*/
280
diff --git a/target/arm/syndrome.h b/target/arm/syndrome.h
281
new file mode 100644
282
index XXXXXXX..XXXXXXX
283
--- /dev/null
284
+++ b/target/arm/syndrome.h
285
@@ -XXX,XX +XXX,XX @@
286
+/*
149
+/*
287
+ * QEMU ARM CPU -- syndrome functions and types
150
+ * Clear elements in a tile slice comprising len bytes.
288
+ *
289
+ * Copyright (c) 2014 Linaro Ltd
290
+ *
291
+ * This program is free software; you can redistribute it and/or
292
+ * modify it under the terms of the GNU General Public License
293
+ * as published by the Free Software Foundation; either version 2
294
+ * of the License, or (at your option) any later version.
295
+ *
296
+ * This program is distributed in the hope that it will be useful,
297
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
298
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
299
+ * GNU General Public License for more details.
300
+ *
301
+ * You should have received a copy of the GNU General Public License
302
+ * along with this program; if not, see
303
+ * <http://www.gnu.org/licenses/gpl-2.0.html>
304
+ *
305
+ * This header defines functions, types, etc which need to be shared
306
+ * between different source files within target/arm/ but which are
307
+ * private to it and not required by the rest of QEMU.
308
+ */
151
+ */
309
+
152
+
310
+#ifndef TARGET_ARM_SYNDROME_H
153
+typedef void ClearFn(void *ptr, size_t off, size_t len);
311
+#define TARGET_ARM_SYNDROME_H
154
+
312
+
155
+static void clear_horizontal(void *ptr, size_t off, size_t len)
313
+/* Valid Syndrome Register EC field values */
156
+{
314
+enum arm_exception_class {
157
+ memset(ptr + off, 0, len);
315
+ EC_UNCATEGORIZED = 0x00,
158
+}
316
+ EC_WFX_TRAP = 0x01,
159
+
317
+ EC_CP15RTTRAP = 0x03,
160
+static void clear_vertical_b(void *vptr, size_t off, size_t len)
318
+ EC_CP15RRTTRAP = 0x04,
161
+{
319
+ EC_CP14RTTRAP = 0x05,
162
+ for (size_t i = 0; i < len; ++i) {
320
+ EC_CP14DTTRAP = 0x06,
163
+ *(uint8_t *)(vptr + tile_vslice_offset(i + off)) = 0;
321
+ EC_ADVSIMDFPACCESSTRAP = 0x07,
164
+ }
322
+ EC_FPIDTRAP = 0x08,
165
+}
323
+ EC_PACTRAP = 0x09,
166
+
324
+ EC_CP14RRTTRAP = 0x0c,
167
+static void clear_vertical_h(void *vptr, size_t off, size_t len)
325
+ EC_BTITRAP = 0x0d,
168
+{
326
+ EC_ILLEGALSTATE = 0x0e,
169
+ for (size_t i = 0; i < len; i += 2) {
327
+ EC_AA32_SVC = 0x11,
170
+ *(uint16_t *)(vptr + tile_vslice_offset(i + off)) = 0;
328
+ EC_AA32_HVC = 0x12,
171
+ }
329
+ EC_AA32_SMC = 0x13,
172
+}
330
+ EC_AA64_SVC = 0x15,
173
+
331
+ EC_AA64_HVC = 0x16,
174
+static void clear_vertical_s(void *vptr, size_t off, size_t len)
332
+ EC_AA64_SMC = 0x17,
175
+{
333
+ EC_SYSTEMREGISTERTRAP = 0x18,
176
+ for (size_t i = 0; i < len; i += 4) {
334
+ EC_SVEACCESSTRAP = 0x19,
177
+ *(uint32_t *)(vptr + tile_vslice_offset(i + off)) = 0;
335
+ EC_INSNABORT = 0x20,
178
+ }
336
+ EC_INSNABORT_SAME_EL = 0x21,
179
+}
337
+ EC_PCALIGNMENT = 0x22,
180
+
338
+ EC_DATAABORT = 0x24,
181
+static void clear_vertical_d(void *vptr, size_t off, size_t len)
339
+ EC_DATAABORT_SAME_EL = 0x25,
182
+{
340
+ EC_SPALIGNMENT = 0x26,
183
+ for (size_t i = 0; i < len; i += 8) {
341
+ EC_AA32_FPTRAP = 0x28,
184
+ *(uint64_t *)(vptr + tile_vslice_offset(i + off)) = 0;
342
+ EC_AA64_FPTRAP = 0x2c,
185
+ }
343
+ EC_SERROR = 0x2f,
186
+}
344
+ EC_BREAKPOINT = 0x30,
187
+
345
+ EC_BREAKPOINT_SAME_EL = 0x31,
188
+static void clear_vertical_q(void *vptr, size_t off, size_t len)
346
+ EC_SOFTWARESTEP = 0x32,
189
+{
347
+ EC_SOFTWARESTEP_SAME_EL = 0x33,
190
+ for (size_t i = 0; i < len; i += 16) {
348
+ EC_WATCHPOINT = 0x34,
191
+ memset(vptr + tile_vslice_offset(i + off), 0, 16);
349
+ EC_WATCHPOINT_SAME_EL = 0x35,
192
+ }
350
+ EC_AA32_BKPT = 0x38,
351
+ EC_VECTORCATCH = 0x3a,
352
+ EC_AA64_BKPT = 0x3c,
353
+};
354
+
355
+#define ARM_EL_EC_SHIFT 26
356
+#define ARM_EL_IL_SHIFT 25
357
+#define ARM_EL_ISV_SHIFT 24
358
+#define ARM_EL_IL (1 << ARM_EL_IL_SHIFT)
359
+#define ARM_EL_ISV (1 << ARM_EL_ISV_SHIFT)
360
+
361
+static inline uint32_t syn_get_ec(uint32_t syn)
362
+{
363
+ return syn >> ARM_EL_EC_SHIFT;
364
+}
193
+}
365
+
194
+
366
+/*
195
+/*
367
+ * Utility functions for constructing various kinds of syndrome value.
196
+ * Copy elements from an array into a tile slice comprising len bytes.
368
+ * Note that in general we follow the AArch64 syndrome values; in a
369
+ * few cases the value in HSR for exceptions taken to AArch32 Hyp
370
+ * mode differs slightly, and we fix this up when populating HSR in
371
+ * arm_cpu_do_interrupt_aarch32_hyp().
372
+ * The exception is FP/SIMD access traps -- these report extra information
373
+ * when taking an exception to AArch32. For those we include the extra coproc
374
+ * and TA fields, and mask them out when taking the exception to AArch64.
375
+ */
197
+ */
376
+static inline uint32_t syn_uncategorized(void)
198
+
377
+{
199
+typedef void CopyFn(void *dst, const void *src, size_t len);
378
+ return (EC_UNCATEGORIZED << ARM_EL_EC_SHIFT) | ARM_EL_IL;
200
+
379
+}
201
+static void copy_horizontal(void *dst, const void *src, size_t len)
380
+
202
+{
381
+static inline uint32_t syn_aa64_svc(uint32_t imm16)
203
+ memcpy(dst, src, len);
382
+{
204
+}
383
+ return (EC_AA64_SVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
205
+
384
+}
206
+static void copy_vertical_b(void *vdst, const void *vsrc, size_t len)
385
+
207
+{
386
+static inline uint32_t syn_aa64_hvc(uint32_t imm16)
208
+ const uint8_t *src = vsrc;
387
+{
209
+ uint8_t *dst = vdst;
388
+ return (EC_AA64_HVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
210
+ size_t i;
389
+}
211
+
390
+
212
+ for (i = 0; i < len; ++i) {
391
+static inline uint32_t syn_aa64_smc(uint32_t imm16)
213
+ dst[tile_vslice_index(i)] = src[i];
392
+{
214
+ }
393
+ return (EC_AA64_SMC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
215
+}
394
+}
216
+
395
+
217
+static void copy_vertical_h(void *vdst, const void *vsrc, size_t len)
396
+static inline uint32_t syn_aa32_svc(uint32_t imm16, bool is_16bit)
218
+{
397
+{
219
+ const uint16_t *src = vsrc;
398
+ return (EC_AA32_SVC << ARM_EL_EC_SHIFT) | (imm16 & 0xffff)
220
+ uint16_t *dst = vdst;
399
+ | (is_16bit ? 0 : ARM_EL_IL);
221
+ size_t i;
400
+}
222
+
401
+
223
+ for (i = 0; i < len / 2; ++i) {
402
+static inline uint32_t syn_aa32_hvc(uint32_t imm16)
224
+ dst[tile_vslice_index(i)] = src[i];
403
+{
225
+ }
404
+ return (EC_AA32_HVC << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
226
+}
405
+}
227
+
406
+
228
+static void copy_vertical_s(void *vdst, const void *vsrc, size_t len)
407
+static inline uint32_t syn_aa32_smc(void)
229
+{
408
+{
230
+ const uint32_t *src = vsrc;
409
+ return (EC_AA32_SMC << ARM_EL_EC_SHIFT) | ARM_EL_IL;
231
+ uint32_t *dst = vdst;
410
+}
232
+ size_t i;
411
+
233
+
412
+static inline uint32_t syn_aa64_bkpt(uint32_t imm16)
234
+ for (i = 0; i < len / 4; ++i) {
413
+{
235
+ dst[tile_vslice_index(i)] = src[i];
414
+ return (EC_AA64_BKPT << ARM_EL_EC_SHIFT) | ARM_EL_IL | (imm16 & 0xffff);
236
+ }
415
+}
237
+}
416
+
238
+
417
+static inline uint32_t syn_aa32_bkpt(uint32_t imm16, bool is_16bit)
239
+static void copy_vertical_d(void *vdst, const void *vsrc, size_t len)
418
+{
240
+{
419
+ return (EC_AA32_BKPT << ARM_EL_EC_SHIFT) | (imm16 & 0xffff)
241
+ const uint64_t *src = vsrc;
420
+ | (is_16bit ? 0 : ARM_EL_IL);
242
+ uint64_t *dst = vdst;
421
+}
243
+ size_t i;
422
+
244
+
423
+static inline uint32_t syn_aa64_sysregtrap(int op0, int op1, int op2,
245
+ for (i = 0; i < len / 8; ++i) {
424
+ int crn, int crm, int rt,
246
+ dst[tile_vslice_index(i)] = src[i];
425
+ int isread)
247
+ }
426
+{
248
+}
427
+ return (EC_SYSTEMREGISTERTRAP << ARM_EL_EC_SHIFT) | ARM_EL_IL
249
+
428
+ | (op0 << 20) | (op2 << 17) | (op1 << 14) | (crn << 10) | (rt << 5)
250
+static void copy_vertical_q(void *vdst, const void *vsrc, size_t len)
429
+ | (crm << 1) | isread;
251
+{
430
+}
252
+ for (size_t i = 0; i < len; i += 16) {
431
+
253
+ memcpy(vdst + tile_vslice_offset(i), vsrc + i, 16);
432
+static inline uint32_t syn_cp14_rt_trap(int cv, int cond, int opc1, int opc2,
254
+ }
433
+ int crn, int crm, int rt, int isread,
255
+}
434
+ bool is_16bit)
256
+
435
+{
257
+/*
436
+ return (EC_CP14RTTRAP << ARM_EL_EC_SHIFT)
258
+ * Host and TLB primitives for vertical tile slice addressing.
437
+ | (is_16bit ? 0 : ARM_EL_IL)
259
+ */
438
+ | (cv << 24) | (cond << 20) | (opc2 << 17) | (opc1 << 14)
260
+
439
+ | (crn << 10) | (rt << 5) | (crm << 1) | isread;
261
+#define DO_LD(NAME, TYPE, HOST, TLB) \
440
+}
262
+static inline void sme_##NAME##_v_host(void *za, intptr_t off, void *host) \
441
+
263
+{ \
442
+static inline uint32_t syn_cp15_rt_trap(int cv, int cond, int opc1, int opc2,
264
+ TYPE val = HOST(host); \
443
+ int crn, int crm, int rt, int isread,
265
+ *(TYPE *)(za + tile_vslice_offset(off)) = val; \
444
+ bool is_16bit)
266
+} \
445
+{
267
+static inline void sme_##NAME##_v_tlb(CPUARMState *env, void *za, \
446
+ return (EC_CP15RTTRAP << ARM_EL_EC_SHIFT)
268
+ intptr_t off, target_ulong addr, uintptr_t ra) \
447
+ | (is_16bit ? 0 : ARM_EL_IL)
269
+{ \
448
+ | (cv << 24) | (cond << 20) | (opc2 << 17) | (opc1 << 14)
270
+ TYPE val = TLB(env, useronly_clean_ptr(addr), ra); \
449
+ | (crn << 10) | (rt << 5) | (crm << 1) | isread;
271
+ *(TYPE *)(za + tile_vslice_offset(off)) = val; \
450
+}
272
+}
451
+
273
+
452
+static inline uint32_t syn_cp14_rrt_trap(int cv, int cond, int opc1, int crm,
274
+#define DO_ST(NAME, TYPE, HOST, TLB) \
453
+ int rt, int rt2, int isread,
275
+static inline void sme_##NAME##_v_host(void *za, intptr_t off, void *host) \
454
+ bool is_16bit)
276
+{ \
455
+{
277
+ TYPE val = *(TYPE *)(za + tile_vslice_offset(off)); \
456
+ return (EC_CP14RRTTRAP << ARM_EL_EC_SHIFT)
278
+ HOST(host, val); \
457
+ | (is_16bit ? 0 : ARM_EL_IL)
279
+} \
458
+ | (cv << 24) | (cond << 20) | (opc1 << 16)
280
+static inline void sme_##NAME##_v_tlb(CPUARMState *env, void *za, \
459
+ | (rt2 << 10) | (rt << 5) | (crm << 1) | isread;
281
+ intptr_t off, target_ulong addr, uintptr_t ra) \
460
+}
282
+{ \
461
+
283
+ TYPE val = *(TYPE *)(za + tile_vslice_offset(off)); \
462
+static inline uint32_t syn_cp15_rrt_trap(int cv, int cond, int opc1, int crm,
284
+ TLB(env, useronly_clean_ptr(addr), val, ra); \
463
+ int rt, int rt2, int isread,
285
+}
464
+ bool is_16bit)
286
+
465
+{
287
+/*
466
+ return (EC_CP15RRTTRAP << ARM_EL_EC_SHIFT)
288
+ * The ARMVectorReg elements are stored in host-endian 64-bit units.
467
+ | (is_16bit ? 0 : ARM_EL_IL)
289
+ * For 128-bit quantities, the sequence defined by the Elem[] pseudocode
468
+ | (cv << 24) | (cond << 20) | (opc1 << 16)
290
+ * corresponds to storing the two 64-bit pieces in little-endian order.
469
+ | (rt2 << 10) | (rt << 5) | (crm << 1) | isread;
291
+ */
470
+}
292
+#define DO_LDQ(HNAME, VNAME, BE, HOST, TLB) \
471
+
293
+static inline void HNAME##_host(void *za, intptr_t off, void *host) \
472
+static inline uint32_t syn_fp_access_trap(int cv, int cond, bool is_16bit)
294
+{ \
473
+{
295
+ uint64_t val0 = HOST(host), val1 = HOST(host + 8); \
474
+ /* AArch32 FP trap or any AArch64 FP/SIMD trap: TA == 0 coproc == 0xa */
296
+ uint64_t *ptr = za + off; \
475
+ return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
297
+ ptr[0] = BE ? val1 : val0, ptr[1] = BE ? val0 : val1; \
476
+ | (is_16bit ? 0 : ARM_EL_IL)
298
+} \
477
+ | (cv << 24) | (cond << 20) | 0xa;
299
+static inline void VNAME##_v_host(void *za, intptr_t off, void *host) \
478
+}
300
+{ \
479
+
301
+ HNAME##_host(za, tile_vslice_offset(off), host); \
480
+static inline uint32_t syn_simd_access_trap(int cv, int cond, bool is_16bit)
302
+} \
481
+{
303
+static inline void HNAME##_tlb(CPUARMState *env, void *za, intptr_t off, \
482
+ /* AArch32 SIMD trap: TA == 1 coproc == 0 */
304
+ target_ulong addr, uintptr_t ra) \
483
+ return (EC_ADVSIMDFPACCESSTRAP << ARM_EL_EC_SHIFT)
305
+{ \
484
+ | (is_16bit ? 0 : ARM_EL_IL)
306
+ uint64_t val0 = TLB(env, useronly_clean_ptr(addr), ra); \
485
+ | (cv << 24) | (cond << 20) | (1 << 5);
307
+ uint64_t val1 = TLB(env, useronly_clean_ptr(addr + 8), ra); \
486
+}
308
+ uint64_t *ptr = za + off; \
487
+
309
+ ptr[0] = BE ? val1 : val0, ptr[1] = BE ? val0 : val1; \
488
+static inline uint32_t syn_sve_access_trap(void)
310
+} \
489
+{
311
+static inline void VNAME##_v_tlb(CPUARMState *env, void *za, intptr_t off, \
490
+ return EC_SVEACCESSTRAP << ARM_EL_EC_SHIFT;
312
+ target_ulong addr, uintptr_t ra) \
491
+}
313
+{ \
492
+
314
+ HNAME##_tlb(env, za, tile_vslice_offset(off), addr, ra); \
493
+static inline uint32_t syn_pactrap(void)
315
+}
494
+{
316
+
495
+ return EC_PACTRAP << ARM_EL_EC_SHIFT;
317
+#define DO_STQ(HNAME, VNAME, BE, HOST, TLB) \
496
+}
318
+static inline void HNAME##_host(void *za, intptr_t off, void *host) \
497
+
319
+{ \
498
+static inline uint32_t syn_btitrap(int btype)
320
+ uint64_t *ptr = za + off; \
499
+{
321
+ HOST(host, ptr[BE]); \
500
+ return (EC_BTITRAP << ARM_EL_EC_SHIFT) | btype;
322
+ HOST(host + 1, ptr[!BE]); \
501
+}
323
+} \
502
+
324
+static inline void VNAME##_v_host(void *za, intptr_t off, void *host) \
503
+static inline uint32_t syn_insn_abort(int same_el, int ea, int s1ptw, int fsc)
325
+{ \
504
+{
326
+ HNAME##_host(za, tile_vslice_offset(off), host); \
505
+ return (EC_INSNABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
327
+} \
506
+ | ARM_EL_IL | (ea << 9) | (s1ptw << 7) | fsc;
328
+static inline void HNAME##_tlb(CPUARMState *env, void *za, intptr_t off, \
507
+}
329
+ target_ulong addr, uintptr_t ra) \
508
+
330
+{ \
509
+static inline uint32_t syn_data_abort_no_iss(int same_el, int fnv,
331
+ uint64_t *ptr = za + off; \
510
+ int ea, int cm, int s1ptw,
332
+ TLB(env, useronly_clean_ptr(addr), ptr[BE], ra); \
511
+ int wnr, int fsc)
333
+ TLB(env, useronly_clean_ptr(addr + 8), ptr[!BE], ra); \
512
+{
334
+} \
513
+ return (EC_DATAABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
335
+static inline void VNAME##_v_tlb(CPUARMState *env, void *za, intptr_t off, \
514
+ | ARM_EL_IL
336
+ target_ulong addr, uintptr_t ra) \
515
+ | (fnv << 10) | (ea << 9) | (cm << 8) | (s1ptw << 7)
337
+{ \
516
+ | (wnr << 6) | fsc;
338
+ HNAME##_tlb(env, za, tile_vslice_offset(off), addr, ra); \
517
+}
339
+}
518
+
340
+
519
+static inline uint32_t syn_data_abort_with_iss(int same_el,
341
+DO_LD(ld1b, uint8_t, ldub_p, cpu_ldub_data_ra)
520
+ int sas, int sse, int srt,
342
+DO_LD(ld1h_be, uint16_t, lduw_be_p, cpu_lduw_be_data_ra)
521
+ int sf, int ar,
343
+DO_LD(ld1h_le, uint16_t, lduw_le_p, cpu_lduw_le_data_ra)
522
+ int ea, int cm, int s1ptw,
344
+DO_LD(ld1s_be, uint32_t, ldl_be_p, cpu_ldl_be_data_ra)
523
+ int wnr, int fsc,
345
+DO_LD(ld1s_le, uint32_t, ldl_le_p, cpu_ldl_le_data_ra)
524
+ bool is_16bit)
346
+DO_LD(ld1d_be, uint64_t, ldq_be_p, cpu_ldq_be_data_ra)
525
+{
347
+DO_LD(ld1d_le, uint64_t, ldq_le_p, cpu_ldq_le_data_ra)
526
+ return (EC_DATAABORT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
348
+
527
+ | (is_16bit ? 0 : ARM_EL_IL)
349
+DO_LDQ(sve_ld1qq_be, sme_ld1q_be, 1, ldq_be_p, cpu_ldq_be_data_ra)
528
+ | ARM_EL_ISV | (sas << 22) | (sse << 21) | (srt << 16)
350
+DO_LDQ(sve_ld1qq_le, sme_ld1q_le, 0, ldq_le_p, cpu_ldq_le_data_ra)
529
+ | (sf << 15) | (ar << 14)
351
+
530
+ | (ea << 9) | (cm << 8) | (s1ptw << 7) | (wnr << 6) | fsc;
352
+DO_ST(st1b, uint8_t, stb_p, cpu_stb_data_ra)
531
+}
353
+DO_ST(st1h_be, uint16_t, stw_be_p, cpu_stw_be_data_ra)
532
+
354
+DO_ST(st1h_le, uint16_t, stw_le_p, cpu_stw_le_data_ra)
533
+static inline uint32_t syn_swstep(int same_el, int isv, int ex)
355
+DO_ST(st1s_be, uint32_t, stl_be_p, cpu_stl_be_data_ra)
534
+{
356
+DO_ST(st1s_le, uint32_t, stl_le_p, cpu_stl_le_data_ra)
535
+ return (EC_SOFTWARESTEP << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
357
+DO_ST(st1d_be, uint64_t, stq_be_p, cpu_stq_be_data_ra)
536
+ | ARM_EL_IL | (isv << 24) | (ex << 6) | 0x22;
358
+DO_ST(st1d_le, uint64_t, stq_le_p, cpu_stq_le_data_ra)
537
+}
359
+
538
+
360
+DO_STQ(sve_st1qq_be, sme_st1q_be, 1, stq_be_p, cpu_stq_be_data_ra)
539
+static inline uint32_t syn_watchpoint(int same_el, int cm, int wnr)
361
+DO_STQ(sve_st1qq_le, sme_st1q_le, 0, stq_le_p, cpu_stq_le_data_ra)
540
+{
362
+
541
+ return (EC_WATCHPOINT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
363
+#undef DO_LD
542
+ | ARM_EL_IL | (cm << 8) | (wnr << 6) | 0x22;
364
+#undef DO_ST
543
+}
365
+#undef DO_LDQ
544
+
366
+#undef DO_STQ
545
+static inline uint32_t syn_breakpoint(int same_el)
367
+
546
+{
368
+/*
547
+ return (EC_BREAKPOINT << ARM_EL_EC_SHIFT) | (same_el << ARM_EL_EC_SHIFT)
369
+ * Common helper for all contiguous predicated loads.
548
+ | ARM_EL_IL | 0x22;
370
+ */
549
+}
371
+
550
+
372
+static inline QEMU_ALWAYS_INLINE
551
+static inline uint32_t syn_wfx(int cv, int cond, int ti, bool is_16bit)
373
+void sme_ld1(CPUARMState *env, void *za, uint64_t *vg,
552
+{
374
+ const target_ulong addr, uint32_t desc, const uintptr_t ra,
553
+ return (EC_WFX_TRAP << ARM_EL_EC_SHIFT) |
375
+ const int esz, uint32_t mtedesc, bool vertical,
554
+ (is_16bit ? 0 : (1 << ARM_EL_IL_SHIFT)) |
376
+ sve_ldst1_host_fn *host_fn,
555
+ (cv << 24) | (cond << 20) | ti;
377
+ sve_ldst1_tlb_fn *tlb_fn,
556
+}
378
+ ClearFn *clr_fn,
557
+
379
+ CopyFn *cpy_fn)
558
+#endif /* TARGET_ARM_SYNDROME_H */
380
+{
381
+ const intptr_t reg_max = simd_oprsz(desc);
382
+ const intptr_t esize = 1 << esz;
383
+ intptr_t reg_off, reg_last;
384
+ SVEContLdSt info;
385
+ void *host;
386
+ int flags;
387
+
388
+ /* Find the active elements. */
389
+ if (!sve_cont_ldst_elements(&info, addr, vg, reg_max, esz, esize)) {
390
+ /* The entire predicate was false; no load occurs. */
391
+ clr_fn(za, 0, reg_max);
392
+ return;
393
+ }
394
+
395
+ /* Probe the page(s). Exit with exception for any invalid page. */
396
+ sve_cont_ldst_pages(&info, FAULT_ALL, env, addr, MMU_DATA_LOAD, ra);
397
+
398
+ /* Handle watchpoints for all active elements. */
399
+ sve_cont_ldst_watchpoints(&info, env, vg, addr, esize, esize,
400
+ BP_MEM_READ, ra);
401
+
402
+ /*
403
+ * Handle mte checks for all active elements.
404
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
405
+ */
406
+ if (mtedesc) {
407
+ sve_cont_ldst_mte_check(&info, env, vg, addr, esize, esize,
408
+ mtedesc, ra);
409
+ }
410
+
411
+ flags = info.page[0].flags | info.page[1].flags;
412
+ if (unlikely(flags != 0)) {
413
+#ifdef CONFIG_USER_ONLY
414
+ g_assert_not_reached();
415
+#else
416
+ /*
417
+ * At least one page includes MMIO.
418
+ * Any bus operation can fail with cpu_transaction_failed,
419
+ * which for ARM will raise SyncExternal. Perform the load
420
+ * into scratch memory to preserve register state until the end.
421
+ */
422
+ ARMVectorReg scratch = { };
423
+
424
+ reg_off = info.reg_off_first[0];
425
+ reg_last = info.reg_off_last[1];
426
+ if (reg_last < 0) {
427
+ reg_last = info.reg_off_split;
428
+ if (reg_last < 0) {
429
+ reg_last = info.reg_off_last[0];
430
+ }
431
+ }
432
+
433
+ do {
434
+ uint64_t pg = vg[reg_off >> 6];
435
+ do {
436
+ if ((pg >> (reg_off & 63)) & 1) {
437
+ tlb_fn(env, &scratch, reg_off, addr + reg_off, ra);
438
+ }
439
+ reg_off += esize;
440
+ } while (reg_off & 63);
441
+ } while (reg_off <= reg_last);
442
+
443
+ cpy_fn(za, &scratch, reg_max);
444
+ return;
445
+#endif
446
+ }
447
+
448
+ /* The entire operation is in RAM, on valid pages. */
449
+
450
+ reg_off = info.reg_off_first[0];
451
+ reg_last = info.reg_off_last[0];
452
+ host = info.page[0].host;
453
+
454
+ if (!vertical) {
455
+ memset(za, 0, reg_max);
456
+ } else if (reg_off) {
457
+ clr_fn(za, 0, reg_off);
458
+ }
459
+
460
+ while (reg_off <= reg_last) {
461
+ uint64_t pg = vg[reg_off >> 6];
462
+ do {
463
+ if ((pg >> (reg_off & 63)) & 1) {
464
+ host_fn(za, reg_off, host + reg_off);
465
+ } else if (vertical) {
466
+ clr_fn(za, reg_off, esize);
467
+ }
468
+ reg_off += esize;
469
+ } while (reg_off <= reg_last && (reg_off & 63));
470
+ }
471
+
472
+ /*
473
+ * Use the slow path to manage the cross-page misalignment.
474
+ * But we know this is RAM and cannot trap.
475
+ */
476
+ reg_off = info.reg_off_split;
477
+ if (unlikely(reg_off >= 0)) {
478
+ tlb_fn(env, za, reg_off, addr + reg_off, ra);
479
+ }
480
+
481
+ reg_off = info.reg_off_first[1];
482
+ if (unlikely(reg_off >= 0)) {
483
+ reg_last = info.reg_off_last[1];
484
+ host = info.page[1].host;
485
+
486
+ do {
487
+ uint64_t pg = vg[reg_off >> 6];
488
+ do {
489
+ if ((pg >> (reg_off & 63)) & 1) {
490
+ host_fn(za, reg_off, host + reg_off);
491
+ } else if (vertical) {
492
+ clr_fn(za, reg_off, esize);
493
+ }
494
+ reg_off += esize;
495
+ } while (reg_off & 63);
496
+ } while (reg_off <= reg_last);
497
+ }
498
+}
499
+
500
+static inline QEMU_ALWAYS_INLINE
501
+void sme_ld1_mte(CPUARMState *env, void *za, uint64_t *vg,
502
+ target_ulong addr, uint32_t desc, uintptr_t ra,
503
+ const int esz, bool vertical,
504
+ sve_ldst1_host_fn *host_fn,
505
+ sve_ldst1_tlb_fn *tlb_fn,
506
+ ClearFn *clr_fn,
507
+ CopyFn *cpy_fn)
508
+{
509
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
510
+ int bit55 = extract64(addr, 55, 1);
511
+
512
+ /* Remove mtedesc from the normal sve descriptor. */
513
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
514
+
515
+ /* Perform gross MTE suppression early. */
516
+ if (!tbi_check(desc, bit55) ||
517
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
518
+ mtedesc = 0;
519
+ }
520
+
521
+ sme_ld1(env, za, vg, addr, desc, ra, esz, mtedesc, vertical,
522
+ host_fn, tlb_fn, clr_fn, cpy_fn);
523
+}
524
+
525
+#define DO_LD(L, END, ESZ) \
526
+void HELPER(sme_ld1##L##END##_h)(CPUARMState *env, void *za, void *vg, \
527
+ target_ulong addr, uint32_t desc) \
528
+{ \
529
+ sme_ld1(env, za, vg, addr, desc, GETPC(), ESZ, 0, false, \
530
+ sve_ld1##L##L##END##_host, sve_ld1##L##L##END##_tlb, \
531
+ clear_horizontal, copy_horizontal); \
532
+} \
533
+void HELPER(sme_ld1##L##END##_v)(CPUARMState *env, void *za, void *vg, \
534
+ target_ulong addr, uint32_t desc) \
535
+{ \
536
+ sme_ld1(env, za, vg, addr, desc, GETPC(), ESZ, 0, true, \
537
+ sme_ld1##L##END##_v_host, sme_ld1##L##END##_v_tlb, \
538
+ clear_vertical_##L, copy_vertical_##L); \
539
+} \
540
+void HELPER(sme_ld1##L##END##_h_mte)(CPUARMState *env, void *za, void *vg, \
541
+ target_ulong addr, uint32_t desc) \
542
+{ \
543
+ sme_ld1_mte(env, za, vg, addr, desc, GETPC(), ESZ, false, \
544
+ sve_ld1##L##L##END##_host, sve_ld1##L##L##END##_tlb, \
545
+ clear_horizontal, copy_horizontal); \
546
+} \
547
+void HELPER(sme_ld1##L##END##_v_mte)(CPUARMState *env, void *za, void *vg, \
548
+ target_ulong addr, uint32_t desc) \
549
+{ \
550
+ sme_ld1_mte(env, za, vg, addr, desc, GETPC(), ESZ, true, \
551
+ sme_ld1##L##END##_v_host, sme_ld1##L##END##_v_tlb, \
552
+ clear_vertical_##L, copy_vertical_##L); \
553
+}
554
+
555
+DO_LD(b, , MO_8)
556
+DO_LD(h, _be, MO_16)
557
+DO_LD(h, _le, MO_16)
558
+DO_LD(s, _be, MO_32)
559
+DO_LD(s, _le, MO_32)
560
+DO_LD(d, _be, MO_64)
561
+DO_LD(d, _le, MO_64)
562
+DO_LD(q, _be, MO_128)
563
+DO_LD(q, _le, MO_128)
564
+
565
+#undef DO_LD
566
+
567
+/*
568
+ * Common helper for all contiguous predicated stores.
569
+ */
570
+
571
+static inline QEMU_ALWAYS_INLINE
572
+void sme_st1(CPUARMState *env, void *za, uint64_t *vg,
573
+ const target_ulong addr, uint32_t desc, const uintptr_t ra,
574
+ const int esz, uint32_t mtedesc, bool vertical,
575
+ sve_ldst1_host_fn *host_fn,
576
+ sve_ldst1_tlb_fn *tlb_fn)
577
+{
578
+ const intptr_t reg_max = simd_oprsz(desc);
579
+ const intptr_t esize = 1 << esz;
580
+ intptr_t reg_off, reg_last;
581
+ SVEContLdSt info;
582
+ void *host;
583
+ int flags;
584
+
585
+ /* Find the active elements. */
586
+ if (!sve_cont_ldst_elements(&info, addr, vg, reg_max, esz, esize)) {
587
+ /* The entire predicate was false; no store occurs. */
588
+ return;
589
+ }
590
+
591
+ /* Probe the page(s). Exit with exception for any invalid page. */
592
+ sve_cont_ldst_pages(&info, FAULT_ALL, env, addr, MMU_DATA_STORE, ra);
593
+
594
+ /* Handle watchpoints for all active elements. */
595
+ sve_cont_ldst_watchpoints(&info, env, vg, addr, esize, esize,
596
+ BP_MEM_WRITE, ra);
597
+
598
+ /*
599
+ * Handle mte checks for all active elements.
600
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
601
+ */
602
+ if (mtedesc) {
603
+ sve_cont_ldst_mte_check(&info, env, vg, addr, esize, esize,
604
+ mtedesc, ra);
605
+ }
606
+
607
+ flags = info.page[0].flags | info.page[1].flags;
608
+ if (unlikely(flags != 0)) {
609
+#ifdef CONFIG_USER_ONLY
610
+ g_assert_not_reached();
611
+#else
612
+ /*
613
+ * At least one page includes MMIO.
614
+ * Any bus operation can fail with cpu_transaction_failed,
615
+ * which for ARM will raise SyncExternal. We cannot avoid
616
+ * this fault and will leave with the store incomplete.
617
+ */
618
+ reg_off = info.reg_off_first[0];
619
+ reg_last = info.reg_off_last[1];
620
+ if (reg_last < 0) {
621
+ reg_last = info.reg_off_split;
622
+ if (reg_last < 0) {
623
+ reg_last = info.reg_off_last[0];
624
+ }
625
+ }
626
+
627
+ do {
628
+ uint64_t pg = vg[reg_off >> 6];
629
+ do {
630
+ if ((pg >> (reg_off & 63)) & 1) {
631
+ tlb_fn(env, za, reg_off, addr + reg_off, ra);
632
+ }
633
+ reg_off += esize;
634
+ } while (reg_off & 63);
635
+ } while (reg_off <= reg_last);
636
+ return;
637
+#endif
638
+ }
639
+
640
+ reg_off = info.reg_off_first[0];
641
+ reg_last = info.reg_off_last[0];
642
+ host = info.page[0].host;
643
+
644
+ while (reg_off <= reg_last) {
645
+ uint64_t pg = vg[reg_off >> 6];
646
+ do {
647
+ if ((pg >> (reg_off & 63)) & 1) {
648
+ host_fn(za, reg_off, host + reg_off);
649
+ }
650
+ reg_off += 1 << esz;
651
+ } while (reg_off <= reg_last && (reg_off & 63));
652
+ }
653
+
654
+ /*
655
+ * Use the slow path to manage the cross-page misalignment.
656
+ * But we know this is RAM and cannot trap.
657
+ */
658
+ reg_off = info.reg_off_split;
659
+ if (unlikely(reg_off >= 0)) {
660
+ tlb_fn(env, za, reg_off, addr + reg_off, ra);
661
+ }
662
+
663
+ reg_off = info.reg_off_first[1];
664
+ if (unlikely(reg_off >= 0)) {
665
+ reg_last = info.reg_off_last[1];
666
+ host = info.page[1].host;
667
+
668
+ do {
669
+ uint64_t pg = vg[reg_off >> 6];
670
+ do {
671
+ if ((pg >> (reg_off & 63)) & 1) {
672
+ host_fn(za, reg_off, host + reg_off);
673
+ }
674
+ reg_off += 1 << esz;
675
+ } while (reg_off & 63);
676
+ } while (reg_off <= reg_last);
677
+ }
678
+}
679
+
680
+static inline QEMU_ALWAYS_INLINE
681
+void sme_st1_mte(CPUARMState *env, void *za, uint64_t *vg, target_ulong addr,
682
+ uint32_t desc, uintptr_t ra, int esz, bool vertical,
683
+ sve_ldst1_host_fn *host_fn,
684
+ sve_ldst1_tlb_fn *tlb_fn)
685
+{
686
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
687
+ int bit55 = extract64(addr, 55, 1);
688
+
689
+ /* Remove mtedesc from the normal sve descriptor. */
690
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
691
+
692
+ /* Perform gross MTE suppression early. */
693
+ if (!tbi_check(desc, bit55) ||
694
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
695
+ mtedesc = 0;
696
+ }
697
+
698
+ sme_st1(env, za, vg, addr, desc, ra, esz, mtedesc,
699
+ vertical, host_fn, tlb_fn);
700
+}
701
+
702
+#define DO_ST(L, END, ESZ) \
703
+void HELPER(sme_st1##L##END##_h)(CPUARMState *env, void *za, void *vg, \
704
+ target_ulong addr, uint32_t desc) \
705
+{ \
706
+ sme_st1(env, za, vg, addr, desc, GETPC(), ESZ, 0, false, \
707
+ sve_st1##L##L##END##_host, sve_st1##L##L##END##_tlb); \
708
+} \
709
+void HELPER(sme_st1##L##END##_v)(CPUARMState *env, void *za, void *vg, \
710
+ target_ulong addr, uint32_t desc) \
711
+{ \
712
+ sme_st1(env, za, vg, addr, desc, GETPC(), ESZ, 0, true, \
713
+ sme_st1##L##END##_v_host, sme_st1##L##END##_v_tlb); \
714
+} \
715
+void HELPER(sme_st1##L##END##_h_mte)(CPUARMState *env, void *za, void *vg, \
716
+ target_ulong addr, uint32_t desc) \
717
+{ \
718
+ sme_st1_mte(env, za, vg, addr, desc, GETPC(), ESZ, false, \
719
+ sve_st1##L##L##END##_host, sve_st1##L##L##END##_tlb); \
720
+} \
721
+void HELPER(sme_st1##L##END##_v_mte)(CPUARMState *env, void *za, void *vg, \
722
+ target_ulong addr, uint32_t desc) \
723
+{ \
724
+ sme_st1_mte(env, za, vg, addr, desc, GETPC(), ESZ, true, \
725
+ sme_st1##L##END##_v_host, sme_st1##L##END##_v_tlb); \
726
+}
727
+
728
+DO_ST(b, , MO_8)
729
+DO_ST(h, _be, MO_16)
730
+DO_ST(h, _le, MO_16)
731
+DO_ST(s, _be, MO_32)
732
+DO_ST(s, _le, MO_32)
733
+DO_ST(d, _be, MO_64)
734
+DO_ST(d, _le, MO_64)
735
+DO_ST(q, _be, MO_128)
736
+DO_ST(q, _le, MO_128)
737
+
738
+#undef DO_ST
739
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
740
index XXXXXXX..XXXXXXX 100644
741
--- a/target/arm/translate-sme.c
742
+++ b/target/arm/translate-sme.c
743
@@ -XXX,XX +XXX,XX @@ static bool trans_MOVA(DisasContext *s, arg_MOVA *a)
744
745
return true;
746
}
747
+
748
+static bool trans_LDST1(DisasContext *s, arg_LDST1 *a)
749
+{
750
+ typedef void GenLdSt1(TCGv_env, TCGv_ptr, TCGv_ptr, TCGv, TCGv_i32);
751
+
752
+ /*
753
+ * Indexed by [esz][be][v][mte][st], which is (except for load/store)
754
+ * also the order in which the elements appear in the function names,
755
+ * and so how we must concatenate the pieces.
756
+ */
757
+
758
+#define FN_LS(F) { gen_helper_sme_ld1##F, gen_helper_sme_st1##F }
759
+#define FN_MTE(F) { FN_LS(F), FN_LS(F##_mte) }
760
+#define FN_HV(F) { FN_MTE(F##_h), FN_MTE(F##_v) }
761
+#define FN_END(L, B) { FN_HV(L), FN_HV(B) }
762
+
763
+ static GenLdSt1 * const fns[5][2][2][2][2] = {
764
+ FN_END(b, b),
765
+ FN_END(h_le, h_be),
766
+ FN_END(s_le, s_be),
767
+ FN_END(d_le, d_be),
768
+ FN_END(q_le, q_be),
769
+ };
770
+
771
+#undef FN_LS
772
+#undef FN_MTE
773
+#undef FN_HV
774
+#undef FN_END
775
+
776
+ TCGv_ptr t_za, t_pg;
777
+ TCGv_i64 addr;
778
+ int svl, desc = 0;
779
+ bool be = s->be_data == MO_BE;
780
+ bool mte = s->mte_active[0];
781
+
782
+ if (!dc_isar_feature(aa64_sme, s)) {
783
+ return false;
784
+ }
785
+ if (!sme_smza_enabled_check(s)) {
786
+ return true;
787
+ }
788
+
789
+ t_za = get_tile_rowcol(s, a->esz, a->rs, a->za_imm, a->v);
790
+ t_pg = pred_full_reg_ptr(s, a->pg);
791
+ addr = tcg_temp_new_i64();
792
+
793
+ tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), a->esz);
794
+ tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
795
+
796
+ if (mte) {
797
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
798
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
799
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
800
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, a->st);
801
+ desc = FIELD_DP32(desc, MTEDESC, SIZEM1, (1 << a->esz) - 1);
802
+ desc <<= SVE_MTEDESC_SHIFT;
803
+ } else {
804
+ addr = clean_data_tbi(s, addr);
805
+ }
806
+ svl = streaming_vec_reg_size(s);
807
+ desc = simd_desc(svl, svl, desc);
808
+
809
+ fns[a->esz][be][a->v][mte][a->st](cpu_env, t_za, t_pg, addr,
810
+ tcg_constant_i32(desc));
811
+
812
+ tcg_temp_free_ptr(t_za);
813
+ tcg_temp_free_ptr(t_pg);
814
+ tcg_temp_free_i64(addr);
815
+ return true;
816
+}
559
--
817
--
560
2.20.1
818
2.25.1
561
562
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Resolve the untagged address once, using thread_cpu.
3
Add a TCGv_ptr base argument, which will be cpu_env for SVE.
4
Tidy the DEBUG_REMAP code using glib routines.
4
We will reuse this for SME save and restore array insns.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-20-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-22-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
linux-user/uaccess.c | 27 ++++++++++++++-------------
11
target/arm/translate-a64.h | 3 +++
12
1 file changed, 14 insertions(+), 13 deletions(-)
12
target/arm/translate-sve.c | 48 ++++++++++++++++++++++++++++----------
13
2 files changed, 39 insertions(+), 12 deletions(-)
13
14
14
diff --git a/linux-user/uaccess.c b/linux-user/uaccess.c
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/uaccess.c
17
--- a/target/arm/translate-a64.h
17
+++ b/linux-user/uaccess.c
18
+++ b/target/arm/translate-a64.h
18
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ void gen_gvec_xar(unsigned vece, uint32_t rd_ofs, uint32_t rn_ofs,
19
20
uint32_t rm_ofs, int64_t shift,
20
void *lock_user(int type, abi_ulong guest_addr, size_t len, bool copy)
21
uint32_t opr_sz, uint32_t max_sz);
22
23
+void gen_sve_ldr(DisasContext *s, TCGv_ptr, int vofs, int len, int rn, int imm);
24
+void gen_sve_str(DisasContext *s, TCGv_ptr, int vofs, int len, int rn, int imm);
25
+
26
#endif /* TARGET_ARM_TRANSLATE_A64_H */
27
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/translate-sve.c
30
+++ b/target/arm/translate-sve.c
31
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(UCVTF_dd, aa64_sve, gen_gvec_fpst_arg_zpz,
32
* The load should begin at the address Rn + IMM.
33
*/
34
35
-static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
36
+void gen_sve_ldr(DisasContext *s, TCGv_ptr base, int vofs,
37
+ int len, int rn, int imm)
21
{
38
{
22
+ void *host_addr;
39
int len_align = QEMU_ALIGN_DOWN(len, 8);
40
int len_remain = len % 8;
41
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
42
t0 = tcg_temp_new_i64();
43
for (i = 0; i < len_align; i += 8) {
44
tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUQ);
45
- tcg_gen_st_i64(t0, cpu_env, vofs + i);
46
+ tcg_gen_st_i64(t0, base, vofs + i);
47
tcg_gen_addi_i64(clean_addr, clean_addr, 8);
48
}
49
tcg_temp_free_i64(t0);
50
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
51
clean_addr = new_tmp_a64_local(s);
52
tcg_gen_mov_i64(clean_addr, t0);
53
54
+ if (base != cpu_env) {
55
+ TCGv_ptr b = tcg_temp_local_new_ptr();
56
+ tcg_gen_mov_ptr(b, base);
57
+ base = b;
58
+ }
23
+
59
+
24
+ guest_addr = cpu_untagged_addr(thread_cpu, guest_addr);
60
gen_set_label(loop);
25
if (!access_ok_untagged(type, guest_addr, len)) {
61
26
return NULL;
62
t0 = tcg_temp_new_i64();
63
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
64
tcg_gen_addi_i64(clean_addr, clean_addr, 8);
65
66
tp = tcg_temp_new_ptr();
67
- tcg_gen_add_ptr(tp, cpu_env, i);
68
+ tcg_gen_add_ptr(tp, base, i);
69
tcg_gen_addi_ptr(i, i, 8);
70
tcg_gen_st_i64(t0, tp, vofs);
71
tcg_temp_free_ptr(tp);
72
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
73
74
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
75
tcg_temp_free_ptr(i);
76
+
77
+ if (base != cpu_env) {
78
+ tcg_temp_free_ptr(base);
79
+ assert(len_remain == 0);
80
+ }
27
}
81
}
28
+ host_addr = g2h_untagged(guest_addr);
82
29
#ifdef DEBUG_REMAP
83
/*
30
- {
84
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
31
- void *addr;
85
default:
32
- addr = g_malloc(len);
86
g_assert_not_reached();
33
- if (copy) {
87
}
34
- memcpy(addr, g2h(guest_addr), len);
88
- tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
35
- } else {
89
+ tcg_gen_st_i64(t0, base, vofs + len_align);
36
- memset(addr, 0, len);
90
tcg_temp_free_i64(t0);
37
- }
38
- return addr;
39
+ if (copy) {
40
+ host_addr = g_memdup(host_addr, len);
41
+ } else {
42
+ host_addr = g_malloc0(len);
43
}
91
}
44
-#else
45
- return g2h_untagged(guest_addr);
46
#endif
47
+ return host_addr;
48
}
92
}
49
93
50
#ifdef DEBUG_REMAP
94
/* Similarly for stores. */
51
void unlock_user(void *host_ptr, abi_ulong guest_addr, size_t len);
95
-static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
96
+void gen_sve_str(DisasContext *s, TCGv_ptr base, int vofs,
97
+ int len, int rn, int imm)
52
{
98
{
53
+ void *host_ptr_conv;
99
int len_align = QEMU_ALIGN_DOWN(len, 8);
100
int len_remain = len % 8;
101
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
102
103
t0 = tcg_temp_new_i64();
104
for (i = 0; i < len_align; i += 8) {
105
- tcg_gen_ld_i64(t0, cpu_env, vofs + i);
106
+ tcg_gen_ld_i64(t0, base, vofs + i);
107
tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUQ);
108
tcg_gen_addi_i64(clean_addr, clean_addr, 8);
109
}
110
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
111
clean_addr = new_tmp_a64_local(s);
112
tcg_gen_mov_i64(clean_addr, t0);
113
114
+ if (base != cpu_env) {
115
+ TCGv_ptr b = tcg_temp_local_new_ptr();
116
+ tcg_gen_mov_ptr(b, base);
117
+ base = b;
118
+ }
54
+
119
+
55
if (!host_ptr) {
120
gen_set_label(loop);
56
return;
121
122
t0 = tcg_temp_new_i64();
123
tp = tcg_temp_new_ptr();
124
- tcg_gen_add_ptr(tp, cpu_env, i);
125
+ tcg_gen_add_ptr(tp, base, i);
126
tcg_gen_ld_i64(t0, tp, vofs);
127
tcg_gen_addi_ptr(i, i, 8);
128
tcg_temp_free_ptr(tp);
129
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
130
131
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
132
tcg_temp_free_ptr(i);
133
+
134
+ if (base != cpu_env) {
135
+ tcg_temp_free_ptr(base);
136
+ assert(len_remain == 0);
137
+ }
57
}
138
}
58
- if (host_ptr == g2h_untagged(guest_addr)) {
139
59
+ host_ptr_conv = g2h(thread_cpu, guest_addr);
140
/* Predicate register stores can be any multiple of 2. */
60
+ if (host_ptr == host_ptr_conv) {
141
if (len_remain) {
61
return;
142
t0 = tcg_temp_new_i64();
143
- tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
144
+ tcg_gen_ld_i64(t0, base, vofs + len_align);
145
146
switch (len_remain) {
147
case 2:
148
@@ -XXX,XX +XXX,XX @@ static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
149
if (sve_access_check(s)) {
150
int size = vec_full_reg_size(s);
151
int off = vec_full_reg_offset(s, a->rd);
152
- do_ldr(s, off, size, a->rn, a->imm * size);
153
+ gen_sve_ldr(s, cpu_env, off, size, a->rn, a->imm * size);
62
}
154
}
63
if (len != 0) {
155
return true;
64
- memcpy(g2h_untagged(guest_addr), host_ptr, len);
156
}
65
+ memcpy(host_ptr_conv, host_ptr, len);
157
@@ -XXX,XX +XXX,XX @@ static bool trans_LDR_pri(DisasContext *s, arg_rri *a)
158
if (sve_access_check(s)) {
159
int size = pred_full_reg_size(s);
160
int off = pred_full_reg_offset(s, a->rd);
161
- do_ldr(s, off, size, a->rn, a->imm * size);
162
+ gen_sve_ldr(s, cpu_env, off, size, a->rn, a->imm * size);
66
}
163
}
67
g_free(host_ptr);
164
return true;
165
}
166
@@ -XXX,XX +XXX,XX @@ static bool trans_STR_zri(DisasContext *s, arg_rri *a)
167
if (sve_access_check(s)) {
168
int size = vec_full_reg_size(s);
169
int off = vec_full_reg_offset(s, a->rd);
170
- do_str(s, off, size, a->rn, a->imm * size);
171
+ gen_sve_str(s, cpu_env, off, size, a->rn, a->imm * size);
172
}
173
return true;
174
}
175
@@ -XXX,XX +XXX,XX @@ static bool trans_STR_pri(DisasContext *s, arg_rri *a)
176
if (sve_access_check(s)) {
177
int size = pred_full_reg_size(s);
178
int off = pred_full_reg_offset(s, a->rd);
179
- do_str(s, off, size, a->rn, a->imm * size);
180
+ gen_sve_str(s, cpu_env, off, size, a->rn, a->imm * size);
181
}
182
return true;
68
}
183
}
69
--
184
--
70
2.20.1
185
2.25.1
71
72
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
These functions are not small, except for unlock_user
3
We can reuse the SVE functions for LDR and STR, passing in the
4
without debugging enabled. Move them out of line, and
4
base of the ZA vector and a zero offset.
5
add missing braces on the way.
6
5
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20220708151540.18136-23-richard.henderson@linaro.org
10
Message-id: 20210212184902.1251044-18-richard.henderson@linaro.org
11
[PMM: fixed the sense of an ifdef test in qemu.h]
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
10
---
14
linux-user/qemu.h | 47 +++++++-------------------------------------
11
target/arm/sme.decode | 7 +++++++
15
linux-user/uaccess.c | 46 +++++++++++++++++++++++++++++++++++++++++++
12
target/arm/translate-sme.c | 24 ++++++++++++++++++++++++
16
2 files changed, 53 insertions(+), 40 deletions(-)
13
2 files changed, 31 insertions(+)
17
14
18
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
15
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
19
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
20
--- a/linux-user/qemu.h
17
--- a/target/arm/sme.decode
21
+++ b/linux-user/qemu.h
18
+++ b/target/arm/sme.decode
22
@@ -XXX,XX +XXX,XX @@ abi_long copy_to_user(abi_ulong gaddr, void *hptr, size_t len);
19
@@ -XXX,XX +XXX,XX @@ LDST1 1110000 0 esz:2 st:1 rm:5 v:1 .. pg:3 rn:5 0 za_imm:4 \
23
20
&ldst rs=%mova_rs
24
/* Lock an area of guest memory into the host. If copy is true then the
21
LDST1 1110000 111 st:1 rm:5 v:1 .. pg:3 rn:5 0 za_imm:4 \
25
host area will have the same contents as the guest. */
22
&ldst esz=4 rs=%mova_rs
26
-static inline void *lock_user(int type, abi_ulong guest_addr, long len, int copy)
23
+
27
-{
24
+&ldstr rv rn imm
28
- if (!access_ok_untagged(type, guest_addr, len)) {
25
+@ldstr ....... ... . ...... .. ... rn:5 . imm:4 \
29
- return NULL;
26
+ &ldstr rv=%mova_rs
30
- }
27
+
31
-#ifdef DEBUG_REMAP
28
+LDR 1110000 100 0 000000 .. 000 ..... 0 .... @ldstr
32
- {
29
+STR 1110000 100 1 000000 .. 000 ..... 0 .... @ldstr
33
- void *addr;
30
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
34
- addr = g_malloc(len);
35
- if (copy)
36
- memcpy(addr, g2h(guest_addr), len);
37
- else
38
- memset(addr, 0, len);
39
- return addr;
40
- }
41
-#else
42
- return g2h_untagged(guest_addr);
43
-#endif
44
-}
45
+void *lock_user(int type, abi_ulong guest_addr, long len, int copy);
46
47
/* Unlock an area of guest memory. The first LEN bytes must be
48
flushed back to guest memory. host_ptr = NULL is explicitly
49
allowed and does nothing. */
50
-static inline void unlock_user(void *host_ptr, abi_ulong guest_addr,
51
- long len)
52
-{
53
-
54
-#ifdef DEBUG_REMAP
55
- if (!host_ptr)
56
- return;
57
- if (host_ptr == g2h_untagged(guest_addr))
58
- return;
59
- if (len > 0)
60
- memcpy(g2h_untagged(guest_addr), host_ptr, len);
61
- g_free(host_ptr);
62
+#ifndef DEBUG_REMAP
63
+static inline void unlock_user(void *host_ptr, abi_ulong guest_addr, long len)
64
+{ }
65
+#else
66
+void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
67
#endif
68
-}
69
70
/* Return the length of a string in target memory or -TARGET_EFAULT if
71
access error. */
72
abi_long target_strlen(abi_ulong gaddr);
73
74
/* Like lock_user but for null terminated strings. */
75
-static inline void *lock_user_string(abi_ulong guest_addr)
76
-{
77
- abi_long len;
78
- len = target_strlen(guest_addr);
79
- if (len < 0)
80
- return NULL;
81
- return lock_user(VERIFY_READ, guest_addr, (long)(len + 1), 1);
82
-}
83
+void *lock_user_string(abi_ulong guest_addr);
84
85
/* Helper macros for locking/unlocking a target struct. */
86
#define lock_user_struct(type, host_ptr, guest_addr, copy)    \
87
diff --git a/linux-user/uaccess.c b/linux-user/uaccess.c
88
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
89
--- a/linux-user/uaccess.c
32
--- a/target/arm/translate-sme.c
90
+++ b/linux-user/uaccess.c
33
+++ b/target/arm/translate-sme.c
91
@@ -XXX,XX +XXX,XX @@
34
@@ -XXX,XX +XXX,XX @@ static bool trans_LDST1(DisasContext *s, arg_LDST1 *a)
92
35
tcg_temp_free_i64(addr);
93
#include "qemu.h"
36
return true;
94
37
}
95
+void *lock_user(int type, abi_ulong guest_addr, long len, int copy)
38
+
39
+typedef void GenLdStR(DisasContext *, TCGv_ptr, int, int, int, int);
40
+
41
+static bool do_ldst_r(DisasContext *s, arg_ldstr *a, GenLdStR *fn)
96
+{
42
+{
97
+ if (!access_ok_untagged(type, guest_addr, len)) {
43
+ int svl = streaming_vec_reg_size(s);
98
+ return NULL;
44
+ int imm = a->imm;
45
+ TCGv_ptr base;
46
+
47
+ if (!sme_za_enabled_check(s)) {
48
+ return true;
99
+ }
49
+ }
100
+#ifdef DEBUG_REMAP
50
+
101
+ {
51
+ /* ZA[n] equates to ZA0H.B[n]. */
102
+ void *addr;
52
+ base = get_tile_rowcol(s, MO_8, a->rv, imm, false);
103
+ addr = g_malloc(len);
53
+
104
+ if (copy) {
54
+ fn(s, base, 0, svl, a->rn, imm * svl);
105
+ memcpy(addr, g2h(guest_addr), len);
55
+
106
+ } else {
56
+ tcg_temp_free_ptr(base);
107
+ memset(addr, 0, len);
57
+ return true;
108
+ }
109
+ return addr;
110
+ }
111
+#else
112
+ return g2h_untagged(guest_addr);
113
+#endif
114
+}
58
+}
115
+
59
+
116
+#ifdef DEBUG_REMAP
60
+TRANS_FEAT(LDR, aa64_sme, do_ldst_r, a, gen_sve_ldr)
117
+void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
61
+TRANS_FEAT(STR, aa64_sme, do_ldst_r, a, gen_sve_str)
118
+{
119
+ if (!host_ptr) {
120
+ return;
121
+ }
122
+ if (host_ptr == g2h_untagged(guest_addr)) {
123
+ return;
124
+ }
125
+ if (len > 0) {
126
+ memcpy(g2h_untagged(guest_addr), host_ptr, len);
127
+ }
128
+ g_free(host_ptr);
129
+}
130
+#endif
131
+
132
+void *lock_user_string(abi_ulong guest_addr)
133
+{
134
+ abi_long len = target_strlen(guest_addr);
135
+ if (len < 0) {
136
+ return NULL;
137
+ }
138
+ return lock_user(VERIFY_READ, guest_addr, (long)(len + 1), 1);
139
+}
140
+
141
/* copy_from_user() and copy_to_user() are usually used to copy data
142
* buffers between the target and host. These internally perform
143
* locking/unlocking of the memory.
144
--
62
--
145
2.20.1
63
2.25.1
146
147
diff view generated by jsdifflib
1
From: Hao Wu <wuhaotsh@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This commit implements the single-byte mode of the SMBus.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Each Nuvoton SoC has 16 System Management Bus (SMBus). These buses
5
Message-id: 20220708151540.18136-24-richard.henderson@linaro.org
6
compliant with SMBus and I2C protocol.
7
8
This patch implements the single-byte mode of the SMBus. In this mode,
9
the user sends or receives a byte each time. The SMBus device transmits
10
it to the underlying i2c device and sends an interrupt back to the QEMU
11
guest.
12
13
Reviewed-by: Doug Evans<dje@google.com>
14
Reviewed-by: Tyrong Ting<kfting@nuvoton.com>
15
Signed-off-by: Hao Wu <wuhaotsh@google.com>
16
Reviewed-by: Corey Minyard <cminyard@mvista.com>
17
Message-id: 20210210220426.3577804-2-wuhaotsh@google.com
18
Acked-by: Corey Minyard <cminyard@mvista.com>
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
---
7
---
21
docs/system/arm/nuvoton.rst | 2 +-
8
target/arm/helper-sme.h | 5 +++
22
include/hw/arm/npcm7xx.h | 2 +
9
target/arm/sme.decode | 11 +++++
23
include/hw/i2c/npcm7xx_smbus.h | 88 ++++
10
target/arm/sme_helper.c | 90 ++++++++++++++++++++++++++++++++++++++
24
hw/arm/npcm7xx.c | 68 ++-
11
target/arm/translate-sme.c | 31 +++++++++++++
25
hw/i2c/npcm7xx_smbus.c | 783 +++++++++++++++++++++++++++++++++
12
4 files changed, 137 insertions(+)
26
hw/i2c/meson.build | 1 +
27
hw/i2c/trace-events | 11 +
28
7 files changed, 938 insertions(+), 17 deletions(-)
29
create mode 100644 include/hw/i2c/npcm7xx_smbus.h
30
create mode 100644 hw/i2c/npcm7xx_smbus.c
31
13
32
diff --git a/docs/system/arm/nuvoton.rst b/docs/system/arm/nuvoton.rst
14
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
33
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
34
--- a/docs/system/arm/nuvoton.rst
16
--- a/target/arm/helper-sme.h
35
+++ b/docs/system/arm/nuvoton.rst
17
+++ b/target/arm/helper-sme.h
36
@@ -XXX,XX +XXX,XX @@ Supported devices
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sme_st1q_be_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i
37
* GPIO controller
19
DEF_HELPER_FLAGS_5(sme_st1q_le_h_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
38
* Analog to Digital Converter (ADC)
20
DEF_HELPER_FLAGS_5(sme_st1q_be_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
39
* Pulse Width Modulation (PWM)
21
DEF_HELPER_FLAGS_5(sme_st1q_le_v_mte, TCG_CALL_NO_WG, void, env, ptr, ptr, tl, i32)
40
+ * SMBus controller (SMBF)
22
+
41
23
+DEF_HELPER_FLAGS_5(sme_addha_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
42
Missing devices
24
+DEF_HELPER_FLAGS_5(sme_addva_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
43
---------------
25
+DEF_HELPER_FLAGS_5(sme_addha_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
44
@@ -XXX,XX +XXX,XX @@ Missing devices
26
+DEF_HELPER_FLAGS_5(sme_addva_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
45
27
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
46
* Ethernet controllers (GMAC and EMC)
47
* USB device (USBD)
48
- * SMBus controller (SMBF)
49
* Peripheral SPI controller (PSPI)
50
* SD/MMC host
51
* PECI interface
52
diff --git a/include/hw/arm/npcm7xx.h b/include/hw/arm/npcm7xx.h
53
index XXXXXXX..XXXXXXX 100644
28
index XXXXXXX..XXXXXXX 100644
54
--- a/include/hw/arm/npcm7xx.h
29
--- a/target/arm/sme.decode
55
+++ b/include/hw/arm/npcm7xx.h
30
+++ b/target/arm/sme.decode
56
@@ -XXX,XX +XXX,XX @@
31
@@ -XXX,XX +XXX,XX @@ LDST1 1110000 111 st:1 rm:5 v:1 .. pg:3 rn:5 0 za_imm:4 \
57
#include "hw/adc/npcm7xx_adc.h"
32
58
#include "hw/cpu/a9mpcore.h"
33
LDR 1110000 100 0 000000 .. 000 ..... 0 .... @ldstr
59
#include "hw/gpio/npcm7xx_gpio.h"
34
STR 1110000 100 1 000000 .. 000 ..... 0 .... @ldstr
60
+#include "hw/i2c/npcm7xx_smbus.h"
61
#include "hw/mem/npcm7xx_mc.h"
62
#include "hw/misc/npcm7xx_clk.h"
63
#include "hw/misc/npcm7xx_gcr.h"
64
@@ -XXX,XX +XXX,XX @@ typedef struct NPCM7xxState {
65
NPCM7xxMCState mc;
66
NPCM7xxRNGState rng;
67
NPCM7xxGPIOState gpio[8];
68
+ NPCM7xxSMBusState smbus[16];
69
EHCISysBusState ehci;
70
OHCISysBusState ohci;
71
NPCM7xxFIUState fiu[2];
72
diff --git a/include/hw/i2c/npcm7xx_smbus.h b/include/hw/i2c/npcm7xx_smbus.h
73
new file mode 100644
74
index XXXXXXX..XXXXXXX
75
--- /dev/null
76
+++ b/include/hw/i2c/npcm7xx_smbus.h
77
@@ -XXX,XX +XXX,XX @@
78
+/*
79
+ * Nuvoton NPCM7xx SMBus Module.
80
+ *
81
+ * Copyright 2020 Google LLC
82
+ *
83
+ * This program is free software; you can redistribute it and/or modify it
84
+ * under the terms of the GNU General Public License as published by the
85
+ * Free Software Foundation; either version 2 of the License, or
86
+ * (at your option) any later version.
87
+ *
88
+ * This program is distributed in the hope that it will be useful, but WITHOUT
89
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
90
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
91
+ * for more details.
92
+ */
93
+#ifndef NPCM7XX_SMBUS_H
94
+#define NPCM7XX_SMBUS_H
95
+
35
+
96
+#include "exec/memory.h"
36
+### SME Add Vector to Array
97
+#include "hw/i2c/i2c.h"
98
+#include "hw/irq.h"
99
+#include "hw/sysbus.h"
100
+
37
+
101
+/*
38
+&adda zad zn pm pn
102
+ * Number of addresses this module contains. Do not change this without
39
+@adda_32 ........ .. ..... . pm:3 pn:3 zn:5 ... zad:2 &adda
103
+ * incrementing the version_id in the vmstate.
40
+@adda_64 ........ .. ..... . pm:3 pn:3 zn:5 .. zad:3 &adda
104
+ */
105
+#define NPCM7XX_SMBUS_NR_ADDRS 10
106
+
41
+
107
+typedef enum NPCM7xxSMBusStatus {
42
+ADDHA_s 11000000 10 01000 0 ... ... ..... 000 .. @adda_32
108
+ NPCM7XX_SMBUS_STATUS_IDLE,
43
+ADDVA_s 11000000 10 01000 1 ... ... ..... 000 .. @adda_32
109
+ NPCM7XX_SMBUS_STATUS_SENDING,
44
+ADDHA_d 11000000 11 01000 0 ... ... ..... 00 ... @adda_64
110
+ NPCM7XX_SMBUS_STATUS_RECEIVING,
45
+ADDVA_d 11000000 11 01000 1 ... ... ..... 00 ... @adda_64
111
+ NPCM7XX_SMBUS_STATUS_NEGACK,
46
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
112
+ NPCM7XX_SMBUS_STATUS_STOPPING_LAST_RECEIVE,
47
index XXXXXXX..XXXXXXX 100644
113
+ NPCM7XX_SMBUS_STATUS_STOPPING_NEGACK,
48
--- a/target/arm/sme_helper.c
114
+} NPCM7xxSMBusStatus;
49
+++ b/target/arm/sme_helper.c
50
@@ -XXX,XX +XXX,XX @@ DO_ST(q, _be, MO_128)
51
DO_ST(q, _le, MO_128)
52
53
#undef DO_ST
115
+
54
+
116
+/*
55
+void HELPER(sme_addha_s)(void *vzda, void *vzn, void *vpn,
117
+ * struct NPCM7xxSMBusState - System Management Bus device state.
56
+ void *vpm, uint32_t desc)
118
+ * @bus: The underlying I2C Bus.
57
+{
119
+ * @irq: GIC interrupt line to fire on events (if enabled).
58
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 4;
120
+ * @sda: The serial data register.
59
+ uint64_t *pn = vpn, *pm = vpm;
121
+ * @st: The status register.
60
+ uint32_t *zda = vzda, *zn = vzn;
122
+ * @cst: The control status register.
123
+ * @cst2: The control status register 2.
124
+ * @cst3: The control status register 3.
125
+ * @ctl1: The control register 1.
126
+ * @ctl2: The control register 2.
127
+ * @ctl3: The control register 3.
128
+ * @ctl4: The control register 4.
129
+ * @ctl5: The control register 5.
130
+ * @addr: The SMBus module's own addresses on the I2C bus.
131
+ * @scllt: The SCL low time register.
132
+ * @sclht: The SCL high time register.
133
+ * @status: The current status of the SMBus.
134
+ */
135
+typedef struct NPCM7xxSMBusState {
136
+ SysBusDevice parent;
137
+
61
+
138
+ MemoryRegion iomem;
62
+ for (row = 0; row < oprsz; ) {
139
+
63
+ uint64_t pa = pn[row >> 4];
140
+ I2CBus *bus;
64
+ do {
141
+ qemu_irq irq;
65
+ if (pa & 1) {
142
+
66
+ for (col = 0; col < oprsz; ) {
143
+ uint8_t sda;
67
+ uint64_t pb = pm[col >> 4];
144
+ uint8_t st;
68
+ do {
145
+ uint8_t cst;
69
+ if (pb & 1) {
146
+ uint8_t cst2;
70
+ zda[tile_vslice_index(row) + H4(col)] += zn[H4(col)];
147
+ uint8_t cst3;
71
+ }
148
+ uint8_t ctl1;
72
+ pb >>= 4;
149
+ uint8_t ctl2;
73
+ } while (++col & 15);
150
+ uint8_t ctl3;
74
+ }
151
+ uint8_t ctl4;
75
+ }
152
+ uint8_t ctl5;
76
+ pa >>= 4;
153
+ uint8_t addr[NPCM7XX_SMBUS_NR_ADDRS];
77
+ } while (++row & 15);
154
+
155
+ uint8_t scllt;
156
+ uint8_t sclht;
157
+
158
+ NPCM7xxSMBusStatus status;
159
+} NPCM7xxSMBusState;
160
+
161
+#define TYPE_NPCM7XX_SMBUS "npcm7xx-smbus"
162
+#define NPCM7XX_SMBUS(obj) OBJECT_CHECK(NPCM7xxSMBusState, (obj), \
163
+ TYPE_NPCM7XX_SMBUS)
164
+
165
+#endif /* NPCM7XX_SMBUS_H */
166
diff --git a/hw/arm/npcm7xx.c b/hw/arm/npcm7xx.c
167
index XXXXXXX..XXXXXXX 100644
168
--- a/hw/arm/npcm7xx.c
169
+++ b/hw/arm/npcm7xx.c
170
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxInterrupt {
171
NPCM7XX_WDG2_IRQ, /* Timer Module 2 Watchdog */
172
NPCM7XX_EHCI_IRQ = 61,
173
NPCM7XX_OHCI_IRQ = 62,
174
+ NPCM7XX_SMBUS0_IRQ = 64,
175
+ NPCM7XX_SMBUS1_IRQ,
176
+ NPCM7XX_SMBUS2_IRQ,
177
+ NPCM7XX_SMBUS3_IRQ,
178
+ NPCM7XX_SMBUS4_IRQ,
179
+ NPCM7XX_SMBUS5_IRQ,
180
+ NPCM7XX_SMBUS6_IRQ,
181
+ NPCM7XX_SMBUS7_IRQ,
182
+ NPCM7XX_SMBUS8_IRQ,
183
+ NPCM7XX_SMBUS9_IRQ,
184
+ NPCM7XX_SMBUS10_IRQ,
185
+ NPCM7XX_SMBUS11_IRQ,
186
+ NPCM7XX_SMBUS12_IRQ,
187
+ NPCM7XX_SMBUS13_IRQ,
188
+ NPCM7XX_SMBUS14_IRQ,
189
+ NPCM7XX_SMBUS15_IRQ,
190
NPCM7XX_PWM0_IRQ = 93, /* PWM module 0 */
191
NPCM7XX_PWM1_IRQ, /* PWM module 1 */
192
NPCM7XX_GPIO0_IRQ = 116,
193
@@ -XXX,XX +XXX,XX @@ static const hwaddr npcm7xx_pwm_addr[] = {
194
0xf0104000,
195
};
196
197
+/* Direct memory-mapped access to each SMBus Module. */
198
+static const hwaddr npcm7xx_smbus_addr[] = {
199
+ 0xf0080000,
200
+ 0xf0081000,
201
+ 0xf0082000,
202
+ 0xf0083000,
203
+ 0xf0084000,
204
+ 0xf0085000,
205
+ 0xf0086000,
206
+ 0xf0087000,
207
+ 0xf0088000,
208
+ 0xf0089000,
209
+ 0xf008a000,
210
+ 0xf008b000,
211
+ 0xf008c000,
212
+ 0xf008d000,
213
+ 0xf008e000,
214
+ 0xf008f000,
215
+};
216
+
217
static const struct {
218
hwaddr regs_addr;
219
uint32_t unconnected_pins;
220
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_init(Object *obj)
221
object_initialize_child(obj, "gpio[*]", &s->gpio[i], TYPE_NPCM7XX_GPIO);
222
}
223
224
+ for (i = 0; i < ARRAY_SIZE(s->smbus); i++) {
225
+ object_initialize_child(obj, "smbus[*]", &s->smbus[i],
226
+ TYPE_NPCM7XX_SMBUS);
227
+ }
228
+
229
object_initialize_child(obj, "ehci", &s->ehci, TYPE_NPCM7XX_EHCI);
230
object_initialize_child(obj, "ohci", &s->ohci, TYPE_SYSBUS_OHCI);
231
232
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_realize(DeviceState *dev, Error **errp)
233
npcm7xx_irq(s, NPCM7XX_GPIO0_IRQ + i));
234
}
235
236
+ /* SMBus modules. Cannot fail. */
237
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(npcm7xx_smbus_addr) != ARRAY_SIZE(s->smbus));
238
+ for (i = 0; i < ARRAY_SIZE(s->smbus); i++) {
239
+ Object *obj = OBJECT(&s->smbus[i]);
240
+
241
+ sysbus_realize(SYS_BUS_DEVICE(obj), &error_abort);
242
+ sysbus_mmio_map(SYS_BUS_DEVICE(obj), 0, npcm7xx_smbus_addr[i]);
243
+ sysbus_connect_irq(SYS_BUS_DEVICE(obj), 0,
244
+ npcm7xx_irq(s, NPCM7XX_SMBUS0_IRQ + i));
245
+ }
246
+
247
/* USB Host */
248
object_property_set_bool(OBJECT(&s->ehci), "companion-enable", true,
249
&error_abort);
250
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_realize(DeviceState *dev, Error **errp)
251
create_unimplemented_device("npcm7xx.pcierc", 0xe1000000, 64 * KiB);
252
create_unimplemented_device("npcm7xx.kcs", 0xf0007000, 4 * KiB);
253
create_unimplemented_device("npcm7xx.gfxi", 0xf000e000, 4 * KiB);
254
- create_unimplemented_device("npcm7xx.smbus[0]", 0xf0080000, 4 * KiB);
255
- create_unimplemented_device("npcm7xx.smbus[1]", 0xf0081000, 4 * KiB);
256
- create_unimplemented_device("npcm7xx.smbus[2]", 0xf0082000, 4 * KiB);
257
- create_unimplemented_device("npcm7xx.smbus[3]", 0xf0083000, 4 * KiB);
258
- create_unimplemented_device("npcm7xx.smbus[4]", 0xf0084000, 4 * KiB);
259
- create_unimplemented_device("npcm7xx.smbus[5]", 0xf0085000, 4 * KiB);
260
- create_unimplemented_device("npcm7xx.smbus[6]", 0xf0086000, 4 * KiB);
261
- create_unimplemented_device("npcm7xx.smbus[7]", 0xf0087000, 4 * KiB);
262
- create_unimplemented_device("npcm7xx.smbus[8]", 0xf0088000, 4 * KiB);
263
- create_unimplemented_device("npcm7xx.smbus[9]", 0xf0089000, 4 * KiB);
264
- create_unimplemented_device("npcm7xx.smbus[10]", 0xf008a000, 4 * KiB);
265
- create_unimplemented_device("npcm7xx.smbus[11]", 0xf008b000, 4 * KiB);
266
- create_unimplemented_device("npcm7xx.smbus[12]", 0xf008c000, 4 * KiB);
267
- create_unimplemented_device("npcm7xx.smbus[13]", 0xf008d000, 4 * KiB);
268
- create_unimplemented_device("npcm7xx.smbus[14]", 0xf008e000, 4 * KiB);
269
- create_unimplemented_device("npcm7xx.smbus[15]", 0xf008f000, 4 * KiB);
270
create_unimplemented_device("npcm7xx.espi", 0xf009f000, 4 * KiB);
271
create_unimplemented_device("npcm7xx.peci", 0xf0100000, 4 * KiB);
272
create_unimplemented_device("npcm7xx.siox[1]", 0xf0101000, 4 * KiB);
273
diff --git a/hw/i2c/npcm7xx_smbus.c b/hw/i2c/npcm7xx_smbus.c
274
new file mode 100644
275
index XXXXXXX..XXXXXXX
276
--- /dev/null
277
+++ b/hw/i2c/npcm7xx_smbus.c
278
@@ -XXX,XX +XXX,XX @@
279
+/*
280
+ * Nuvoton NPCM7xx SMBus Module.
281
+ *
282
+ * Copyright 2020 Google LLC
283
+ *
284
+ * This program is free software; you can redistribute it and/or modify it
285
+ * under the terms of the GNU General Public License as published by the
286
+ * Free Software Foundation; either version 2 of the License, or
287
+ * (at your option) any later version.
288
+ *
289
+ * This program is distributed in the hope that it will be useful, but WITHOUT
290
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
291
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
292
+ * for more details.
293
+ */
294
+
295
+#include "qemu/osdep.h"
296
+
297
+#include "hw/i2c/npcm7xx_smbus.h"
298
+#include "migration/vmstate.h"
299
+#include "qemu/bitops.h"
300
+#include "qemu/guest-random.h"
301
+#include "qemu/log.h"
302
+#include "qemu/module.h"
303
+#include "qemu/units.h"
304
+
305
+#include "trace.h"
306
+
307
+enum NPCM7xxSMBusCommonRegister {
308
+ NPCM7XX_SMB_SDA = 0x0,
309
+ NPCM7XX_SMB_ST = 0x2,
310
+ NPCM7XX_SMB_CST = 0x4,
311
+ NPCM7XX_SMB_CTL1 = 0x6,
312
+ NPCM7XX_SMB_ADDR1 = 0x8,
313
+ NPCM7XX_SMB_CTL2 = 0xa,
314
+ NPCM7XX_SMB_ADDR2 = 0xc,
315
+ NPCM7XX_SMB_CTL3 = 0xe,
316
+ NPCM7XX_SMB_CST2 = 0x18,
317
+ NPCM7XX_SMB_CST3 = 0x19,
318
+ NPCM7XX_SMB_VER = 0x1f,
319
+};
320
+
321
+enum NPCM7xxSMBusBank0Register {
322
+ NPCM7XX_SMB_ADDR3 = 0x10,
323
+ NPCM7XX_SMB_ADDR7 = 0x11,
324
+ NPCM7XX_SMB_ADDR4 = 0x12,
325
+ NPCM7XX_SMB_ADDR8 = 0x13,
326
+ NPCM7XX_SMB_ADDR5 = 0x14,
327
+ NPCM7XX_SMB_ADDR9 = 0x15,
328
+ NPCM7XX_SMB_ADDR6 = 0x16,
329
+ NPCM7XX_SMB_ADDR10 = 0x17,
330
+ NPCM7XX_SMB_CTL4 = 0x1a,
331
+ NPCM7XX_SMB_CTL5 = 0x1b,
332
+ NPCM7XX_SMB_SCLLT = 0x1c,
333
+ NPCM7XX_SMB_FIF_CTL = 0x1d,
334
+ NPCM7XX_SMB_SCLHT = 0x1e,
335
+};
336
+
337
+enum NPCM7xxSMBusBank1Register {
338
+ NPCM7XX_SMB_FIF_CTS = 0x10,
339
+ NPCM7XX_SMB_FAIR_PER = 0x11,
340
+ NPCM7XX_SMB_TXF_CTL = 0x12,
341
+ NPCM7XX_SMB_T_OUT = 0x14,
342
+ NPCM7XX_SMB_TXF_STS = 0x1a,
343
+ NPCM7XX_SMB_RXF_STS = 0x1c,
344
+ NPCM7XX_SMB_RXF_CTL = 0x1e,
345
+};
346
+
347
+/* ST fields */
348
+#define NPCM7XX_SMBST_STP BIT(7)
349
+#define NPCM7XX_SMBST_SDAST BIT(6)
350
+#define NPCM7XX_SMBST_BER BIT(5)
351
+#define NPCM7XX_SMBST_NEGACK BIT(4)
352
+#define NPCM7XX_SMBST_STASTR BIT(3)
353
+#define NPCM7XX_SMBST_NMATCH BIT(2)
354
+#define NPCM7XX_SMBST_MODE BIT(1)
355
+#define NPCM7XX_SMBST_XMIT BIT(0)
356
+
357
+/* CST fields */
358
+#define NPCM7XX_SMBCST_ARPMATCH BIT(7)
359
+#define NPCM7XX_SMBCST_MATCHAF BIT(6)
360
+#define NPCM7XX_SMBCST_TGSCL BIT(5)
361
+#define NPCM7XX_SMBCST_TSDA BIT(4)
362
+#define NPCM7XX_SMBCST_GCMATCH BIT(3)
363
+#define NPCM7XX_SMBCST_MATCH BIT(2)
364
+#define NPCM7XX_SMBCST_BB BIT(1)
365
+#define NPCM7XX_SMBCST_BUSY BIT(0)
366
+
367
+/* CST2 fields */
368
+#define NPCM7XX_SMBCST2_INTSTS BIT(7)
369
+#define NPCM7XX_SMBCST2_MATCH7F BIT(6)
370
+#define NPCM7XX_SMBCST2_MATCH6F BIT(5)
371
+#define NPCM7XX_SMBCST2_MATCH5F BIT(4)
372
+#define NPCM7XX_SMBCST2_MATCH4F BIT(3)
373
+#define NPCM7XX_SMBCST2_MATCH3F BIT(2)
374
+#define NPCM7XX_SMBCST2_MATCH2F BIT(1)
375
+#define NPCM7XX_SMBCST2_MATCH1F BIT(0)
376
+
377
+/* CST3 fields */
378
+#define NPCM7XX_SMBCST3_EO_BUSY BIT(7)
379
+#define NPCM7XX_SMBCST3_MATCH10F BIT(2)
380
+#define NPCM7XX_SMBCST3_MATCH9F BIT(1)
381
+#define NPCM7XX_SMBCST3_MATCH8F BIT(0)
382
+
383
+/* CTL1 fields */
384
+#define NPCM7XX_SMBCTL1_STASTRE BIT(7)
385
+#define NPCM7XX_SMBCTL1_NMINTE BIT(6)
386
+#define NPCM7XX_SMBCTL1_GCMEN BIT(5)
387
+#define NPCM7XX_SMBCTL1_ACK BIT(4)
388
+#define NPCM7XX_SMBCTL1_EOBINTE BIT(3)
389
+#define NPCM7XX_SMBCTL1_INTEN BIT(2)
390
+#define NPCM7XX_SMBCTL1_STOP BIT(1)
391
+#define NPCM7XX_SMBCTL1_START BIT(0)
392
+
393
+/* CTL2 fields */
394
+#define NPCM7XX_SMBCTL2_SCLFRQ(rv) extract8((rv), 1, 6)
395
+#define NPCM7XX_SMBCTL2_ENABLE BIT(0)
396
+
397
+/* CTL3 fields */
398
+#define NPCM7XX_SMBCTL3_SCL_LVL BIT(7)
399
+#define NPCM7XX_SMBCTL3_SDA_LVL BIT(6)
400
+#define NPCM7XX_SMBCTL3_BNK_SEL BIT(5)
401
+#define NPCM7XX_SMBCTL3_400K_MODE BIT(4)
402
+#define NPCM7XX_SMBCTL3_IDL_START BIT(3)
403
+#define NPCM7XX_SMBCTL3_ARPMEN BIT(2)
404
+#define NPCM7XX_SMBCTL3_SCLFRQ(rv) extract8((rv), 0, 2)
405
+
406
+/* ADDR fields */
407
+#define NPCM7XX_ADDR_EN BIT(7)
408
+#define NPCM7XX_ADDR_A(rv) extract8((rv), 0, 6)
409
+
410
+#define KEEP_OLD_BIT(o, n, b) (((n) & (~(b))) | ((o) & (b)))
411
+#define WRITE_ONE_CLEAR(o, n, b) ((n) & (b) ? (o) & (~(b)) : (o))
412
+
413
+#define NPCM7XX_SMBUS_ENABLED(s) ((s)->ctl2 & NPCM7XX_SMBCTL2_ENABLE)
414
+
415
+/* VERSION fields values, read-only. */
416
+#define NPCM7XX_SMBUS_VERSION_NUMBER 1
417
+#define NPCM7XX_SMBUS_VERSION_FIFO_SUPPORTED 0
418
+
419
+/* Reset values */
420
+#define NPCM7XX_SMB_ST_INIT_VAL 0x00
421
+#define NPCM7XX_SMB_CST_INIT_VAL 0x10
422
+#define NPCM7XX_SMB_CST2_INIT_VAL 0x00
423
+#define NPCM7XX_SMB_CST3_INIT_VAL 0x00
424
+#define NPCM7XX_SMB_CTL1_INIT_VAL 0x00
425
+#define NPCM7XX_SMB_CTL2_INIT_VAL 0x00
426
+#define NPCM7XX_SMB_CTL3_INIT_VAL 0xc0
427
+#define NPCM7XX_SMB_CTL4_INIT_VAL 0x07
428
+#define NPCM7XX_SMB_CTL5_INIT_VAL 0x00
429
+#define NPCM7XX_SMB_ADDR_INIT_VAL 0x00
430
+#define NPCM7XX_SMB_SCLLT_INIT_VAL 0x00
431
+#define NPCM7XX_SMB_SCLHT_INIT_VAL 0x00
432
+
433
+static uint8_t npcm7xx_smbus_get_version(void)
434
+{
435
+ return NPCM7XX_SMBUS_VERSION_FIFO_SUPPORTED << 7 |
436
+ NPCM7XX_SMBUS_VERSION_NUMBER;
437
+}
438
+
439
+static void npcm7xx_smbus_update_irq(NPCM7xxSMBusState *s)
440
+{
441
+ int level;
442
+
443
+ if (s->ctl1 & NPCM7XX_SMBCTL1_INTEN) {
444
+ level = !!((s->ctl1 & NPCM7XX_SMBCTL1_NMINTE &&
445
+ s->st & NPCM7XX_SMBST_NMATCH) ||
446
+ (s->st & NPCM7XX_SMBST_BER) ||
447
+ (s->st & NPCM7XX_SMBST_NEGACK) ||
448
+ (s->st & NPCM7XX_SMBST_SDAST) ||
449
+ (s->ctl1 & NPCM7XX_SMBCTL1_STASTRE &&
450
+ s->st & NPCM7XX_SMBST_SDAST) ||
451
+ (s->ctl1 & NPCM7XX_SMBCTL1_EOBINTE &&
452
+ s->cst3 & NPCM7XX_SMBCST3_EO_BUSY));
453
+
454
+ if (level) {
455
+ s->cst2 |= NPCM7XX_SMBCST2_INTSTS;
456
+ } else {
457
+ s->cst2 &= ~NPCM7XX_SMBCST2_INTSTS;
458
+ }
459
+ qemu_set_irq(s->irq, level);
460
+ }
78
+ }
461
+}
79
+}
462
+
80
+
463
+static void npcm7xx_smbus_nack(NPCM7xxSMBusState *s)
81
+void HELPER(sme_addha_d)(void *vzda, void *vzn, void *vpn,
82
+ void *vpm, uint32_t desc)
464
+{
83
+{
465
+ s->st &= ~NPCM7XX_SMBST_SDAST;
84
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 8;
466
+ s->st |= NPCM7XX_SMBST_NEGACK;
85
+ uint8_t *pn = vpn, *pm = vpm;
467
+ s->status = NPCM7XX_SMBUS_STATUS_NEGACK;
86
+ uint64_t *zda = vzda, *zn = vzn;
468
+}
469
+
87
+
470
+static void npcm7xx_smbus_send_byte(NPCM7xxSMBusState *s, uint8_t value)
88
+ for (row = 0; row < oprsz; ++row) {
471
+{
89
+ if (pn[H1(row)] & 1) {
472
+ int rv = i2c_send(s->bus, value);
90
+ for (col = 0; col < oprsz; ++col) {
473
+
91
+ if (pm[H1(col)] & 1) {
474
+ if (rv) {
92
+ zda[tile_vslice_index(row) + col] += zn[col];
475
+ npcm7xx_smbus_nack(s);
93
+ }
476
+ } else {
94
+ }
477
+ s->st |= NPCM7XX_SMBST_SDAST;
478
+ }
479
+ trace_npcm7xx_smbus_send_byte((DEVICE(s)->canonical_path), value, !rv);
480
+ npcm7xx_smbus_update_irq(s);
481
+}
482
+
483
+static void npcm7xx_smbus_recv_byte(NPCM7xxSMBusState *s)
484
+{
485
+ s->sda = i2c_recv(s->bus);
486
+ s->st |= NPCM7XX_SMBST_SDAST;
487
+ if (s->st & NPCM7XX_SMBCTL1_ACK) {
488
+ trace_npcm7xx_smbus_nack(DEVICE(s)->canonical_path);
489
+ i2c_nack(s->bus);
490
+ s->st &= NPCM7XX_SMBCTL1_ACK;
491
+ }
492
+ trace_npcm7xx_smbus_recv_byte((DEVICE(s)->canonical_path), s->sda);
493
+ npcm7xx_smbus_update_irq(s);
494
+}
495
+
496
+static void npcm7xx_smbus_start(NPCM7xxSMBusState *s)
497
+{
498
+ /*
499
+ * We can start the bus if one of these is true:
500
+ * 1. The bus is idle (so we can request it)
501
+ * 2. We are the occupier (it's a repeated start condition.)
502
+ */
503
+ int available = !i2c_bus_busy(s->bus) ||
504
+ s->status != NPCM7XX_SMBUS_STATUS_IDLE;
505
+
506
+ if (available) {
507
+ s->st |= NPCM7XX_SMBST_MODE | NPCM7XX_SMBST_XMIT | NPCM7XX_SMBST_SDAST;
508
+ s->cst |= NPCM7XX_SMBCST_BUSY;
509
+ } else {
510
+ s->st &= ~NPCM7XX_SMBST_MODE;
511
+ s->cst &= ~NPCM7XX_SMBCST_BUSY;
512
+ s->st |= NPCM7XX_SMBST_BER;
513
+ }
514
+
515
+ trace_npcm7xx_smbus_start(DEVICE(s)->canonical_path, available);
516
+ s->cst |= NPCM7XX_SMBCST_BB;
517
+ s->status = NPCM7XX_SMBUS_STATUS_IDLE;
518
+ npcm7xx_smbus_update_irq(s);
519
+}
520
+
521
+static void npcm7xx_smbus_send_address(NPCM7xxSMBusState *s, uint8_t value)
522
+{
523
+ int recv;
524
+ int rv;
525
+
526
+ recv = value & BIT(0);
527
+ rv = i2c_start_transfer(s->bus, value >> 1, recv);
528
+ trace_npcm7xx_smbus_send_address(DEVICE(s)->canonical_path,
529
+ value >> 1, recv, !rv);
530
+ if (rv) {
531
+ qemu_log_mask(LOG_GUEST_ERROR,
532
+ "%s: requesting i2c bus for 0x%02x failed: %d\n",
533
+ DEVICE(s)->canonical_path, value, rv);
534
+ /* Failed to start transfer. NACK to reject.*/
535
+ if (recv) {
536
+ s->st &= ~NPCM7XX_SMBST_XMIT;
537
+ } else {
538
+ s->st |= NPCM7XX_SMBST_XMIT;
539
+ }
540
+ npcm7xx_smbus_nack(s);
541
+ npcm7xx_smbus_update_irq(s);
542
+ return;
543
+ }
544
+
545
+ s->st &= ~NPCM7XX_SMBST_NEGACK;
546
+ if (recv) {
547
+ s->status = NPCM7XX_SMBUS_STATUS_RECEIVING;
548
+ s->st &= ~NPCM7XX_SMBST_XMIT;
549
+ } else {
550
+ s->status = NPCM7XX_SMBUS_STATUS_SENDING;
551
+ s->st |= NPCM7XX_SMBST_XMIT;
552
+ }
553
+
554
+ if (s->ctl1 & NPCM7XX_SMBCTL1_STASTRE) {
555
+ s->st |= NPCM7XX_SMBST_STASTR;
556
+ if (!recv) {
557
+ s->st |= NPCM7XX_SMBST_SDAST;
558
+ }
559
+ } else if (recv) {
560
+ npcm7xx_smbus_recv_byte(s);
561
+ }
562
+ npcm7xx_smbus_update_irq(s);
563
+}
564
+
565
+static void npcm7xx_smbus_execute_stop(NPCM7xxSMBusState *s)
566
+{
567
+ i2c_end_transfer(s->bus);
568
+ s->st = 0;
569
+ s->cst = 0;
570
+ s->status = NPCM7XX_SMBUS_STATUS_IDLE;
571
+ s->cst3 |= NPCM7XX_SMBCST3_EO_BUSY;
572
+ trace_npcm7xx_smbus_stop(DEVICE(s)->canonical_path);
573
+ npcm7xx_smbus_update_irq(s);
574
+}
575
+
576
+
577
+static void npcm7xx_smbus_stop(NPCM7xxSMBusState *s)
578
+{
579
+ if (s->st & NPCM7XX_SMBST_MODE) {
580
+ switch (s->status) {
581
+ case NPCM7XX_SMBUS_STATUS_RECEIVING:
582
+ case NPCM7XX_SMBUS_STATUS_STOPPING_LAST_RECEIVE:
583
+ s->status = NPCM7XX_SMBUS_STATUS_STOPPING_LAST_RECEIVE;
584
+ break;
585
+
586
+ case NPCM7XX_SMBUS_STATUS_NEGACK:
587
+ s->status = NPCM7XX_SMBUS_STATUS_STOPPING_NEGACK;
588
+ break;
589
+
590
+ default:
591
+ npcm7xx_smbus_execute_stop(s);
592
+ break;
593
+ }
95
+ }
594
+ }
96
+ }
595
+}
97
+}
596
+
98
+
597
+static uint8_t npcm7xx_smbus_read_sda(NPCM7xxSMBusState *s)
99
+void HELPER(sme_addva_s)(void *vzda, void *vzn, void *vpn,
100
+ void *vpm, uint32_t desc)
598
+{
101
+{
599
+ uint8_t value = s->sda;
102
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 4;
103
+ uint64_t *pn = vpn, *pm = vpm;
104
+ uint32_t *zda = vzda, *zn = vzn;
600
+
105
+
601
+ switch (s->status) {
106
+ for (row = 0; row < oprsz; ) {
602
+ case NPCM7XX_SMBUS_STATUS_STOPPING_LAST_RECEIVE:
107
+ uint64_t pa = pn[row >> 4];
603
+ npcm7xx_smbus_execute_stop(s);
108
+ do {
604
+ break;
109
+ if (pa & 1) {
605
+
110
+ uint32_t zn_row = zn[H4(row)];
606
+ case NPCM7XX_SMBUS_STATUS_RECEIVING:
111
+ for (col = 0; col < oprsz; ) {
607
+ npcm7xx_smbus_recv_byte(s);
112
+ uint64_t pb = pm[col >> 4];
608
+ break;
113
+ do {
609
+
114
+ if (pb & 1) {
610
+ default:
115
+ zda[tile_vslice_index(row) + H4(col)] += zn_row;
611
+ /* Do nothing */
116
+ }
612
+ break;
117
+ pb >>= 4;
118
+ } while (++col & 15);
119
+ }
120
+ }
121
+ pa >>= 4;
122
+ } while (++row & 15);
613
+ }
123
+ }
614
+
615
+ return value;
616
+}
124
+}
617
+
125
+
618
+static void npcm7xx_smbus_write_sda(NPCM7xxSMBusState *s, uint8_t value)
126
+void HELPER(sme_addva_d)(void *vzda, void *vzn, void *vpn,
127
+ void *vpm, uint32_t desc)
619
+{
128
+{
620
+ s->sda = value;
129
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 8;
621
+ if (s->st & NPCM7XX_SMBST_MODE) {
130
+ uint8_t *pn = vpn, *pm = vpm;
622
+ switch (s->status) {
131
+ uint64_t *zda = vzda, *zn = vzn;
623
+ case NPCM7XX_SMBUS_STATUS_IDLE:
132
+
624
+ npcm7xx_smbus_send_address(s, value);
133
+ for (row = 0; row < oprsz; ++row) {
625
+ break;
134
+ if (pn[H1(row)] & 1) {
626
+ case NPCM7XX_SMBUS_STATUS_SENDING:
135
+ uint64_t zn_row = zn[row];
627
+ npcm7xx_smbus_send_byte(s, value);
136
+ for (col = 0; col < oprsz; ++col) {
628
+ break;
137
+ if (pm[H1(col)] & 1) {
629
+ default:
138
+ zda[tile_vslice_index(row) + col] += zn_row;
630
+ qemu_log_mask(LOG_GUEST_ERROR,
139
+ }
631
+ "%s: write to SDA in invalid status %d: %u\n",
140
+ }
632
+ DEVICE(s)->canonical_path, s->status, value);
633
+ break;
634
+ }
141
+ }
635
+ }
142
+ }
636
+}
143
+}
144
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
145
index XXXXXXX..XXXXXXX 100644
146
--- a/target/arm/translate-sme.c
147
+++ b/target/arm/translate-sme.c
148
@@ -XXX,XX +XXX,XX @@ static bool do_ldst_r(DisasContext *s, arg_ldstr *a, GenLdStR *fn)
149
150
TRANS_FEAT(LDR, aa64_sme, do_ldst_r, a, gen_sve_ldr)
151
TRANS_FEAT(STR, aa64_sme, do_ldst_r, a, gen_sve_str)
637
+
152
+
638
+static void npcm7xx_smbus_write_st(NPCM7xxSMBusState *s, uint8_t value)
153
+static bool do_adda(DisasContext *s, arg_adda *a, MemOp esz,
154
+ gen_helper_gvec_4 *fn)
639
+{
155
+{
640
+ s->st = WRITE_ONE_CLEAR(s->st, value, NPCM7XX_SMBST_STP);
156
+ int svl = streaming_vec_reg_size(s);
641
+ s->st = WRITE_ONE_CLEAR(s->st, value, NPCM7XX_SMBST_BER);
157
+ uint32_t desc = simd_desc(svl, svl, 0);
642
+ s->st = WRITE_ONE_CLEAR(s->st, value, NPCM7XX_SMBST_STASTR);
158
+ TCGv_ptr za, zn, pn, pm;
643
+ s->st = WRITE_ONE_CLEAR(s->st, value, NPCM7XX_SMBST_NMATCH);
644
+
159
+
645
+ if (value & NPCM7XX_SMBST_NEGACK) {
160
+ if (!sme_smza_enabled_check(s)) {
646
+ s->st &= ~NPCM7XX_SMBST_NEGACK;
161
+ return true;
647
+ if (s->status == NPCM7XX_SMBUS_STATUS_STOPPING_NEGACK) {
648
+ npcm7xx_smbus_execute_stop(s);
649
+ }
650
+ }
162
+ }
651
+
163
+
652
+ if (value & NPCM7XX_SMBST_STASTR &&
164
+ /* Sum XZR+zad to find ZAd. */
653
+ s->status == NPCM7XX_SMBUS_STATUS_RECEIVING) {
165
+ za = get_tile_rowcol(s, esz, 31, a->zad, false);
654
+ npcm7xx_smbus_recv_byte(s);
166
+ zn = vec_full_reg_ptr(s, a->zn);
655
+ }
167
+ pn = pred_full_reg_ptr(s, a->pn);
168
+ pm = pred_full_reg_ptr(s, a->pm);
656
+
169
+
657
+ npcm7xx_smbus_update_irq(s);
170
+ fn(za, zn, pn, pm, tcg_constant_i32(desc));
171
+
172
+ tcg_temp_free_ptr(za);
173
+ tcg_temp_free_ptr(zn);
174
+ tcg_temp_free_ptr(pn);
175
+ tcg_temp_free_ptr(pm);
176
+ return true;
658
+}
177
+}
659
+
178
+
660
+static void npcm7xx_smbus_write_cst(NPCM7xxSMBusState *s, uint8_t value)
179
+TRANS_FEAT(ADDHA_s, aa64_sme, do_adda, a, MO_32, gen_helper_sme_addha_s)
661
+{
180
+TRANS_FEAT(ADDVA_s, aa64_sme, do_adda, a, MO_32, gen_helper_sme_addva_s)
662
+ uint8_t new_value = s->cst;
181
+TRANS_FEAT(ADDHA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addha_d)
663
+
182
+TRANS_FEAT(ADDVA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addva_d)
664
+ s->cst = WRITE_ONE_CLEAR(new_value, value, NPCM7XX_SMBCST_BB);
665
+ npcm7xx_smbus_update_irq(s);
666
+}
667
+
668
+static void npcm7xx_smbus_write_cst3(NPCM7xxSMBusState *s, uint8_t value)
669
+{
670
+ s->cst3 = WRITE_ONE_CLEAR(s->cst3, value, NPCM7XX_SMBCST3_EO_BUSY);
671
+ npcm7xx_smbus_update_irq(s);
672
+}
673
+
674
+static void npcm7xx_smbus_write_ctl1(NPCM7xxSMBusState *s, uint8_t value)
675
+{
676
+ s->ctl1 = KEEP_OLD_BIT(s->ctl1, value,
677
+ NPCM7XX_SMBCTL1_START | NPCM7XX_SMBCTL1_STOP | NPCM7XX_SMBCTL1_ACK);
678
+
679
+ if (value & NPCM7XX_SMBCTL1_START) {
680
+ npcm7xx_smbus_start(s);
681
+ }
682
+
683
+ if (value & NPCM7XX_SMBCTL1_STOP) {
684
+ npcm7xx_smbus_stop(s);
685
+ }
686
+
687
+ npcm7xx_smbus_update_irq(s);
688
+}
689
+
690
+static void npcm7xx_smbus_write_ctl2(NPCM7xxSMBusState *s, uint8_t value)
691
+{
692
+ s->ctl2 = value;
693
+
694
+ if (!NPCM7XX_SMBUS_ENABLED(s)) {
695
+ /* Disable this SMBus module. */
696
+ s->ctl1 = 0;
697
+ s->st = 0;
698
+ s->cst3 = s->cst3 & (~NPCM7XX_SMBCST3_EO_BUSY);
699
+ s->cst = 0;
700
+ }
701
+}
702
+
703
+static void npcm7xx_smbus_write_ctl3(NPCM7xxSMBusState *s, uint8_t value)
704
+{
705
+ uint8_t old_ctl3 = s->ctl3;
706
+
707
+ /* Write to SDA and SCL bits are ignored. */
708
+ s->ctl3 = KEEP_OLD_BIT(old_ctl3, value,
709
+ NPCM7XX_SMBCTL3_SCL_LVL | NPCM7XX_SMBCTL3_SDA_LVL);
710
+}
711
+
712
+static uint64_t npcm7xx_smbus_read(void *opaque, hwaddr offset, unsigned size)
713
+{
714
+ NPCM7xxSMBusState *s = opaque;
715
+ uint64_t value = 0;
716
+ uint8_t bank = s->ctl3 & NPCM7XX_SMBCTL3_BNK_SEL;
717
+
718
+ /* The order of the registers are their order in memory. */
719
+ switch (offset) {
720
+ case NPCM7XX_SMB_SDA:
721
+ value = npcm7xx_smbus_read_sda(s);
722
+ break;
723
+
724
+ case NPCM7XX_SMB_ST:
725
+ value = s->st;
726
+ break;
727
+
728
+ case NPCM7XX_SMB_CST:
729
+ value = s->cst;
730
+ break;
731
+
732
+ case NPCM7XX_SMB_CTL1:
733
+ value = s->ctl1;
734
+ break;
735
+
736
+ case NPCM7XX_SMB_ADDR1:
737
+ value = s->addr[0];
738
+ break;
739
+
740
+ case NPCM7XX_SMB_CTL2:
741
+ value = s->ctl2;
742
+ break;
743
+
744
+ case NPCM7XX_SMB_ADDR2:
745
+ value = s->addr[1];
746
+ break;
747
+
748
+ case NPCM7XX_SMB_CTL3:
749
+ value = s->ctl3;
750
+ break;
751
+
752
+ case NPCM7XX_SMB_CST2:
753
+ value = s->cst2;
754
+ break;
755
+
756
+ case NPCM7XX_SMB_CST3:
757
+ value = s->cst3;
758
+ break;
759
+
760
+ case NPCM7XX_SMB_VER:
761
+ value = npcm7xx_smbus_get_version();
762
+ break;
763
+
764
+ /* This register is either invalid or banked at this point. */
765
+ default:
766
+ if (bank) {
767
+ /* Bank 1 */
768
+ qemu_log_mask(LOG_GUEST_ERROR,
769
+ "%s: read from invalid offset 0x%" HWADDR_PRIx "\n",
770
+ DEVICE(s)->canonical_path, offset);
771
+ } else {
772
+ /* Bank 0 */
773
+ switch (offset) {
774
+ case NPCM7XX_SMB_ADDR3:
775
+ value = s->addr[2];
776
+ break;
777
+
778
+ case NPCM7XX_SMB_ADDR7:
779
+ value = s->addr[6];
780
+ break;
781
+
782
+ case NPCM7XX_SMB_ADDR4:
783
+ value = s->addr[3];
784
+ break;
785
+
786
+ case NPCM7XX_SMB_ADDR8:
787
+ value = s->addr[7];
788
+ break;
789
+
790
+ case NPCM7XX_SMB_ADDR5:
791
+ value = s->addr[4];
792
+ break;
793
+
794
+ case NPCM7XX_SMB_ADDR9:
795
+ value = s->addr[8];
796
+ break;
797
+
798
+ case NPCM7XX_SMB_ADDR6:
799
+ value = s->addr[5];
800
+ break;
801
+
802
+ case NPCM7XX_SMB_ADDR10:
803
+ value = s->addr[9];
804
+ break;
805
+
806
+ case NPCM7XX_SMB_CTL4:
807
+ value = s->ctl4;
808
+ break;
809
+
810
+ case NPCM7XX_SMB_CTL5:
811
+ value = s->ctl5;
812
+ break;
813
+
814
+ case NPCM7XX_SMB_SCLLT:
815
+ value = s->scllt;
816
+ break;
817
+
818
+ case NPCM7XX_SMB_SCLHT:
819
+ value = s->sclht;
820
+ break;
821
+
822
+ default:
823
+ qemu_log_mask(LOG_GUEST_ERROR,
824
+ "%s: read from invalid offset 0x%" HWADDR_PRIx "\n",
825
+ DEVICE(s)->canonical_path, offset);
826
+ break;
827
+ }
828
+ }
829
+ break;
830
+ }
831
+
832
+ trace_npcm7xx_smbus_read(DEVICE(s)->canonical_path, offset, value, size);
833
+
834
+ return value;
835
+}
836
+
837
+static void npcm7xx_smbus_write(void *opaque, hwaddr offset, uint64_t value,
838
+ unsigned size)
839
+{
840
+ NPCM7xxSMBusState *s = opaque;
841
+ uint8_t bank = s->ctl3 & NPCM7XX_SMBCTL3_BNK_SEL;
842
+
843
+ trace_npcm7xx_smbus_write(DEVICE(s)->canonical_path, offset, value, size);
844
+
845
+ /* The order of the registers are their order in memory. */
846
+ switch (offset) {
847
+ case NPCM7XX_SMB_SDA:
848
+ npcm7xx_smbus_write_sda(s, value);
849
+ break;
850
+
851
+ case NPCM7XX_SMB_ST:
852
+ npcm7xx_smbus_write_st(s, value);
853
+ break;
854
+
855
+ case NPCM7XX_SMB_CST:
856
+ npcm7xx_smbus_write_cst(s, value);
857
+ break;
858
+
859
+ case NPCM7XX_SMB_CTL1:
860
+ npcm7xx_smbus_write_ctl1(s, value);
861
+ break;
862
+
863
+ case NPCM7XX_SMB_ADDR1:
864
+ s->addr[0] = value;
865
+ break;
866
+
867
+ case NPCM7XX_SMB_CTL2:
868
+ npcm7xx_smbus_write_ctl2(s, value);
869
+ break;
870
+
871
+ case NPCM7XX_SMB_ADDR2:
872
+ s->addr[1] = value;
873
+ break;
874
+
875
+ case NPCM7XX_SMB_CTL3:
876
+ npcm7xx_smbus_write_ctl3(s, value);
877
+ break;
878
+
879
+ case NPCM7XX_SMB_CST2:
880
+ qemu_log_mask(LOG_GUEST_ERROR,
881
+ "%s: write to read-only reg: offset 0x%" HWADDR_PRIx "\n",
882
+ DEVICE(s)->canonical_path, offset);
883
+ break;
884
+
885
+ case NPCM7XX_SMB_CST3:
886
+ npcm7xx_smbus_write_cst3(s, value);
887
+ break;
888
+
889
+ case NPCM7XX_SMB_VER:
890
+ qemu_log_mask(LOG_GUEST_ERROR,
891
+ "%s: write to read-only reg: offset 0x%" HWADDR_PRIx "\n",
892
+ DEVICE(s)->canonical_path, offset);
893
+ break;
894
+
895
+ /* This register is either invalid or banked at this point. */
896
+ default:
897
+ if (bank) {
898
+ /* Bank 1 */
899
+ qemu_log_mask(LOG_GUEST_ERROR,
900
+ "%s: write to invalid offset 0x%" HWADDR_PRIx "\n",
901
+ DEVICE(s)->canonical_path, offset);
902
+ } else {
903
+ /* Bank 0 */
904
+ switch (offset) {
905
+ case NPCM7XX_SMB_ADDR3:
906
+ s->addr[2] = value;
907
+ break;
908
+
909
+ case NPCM7XX_SMB_ADDR7:
910
+ s->addr[6] = value;
911
+ break;
912
+
913
+ case NPCM7XX_SMB_ADDR4:
914
+ s->addr[3] = value;
915
+ break;
916
+
917
+ case NPCM7XX_SMB_ADDR8:
918
+ s->addr[7] = value;
919
+ break;
920
+
921
+ case NPCM7XX_SMB_ADDR5:
922
+ s->addr[4] = value;
923
+ break;
924
+
925
+ case NPCM7XX_SMB_ADDR9:
926
+ s->addr[8] = value;
927
+ break;
928
+
929
+ case NPCM7XX_SMB_ADDR6:
930
+ s->addr[5] = value;
931
+ break;
932
+
933
+ case NPCM7XX_SMB_ADDR10:
934
+ s->addr[9] = value;
935
+ break;
936
+
937
+ case NPCM7XX_SMB_CTL4:
938
+ s->ctl4 = value;
939
+ break;
940
+
941
+ case NPCM7XX_SMB_CTL5:
942
+ s->ctl5 = value;
943
+ break;
944
+
945
+ case NPCM7XX_SMB_SCLLT:
946
+ s->scllt = value;
947
+ break;
948
+
949
+ case NPCM7XX_SMB_SCLHT:
950
+ s->sclht = value;
951
+ break;
952
+
953
+ default:
954
+ qemu_log_mask(LOG_GUEST_ERROR,
955
+ "%s: write to invalid offset 0x%" HWADDR_PRIx "\n",
956
+ DEVICE(s)->canonical_path, offset);
957
+ break;
958
+ }
959
+ }
960
+ break;
961
+ }
962
+}
963
+
964
+static const MemoryRegionOps npcm7xx_smbus_ops = {
965
+ .read = npcm7xx_smbus_read,
966
+ .write = npcm7xx_smbus_write,
967
+ .endianness = DEVICE_LITTLE_ENDIAN,
968
+ .valid = {
969
+ .min_access_size = 1,
970
+ .max_access_size = 1,
971
+ .unaligned = false,
972
+ },
973
+};
974
+
975
+static void npcm7xx_smbus_enter_reset(Object *obj, ResetType type)
976
+{
977
+ NPCM7xxSMBusState *s = NPCM7XX_SMBUS(obj);
978
+
979
+ s->st = NPCM7XX_SMB_ST_INIT_VAL;
980
+ s->cst = NPCM7XX_SMB_CST_INIT_VAL;
981
+ s->cst2 = NPCM7XX_SMB_CST2_INIT_VAL;
982
+ s->cst3 = NPCM7XX_SMB_CST3_INIT_VAL;
983
+ s->ctl1 = NPCM7XX_SMB_CTL1_INIT_VAL;
984
+ s->ctl2 = NPCM7XX_SMB_CTL2_INIT_VAL;
985
+ s->ctl3 = NPCM7XX_SMB_CTL3_INIT_VAL;
986
+ s->ctl4 = NPCM7XX_SMB_CTL4_INIT_VAL;
987
+ s->ctl5 = NPCM7XX_SMB_CTL5_INIT_VAL;
988
+
989
+ for (int i = 0; i < NPCM7XX_SMBUS_NR_ADDRS; ++i) {
990
+ s->addr[i] = NPCM7XX_SMB_ADDR_INIT_VAL;
991
+ }
992
+ s->scllt = NPCM7XX_SMB_SCLLT_INIT_VAL;
993
+ s->sclht = NPCM7XX_SMB_SCLHT_INIT_VAL;
994
+
995
+ s->status = NPCM7XX_SMBUS_STATUS_IDLE;
996
+}
997
+
998
+static void npcm7xx_smbus_hold_reset(Object *obj)
999
+{
1000
+ NPCM7xxSMBusState *s = NPCM7XX_SMBUS(obj);
1001
+
1002
+ qemu_irq_lower(s->irq);
1003
+}
1004
+
1005
+static void npcm7xx_smbus_init(Object *obj)
1006
+{
1007
+ NPCM7xxSMBusState *s = NPCM7XX_SMBUS(obj);
1008
+ SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
1009
+
1010
+ sysbus_init_irq(sbd, &s->irq);
1011
+ memory_region_init_io(&s->iomem, obj, &npcm7xx_smbus_ops, s,
1012
+ "regs", 4 * KiB);
1013
+ sysbus_init_mmio(sbd, &s->iomem);
1014
+
1015
+ s->bus = i2c_init_bus(DEVICE(s), "i2c-bus");
1016
+ s->status = NPCM7XX_SMBUS_STATUS_IDLE;
1017
+}
1018
+
1019
+static const VMStateDescription vmstate_npcm7xx_smbus = {
1020
+ .name = "npcm7xx-smbus",
1021
+ .version_id = 0,
1022
+ .minimum_version_id = 0,
1023
+ .fields = (VMStateField[]) {
1024
+ VMSTATE_UINT8(sda, NPCM7xxSMBusState),
1025
+ VMSTATE_UINT8(st, NPCM7xxSMBusState),
1026
+ VMSTATE_UINT8(cst, NPCM7xxSMBusState),
1027
+ VMSTATE_UINT8(cst2, NPCM7xxSMBusState),
1028
+ VMSTATE_UINT8(cst3, NPCM7xxSMBusState),
1029
+ VMSTATE_UINT8(ctl1, NPCM7xxSMBusState),
1030
+ VMSTATE_UINT8(ctl2, NPCM7xxSMBusState),
1031
+ VMSTATE_UINT8(ctl3, NPCM7xxSMBusState),
1032
+ VMSTATE_UINT8(ctl4, NPCM7xxSMBusState),
1033
+ VMSTATE_UINT8(ctl5, NPCM7xxSMBusState),
1034
+ VMSTATE_UINT8_ARRAY(addr, NPCM7xxSMBusState, NPCM7XX_SMBUS_NR_ADDRS),
1035
+ VMSTATE_UINT8(scllt, NPCM7xxSMBusState),
1036
+ VMSTATE_UINT8(sclht, NPCM7xxSMBusState),
1037
+ VMSTATE_END_OF_LIST(),
1038
+ },
1039
+};
1040
+
1041
+static void npcm7xx_smbus_class_init(ObjectClass *klass, void *data)
1042
+{
1043
+ ResettableClass *rc = RESETTABLE_CLASS(klass);
1044
+ DeviceClass *dc = DEVICE_CLASS(klass);
1045
+
1046
+ dc->desc = "NPCM7xx System Management Bus";
1047
+ dc->vmsd = &vmstate_npcm7xx_smbus;
1048
+ rc->phases.enter = npcm7xx_smbus_enter_reset;
1049
+ rc->phases.hold = npcm7xx_smbus_hold_reset;
1050
+}
1051
+
1052
+static const TypeInfo npcm7xx_smbus_types[] = {
1053
+ {
1054
+ .name = TYPE_NPCM7XX_SMBUS,
1055
+ .parent = TYPE_SYS_BUS_DEVICE,
1056
+ .instance_size = sizeof(NPCM7xxSMBusState),
1057
+ .class_init = npcm7xx_smbus_class_init,
1058
+ .instance_init = npcm7xx_smbus_init,
1059
+ },
1060
+};
1061
+DEFINE_TYPES(npcm7xx_smbus_types);
1062
diff --git a/hw/i2c/meson.build b/hw/i2c/meson.build
1063
index XXXXXXX..XXXXXXX 100644
1064
--- a/hw/i2c/meson.build
1065
+++ b/hw/i2c/meson.build
1066
@@ -XXX,XX +XXX,XX @@ i2c_ss.add(when: 'CONFIG_EXYNOS4', if_true: files('exynos4210_i2c.c'))
1067
i2c_ss.add(when: 'CONFIG_IMX_I2C', if_true: files('imx_i2c.c'))
1068
i2c_ss.add(when: 'CONFIG_MPC_I2C', if_true: files('mpc_i2c.c'))
1069
i2c_ss.add(when: 'CONFIG_NRF51_SOC', if_true: files('microbit_i2c.c'))
1070
+i2c_ss.add(when: 'CONFIG_NPCM7XX', if_true: files('npcm7xx_smbus.c'))
1071
i2c_ss.add(when: 'CONFIG_SMBUS_EEPROM', if_true: files('smbus_eeprom.c'))
1072
i2c_ss.add(when: 'CONFIG_VERSATILE_I2C', if_true: files('versatile_i2c.c'))
1073
i2c_ss.add(when: 'CONFIG_OMAP', if_true: files('omap_i2c.c'))
1074
diff --git a/hw/i2c/trace-events b/hw/i2c/trace-events
1075
index XXXXXXX..XXXXXXX 100644
1076
--- a/hw/i2c/trace-events
1077
+++ b/hw/i2c/trace-events
1078
@@ -XXX,XX +XXX,XX @@ aspeed_i2c_bus_read(uint32_t busid, uint64_t offset, unsigned size, uint64_t val
1079
aspeed_i2c_bus_write(uint32_t busid, uint64_t offset, unsigned size, uint64_t value) "bus[%d]: To 0x%" PRIx64 " of size %u: 0x%" PRIx64
1080
aspeed_i2c_bus_send(const char *mode, int i, int count, uint8_t byte) "%s send %d/%d 0x%02x"
1081
aspeed_i2c_bus_recv(const char *mode, int i, int count, uint8_t byte) "%s recv %d/%d 0x%02x"
1082
+
1083
+# npcm7xx_smbus.c
1084
+
1085
+npcm7xx_smbus_read(const char *id, uint64_t offset, uint64_t value, unsigned size) "%s offset: 0x%04" PRIx64 " value: 0x%02" PRIx64 " size: %u"
1086
+npcm7xx_smbus_write(const char *id, uint64_t offset, uint64_t value, unsigned size) "%s offset: 0x%04" PRIx64 " value: 0x%02" PRIx64 " size: %u"
1087
+npcm7xx_smbus_start(const char *id, int success) "%s starting, success: %d"
1088
+npcm7xx_smbus_send_address(const char *id, uint8_t addr, int recv, int success) "%s sending address: 0x%02x, recv: %d, success: %d"
1089
+npcm7xx_smbus_send_byte(const char *id, uint8_t value, int success) "%s send byte: 0x%02x, success: %d"
1090
+npcm7xx_smbus_recv_byte(const char *id, uint8_t value) "%s recv byte: 0x%02x"
1091
+npcm7xx_smbus_stop(const char *id) "%s stopping"
1092
+npcm7xx_smbus_nack(const char *id) "%s nacking"
1093
--
183
--
1094
2.20.1
184
2.25.1
1095
1096
diff view generated by jsdifflib
1
From: Doug Evans <dje@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Hao Wu <wuhaotsh@google.com>
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Reviewed-by: Avi Fishman <avi.fishman@nuvoton.com>
4
Message-id: 20220708151540.18136-25-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Doug Evans <dje@google.com>
7
Message-id: 20210213002520.1374134-4-dje@google.com
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
7
---
10
tests/qtest/npcm7xx_emc-test.c | 862 +++++++++++++++++++++++++++++++++
8
target/arm/helper-sme.h | 5 +++
11
tests/qtest/meson.build | 1 +
9
target/arm/sme.decode | 9 +++++
12
2 files changed, 863 insertions(+)
10
target/arm/sme_helper.c | 69 ++++++++++++++++++++++++++++++++++++++
13
create mode 100644 tests/qtest/npcm7xx_emc-test.c
11
target/arm/translate-sme.c | 32 ++++++++++++++++++
12
4 files changed, 115 insertions(+)
14
13
15
diff --git a/tests/qtest/npcm7xx_emc-test.c b/tests/qtest/npcm7xx_emc-test.c
14
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
16
new file mode 100644
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX
16
--- a/target/arm/helper-sme.h
18
--- /dev/null
17
+++ b/target/arm/helper-sme.h
19
+++ b/tests/qtest/npcm7xx_emc-test.c
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sme_addha_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
19
DEF_HELPER_FLAGS_5(sme_addva_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
20
DEF_HELPER_FLAGS_5(sme_addha_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
21
DEF_HELPER_FLAGS_5(sme_addva_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
22
+
23
+DEF_HELPER_FLAGS_7(sme_fmopa_s, TCG_CALL_NO_RWG,
24
+ void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
25
+DEF_HELPER_FLAGS_7(sme_fmopa_d, TCG_CALL_NO_RWG,
26
+ void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
27
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/sme.decode
30
+++ b/target/arm/sme.decode
31
@@ -XXX,XX +XXX,XX @@ ADDHA_s 11000000 10 01000 0 ... ... ..... 000 .. @adda_32
32
ADDVA_s 11000000 10 01000 1 ... ... ..... 000 .. @adda_32
33
ADDHA_d 11000000 11 01000 0 ... ... ..... 00 ... @adda_64
34
ADDVA_d 11000000 11 01000 1 ... ... ..... 00 ... @adda_64
35
+
36
+### SME Outer Product
37
+
38
+&op zad zn zm pm pn sub:bool
39
+@op_32 ........ ... zm:5 pm:3 pn:3 zn:5 sub:1 .. zad:2 &op
40
+@op_64 ........ ... zm:5 pm:3 pn:3 zn:5 sub:1 . zad:3 &op
41
+
42
+FMOPA_s 10000000 100 ..... ... ... ..... . 00 .. @op_32
43
+FMOPA_d 10000000 110 ..... ... ... ..... . 0 ... @op_64
44
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/target/arm/sme_helper.c
47
+++ b/target/arm/sme_helper.c
20
@@ -XXX,XX +XXX,XX @@
48
@@ -XXX,XX +XXX,XX @@
21
+/*
49
#include "exec/cpu_ldst.h"
22
+ * QTests for Nuvoton NPCM7xx EMC Modules.
50
#include "exec/exec-all.h"
23
+ *
51
#include "qemu/int128.h"
24
+ * Copyright 2020 Google LLC
52
+#include "fpu/softfloat.h"
25
+ *
53
#include "vec_internal.h"
26
+ * This program is free software; you can redistribute it and/or modify it
54
#include "sve_ldst_internal.h"
27
+ * under the terms of the GNU General Public License as published by the
55
28
+ * Free Software Foundation; either version 2 of the License, or
56
@@ -XXX,XX +XXX,XX @@ void HELPER(sme_addva_d)(void *vzda, void *vzn, void *vpn,
29
+ * (at your option) any later version.
57
}
30
+ *
58
}
31
+ * This program is distributed in the hope that it will be useful, but WITHOUT
59
}
32
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
33
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
34
+ * for more details.
35
+ */
36
+
60
+
37
+#include "qemu/osdep.h"
61
+void HELPER(sme_fmopa_s)(void *vza, void *vzn, void *vzm, void *vpn,
38
+#include "qemu-common.h"
62
+ void *vpm, void *vst, uint32_t desc)
39
+#include "libqos/libqos.h"
40
+#include "qapi/qmp/qdict.h"
41
+#include "qapi/qmp/qnum.h"
42
+#include "qemu/bitops.h"
43
+#include "qemu/iov.h"
44
+
45
+/* Name of the emc device. */
46
+#define TYPE_NPCM7XX_EMC "npcm7xx-emc"
47
+
48
+/* Timeout for various operations, in seconds. */
49
+#define TIMEOUT_SECONDS 10
50
+
51
+/* Address in memory of the descriptor. */
52
+#define DESC_ADDR (1 << 20) /* 1 MiB */
53
+
54
+/* Address in memory of the data packet. */
55
+#define DATA_ADDR (DESC_ADDR + 4096)
56
+
57
+#define CRC_LENGTH 4
58
+
59
+#define NUM_TX_DESCRIPTORS 3
60
+#define NUM_RX_DESCRIPTORS 2
61
+
62
+/* Size of tx,rx test buffers. */
63
+#define TX_DATA_LEN 64
64
+#define RX_DATA_LEN 64
65
+
66
+#define TX_STEP_COUNT 10000
67
+#define RX_STEP_COUNT 10000
68
+
69
+/* 32-bit register indices. */
70
+typedef enum NPCM7xxPWMRegister {
71
+ /* Control registers. */
72
+ REG_CAMCMR,
73
+ REG_CAMEN,
74
+
75
+ /* There are 16 CAMn[ML] registers. */
76
+ REG_CAMM_BASE,
77
+ REG_CAML_BASE,
78
+
79
+ REG_TXDLSA = 0x22,
80
+ REG_RXDLSA,
81
+ REG_MCMDR,
82
+ REG_MIID,
83
+ REG_MIIDA,
84
+ REG_FFTCR,
85
+ REG_TSDR,
86
+ REG_RSDR,
87
+ REG_DMARFC,
88
+ REG_MIEN,
89
+
90
+ /* Status registers. */
91
+ REG_MISTA,
92
+ REG_MGSTA,
93
+ REG_MPCNT,
94
+ REG_MRPC,
95
+ REG_MRPCC,
96
+ REG_MREPC,
97
+ REG_DMARFS,
98
+ REG_CTXDSA,
99
+ REG_CTXBSA,
100
+ REG_CRXDSA,
101
+ REG_CRXBSA,
102
+
103
+ NPCM7XX_NUM_EMC_REGS,
104
+} NPCM7xxPWMRegister;
105
+
106
+enum { NUM_CAMML_REGS = 16 };
107
+
108
+/* REG_CAMCMR fields */
109
+/* Enable CAM Compare */
110
+#define REG_CAMCMR_ECMP (1 << 4)
111
+/* Accept Unicast Packet */
112
+#define REG_CAMCMR_AUP (1 << 0)
113
+
114
+/* REG_MCMDR fields */
115
+/* Software Reset */
116
+#define REG_MCMDR_SWR (1 << 24)
117
+/* Frame Transmission On */
118
+#define REG_MCMDR_TXON (1 << 8)
119
+/* Accept Long Packet */
120
+#define REG_MCMDR_ALP (1 << 1)
121
+/* Frame Reception On */
122
+#define REG_MCMDR_RXON (1 << 0)
123
+
124
+/* REG_MIEN fields */
125
+/* Enable Transmit Completion Interrupt */
126
+#define REG_MIEN_ENTXCP (1 << 18)
127
+/* Enable Transmit Interrupt */
128
+#define REG_MIEN_ENTXINTR (1 << 16)
129
+/* Enable Receive Good Interrupt */
130
+#define REG_MIEN_ENRXGD (1 << 4)
131
+/* ENable Receive Interrupt */
132
+#define REG_MIEN_ENRXINTR (1 << 0)
133
+
134
+/* REG_MISTA fields */
135
+/* Transmit Bus Error Interrupt */
136
+#define REG_MISTA_TXBERR (1 << 24)
137
+/* Transmit Descriptor Unavailable Interrupt */
138
+#define REG_MISTA_TDU (1 << 23)
139
+/* Transmit Completion Interrupt */
140
+#define REG_MISTA_TXCP (1 << 18)
141
+/* Transmit Interrupt */
142
+#define REG_MISTA_TXINTR (1 << 16)
143
+/* Receive Bus Error Interrupt */
144
+#define REG_MISTA_RXBERR (1 << 11)
145
+/* Receive Descriptor Unavailable Interrupt */
146
+#define REG_MISTA_RDU (1 << 10)
147
+/* DMA Early Notification Interrupt */
148
+#define REG_MISTA_DENI (1 << 9)
149
+/* Maximum Frame Length Interrupt */
150
+#define REG_MISTA_DFOI (1 << 8)
151
+/* Receive Good Interrupt */
152
+#define REG_MISTA_RXGD (1 << 4)
153
+/* Packet Too Long Interrupt */
154
+#define REG_MISTA_PTLE (1 << 3)
155
+/* Receive Interrupt */
156
+#define REG_MISTA_RXINTR (1 << 0)
157
+
158
+typedef struct NPCM7xxEMCTxDesc NPCM7xxEMCTxDesc;
159
+typedef struct NPCM7xxEMCRxDesc NPCM7xxEMCRxDesc;
160
+
161
+struct NPCM7xxEMCTxDesc {
162
+ uint32_t flags;
163
+ uint32_t txbsa;
164
+ uint32_t status_and_length;
165
+ uint32_t ntxdsa;
166
+};
167
+
168
+struct NPCM7xxEMCRxDesc {
169
+ uint32_t status_and_length;
170
+ uint32_t rxbsa;
171
+ uint32_t reserved;
172
+ uint32_t nrxdsa;
173
+};
174
+
175
+/* NPCM7xxEMCTxDesc.flags values */
176
+/* Owner: 0 = cpu, 1 = emc */
177
+#define TX_DESC_FLAG_OWNER_MASK (1 << 31)
178
+/* Transmit interrupt enable */
179
+#define TX_DESC_FLAG_INTEN (1 << 2)
180
+
181
+/* NPCM7xxEMCTxDesc.status_and_length values */
182
+/* Transmission complete */
183
+#define TX_DESC_STATUS_TXCP (1 << 19)
184
+/* Transmit interrupt */
185
+#define TX_DESC_STATUS_TXINTR (1 << 16)
186
+
187
+/* NPCM7xxEMCRxDesc.status_and_length values */
188
+/* Owner: 0b00 = cpu, 0b10 = emc */
189
+#define RX_DESC_STATUS_OWNER_SHIFT 30
190
+#define RX_DESC_STATUS_OWNER_MASK 0xc0000000
191
+/* Frame Reception Complete */
192
+#define RX_DESC_STATUS_RXGD (1 << 20)
193
+/* Packet too long */
194
+#define RX_DESC_STATUS_PTLE (1 << 19)
195
+/* Receive Interrupt */
196
+#define RX_DESC_STATUS_RXINTR (1 << 16)
197
+
198
+#define RX_DESC_PKT_LEN(word) ((uint32_t) (word) & 0xffff)
199
+
200
+typedef struct EMCModule {
201
+ int rx_irq;
202
+ int tx_irq;
203
+ uint64_t base_addr;
204
+} EMCModule;
205
+
206
+typedef struct TestData {
207
+ const EMCModule *module;
208
+} TestData;
209
+
210
+static const EMCModule emc_module_list[] = {
211
+ {
212
+ .rx_irq = 15,
213
+ .tx_irq = 16,
214
+ .base_addr = 0xf0825000
215
+ },
216
+ {
217
+ .rx_irq = 114,
218
+ .tx_irq = 115,
219
+ .base_addr = 0xf0826000
220
+ }
221
+};
222
+
223
+/* Returns the index of the EMC module. */
224
+static int emc_module_index(const EMCModule *mod)
225
+{
63
+{
226
+ ptrdiff_t diff = mod - emc_module_list;
64
+ intptr_t row, col, oprsz = simd_maxsz(desc);
227
+
65
+ uint32_t neg = simd_data(desc) << 31;
228
+ g_assert_true(diff >= 0 && diff < ARRAY_SIZE(emc_module_list));
66
+ uint16_t *pn = vpn, *pm = vpm;
229
+
67
+ float_status fpst;
230
+ return diff;
231
+}
232
+
233
+static void packet_test_clear(void *sockets)
234
+{
235
+ int *test_sockets = sockets;
236
+
237
+ close(test_sockets[0]);
238
+ g_free(test_sockets);
239
+}
240
+
241
+static int *packet_test_init(int module_num, GString *cmd_line)
242
+{
243
+ int *test_sockets = g_new(int, 2);
244
+ int ret = socketpair(PF_UNIX, SOCK_STREAM, 0, test_sockets);
245
+ g_assert_cmpint(ret, != , -1);
246
+
68
+
247
+ /*
69
+ /*
248
+ * KISS and use -nic. We specify two nics (both emc{0,1}) because there's
70
+ * Make a copy of float_status because this operation does not
249
+ * currently no way to specify only emc1: The driver implicitly relies on
71
+ * update the cumulative fp exception status. It also produces
250
+ * emc[i] == nd_table[i].
72
+ * default nans.
251
+ */
73
+ */
252
+ if (module_num == 0) {
74
+ fpst = *(float_status *)vst;
253
+ g_string_append_printf(cmd_line,
75
+ set_default_nan_mode(true, &fpst);
254
+ " -nic socket,fd=%d,model=" TYPE_NPCM7XX_EMC " "
255
+ " -nic user,model=" TYPE_NPCM7XX_EMC " ",
256
+ test_sockets[1]);
257
+ } else {
258
+ g_string_append_printf(cmd_line,
259
+ " -nic user,model=" TYPE_NPCM7XX_EMC " "
260
+ " -nic socket,fd=%d,model=" TYPE_NPCM7XX_EMC " ",
261
+ test_sockets[1]);
262
+ }
263
+
76
+
264
+ g_test_queue_destroy(packet_test_clear, test_sockets);
77
+ for (row = 0; row < oprsz; ) {
265
+ return test_sockets;
78
+ uint16_t pa = pn[H2(row >> 4)];
266
+}
79
+ do {
80
+ if (pa & 1) {
81
+ void *vza_row = vza + tile_vslice_offset(row);
82
+ uint32_t n = *(uint32_t *)(vzn + H1_4(row)) ^ neg;
267
+
83
+
268
+static uint32_t emc_read(QTestState *qts, const EMCModule *mod,
84
+ for (col = 0; col < oprsz; ) {
269
+ NPCM7xxPWMRegister regno)
85
+ uint16_t pb = pm[H2(col >> 4)];
270
+{
86
+ do {
271
+ return qtest_readl(qts, mod->base_addr + regno * sizeof(uint32_t));
87
+ if (pb & 1) {
272
+}
88
+ uint32_t *a = vza_row + H1_4(col);
273
+
89
+ uint32_t *m = vzm + H1_4(col);
274
+static void emc_write(QTestState *qts, const EMCModule *mod,
90
+ *a = float32_muladd(n, *m, *a, 0, vst);
275
+ NPCM7xxPWMRegister regno, uint32_t value)
91
+ }
276
+{
92
+ col += 4;
277
+ qtest_writel(qts, mod->base_addr + regno * sizeof(uint32_t), value);
93
+ pb >>= 4;
278
+}
94
+ } while (col & 15);
279
+
95
+ }
280
+static void emc_read_tx_desc(QTestState *qts, uint32_t addr,
281
+ NPCM7xxEMCTxDesc *desc)
282
+{
283
+ qtest_memread(qts, addr, desc, sizeof(*desc));
284
+ desc->flags = le32_to_cpu(desc->flags);
285
+ desc->txbsa = le32_to_cpu(desc->txbsa);
286
+ desc->status_and_length = le32_to_cpu(desc->status_and_length);
287
+ desc->ntxdsa = le32_to_cpu(desc->ntxdsa);
288
+}
289
+
290
+static void emc_write_tx_desc(QTestState *qts, const NPCM7xxEMCTxDesc *desc,
291
+ uint32_t addr)
292
+{
293
+ NPCM7xxEMCTxDesc le_desc;
294
+
295
+ le_desc.flags = cpu_to_le32(desc->flags);
296
+ le_desc.txbsa = cpu_to_le32(desc->txbsa);
297
+ le_desc.status_and_length = cpu_to_le32(desc->status_and_length);
298
+ le_desc.ntxdsa = cpu_to_le32(desc->ntxdsa);
299
+ qtest_memwrite(qts, addr, &le_desc, sizeof(le_desc));
300
+}
301
+
302
+static void emc_read_rx_desc(QTestState *qts, uint32_t addr,
303
+ NPCM7xxEMCRxDesc *desc)
304
+{
305
+ qtest_memread(qts, addr, desc, sizeof(*desc));
306
+ desc->status_and_length = le32_to_cpu(desc->status_and_length);
307
+ desc->rxbsa = le32_to_cpu(desc->rxbsa);
308
+ desc->reserved = le32_to_cpu(desc->reserved);
309
+ desc->nrxdsa = le32_to_cpu(desc->nrxdsa);
310
+}
311
+
312
+static void emc_write_rx_desc(QTestState *qts, const NPCM7xxEMCRxDesc *desc,
313
+ uint32_t addr)
314
+{
315
+ NPCM7xxEMCRxDesc le_desc;
316
+
317
+ le_desc.status_and_length = cpu_to_le32(desc->status_and_length);
318
+ le_desc.rxbsa = cpu_to_le32(desc->rxbsa);
319
+ le_desc.reserved = cpu_to_le32(desc->reserved);
320
+ le_desc.nrxdsa = cpu_to_le32(desc->nrxdsa);
321
+ qtest_memwrite(qts, addr, &le_desc, sizeof(le_desc));
322
+}
323
+
324
+/*
325
+ * Reset the EMC module.
326
+ * The module must be reset before, e.g., TXDLSA,RXDLSA are changed.
327
+ */
328
+static bool emc_soft_reset(QTestState *qts, const EMCModule *mod)
329
+{
330
+ uint32_t val;
331
+ uint64_t end_time;
332
+
333
+ emc_write(qts, mod, REG_MCMDR, REG_MCMDR_SWR);
334
+
335
+ /*
336
+ * Wait for device to reset as the linux driver does.
337
+ * During reset the AHB reads 0 for all registers. So first wait for
338
+ * something that resets to non-zero, and then wait for SWR becoming 0.
339
+ */
340
+ end_time = g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
341
+
342
+ do {
343
+ qtest_clock_step(qts, 100);
344
+ val = emc_read(qts, mod, REG_FFTCR);
345
+ } while (val == 0 && g_get_monotonic_time() < end_time);
346
+ if (val != 0) {
347
+ do {
348
+ qtest_clock_step(qts, 100);
349
+ val = emc_read(qts, mod, REG_MCMDR);
350
+ if ((val & REG_MCMDR_SWR) == 0) {
351
+ /*
352
+ * N.B. The CAMs have been reset here, so macaddr matching of
353
+ * incoming packets will not work.
354
+ */
355
+ return true;
356
+ }
96
+ }
357
+ } while (g_get_monotonic_time() < end_time);
97
+ row += 4;
358
+ }
98
+ pa >>= 4;
359
+
99
+ } while (row & 15);
360
+ g_message("%s: Timeout expired", __func__);
361
+ return false;
362
+}
363
+
364
+/* Check emc registers are reset to default value. */
365
+static void test_init(gconstpointer test_data)
366
+{
367
+ const TestData *td = test_data;
368
+ const EMCModule *mod = td->module;
369
+ QTestState *qts = qtest_init("-machine quanta-gsj");
370
+ int i;
371
+
372
+#define CHECK_REG(regno, value) \
373
+ do { \
374
+ g_assert_cmphex(emc_read(qts, mod, (regno)), ==, (value)); \
375
+ } while (0)
376
+
377
+ CHECK_REG(REG_CAMCMR, 0);
378
+ CHECK_REG(REG_CAMEN, 0);
379
+ CHECK_REG(REG_TXDLSA, 0xfffffffc);
380
+ CHECK_REG(REG_RXDLSA, 0xfffffffc);
381
+ CHECK_REG(REG_MCMDR, 0);
382
+ CHECK_REG(REG_MIID, 0);
383
+ CHECK_REG(REG_MIIDA, 0x00900000);
384
+ CHECK_REG(REG_FFTCR, 0x0101);
385
+ CHECK_REG(REG_DMARFC, 0x0800);
386
+ CHECK_REG(REG_MIEN, 0);
387
+ CHECK_REG(REG_MISTA, 0);
388
+ CHECK_REG(REG_MGSTA, 0);
389
+ CHECK_REG(REG_MPCNT, 0x7fff);
390
+ CHECK_REG(REG_MRPC, 0);
391
+ CHECK_REG(REG_MRPCC, 0);
392
+ CHECK_REG(REG_MREPC, 0);
393
+ CHECK_REG(REG_DMARFS, 0);
394
+ CHECK_REG(REG_CTXDSA, 0);
395
+ CHECK_REG(REG_CTXBSA, 0);
396
+ CHECK_REG(REG_CRXDSA, 0);
397
+ CHECK_REG(REG_CRXBSA, 0);
398
+
399
+#undef CHECK_REG
400
+
401
+ for (i = 0; i < NUM_CAMML_REGS; ++i) {
402
+ g_assert_cmpuint(emc_read(qts, mod, REG_CAMM_BASE + i * 2), ==,
403
+ 0);
404
+ g_assert_cmpuint(emc_read(qts, mod, REG_CAML_BASE + i * 2), ==,
405
+ 0);
406
+ }
407
+
408
+ qtest_quit(qts);
409
+}
410
+
411
+static bool emc_wait_irq(QTestState *qts, const EMCModule *mod, int step,
412
+ bool is_tx)
413
+{
414
+ uint64_t end_time =
415
+ g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
416
+
417
+ do {
418
+ if (qtest_get_irq(qts, is_tx ? mod->tx_irq : mod->rx_irq)) {
419
+ return true;
420
+ }
421
+ qtest_clock_step(qts, step);
422
+ } while (g_get_monotonic_time() < end_time);
423
+
424
+ g_message("%s: Timeout expired", __func__);
425
+ return false;
426
+}
427
+
428
+static bool emc_wait_mista(QTestState *qts, const EMCModule *mod, int step,
429
+ uint32_t flag)
430
+{
431
+ uint64_t end_time =
432
+ g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
433
+
434
+ do {
435
+ uint32_t mista = emc_read(qts, mod, REG_MISTA);
436
+ if (mista & flag) {
437
+ return true;
438
+ }
439
+ qtest_clock_step(qts, step);
440
+ } while (g_get_monotonic_time() < end_time);
441
+
442
+ g_message("%s: Timeout expired", __func__);
443
+ return false;
444
+}
445
+
446
+static bool wait_socket_readable(int fd)
447
+{
448
+ fd_set read_fds;
449
+ struct timeval tv;
450
+ int rv;
451
+
452
+ FD_ZERO(&read_fds);
453
+ FD_SET(fd, &read_fds);
454
+ tv.tv_sec = TIMEOUT_SECONDS;
455
+ tv.tv_usec = 0;
456
+ rv = select(fd + 1, &read_fds, NULL, NULL, &tv);
457
+ if (rv == -1) {
458
+ perror("select");
459
+ } else if (rv == 0) {
460
+ g_message("%s: Timeout expired", __func__);
461
+ }
462
+ return rv == 1;
463
+}
464
+
465
+/* Initialize *desc (in host endian format). */
466
+static void init_tx_desc(NPCM7xxEMCTxDesc *desc, size_t count,
467
+ uint32_t desc_addr)
468
+{
469
+ g_assert(count >= 2);
470
+ memset(&desc[0], 0, sizeof(*desc) * count);
471
+ /* Leave the last one alone, owned by the cpu -> stops transmission. */
472
+ for (size_t i = 0; i < count - 1; ++i) {
473
+ desc[i].flags =
474
+ (TX_DESC_FLAG_OWNER_MASK | /* owner = 1: emc */
475
+ TX_DESC_FLAG_INTEN |
476
+ 0 | /* crc append = 0 */
477
+ 0 /* padding enable = 0 */);
478
+ desc[i].status_and_length =
479
+ (0 | /* collision count = 0 */
480
+ 0 | /* SQE = 0 */
481
+ 0 | /* PAU = 0 */
482
+ 0 | /* TXHA = 0 */
483
+ 0 | /* LC = 0 */
484
+ 0 | /* TXABT = 0 */
485
+ 0 | /* NCS = 0 */
486
+ 0 | /* EXDEF = 0 */
487
+ 0 | /* TXCP = 0 */
488
+ 0 | /* DEF = 0 */
489
+ 0 | /* TXINTR = 0 */
490
+ 0 /* length filled in later */);
491
+ desc[i].ntxdsa = desc_addr + (i + 1) * sizeof(*desc);
492
+ }
100
+ }
493
+}
101
+}
494
+
102
+
495
+static void enable_tx(QTestState *qts, const EMCModule *mod,
103
+void HELPER(sme_fmopa_d)(void *vza, void *vzn, void *vzm, void *vpn,
496
+ const NPCM7xxEMCTxDesc *desc, size_t count,
104
+ void *vpm, void *vst, uint32_t desc)
497
+ uint32_t desc_addr, uint32_t mien_flags)
498
+{
105
+{
499
+ /* Write the descriptors to guest memory. */
106
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 8;
500
+ for (size_t i = 0; i < count; ++i) {
107
+ uint64_t neg = (uint64_t)simd_data(desc) << 63;
501
+ emc_write_tx_desc(qts, desc + i, desc_addr + i * sizeof(*desc));
108
+ uint64_t *za = vza, *zn = vzn, *zm = vzm;
109
+ uint8_t *pn = vpn, *pm = vpm;
110
+ float_status fpst = *(float_status *)vst;
111
+
112
+ set_default_nan_mode(true, &fpst);
113
+
114
+ for (row = 0; row < oprsz; ++row) {
115
+ if (pn[H1(row)] & 1) {
116
+ uint64_t *za_row = &za[tile_vslice_index(row)];
117
+ uint64_t n = zn[row] ^ neg;
118
+
119
+ for (col = 0; col < oprsz; ++col) {
120
+ if (pm[H1(col)] & 1) {
121
+ uint64_t *a = &za_row[col];
122
+ *a = float64_muladd(n, zm[col], *a, 0, &fpst);
123
+ }
124
+ }
125
+ }
126
+ }
127
+}
128
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
129
index XXXXXXX..XXXXXXX 100644
130
--- a/target/arm/translate-sme.c
131
+++ b/target/arm/translate-sme.c
132
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(ADDHA_s, aa64_sme, do_adda, a, MO_32, gen_helper_sme_addha_s)
133
TRANS_FEAT(ADDVA_s, aa64_sme, do_adda, a, MO_32, gen_helper_sme_addva_s)
134
TRANS_FEAT(ADDHA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addha_d)
135
TRANS_FEAT(ADDVA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addva_d)
136
+
137
+static bool do_outprod_fpst(DisasContext *s, arg_op *a, MemOp esz,
138
+ gen_helper_gvec_5_ptr *fn)
139
+{
140
+ int svl = streaming_vec_reg_size(s);
141
+ uint32_t desc = simd_desc(svl, svl, a->sub);
142
+ TCGv_ptr za, zn, zm, pn, pm, fpst;
143
+
144
+ if (!sme_smza_enabled_check(s)) {
145
+ return true;
502
+ }
146
+ }
503
+
147
+
504
+ /* Trigger sending the packet. */
148
+ /* Sum XZR+zad to find ZAd. */
505
+ /* The module must be reset before changing TXDLSA. */
149
+ za = get_tile_rowcol(s, esz, 31, a->zad, false);
506
+ g_assert(emc_soft_reset(qts, mod));
150
+ zn = vec_full_reg_ptr(s, a->zn);
507
+ emc_write(qts, mod, REG_TXDLSA, desc_addr);
151
+ zm = vec_full_reg_ptr(s, a->zm);
508
+ emc_write(qts, mod, REG_CTXDSA, ~0);
152
+ pn = pred_full_reg_ptr(s, a->pn);
509
+ emc_write(qts, mod, REG_MIEN, REG_MIEN_ENTXCP | mien_flags);
153
+ pm = pred_full_reg_ptr(s, a->pm);
510
+ {
154
+ fpst = fpstatus_ptr(FPST_FPCR);
511
+ uint32_t mcmdr = emc_read(qts, mod, REG_MCMDR);
512
+ mcmdr |= REG_MCMDR_TXON;
513
+ emc_write(qts, mod, REG_MCMDR, mcmdr);
514
+ }
515
+
155
+
516
+ /* Prod the device to send the packet. */
156
+ fn(za, zn, zm, pn, pm, fpst, tcg_constant_i32(desc));
517
+ emc_write(qts, mod, REG_TSDR, 1);
157
+
158
+ tcg_temp_free_ptr(za);
159
+ tcg_temp_free_ptr(zn);
160
+ tcg_temp_free_ptr(pn);
161
+ tcg_temp_free_ptr(pm);
162
+ tcg_temp_free_ptr(fpst);
163
+ return true;
518
+}
164
+}
519
+
165
+
520
+static void emc_send_verify1(QTestState *qts, const EMCModule *mod, int fd,
166
+TRANS_FEAT(FMOPA_s, aa64_sme, do_outprod_fpst, a, MO_32, gen_helper_sme_fmopa_s)
521
+ bool with_irq, uint32_t desc_addr,
167
+TRANS_FEAT(FMOPA_d, aa64_sme_f64f64, do_outprod_fpst, a, MO_64, gen_helper_sme_fmopa_d)
522
+ uint32_t next_desc_addr,
523
+ const char *test_data, int test_size)
524
+{
525
+ NPCM7xxEMCTxDesc result_desc;
526
+ uint32_t expected_mask, expected_value, recv_len;
527
+ int ret;
528
+ char buffer[TX_DATA_LEN];
529
+
530
+ g_assert(wait_socket_readable(fd));
531
+
532
+ /* Read the descriptor back. */
533
+ emc_read_tx_desc(qts, desc_addr, &result_desc);
534
+ /* Descriptor should be owned by cpu now. */
535
+ g_assert((result_desc.flags & TX_DESC_FLAG_OWNER_MASK) == 0);
536
+ /* Test the status bits, ignoring the length field. */
537
+ expected_mask = 0xffff << 16;
538
+ expected_value = TX_DESC_STATUS_TXCP;
539
+ if (with_irq) {
540
+ expected_value |= TX_DESC_STATUS_TXINTR;
541
+ }
542
+ g_assert_cmphex((result_desc.status_and_length & expected_mask), ==,
543
+ expected_value);
544
+
545
+ /* Check data sent to the backend. */
546
+ recv_len = ~0;
547
+ ret = qemu_recv(fd, &recv_len, sizeof(recv_len), MSG_DONTWAIT);
548
+ g_assert_cmpint(ret, == , sizeof(recv_len));
549
+
550
+ g_assert(wait_socket_readable(fd));
551
+ memset(buffer, 0xff, sizeof(buffer));
552
+ ret = qemu_recv(fd, buffer, test_size, MSG_DONTWAIT);
553
+ g_assert_cmpmem(buffer, ret, test_data, test_size);
554
+}
555
+
556
+static void emc_send_verify(QTestState *qts, const EMCModule *mod, int fd,
557
+ bool with_irq)
558
+{
559
+ NPCM7xxEMCTxDesc desc[NUM_TX_DESCRIPTORS];
560
+ uint32_t desc_addr = DESC_ADDR;
561
+ static const char test1_data[] = "TEST1";
562
+ static const char test2_data[] = "Testing 1 2 3 ...";
563
+ uint32_t data1_addr = DATA_ADDR;
564
+ uint32_t data2_addr = data1_addr + sizeof(test1_data);
565
+ bool got_tdu;
566
+ uint32_t end_desc_addr;
567
+
568
+ /* Prepare test data buffer. */
569
+ qtest_memwrite(qts, data1_addr, test1_data, sizeof(test1_data));
570
+ qtest_memwrite(qts, data2_addr, test2_data, sizeof(test2_data));
571
+
572
+ init_tx_desc(&desc[0], NUM_TX_DESCRIPTORS, desc_addr);
573
+ desc[0].txbsa = data1_addr;
574
+ desc[0].status_and_length |= sizeof(test1_data);
575
+ desc[1].txbsa = data2_addr;
576
+ desc[1].status_and_length |= sizeof(test2_data);
577
+
578
+ enable_tx(qts, mod, &desc[0], NUM_TX_DESCRIPTORS, desc_addr,
579
+ with_irq ? REG_MIEN_ENTXINTR : 0);
580
+
581
+ /*
582
+ * It's problematic to observe the interrupt for each packet.
583
+ * Instead just wait until all the packets go out.
584
+ */
585
+ got_tdu = false;
586
+ while (!got_tdu) {
587
+ if (with_irq) {
588
+ g_assert_true(emc_wait_irq(qts, mod, TX_STEP_COUNT,
589
+ /*is_tx=*/true));
590
+ } else {
591
+ g_assert_true(emc_wait_mista(qts, mod, TX_STEP_COUNT,
592
+ REG_MISTA_TXINTR));
593
+ }
594
+ got_tdu = !!(emc_read(qts, mod, REG_MISTA) & REG_MISTA_TDU);
595
+ /* If we don't have TDU yet, reset the interrupt. */
596
+ if (!got_tdu) {
597
+ emc_write(qts, mod, REG_MISTA,
598
+ emc_read(qts, mod, REG_MISTA) & 0xffff0000);
599
+ }
600
+ }
601
+
602
+ end_desc_addr = desc_addr + 2 * sizeof(desc[0]);
603
+ g_assert_cmphex(emc_read(qts, mod, REG_CTXDSA), ==, end_desc_addr);
604
+ g_assert_cmphex(emc_read(qts, mod, REG_MISTA), ==,
605
+ REG_MISTA_TXCP | REG_MISTA_TXINTR | REG_MISTA_TDU);
606
+
607
+ emc_send_verify1(qts, mod, fd, with_irq,
608
+ desc_addr, end_desc_addr,
609
+ test1_data, sizeof(test1_data));
610
+ emc_send_verify1(qts, mod, fd, with_irq,
611
+ desc_addr + sizeof(desc[0]), end_desc_addr,
612
+ test2_data, sizeof(test2_data));
613
+}
614
+
615
+/* Initialize *desc (in host endian format). */
616
+static void init_rx_desc(NPCM7xxEMCRxDesc *desc, size_t count,
617
+ uint32_t desc_addr, uint32_t data_addr)
618
+{
619
+ g_assert_true(count >= 2);
620
+ memset(desc, 0, sizeof(*desc) * count);
621
+ desc[0].rxbsa = data_addr;
622
+ desc[0].status_and_length =
623
+ (0b10 << RX_DESC_STATUS_OWNER_SHIFT | /* owner = 10: emc */
624
+ 0 | /* RP = 0 */
625
+ 0 | /* ALIE = 0 */
626
+ 0 | /* RXGD = 0 */
627
+ 0 | /* PTLE = 0 */
628
+ 0 | /* CRCE = 0 */
629
+ 0 | /* RXINTR = 0 */
630
+ 0 /* length (filled in later) */);
631
+ /* Leave the last one alone, owned by the cpu -> stops transmission. */
632
+ desc[0].nrxdsa = desc_addr + sizeof(*desc);
633
+}
634
+
635
+static void enable_rx(QTestState *qts, const EMCModule *mod,
636
+ const NPCM7xxEMCRxDesc *desc, size_t count,
637
+ uint32_t desc_addr, uint32_t mien_flags,
638
+ uint32_t mcmdr_flags)
639
+{
640
+ /*
641
+ * Write the descriptor to guest memory.
642
+ * FWIW, IWBN if the docs said the buffer needs to be at least DMARFC
643
+ * bytes.
644
+ */
645
+ for (size_t i = 0; i < count; ++i) {
646
+ emc_write_rx_desc(qts, desc + i, desc_addr + i * sizeof(*desc));
647
+ }
648
+
649
+ /* Trigger receiving the packet. */
650
+ /* The module must be reset before changing RXDLSA. */
651
+ g_assert(emc_soft_reset(qts, mod));
652
+ emc_write(qts, mod, REG_RXDLSA, desc_addr);
653
+ emc_write(qts, mod, REG_MIEN, REG_MIEN_ENRXGD | mien_flags);
654
+
655
+ /*
656
+ * We don't know what the device's macaddr is, so just accept all
657
+ * unicast packets (AUP).
658
+ */
659
+ emc_write(qts, mod, REG_CAMCMR, REG_CAMCMR_AUP);
660
+ emc_write(qts, mod, REG_CAMEN, 1 << 0);
661
+ {
662
+ uint32_t mcmdr = emc_read(qts, mod, REG_MCMDR);
663
+ mcmdr |= REG_MCMDR_RXON | mcmdr_flags;
664
+ emc_write(qts, mod, REG_MCMDR, mcmdr);
665
+ }
666
+
667
+ /* Prod the device to accept a packet. */
668
+ emc_write(qts, mod, REG_RSDR, 1);
669
+}
670
+
671
+static void emc_recv_verify(QTestState *qts, const EMCModule *mod, int fd,
672
+ bool with_irq)
673
+{
674
+ NPCM7xxEMCRxDesc desc[NUM_RX_DESCRIPTORS];
675
+ uint32_t desc_addr = DESC_ADDR;
676
+ uint32_t data_addr = DATA_ADDR;
677
+ int ret;
678
+ uint32_t expected_mask, expected_value;
679
+ NPCM7xxEMCRxDesc result_desc;
680
+
681
+ /* Prepare test data buffer. */
682
+ const char test[RX_DATA_LEN] = "TEST";
683
+ int len = htonl(sizeof(test));
684
+ const struct iovec iov[] = {
685
+ {
686
+ .iov_base = &len,
687
+ .iov_len = sizeof(len),
688
+ },{
689
+ .iov_base = (char *) test,
690
+ .iov_len = sizeof(test),
691
+ },
692
+ };
693
+
694
+ /*
695
+ * Reset the device BEFORE sending a test packet, otherwise the packet
696
+ * may get swallowed by an active device of an earlier test.
697
+ */
698
+ init_rx_desc(&desc[0], NUM_RX_DESCRIPTORS, desc_addr, data_addr);
699
+ enable_rx(qts, mod, &desc[0], NUM_RX_DESCRIPTORS, desc_addr,
700
+ with_irq ? REG_MIEN_ENRXINTR : 0, 0);
701
+
702
+ /* Send test packet to device's socket. */
703
+ ret = iov_send(fd, iov, 2, 0, sizeof(len) + sizeof(test));
704
+ g_assert_cmpint(ret, == , sizeof(test) + sizeof(len));
705
+
706
+ /* Wait for RX interrupt. */
707
+ if (with_irq) {
708
+ g_assert_true(emc_wait_irq(qts, mod, RX_STEP_COUNT, /*is_tx=*/false));
709
+ } else {
710
+ g_assert_true(emc_wait_mista(qts, mod, RX_STEP_COUNT, REG_MISTA_RXGD));
711
+ }
712
+
713
+ g_assert_cmphex(emc_read(qts, mod, REG_CRXDSA), ==,
714
+ desc_addr + sizeof(desc[0]));
715
+
716
+ expected_mask = 0xffff;
717
+ expected_value = (REG_MISTA_DENI |
718
+ REG_MISTA_RXGD |
719
+ REG_MISTA_RXINTR);
720
+ g_assert_cmphex((emc_read(qts, mod, REG_MISTA) & expected_mask),
721
+ ==, expected_value);
722
+
723
+ /* Read the descriptor back. */
724
+ emc_read_rx_desc(qts, desc_addr, &result_desc);
725
+ /* Descriptor should be owned by cpu now. */
726
+ g_assert((result_desc.status_and_length & RX_DESC_STATUS_OWNER_MASK) == 0);
727
+ /* Test the status bits, ignoring the length field. */
728
+ expected_mask = 0xffff << 16;
729
+ expected_value = RX_DESC_STATUS_RXGD;
730
+ if (with_irq) {
731
+ expected_value |= RX_DESC_STATUS_RXINTR;
732
+ }
733
+ g_assert_cmphex((result_desc.status_and_length & expected_mask), ==,
734
+ expected_value);
735
+ g_assert_cmpint(RX_DESC_PKT_LEN(result_desc.status_and_length), ==,
736
+ RX_DATA_LEN + CRC_LENGTH);
737
+
738
+ {
739
+ char buffer[RX_DATA_LEN];
740
+ qtest_memread(qts, data_addr, buffer, sizeof(buffer));
741
+ g_assert_cmpstr(buffer, == , "TEST");
742
+ }
743
+}
744
+
745
+static void emc_test_ptle(QTestState *qts, const EMCModule *mod, int fd)
746
+{
747
+ NPCM7xxEMCRxDesc desc[NUM_RX_DESCRIPTORS];
748
+ uint32_t desc_addr = DESC_ADDR;
749
+ uint32_t data_addr = DATA_ADDR;
750
+ int ret;
751
+ NPCM7xxEMCRxDesc result_desc;
752
+ uint32_t expected_mask, expected_value;
753
+
754
+ /* Prepare test data buffer. */
755
+#define PTLE_DATA_LEN 1600
756
+ char test_data[PTLE_DATA_LEN];
757
+ int len = htonl(sizeof(test_data));
758
+ const struct iovec iov[] = {
759
+ {
760
+ .iov_base = &len,
761
+ .iov_len = sizeof(len),
762
+ },{
763
+ .iov_base = (char *) test_data,
764
+ .iov_len = sizeof(test_data),
765
+ },
766
+ };
767
+ memset(test_data, 42, sizeof(test_data));
768
+
769
+ /*
770
+ * Reset the device BEFORE sending a test packet, otherwise the packet
771
+ * may get swallowed by an active device of an earlier test.
772
+ */
773
+ init_rx_desc(&desc[0], NUM_RX_DESCRIPTORS, desc_addr, data_addr);
774
+ enable_rx(qts, mod, &desc[0], NUM_RX_DESCRIPTORS, desc_addr,
775
+ REG_MIEN_ENRXINTR, REG_MCMDR_ALP);
776
+
777
+ /* Send test packet to device's socket. */
778
+ ret = iov_send(fd, iov, 2, 0, sizeof(len) + sizeof(test_data));
779
+ g_assert_cmpint(ret, == , sizeof(test_data) + sizeof(len));
780
+
781
+ /* Wait for RX interrupt. */
782
+ g_assert_true(emc_wait_irq(qts, mod, RX_STEP_COUNT, /*is_tx=*/false));
783
+
784
+ /* Read the descriptor back. */
785
+ emc_read_rx_desc(qts, desc_addr, &result_desc);
786
+ /* Descriptor should be owned by cpu now. */
787
+ g_assert((result_desc.status_and_length & RX_DESC_STATUS_OWNER_MASK) == 0);
788
+ /* Test the status bits, ignoring the length field. */
789
+ expected_mask = 0xffff << 16;
790
+ expected_value = (RX_DESC_STATUS_RXGD |
791
+ RX_DESC_STATUS_PTLE |
792
+ RX_DESC_STATUS_RXINTR);
793
+ g_assert_cmphex((result_desc.status_and_length & expected_mask), ==,
794
+ expected_value);
795
+ g_assert_cmpint(RX_DESC_PKT_LEN(result_desc.status_and_length), ==,
796
+ PTLE_DATA_LEN + CRC_LENGTH);
797
+
798
+ {
799
+ char buffer[PTLE_DATA_LEN];
800
+ qtest_memread(qts, data_addr, buffer, sizeof(buffer));
801
+ g_assert(memcmp(buffer, test_data, PTLE_DATA_LEN) == 0);
802
+ }
803
+}
804
+
805
+static void test_tx(gconstpointer test_data)
806
+{
807
+ const TestData *td = test_data;
808
+ GString *cmd_line = g_string_new("-machine quanta-gsj");
809
+ int *test_sockets = packet_test_init(emc_module_index(td->module),
810
+ cmd_line);
811
+ QTestState *qts = qtest_init(cmd_line->str);
812
+
813
+ /*
814
+ * TODO: For pedantic correctness test_sockets[0] should be closed after
815
+ * the fork and before the exec, but that will require some harness
816
+ * improvements.
817
+ */
818
+ close(test_sockets[1]);
819
+ /* Defensive programming */
820
+ test_sockets[1] = -1;
821
+
822
+ qtest_irq_intercept_in(qts, "/machine/soc/a9mpcore/gic");
823
+
824
+ emc_send_verify(qts, td->module, test_sockets[0], /*with_irq=*/false);
825
+ emc_send_verify(qts, td->module, test_sockets[0], /*with_irq=*/true);
826
+
827
+ qtest_quit(qts);
828
+}
829
+
830
+static void test_rx(gconstpointer test_data)
831
+{
832
+ const TestData *td = test_data;
833
+ GString *cmd_line = g_string_new("-machine quanta-gsj");
834
+ int *test_sockets = packet_test_init(emc_module_index(td->module),
835
+ cmd_line);
836
+ QTestState *qts = qtest_init(cmd_line->str);
837
+
838
+ /*
839
+ * TODO: For pedantic correctness test_sockets[0] should be closed after
840
+ * the fork and before the exec, but that will require some harness
841
+ * improvements.
842
+ */
843
+ close(test_sockets[1]);
844
+ /* Defensive programming */
845
+ test_sockets[1] = -1;
846
+
847
+ qtest_irq_intercept_in(qts, "/machine/soc/a9mpcore/gic");
848
+
849
+ emc_recv_verify(qts, td->module, test_sockets[0], /*with_irq=*/false);
850
+ emc_recv_verify(qts, td->module, test_sockets[0], /*with_irq=*/true);
851
+ emc_test_ptle(qts, td->module, test_sockets[0]);
852
+
853
+ qtest_quit(qts);
854
+}
855
+
856
+static void emc_add_test(const char *name, const TestData* td,
857
+ GTestDataFunc fn)
858
+{
859
+ g_autofree char *full_name = g_strdup_printf(
860
+ "npcm7xx_emc/emc[%d]/%s", emc_module_index(td->module), name);
861
+ qtest_add_data_func(full_name, td, fn);
862
+}
863
+#define add_test(name, td) emc_add_test(#name, td, test_##name)
864
+
865
+int main(int argc, char **argv)
866
+{
867
+ TestData test_data_list[ARRAY_SIZE(emc_module_list)];
868
+
869
+ g_test_init(&argc, &argv, NULL);
870
+
871
+ for (int i = 0; i < ARRAY_SIZE(emc_module_list); ++i) {
872
+ TestData *td = &test_data_list[i];
873
+
874
+ td->module = &emc_module_list[i];
875
+
876
+ add_test(init, td);
877
+ add_test(tx, td);
878
+ add_test(rx, td);
879
+ }
880
+
881
+ return g_test_run();
882
+}
883
diff --git a/tests/qtest/meson.build b/tests/qtest/meson.build
884
index XXXXXXX..XXXXXXX 100644
885
--- a/tests/qtest/meson.build
886
+++ b/tests/qtest/meson.build
887
@@ -XXX,XX +XXX,XX @@ qtests_sparc64 = \
888
889
qtests_npcm7xx = \
890
['npcm7xx_adc-test',
891
+ 'npcm7xx_emc-test',
892
'npcm7xx_gpio-test',
893
'npcm7xx_pwm-test',
894
'npcm7xx_rng-test',
895
--
168
--
896
2.20.1
169
2.25.1
897
898
diff view generated by jsdifflib
New patch
1
From: Richard Henderson <richard.henderson@linaro.org>
1
2
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20220708151540.18136-26-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-sme.h | 2 ++
9
target/arm/sme.decode | 2 ++
10
target/arm/sme_helper.c | 56 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-sme.c | 30 ++++++++++++++++++++
12
4 files changed, 90 insertions(+)
13
14
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-sme.h
17
+++ b/target/arm/helper-sme.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_7(sme_fmopa_s, TCG_CALL_NO_RWG,
19
void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
20
DEF_HELPER_FLAGS_7(sme_fmopa_d, TCG_CALL_NO_RWG,
21
void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
22
+DEF_HELPER_FLAGS_6(sme_bfmopa, TCG_CALL_NO_RWG,
23
+ void, ptr, ptr, ptr, ptr, ptr, i32)
24
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
25
index XXXXXXX..XXXXXXX 100644
26
--- a/target/arm/sme.decode
27
+++ b/target/arm/sme.decode
28
@@ -XXX,XX +XXX,XX @@ ADDVA_d 11000000 11 01000 1 ... ... ..... 00 ... @adda_64
29
30
FMOPA_s 10000000 100 ..... ... ... ..... . 00 .. @op_32
31
FMOPA_d 10000000 110 ..... ... ... ..... . 0 ... @op_64
32
+
33
+BFMOPA 10000001 100 ..... ... ... ..... . 00 .. @op_32
34
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
35
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/sme_helper.c
37
+++ b/target/arm/sme_helper.c
38
@@ -XXX,XX +XXX,XX @@ void HELPER(sme_fmopa_d)(void *vza, void *vzn, void *vzm, void *vpn,
39
}
40
}
41
}
42
+
43
+/*
44
+ * Alter PAIR as needed for controlling predicates being false,
45
+ * and for NEG on an enabled row element.
46
+ */
47
+static inline uint32_t f16mop_adj_pair(uint32_t pair, uint32_t pg, uint32_t neg)
48
+{
49
+ /*
50
+ * The pseudocode uses a conditional negate after the conditional zero.
51
+ * It is simpler here to unconditionally negate before conditional zero.
52
+ */
53
+ pair ^= neg;
54
+ if (!(pg & 1)) {
55
+ pair &= 0xffff0000u;
56
+ }
57
+ if (!(pg & 4)) {
58
+ pair &= 0x0000ffffu;
59
+ }
60
+ return pair;
61
+}
62
+
63
+void HELPER(sme_bfmopa)(void *vza, void *vzn, void *vzm, void *vpn,
64
+ void *vpm, uint32_t desc)
65
+{
66
+ intptr_t row, col, oprsz = simd_maxsz(desc);
67
+ uint32_t neg = simd_data(desc) * 0x80008000u;
68
+ uint16_t *pn = vpn, *pm = vpm;
69
+
70
+ for (row = 0; row < oprsz; ) {
71
+ uint16_t prow = pn[H2(row >> 4)];
72
+ do {
73
+ void *vza_row = vza + tile_vslice_offset(row);
74
+ uint32_t n = *(uint32_t *)(vzn + H1_4(row));
75
+
76
+ n = f16mop_adj_pair(n, prow, neg);
77
+
78
+ for (col = 0; col < oprsz; ) {
79
+ uint16_t pcol = pm[H2(col >> 4)];
80
+ do {
81
+ if (prow & pcol & 0b0101) {
82
+ uint32_t *a = vza_row + H1_4(col);
83
+ uint32_t m = *(uint32_t *)(vzm + H1_4(col));
84
+
85
+ m = f16mop_adj_pair(m, pcol, 0);
86
+ *a = bfdotadd(*a, n, m);
87
+
88
+ col += 4;
89
+ pcol >>= 4;
90
+ }
91
+ } while (col & 15);
92
+ }
93
+ row += 4;
94
+ prow >>= 4;
95
+ } while (row & 15);
96
+ }
97
+}
98
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
99
index XXXXXXX..XXXXXXX 100644
100
--- a/target/arm/translate-sme.c
101
+++ b/target/arm/translate-sme.c
102
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(ADDVA_s, aa64_sme, do_adda, a, MO_32, gen_helper_sme_addva_s)
103
TRANS_FEAT(ADDHA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addha_d)
104
TRANS_FEAT(ADDVA_d, aa64_sme_i16i64, do_adda, a, MO_64, gen_helper_sme_addva_d)
105
106
+static bool do_outprod(DisasContext *s, arg_op *a, MemOp esz,
107
+ gen_helper_gvec_5 *fn)
108
+{
109
+ int svl = streaming_vec_reg_size(s);
110
+ uint32_t desc = simd_desc(svl, svl, a->sub);
111
+ TCGv_ptr za, zn, zm, pn, pm;
112
+
113
+ if (!sme_smza_enabled_check(s)) {
114
+ return true;
115
+ }
116
+
117
+ /* Sum XZR+zad to find ZAd. */
118
+ za = get_tile_rowcol(s, esz, 31, a->zad, false);
119
+ zn = vec_full_reg_ptr(s, a->zn);
120
+ zm = vec_full_reg_ptr(s, a->zm);
121
+ pn = pred_full_reg_ptr(s, a->pn);
122
+ pm = pred_full_reg_ptr(s, a->pm);
123
+
124
+ fn(za, zn, zm, pn, pm, tcg_constant_i32(desc));
125
+
126
+ tcg_temp_free_ptr(za);
127
+ tcg_temp_free_ptr(zn);
128
+ tcg_temp_free_ptr(pn);
129
+ tcg_temp_free_ptr(pm);
130
+ return true;
131
+}
132
+
133
static bool do_outprod_fpst(DisasContext *s, arg_op *a, MemOp esz,
134
gen_helper_gvec_5_ptr *fn)
135
{
136
@@ -XXX,XX +XXX,XX @@ static bool do_outprod_fpst(DisasContext *s, arg_op *a, MemOp esz,
137
138
TRANS_FEAT(FMOPA_s, aa64_sme, do_outprod_fpst, a, MO_32, gen_helper_sme_fmopa_s)
139
TRANS_FEAT(FMOPA_d, aa64_sme_f64f64, do_outprod_fpst, a, MO_64, gen_helper_sme_fmopa_d)
140
+
141
+/* TODO: FEAT_EBF16 */
142
+TRANS_FEAT(BFMOPA, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_bfmopa)
143
--
144
2.25.1
diff view generated by jsdifflib
1
From: Hao Wu <wuhaotsh@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Add I2C temperature sensors for NPCM750 eval board.
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Message-id: 20220708151540.18136-27-richard.henderson@linaro.org
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-sme.h | 2 ++
9
target/arm/sme.decode | 1 +
10
target/arm/sme_helper.c | 74 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-sme.c | 1 +
12
4 files changed, 78 insertions(+)
4
13
5
Reviewed-by: Doug Evans<dje@google.com>
14
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
6
Reviewed-by: Tyrong Ting<kfting@nuvoton.com>
7
Signed-off-by: Hao Wu <wuhaotsh@google.com>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Message-id: 20210210220426.3577804-3-wuhaotsh@google.com
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
hw/arm/npcm7xx_boards.c | 19 +++++++++++++++++++
13
1 file changed, 19 insertions(+)
14
15
diff --git a/hw/arm/npcm7xx_boards.c b/hw/arm/npcm7xx_boards.c
16
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
17
--- a/hw/arm/npcm7xx_boards.c
16
--- a/target/arm/helper-sme.h
18
+++ b/hw/arm/npcm7xx_boards.c
17
+++ b/target/arm/helper-sme.h
19
@@ -XXX,XX +XXX,XX @@ static NPCM7xxState *npcm7xx_create_soc(MachineState *machine,
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_5(sme_addva_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
20
return NPCM7XX(obj);
19
DEF_HELPER_FLAGS_5(sme_addha_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
20
DEF_HELPER_FLAGS_5(sme_addva_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, ptr, i32)
21
22
+DEF_HELPER_FLAGS_7(sme_fmopa_h, TCG_CALL_NO_RWG,
23
+ void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
24
DEF_HELPER_FLAGS_7(sme_fmopa_s, TCG_CALL_NO_RWG,
25
void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
26
DEF_HELPER_FLAGS_7(sme_fmopa_d, TCG_CALL_NO_RWG,
27
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
28
index XXXXXXX..XXXXXXX 100644
29
--- a/target/arm/sme.decode
30
+++ b/target/arm/sme.decode
31
@@ -XXX,XX +XXX,XX @@ FMOPA_s 10000000 100 ..... ... ... ..... . 00 .. @op_32
32
FMOPA_d 10000000 110 ..... ... ... ..... . 0 ... @op_64
33
34
BFMOPA 10000001 100 ..... ... ... ..... . 00 .. @op_32
35
+FMOPA_h 10000001 101 ..... ... ... ..... . 00 .. @op_32
36
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/sme_helper.c
39
+++ b/target/arm/sme_helper.c
40
@@ -XXX,XX +XXX,XX @@ static inline uint32_t f16mop_adj_pair(uint32_t pair, uint32_t pg, uint32_t neg)
41
return pair;
21
}
42
}
22
43
23
+static I2CBus *npcm7xx_i2c_get_bus(NPCM7xxState *soc, uint32_t num)
44
+static float32 f16_dotadd(float32 sum, uint32_t e1, uint32_t e2,
45
+ float_status *s_std, float_status *s_odd)
24
+{
46
+{
25
+ g_assert(num < ARRAY_SIZE(soc->smbus));
47
+ float64 e1r = float16_to_float64(e1 & 0xffff, true, s_std);
26
+ return I2C_BUS(qdev_get_child_bus(DEVICE(&soc->smbus[num]), "i2c-bus"));
48
+ float64 e1c = float16_to_float64(e1 >> 16, true, s_std);
49
+ float64 e2r = float16_to_float64(e2 & 0xffff, true, s_std);
50
+ float64 e2c = float16_to_float64(e2 >> 16, true, s_std);
51
+ float64 t64;
52
+ float32 t32;
53
+
54
+ /*
55
+ * The ARM pseudocode function FPDot performs both multiplies
56
+ * and the add with a single rounding operation. Emulate this
57
+ * by performing the first multiply in round-to-odd, then doing
58
+ * the second multiply as fused multiply-add, and rounding to
59
+ * float32 all in one step.
60
+ */
61
+ t64 = float64_mul(e1r, e2r, s_odd);
62
+ t64 = float64r32_muladd(e1c, e2c, t64, 0, s_std);
63
+
64
+ /* This conversion is exact, because we've already rounded. */
65
+ t32 = float64_to_float32(t64, s_std);
66
+
67
+ /* The final accumulation step is not fused. */
68
+ return float32_add(sum, t32, s_std);
27
+}
69
+}
28
+
70
+
29
+static void npcm750_evb_i2c_init(NPCM7xxState *soc)
71
+void HELPER(sme_fmopa_h)(void *vza, void *vzn, void *vzm, void *vpn,
72
+ void *vpm, void *vst, uint32_t desc)
30
+{
73
+{
31
+ /* lm75 temperature sensor on SVB, tmp105 is compatible */
74
+ intptr_t row, col, oprsz = simd_maxsz(desc);
32
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 0), "tmp105", 0x48);
75
+ uint32_t neg = simd_data(desc) * 0x80008000u;
33
+ /* lm75 temperature sensor on EB, tmp105 is compatible */
76
+ uint16_t *pn = vpn, *pm = vpm;
34
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 1), "tmp105", 0x48);
77
+ float_status fpst_odd, fpst_std;
35
+ /* tmp100 temperature sensor on EB, tmp105 is compatible */
78
+
36
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 2), "tmp105", 0x48);
79
+ /*
37
+ /* tmp100 temperature sensor on SVB, tmp105 is compatible */
80
+ * Make a copy of float_status because this operation does not
38
+ i2c_slave_create_simple(npcm7xx_i2c_get_bus(soc, 6), "tmp105", 0x48);
81
+ * update the cumulative fp exception status. It also produces
82
+ * default nans. Make a second copy with round-to-odd -- see above.
83
+ */
84
+ fpst_std = *(float_status *)vst;
85
+ set_default_nan_mode(true, &fpst_std);
86
+ fpst_odd = fpst_std;
87
+ set_float_rounding_mode(float_round_to_odd, &fpst_odd);
88
+
89
+ for (row = 0; row < oprsz; ) {
90
+ uint16_t prow = pn[H2(row >> 4)];
91
+ do {
92
+ void *vza_row = vza + tile_vslice_offset(row);
93
+ uint32_t n = *(uint32_t *)(vzn + H1_4(row));
94
+
95
+ n = f16mop_adj_pair(n, prow, neg);
96
+
97
+ for (col = 0; col < oprsz; ) {
98
+ uint16_t pcol = pm[H2(col >> 4)];
99
+ do {
100
+ if (prow & pcol & 0b0101) {
101
+ uint32_t *a = vza_row + H1_4(col);
102
+ uint32_t m = *(uint32_t *)(vzm + H1_4(col));
103
+
104
+ m = f16mop_adj_pair(m, pcol, 0);
105
+ *a = f16_dotadd(*a, n, m, &fpst_std, &fpst_odd);
106
+
107
+ col += 4;
108
+ pcol >>= 4;
109
+ }
110
+ } while (col & 15);
111
+ }
112
+ row += 4;
113
+ prow >>= 4;
114
+ } while (row & 15);
115
+ }
39
+}
116
+}
40
+
117
+
41
static void npcm750_evb_init(MachineState *machine)
118
void HELPER(sme_bfmopa)(void *vza, void *vzn, void *vzm, void *vpn,
119
void *vpm, uint32_t desc)
42
{
120
{
43
NPCM7xxState *soc;
121
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
44
@@ -XXX,XX +XXX,XX @@ static void npcm750_evb_init(MachineState *machine)
122
index XXXXXXX..XXXXXXX 100644
45
123
--- a/target/arm/translate-sme.c
46
npcm7xx_load_bootrom(machine, soc);
124
+++ b/target/arm/translate-sme.c
47
npcm7xx_connect_flash(&soc->fiu[0], 0, "w25q256", drive_get(IF_MTD, 0, 0));
125
@@ -XXX,XX +XXX,XX @@ static bool do_outprod_fpst(DisasContext *s, arg_op *a, MemOp esz,
48
+ npcm750_evb_i2c_init(soc);
126
return true;
49
npcm7xx_load_kernel(machine, soc);
50
}
127
}
51
128
129
+TRANS_FEAT(FMOPA_h, aa64_sme, do_outprod_fpst, a, MO_32, gen_helper_sme_fmopa_h)
130
TRANS_FEAT(FMOPA_s, aa64_sme, do_outprod_fpst, a, MO_32, gen_helper_sme_fmopa_s)
131
TRANS_FEAT(FMOPA_d, aa64_sme_f64f64, do_outprod_fpst, a, MO_64, gen_helper_sme_fmopa_d)
132
52
--
133
--
53
2.20.1
134
2.25.1
54
55
diff view generated by jsdifflib
1
From: Hao Wu <wuhaotsh@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This patch adds a QTest for NPCM7XX SMBus's single byte mode. It sends a
3
This is SMOPA, SUMOPA, USMOPA_s, UMOPA, for both Int8 and Int16.
4
byte to a device in the evaluation board, and verify the retrieved value
5
is equivalent to the sent value.
6
4
7
Reviewed-by: Doug Evans<dje@google.com>
8
Reviewed-by: Tyrong Ting<kfting@nuvoton.com>
9
Signed-off-by: Hao Wu <wuhaotsh@google.com>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Message-id: 20210210220426.3577804-5-wuhaotsh@google.com
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20220708151540.18136-28-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
9
---
14
tests/qtest/npcm7xx_smbus-test.c | 352 +++++++++++++++++++++++++++++++
10
target/arm/helper-sme.h | 16 ++++++++
15
tests/qtest/meson.build | 1 +
11
target/arm/sme.decode | 10 +++++
16
2 files changed, 353 insertions(+)
12
target/arm/sme_helper.c | 82 ++++++++++++++++++++++++++++++++++++++
17
create mode 100644 tests/qtest/npcm7xx_smbus-test.c
13
target/arm/translate-sme.c | 10 +++++
14
4 files changed, 118 insertions(+)
18
15
19
diff --git a/tests/qtest/npcm7xx_smbus-test.c b/tests/qtest/npcm7xx_smbus-test.c
16
diff --git a/target/arm/helper-sme.h b/target/arm/helper-sme.h
20
new file mode 100644
17
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX
18
--- a/target/arm/helper-sme.h
22
--- /dev/null
19
+++ b/target/arm/helper-sme.h
23
+++ b/tests/qtest/npcm7xx_smbus-test.c
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_7(sme_fmopa_d, TCG_CALL_NO_RWG,
24
@@ -XXX,XX +XXX,XX @@
21
void, ptr, ptr, ptr, ptr, ptr, ptr, i32)
25
+/*
22
DEF_HELPER_FLAGS_6(sme_bfmopa, TCG_CALL_NO_RWG,
26
+ * QTests for Nuvoton NPCM7xx SMBus Modules.
23
void, ptr, ptr, ptr, ptr, ptr, i32)
27
+ *
24
+DEF_HELPER_FLAGS_6(sme_smopa_s, TCG_CALL_NO_RWG,
28
+ * Copyright 2020 Google LLC
25
+ void, ptr, ptr, ptr, ptr, ptr, i32)
29
+ *
26
+DEF_HELPER_FLAGS_6(sme_umopa_s, TCG_CALL_NO_RWG,
30
+ * This program is free software; you can redistribute it and/or modify it
27
+ void, ptr, ptr, ptr, ptr, ptr, i32)
31
+ * under the terms of the GNU General Public License as published by the
28
+DEF_HELPER_FLAGS_6(sme_sumopa_s, TCG_CALL_NO_RWG,
32
+ * Free Software Foundation; either version 2 of the License, or
29
+ void, ptr, ptr, ptr, ptr, ptr, i32)
33
+ * (at your option) any later version.
30
+DEF_HELPER_FLAGS_6(sme_usmopa_s, TCG_CALL_NO_RWG,
34
+ *
31
+ void, ptr, ptr, ptr, ptr, ptr, i32)
35
+ * This program is distributed in the hope that it will be useful, but WITHOUT
32
+DEF_HELPER_FLAGS_6(sme_smopa_d, TCG_CALL_NO_RWG,
36
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
33
+ void, ptr, ptr, ptr, ptr, ptr, i32)
37
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
34
+DEF_HELPER_FLAGS_6(sme_umopa_d, TCG_CALL_NO_RWG,
38
+ * for more details.
35
+ void, ptr, ptr, ptr, ptr, ptr, i32)
39
+ */
36
+DEF_HELPER_FLAGS_6(sme_sumopa_d, TCG_CALL_NO_RWG,
37
+ void, ptr, ptr, ptr, ptr, ptr, i32)
38
+DEF_HELPER_FLAGS_6(sme_usmopa_d, TCG_CALL_NO_RWG,
39
+ void, ptr, ptr, ptr, ptr, ptr, i32)
40
diff --git a/target/arm/sme.decode b/target/arm/sme.decode
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/sme.decode
43
+++ b/target/arm/sme.decode
44
@@ -XXX,XX +XXX,XX @@ FMOPA_d 10000000 110 ..... ... ... ..... . 0 ... @op_64
45
46
BFMOPA 10000001 100 ..... ... ... ..... . 00 .. @op_32
47
FMOPA_h 10000001 101 ..... ... ... ..... . 00 .. @op_32
40
+
48
+
41
+#include "qemu/osdep.h"
49
+SMOPA_s 1010000 0 10 0 ..... ... ... ..... . 00 .. @op_32
42
+#include "qemu/bitops.h"
50
+SUMOPA_s 1010000 0 10 1 ..... ... ... ..... . 00 .. @op_32
43
+#include "libqos/i2c.h"
51
+USMOPA_s 1010000 1 10 0 ..... ... ... ..... . 00 .. @op_32
44
+#include "libqos/libqtest.h"
52
+UMOPA_s 1010000 1 10 1 ..... ... ... ..... . 00 .. @op_32
45
+#include "hw/misc/tmp105_regs.h"
46
+
53
+
47
+#define NR_SMBUS_DEVICES 16
54
+SMOPA_d 1010000 0 11 0 ..... ... ... ..... . 0 ... @op_64
48
+#define SMBUS_ADDR(x) (0xf0080000 + 0x1000 * (x))
55
+SUMOPA_d 1010000 0 11 1 ..... ... ... ..... . 0 ... @op_64
49
+#define SMBUS_IRQ(x) (64 + (x))
56
+USMOPA_d 1010000 1 11 0 ..... ... ... ..... . 0 ... @op_64
57
+UMOPA_d 1010000 1 11 1 ..... ... ... ..... . 0 ... @op_64
58
diff --git a/target/arm/sme_helper.c b/target/arm/sme_helper.c
59
index XXXXXXX..XXXXXXX 100644
60
--- a/target/arm/sme_helper.c
61
+++ b/target/arm/sme_helper.c
62
@@ -XXX,XX +XXX,XX @@ void HELPER(sme_bfmopa)(void *vza, void *vzn, void *vzm, void *vpn,
63
} while (row & 15);
64
}
65
}
50
+
66
+
51
+#define EVB_DEVICE_ADDR 0x48
67
+typedef uint64_t IMOPFn(uint64_t, uint64_t, uint64_t, uint8_t, bool);
52
+#define INVALID_DEVICE_ADDR 0x01
53
+
68
+
54
+const int evb_bus_list[] = {0, 1, 2, 6};
69
+static inline void do_imopa(uint64_t *za, uint64_t *zn, uint64_t *zm,
70
+ uint8_t *pn, uint8_t *pm,
71
+ uint32_t desc, IMOPFn *fn)
72
+{
73
+ intptr_t row, col, oprsz = simd_oprsz(desc) / 8;
74
+ bool neg = simd_data(desc);
55
+
75
+
56
+/* Offsets */
76
+ for (row = 0; row < oprsz; ++row) {
57
+enum CommonRegister {
77
+ uint8_t pa = pn[H1(row)];
58
+ OFFSET_SDA = 0x0,
78
+ uint64_t *za_row = &za[tile_vslice_index(row)];
59
+ OFFSET_ST = 0x2,
79
+ uint64_t n = zn[row];
60
+ OFFSET_CST = 0x4,
61
+ OFFSET_CTL1 = 0x6,
62
+ OFFSET_ADDR1 = 0x8,
63
+ OFFSET_CTL2 = 0xa,
64
+ OFFSET_ADDR2 = 0xc,
65
+ OFFSET_CTL3 = 0xe,
66
+ OFFSET_CST2 = 0x18,
67
+ OFFSET_CST3 = 0x19,
68
+};
69
+
80
+
70
+enum NPCM7xxSMBusBank0Register {
81
+ for (col = 0; col < oprsz; ++col) {
71
+ OFFSET_ADDR3 = 0x10,
82
+ uint8_t pb = pm[H1(col)];
72
+ OFFSET_ADDR7 = 0x11,
83
+ uint64_t *a = &za_row[col];
73
+ OFFSET_ADDR4 = 0x12,
74
+ OFFSET_ADDR8 = 0x13,
75
+ OFFSET_ADDR5 = 0x14,
76
+ OFFSET_ADDR9 = 0x15,
77
+ OFFSET_ADDR6 = 0x16,
78
+ OFFSET_ADDR10 = 0x17,
79
+ OFFSET_CTL4 = 0x1a,
80
+ OFFSET_CTL5 = 0x1b,
81
+ OFFSET_SCLLT = 0x1c,
82
+ OFFSET_FIF_CTL = 0x1d,
83
+ OFFSET_SCLHT = 0x1e,
84
+};
85
+
84
+
86
+enum NPCM7xxSMBusBank1Register {
85
+ *a = fn(n, zm[col], *a, pa & pb, neg);
87
+ OFFSET_FIF_CTS = 0x10,
88
+ OFFSET_FAIR_PER = 0x11,
89
+ OFFSET_TXF_CTL = 0x12,
90
+ OFFSET_T_OUT = 0x14,
91
+ OFFSET_TXF_STS = 0x1a,
92
+ OFFSET_RXF_STS = 0x1c,
93
+ OFFSET_RXF_CTL = 0x1e,
94
+};
95
+
96
+/* ST fields */
97
+#define ST_STP BIT(7)
98
+#define ST_SDAST BIT(6)
99
+#define ST_BER BIT(5)
100
+#define ST_NEGACK BIT(4)
101
+#define ST_STASTR BIT(3)
102
+#define ST_NMATCH BIT(2)
103
+#define ST_MODE BIT(1)
104
+#define ST_XMIT BIT(0)
105
+
106
+/* CST fields */
107
+#define CST_ARPMATCH BIT(7)
108
+#define CST_MATCHAF BIT(6)
109
+#define CST_TGSCL BIT(5)
110
+#define CST_TSDA BIT(4)
111
+#define CST_GCMATCH BIT(3)
112
+#define CST_MATCH BIT(2)
113
+#define CST_BB BIT(1)
114
+#define CST_BUSY BIT(0)
115
+
116
+/* CST2 fields */
117
+#define CST2_INSTTS BIT(7)
118
+#define CST2_MATCH7F BIT(6)
119
+#define CST2_MATCH6F BIT(5)
120
+#define CST2_MATCH5F BIT(4)
121
+#define CST2_MATCH4F BIT(3)
122
+#define CST2_MATCH3F BIT(2)
123
+#define CST2_MATCH2F BIT(1)
124
+#define CST2_MATCH1F BIT(0)
125
+
126
+/* CST3 fields */
127
+#define CST3_EO_BUSY BIT(7)
128
+#define CST3_MATCH10F BIT(2)
129
+#define CST3_MATCH9F BIT(1)
130
+#define CST3_MATCH8F BIT(0)
131
+
132
+/* CTL1 fields */
133
+#define CTL1_STASTRE BIT(7)
134
+#define CTL1_NMINTE BIT(6)
135
+#define CTL1_GCMEN BIT(5)
136
+#define CTL1_ACK BIT(4)
137
+#define CTL1_EOBINTE BIT(3)
138
+#define CTL1_INTEN BIT(2)
139
+#define CTL1_STOP BIT(1)
140
+#define CTL1_START BIT(0)
141
+
142
+/* CTL2 fields */
143
+#define CTL2_SCLFRQ(rv) extract8((rv), 1, 6)
144
+#define CTL2_ENABLE BIT(0)
145
+
146
+/* CTL3 fields */
147
+#define CTL3_SCL_LVL BIT(7)
148
+#define CTL3_SDA_LVL BIT(6)
149
+#define CTL3_BNK_SEL BIT(5)
150
+#define CTL3_400K_MODE BIT(4)
151
+#define CTL3_IDL_START BIT(3)
152
+#define CTL3_ARPMEN BIT(2)
153
+#define CTL3_SCLFRQ(rv) extract8((rv), 0, 2)
154
+
155
+/* ADDR fields */
156
+#define ADDR_EN BIT(7)
157
+#define ADDR_A(rv) extract8((rv), 0, 6)
158
+
159
+
160
+static void check_running(QTestState *qts, uint64_t base_addr)
161
+{
162
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_CST) & CST_BUSY);
163
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_CST) & CST_BB);
164
+}
165
+
166
+static void check_stopped(QTestState *qts, uint64_t base_addr)
167
+{
168
+ uint8_t cst3;
169
+
170
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==, 0);
171
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_CST) & CST_BUSY);
172
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_CST) & CST_BB);
173
+
174
+ cst3 = qtest_readb(qts, base_addr + OFFSET_CST3);
175
+ g_assert_true(cst3 & CST3_EO_BUSY);
176
+ qtest_writeb(qts, base_addr + OFFSET_CST3, cst3);
177
+ cst3 = qtest_readb(qts, base_addr + OFFSET_CST3);
178
+ g_assert_false(cst3 & CST3_EO_BUSY);
179
+}
180
+
181
+static void enable_bus(QTestState *qts, uint64_t base_addr)
182
+{
183
+ uint8_t ctl2 = qtest_readb(qts, base_addr + OFFSET_CTL2);
184
+
185
+ ctl2 |= CTL2_ENABLE;
186
+ qtest_writeb(qts, base_addr + OFFSET_CTL2, ctl2);
187
+ g_assert_true(qtest_readb(qts, base_addr + OFFSET_CTL2) & CTL2_ENABLE);
188
+}
189
+
190
+static void disable_bus(QTestState *qts, uint64_t base_addr)
191
+{
192
+ uint8_t ctl2 = qtest_readb(qts, base_addr + OFFSET_CTL2);
193
+
194
+ ctl2 &= ~CTL2_ENABLE;
195
+ qtest_writeb(qts, base_addr + OFFSET_CTL2, ctl2);
196
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_CTL2) & CTL2_ENABLE);
197
+}
198
+
199
+static void start_transfer(QTestState *qts, uint64_t base_addr)
200
+{
201
+ uint8_t ctl1;
202
+
203
+ ctl1 = CTL1_START | CTL1_INTEN | CTL1_STASTRE;
204
+ qtest_writeb(qts, base_addr + OFFSET_CTL1, ctl1);
205
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_CTL1), ==,
206
+ CTL1_INTEN | CTL1_STASTRE | CTL1_INTEN);
207
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==,
208
+ ST_MODE | ST_XMIT | ST_SDAST);
209
+ check_running(qts, base_addr);
210
+}
211
+
212
+static void stop_transfer(QTestState *qts, uint64_t base_addr)
213
+{
214
+ uint8_t ctl1 = qtest_readb(qts, base_addr + OFFSET_CTL1);
215
+
216
+ ctl1 &= ~(CTL1_START | CTL1_ACK);
217
+ ctl1 |= CTL1_STOP | CTL1_INTEN | CTL1_EOBINTE;
218
+ qtest_writeb(qts, base_addr + OFFSET_CTL1, ctl1);
219
+ ctl1 = qtest_readb(qts, base_addr + OFFSET_CTL1);
220
+ g_assert_false(ctl1 & CTL1_STOP);
221
+}
222
+
223
+static void send_byte(QTestState *qts, uint64_t base_addr, uint8_t byte)
224
+{
225
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==,
226
+ ST_MODE | ST_XMIT | ST_SDAST);
227
+ qtest_writeb(qts, base_addr + OFFSET_SDA, byte);
228
+}
229
+
230
+static uint8_t recv_byte(QTestState *qts, uint64_t base_addr)
231
+{
232
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==,
233
+ ST_MODE | ST_SDAST);
234
+ return qtest_readb(qts, base_addr + OFFSET_SDA);
235
+}
236
+
237
+static void send_address(QTestState *qts, uint64_t base_addr, uint8_t addr,
238
+ bool recv, bool valid)
239
+{
240
+ uint8_t encoded_addr = (addr << 1) | (recv ? 1 : 0);
241
+ uint8_t st;
242
+
243
+ qtest_writeb(qts, base_addr + OFFSET_SDA, encoded_addr);
244
+ st = qtest_readb(qts, base_addr + OFFSET_ST);
245
+
246
+ if (valid) {
247
+ if (recv) {
248
+ g_assert_cmphex(st, ==, ST_MODE | ST_SDAST | ST_STASTR);
249
+ } else {
250
+ g_assert_cmphex(st, ==, ST_MODE | ST_XMIT | ST_SDAST | ST_STASTR);
251
+ }
252
+
253
+ qtest_writeb(qts, base_addr + OFFSET_ST, ST_STASTR);
254
+ st = qtest_readb(qts, base_addr + OFFSET_ST);
255
+ if (recv) {
256
+ g_assert_cmphex(st, ==, ST_MODE | ST_SDAST);
257
+ } else {
258
+ g_assert_cmphex(st, ==, ST_MODE | ST_XMIT | ST_SDAST);
259
+ }
260
+ } else {
261
+ if (recv) {
262
+ g_assert_cmphex(st, ==, ST_MODE | ST_NEGACK);
263
+ } else {
264
+ g_assert_cmphex(st, ==, ST_MODE | ST_XMIT | ST_NEGACK);
265
+ }
86
+ }
266
+ }
87
+ }
267
+}
88
+}
268
+
89
+
269
+static void send_nack(QTestState *qts, uint64_t base_addr)
90
+#define DEF_IMOP_32(NAME, NTYPE, MTYPE) \
270
+{
91
+static uint64_t NAME(uint64_t n, uint64_t m, uint64_t a, uint8_t p, bool neg) \
271
+ uint8_t ctl1 = qtest_readb(qts, base_addr + OFFSET_CTL1);
92
+{ \
272
+
93
+ uint32_t sum0 = 0, sum1 = 0; \
273
+ ctl1 &= ~(CTL1_START | CTL1_STOP);
94
+ /* Apply P to N as a mask, making the inactive elements 0. */ \
274
+ ctl1 |= CTL1_ACK | CTL1_INTEN;
95
+ n &= expand_pred_b(p); \
275
+ qtest_writeb(qts, base_addr + OFFSET_CTL1, ctl1);
96
+ sum0 += (NTYPE)(n >> 0) * (MTYPE)(m >> 0); \
97
+ sum0 += (NTYPE)(n >> 8) * (MTYPE)(m >> 8); \
98
+ sum0 += (NTYPE)(n >> 16) * (MTYPE)(m >> 16); \
99
+ sum0 += (NTYPE)(n >> 24) * (MTYPE)(m >> 24); \
100
+ sum1 += (NTYPE)(n >> 32) * (MTYPE)(m >> 32); \
101
+ sum1 += (NTYPE)(n >> 40) * (MTYPE)(m >> 40); \
102
+ sum1 += (NTYPE)(n >> 48) * (MTYPE)(m >> 48); \
103
+ sum1 += (NTYPE)(n >> 56) * (MTYPE)(m >> 56); \
104
+ if (neg) { \
105
+ sum0 = (uint32_t)a - sum0, sum1 = (uint32_t)(a >> 32) - sum1; \
106
+ } else { \
107
+ sum0 = (uint32_t)a + sum0, sum1 = (uint32_t)(a >> 32) + sum1; \
108
+ } \
109
+ return ((uint64_t)sum1 << 32) | sum0; \
276
+}
110
+}
277
+
111
+
278
+/* Check the SMBus's status is set correctly when disabled. */
112
+#define DEF_IMOP_64(NAME, NTYPE, MTYPE) \
279
+static void test_disable_bus(gconstpointer data)
113
+static uint64_t NAME(uint64_t n, uint64_t m, uint64_t a, uint8_t p, bool neg) \
280
+{
114
+{ \
281
+ intptr_t index = (intptr_t)data;
115
+ uint64_t sum = 0; \
282
+ uint64_t base_addr = SMBUS_ADDR(index);
116
+ /* Apply P to N as a mask, making the inactive elements 0. */ \
283
+ QTestState *qts = qtest_init("-machine npcm750-evb");
117
+ n &= expand_pred_h(p); \
284
+
118
+ sum += (NTYPE)(n >> 0) * (MTYPE)(m >> 0); \
285
+ disable_bus(qts, base_addr);
119
+ sum += (NTYPE)(n >> 16) * (MTYPE)(m >> 16); \
286
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_CTL1), ==, 0);
120
+ sum += (NTYPE)(n >> 32) * (MTYPE)(m >> 32); \
287
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_ST), ==, 0);
121
+ sum += (NTYPE)(n >> 48) * (MTYPE)(m >> 48); \
288
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_CST3) & CST3_EO_BUSY);
122
+ return neg ? a - sum : a + sum; \
289
+ g_assert_cmphex(qtest_readb(qts, base_addr + OFFSET_CST), ==, 0);
290
+ qtest_quit(qts);
291
+}
123
+}
292
+
124
+
293
+/* Check the SMBus returns a NACK for an invalid address. */
125
+DEF_IMOP_32(smopa_s, int8_t, int8_t)
294
+static void test_invalid_addr(gconstpointer data)
126
+DEF_IMOP_32(umopa_s, uint8_t, uint8_t)
295
+{
127
+DEF_IMOP_32(sumopa_s, int8_t, uint8_t)
296
+ intptr_t index = (intptr_t)data;
128
+DEF_IMOP_32(usmopa_s, uint8_t, int8_t)
297
+ uint64_t base_addr = SMBUS_ADDR(index);
298
+ int irq = SMBUS_IRQ(index);
299
+ QTestState *qts = qtest_init("-machine npcm750-evb");
300
+
129
+
301
+ qtest_irq_intercept_in(qts, "/machine/soc/a9mpcore/gic");
130
+DEF_IMOP_64(smopa_d, int16_t, int16_t)
302
+ enable_bus(qts, base_addr);
131
+DEF_IMOP_64(umopa_d, uint16_t, uint16_t)
303
+ g_assert_false(qtest_get_irq(qts, irq));
132
+DEF_IMOP_64(sumopa_d, int16_t, uint16_t)
304
+ start_transfer(qts, base_addr);
133
+DEF_IMOP_64(usmopa_d, uint16_t, int16_t)
305
+ send_address(qts, base_addr, INVALID_DEVICE_ADDR, false, false);
306
+ g_assert_true(qtest_get_irq(qts, irq));
307
+ stop_transfer(qts, base_addr);
308
+ check_running(qts, base_addr);
309
+ qtest_writeb(qts, base_addr + OFFSET_ST, ST_NEGACK);
310
+ g_assert_false(qtest_readb(qts, base_addr + OFFSET_ST) & ST_NEGACK);
311
+ check_stopped(qts, base_addr);
312
+ qtest_quit(qts);
313
+}
314
+
134
+
315
+/* Check the SMBus can send and receive bytes to a device in single mode. */
135
+#define DEF_IMOPH(NAME) \
316
+static void test_single_mode(gconstpointer data)
136
+ void HELPER(sme_##NAME)(void *vza, void *vzn, void *vzm, void *vpn, \
317
+{
137
+ void *vpm, uint32_t desc) \
318
+ intptr_t index = (intptr_t)data;
138
+ { do_imopa(vza, vzn, vzm, vpn, vpm, desc, NAME); }
319
+ uint64_t base_addr = SMBUS_ADDR(index);
320
+ int irq = SMBUS_IRQ(index);
321
+ uint8_t value = 0x60;
322
+ QTestState *qts = qtest_init("-machine npcm750-evb");
323
+
139
+
324
+ qtest_irq_intercept_in(qts, "/machine/soc/a9mpcore/gic");
140
+DEF_IMOPH(smopa_s)
325
+ enable_bus(qts, base_addr);
141
+DEF_IMOPH(umopa_s)
142
+DEF_IMOPH(sumopa_s)
143
+DEF_IMOPH(usmopa_s)
144
+DEF_IMOPH(smopa_d)
145
+DEF_IMOPH(umopa_d)
146
+DEF_IMOPH(sumopa_d)
147
+DEF_IMOPH(usmopa_d)
148
diff --git a/target/arm/translate-sme.c b/target/arm/translate-sme.c
149
index XXXXXXX..XXXXXXX 100644
150
--- a/target/arm/translate-sme.c
151
+++ b/target/arm/translate-sme.c
152
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(FMOPA_d, aa64_sme_f64f64, do_outprod_fpst, a, MO_64, gen_helper_sme_f
153
154
/* TODO: FEAT_EBF16 */
155
TRANS_FEAT(BFMOPA, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_bfmopa)
326
+
156
+
327
+ /* Sending */
157
+TRANS_FEAT(SMOPA_s, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_smopa_s)
328
+ g_assert_false(qtest_get_irq(qts, irq));
158
+TRANS_FEAT(UMOPA_s, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_umopa_s)
329
+ start_transfer(qts, base_addr);
159
+TRANS_FEAT(SUMOPA_s, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_sumopa_s)
330
+ g_assert_true(qtest_get_irq(qts, irq));
160
+TRANS_FEAT(USMOPA_s, aa64_sme, do_outprod, a, MO_32, gen_helper_sme_usmopa_s)
331
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, false, true);
332
+ send_byte(qts, base_addr, TMP105_REG_CONFIG);
333
+ send_byte(qts, base_addr, value);
334
+ stop_transfer(qts, base_addr);
335
+ check_stopped(qts, base_addr);
336
+
161
+
337
+ /* Receiving */
162
+TRANS_FEAT(SMOPA_d, aa64_sme_i16i64, do_outprod, a, MO_64, gen_helper_sme_smopa_d)
338
+ start_transfer(qts, base_addr);
163
+TRANS_FEAT(UMOPA_d, aa64_sme_i16i64, do_outprod, a, MO_64, gen_helper_sme_umopa_d)
339
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, false, true);
164
+TRANS_FEAT(SUMOPA_d, aa64_sme_i16i64, do_outprod, a, MO_64, gen_helper_sme_sumopa_d)
340
+ send_byte(qts, base_addr, TMP105_REG_CONFIG);
165
+TRANS_FEAT(USMOPA_d, aa64_sme_i16i64, do_outprod, a, MO_64, gen_helper_sme_usmopa_d)
341
+ start_transfer(qts, base_addr);
342
+ send_address(qts, base_addr, EVB_DEVICE_ADDR, true, true);
343
+ send_nack(qts, base_addr);
344
+ stop_transfer(qts, base_addr);
345
+ check_running(qts, base_addr);
346
+ g_assert_cmphex(recv_byte(qts, base_addr), ==, value);
347
+ check_stopped(qts, base_addr);
348
+ qtest_quit(qts);
349
+}
350
+
351
+static void smbus_add_test(const char *name, int index, GTestDataFunc fn)
352
+{
353
+ g_autofree char *full_name = g_strdup_printf(
354
+ "npcm7xx_smbus[%d]/%s", index, name);
355
+ qtest_add_data_func(full_name, (void *)(intptr_t)index, fn);
356
+}
357
+#define add_test(name, td) smbus_add_test(#name, td, test_##name)
358
+
359
+int main(int argc, char **argv)
360
+{
361
+ int i;
362
+
363
+ g_test_init(&argc, &argv, NULL);
364
+ g_test_set_nonfatal_assertions();
365
+
366
+ for (i = 0; i < NR_SMBUS_DEVICES; ++i) {
367
+ add_test(disable_bus, i);
368
+ add_test(invalid_addr, i);
369
+ }
370
+
371
+ for (i = 0; i < ARRAY_SIZE(evb_bus_list); ++i) {
372
+ add_test(single_mode, evb_bus_list[i]);
373
+ }
374
+
375
+ return g_test_run();
376
+}
377
diff --git a/tests/qtest/meson.build b/tests/qtest/meson.build
378
index XXXXXXX..XXXXXXX 100644
379
--- a/tests/qtest/meson.build
380
+++ b/tests/qtest/meson.build
381
@@ -XXX,XX +XXX,XX @@ qtests_npcm7xx = \
382
'npcm7xx_gpio-test',
383
'npcm7xx_pwm-test',
384
'npcm7xx_rng-test',
385
+ 'npcm7xx_smbus-test',
386
'npcm7xx_timer-test',
387
'npcm7xx_watchdog_timer-test']
388
qtests_arm = \
389
--
166
--
390
2.20.1
167
2.25.1
391
392
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Use the now-saved PAGE_ANON and PAGE_MTE bits,
3
This is an SVE instruction that operates using the SVE vector
4
and the per-page saved data.
4
length but that it is present only if SME is implemented.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-30-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-29-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
target/arm/mte_helper.c | 29 +++++++++++++++++++++++++++--
11
target/arm/sve.decode | 20 +++++++++++++
12
1 file changed, 27 insertions(+), 2 deletions(-)
12
target/arm/translate-sve.c | 57 ++++++++++++++++++++++++++++++++++++++
13
2 files changed, 77 insertions(+)
13
14
14
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
15
diff --git a/target/arm/sve.decode b/target/arm/sve.decode
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/mte_helper.c
17
--- a/target/arm/sve.decode
17
+++ b/target/arm/mte_helper.c
18
+++ b/target/arm/sve.decode
18
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
19
@@ -XXX,XX +XXX,XX @@ BFMLALT_zzxw 01100100 11 1 ..... 0100.1 ..... ..... @rrxr_3a esz=2
19
int tag_size, uintptr_t ra)
20
20
{
21
### SVE2 floating-point bfloat16 dot-product (indexed)
21
#ifdef CONFIG_USER_ONLY
22
BFDOT_zzxz 01100100 01 1 ..... 010000 ..... ..... @rrxr_2 esz=2
22
- /* Tag storage not implemented. */
23
- return NULL;
24
+ uint64_t clean_ptr = useronly_clean_ptr(ptr);
25
+ int flags = page_get_flags(clean_ptr);
26
+ uint8_t *tags;
27
+ uintptr_t index;
28
+
23
+
29
+ if (!(flags & (ptr_access == MMU_DATA_STORE ? PAGE_WRITE : PAGE_READ))) {
24
+### SVE broadcast predicate element
30
+ /* SIGSEGV */
25
+
31
+ arm_cpu_tlb_fill(env_cpu(env), ptr, ptr_size, ptr_access,
26
+&psel esz pd pn pm rv imm
32
+ ptr_mmu_idx, false, ra);
27
+%psel_rv 16:2 !function=plus_12
33
+ g_assert_not_reached();
28
+%psel_imm_b 22:2 19:2
29
+%psel_imm_h 22:2 20:1
30
+%psel_imm_s 22:2
31
+%psel_imm_d 23:1
32
+@psel ........ .. . ... .. .. pn:4 . pm:4 . pd:4 \
33
+ &psel rv=%psel_rv
34
+
35
+PSEL 00100101 .. 1 ..1 .. 01 .... 0 .... 0 .... \
36
+ @psel esz=0 imm=%psel_imm_b
37
+PSEL 00100101 .. 1 .10 .. 01 .... 0 .... 0 .... \
38
+ @psel esz=1 imm=%psel_imm_h
39
+PSEL 00100101 .. 1 100 .. 01 .... 0 .... 0 .... \
40
+ @psel esz=2 imm=%psel_imm_s
41
+PSEL 00100101 .1 1 000 .. 01 .... 0 .... 0 .... \
42
+ @psel esz=3 imm=%psel_imm_d
43
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/translate-sve.c
46
+++ b/target/arm/translate-sve.c
47
@@ -XXX,XX +XXX,XX @@ static bool do_BFMLAL_zzxw(DisasContext *s, arg_rrxr_esz *a, bool sel)
48
49
TRANS_FEAT(BFMLALB_zzxw, aa64_sve_bf16, do_BFMLAL_zzxw, a, false)
50
TRANS_FEAT(BFMLALT_zzxw, aa64_sve_bf16, do_BFMLAL_zzxw, a, true)
51
+
52
+static bool trans_PSEL(DisasContext *s, arg_psel *a)
53
+{
54
+ int vl = vec_full_reg_size(s);
55
+ int pl = pred_gvec_reg_size(s);
56
+ int elements = vl >> a->esz;
57
+ TCGv_i64 tmp, didx, dbit;
58
+ TCGv_ptr ptr;
59
+
60
+ if (!dc_isar_feature(aa64_sme, s)) {
61
+ return false;
62
+ }
63
+ if (!sve_access_check(s)) {
64
+ return true;
34
+ }
65
+ }
35
+
66
+
36
+ /* Require both MAP_ANON and PROT_MTE for the page. */
67
+ tmp = tcg_temp_new_i64();
37
+ if (!(flags & PAGE_ANON) || !(flags & PAGE_MTE)) {
68
+ dbit = tcg_temp_new_i64();
38
+ return NULL;
69
+ didx = tcg_temp_new_i64();
70
+ ptr = tcg_temp_new_ptr();
71
+
72
+ /* Compute the predicate element. */
73
+ tcg_gen_addi_i64(tmp, cpu_reg(s, a->rv), a->imm);
74
+ if (is_power_of_2(elements)) {
75
+ tcg_gen_andi_i64(tmp, tmp, elements - 1);
76
+ } else {
77
+ tcg_gen_remu_i64(tmp, tmp, tcg_constant_i64(elements));
39
+ }
78
+ }
40
+
79
+
41
+ tags = page_get_target_data(clean_ptr);
80
+ /* Extract the predicate byte and bit indices. */
42
+ if (tags == NULL) {
81
+ tcg_gen_shli_i64(tmp, tmp, a->esz);
43
+ size_t alloc_size = TARGET_PAGE_SIZE >> (LOG2_TAG_GRANULE + 1);
82
+ tcg_gen_andi_i64(dbit, tmp, 7);
44
+ tags = page_alloc_target_data(clean_ptr, alloc_size);
83
+ tcg_gen_shri_i64(didx, tmp, 3);
45
+ assert(tags != NULL);
84
+ if (HOST_BIG_ENDIAN) {
85
+ tcg_gen_xori_i64(didx, didx, 7);
46
+ }
86
+ }
47
+
87
+
48
+ index = extract32(ptr, LOG2_TAG_GRANULE + 1,
88
+ /* Load the predicate word. */
49
+ TARGET_PAGE_BITS - LOG2_TAG_GRANULE - 1);
89
+ tcg_gen_trunc_i64_ptr(ptr, didx);
50
+ return tags + index;
90
+ tcg_gen_add_ptr(ptr, ptr, cpu_env);
51
#else
91
+ tcg_gen_ld8u_i64(tmp, ptr, pred_full_reg_offset(s, a->pm));
52
uintptr_t index;
92
+
53
CPUIOTLBEntry *iotlbentry;
93
+ /* Extract the predicate bit and replicate to MO_64. */
94
+ tcg_gen_shr_i64(tmp, tmp, dbit);
95
+ tcg_gen_andi_i64(tmp, tmp, 1);
96
+ tcg_gen_neg_i64(tmp, tmp);
97
+
98
+ /* Apply to either copy the source, or write zeros. */
99
+ tcg_gen_gvec_ands(MO_64, pred_full_reg_offset(s, a->pd),
100
+ pred_full_reg_offset(s, a->pn), tmp, pl, pl);
101
+
102
+ tcg_temp_free_i64(tmp);
103
+ tcg_temp_free_i64(dbit);
104
+ tcg_temp_free_i64(didx);
105
+ tcg_temp_free_ptr(ptr);
106
+ return true;
107
+}
54
--
108
--
55
2.20.1
109
2.25.1
56
57
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Provide both tagged and untagged versions of access_ok.
3
This is an SVE instruction that operates using the SVE vector
4
In a few places use thread_cpu, as the user is several
4
length but that it is present only if SME is implemented.
5
callees removed from do_syscall1.
6
5
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20210212184902.1251044-17-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-30-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
linux-user/qemu.h | 11 +++++++++--
11
target/arm/helper-sve.h | 2 ++
13
linux-user/elfload.c | 2 +-
12
target/arm/sve.decode | 1 +
14
linux-user/hppa/cpu_loop.c | 8 ++++----
13
target/arm/sve_helper.c | 16 ++++++++++++++++
15
linux-user/i386/cpu_loop.c | 2 +-
14
target/arm/translate-sve.c | 2 ++
16
linux-user/i386/signal.c | 5 +++--
15
4 files changed, 21 insertions(+)
17
linux-user/syscall.c | 9 ++++++---
18
6 files changed, 24 insertions(+), 13 deletions(-)
19
16
20
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
21
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
22
--- a/linux-user/qemu.h
19
--- a/target/arm/helper-sve.h
23
+++ b/linux-user/qemu.h
20
+++ b/target/arm/helper-sve.h
24
@@ -XXX,XX +XXX,XX @@ extern unsigned long guest_stack_size;
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_revh_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
25
#define VERIFY_READ PAGE_READ
22
26
#define VERIFY_WRITE (PAGE_READ | PAGE_WRITE)
23
DEF_HELPER_FLAGS_4(sve_revw_d, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
27
24
28
-static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
25
+DEF_HELPER_FLAGS_4(sme_revd_q, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
29
+static inline bool access_ok_untagged(int type, abi_ulong addr, abi_ulong size)
26
+
30
{
27
DEF_HELPER_FLAGS_4(sve_rbit_b, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
31
if (size == 0
28
DEF_HELPER_FLAGS_4(sve_rbit_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
32
? !guest_addr_valid_untagged(addr)
29
DEF_HELPER_FLAGS_4(sve_rbit_s, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
33
@@ -XXX,XX +XXX,XX @@ static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
30
diff --git a/target/arm/sve.decode b/target/arm/sve.decode
34
return page_check_range((target_ulong)addr, size, type) == 0;
31
index XXXXXXX..XXXXXXX 100644
35
}
32
--- a/target/arm/sve.decode
36
33
+++ b/target/arm/sve.decode
37
+static inline bool access_ok(CPUState *cpu, int type,
34
@@ -XXX,XX +XXX,XX @@ REVB 00000101 .. 1001 00 100 ... ..... ..... @rd_pg_rn
38
+ abi_ulong addr, abi_ulong size)
35
REVH 00000101 .. 1001 01 100 ... ..... ..... @rd_pg_rn
36
REVW 00000101 .. 1001 10 100 ... ..... ..... @rd_pg_rn
37
RBIT 00000101 .. 1001 11 100 ... ..... ..... @rd_pg_rn
38
+REVD 00000101 00 1011 10 100 ... ..... ..... @rd_pg_rn_e0
39
40
# SVE vector splice (predicated, destructive)
41
SPLICE 00000101 .. 101 100 100 ... ..... ..... @rdn_pg_rm
42
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
43
index XXXXXXX..XXXXXXX 100644
44
--- a/target/arm/sve_helper.c
45
+++ b/target/arm/sve_helper.c
46
@@ -XXX,XX +XXX,XX @@ DO_ZPZ_D(sve_revh_d, uint64_t, hswap64)
47
48
DO_ZPZ_D(sve_revw_d, uint64_t, wswap64)
49
50
+void HELPER(sme_revd_q)(void *vd, void *vn, void *vg, uint32_t desc)
39
+{
51
+{
40
+ return access_ok_untagged(type, cpu_untagged_addr(cpu, addr), size);
52
+ intptr_t i, opr_sz = simd_oprsz(desc) / 8;
53
+ uint64_t *d = vd, *n = vn;
54
+ uint8_t *pg = vg;
55
+
56
+ for (i = 0; i < opr_sz; i += 2) {
57
+ if (pg[H1(i)] & 1) {
58
+ uint64_t n0 = n[i + 0];
59
+ uint64_t n1 = n[i + 1];
60
+ d[i + 0] = n1;
61
+ d[i + 1] = n0;
62
+ }
63
+ }
41
+}
64
+}
42
+
65
+
43
/* NOTE __get_user and __put_user use host pointers and don't check access.
66
DO_ZPZ(sve_rbit_b, uint8_t, H1, revbit8)
44
These are usually used to access struct data members once the struct has
67
DO_ZPZ(sve_rbit_h, uint16_t, H1_2, revbit16)
45
been locked - usually with lock_user_struct. */
68
DO_ZPZ(sve_rbit_s, uint32_t, H1_4, revbit32)
46
@@ -XXX,XX +XXX,XX @@ abi_long copy_to_user(abi_ulong gaddr, void *hptr, size_t len);
69
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
47
host area will have the same contents as the guest. */
48
static inline void *lock_user(int type, abi_ulong guest_addr, long len, int copy)
49
{
50
- if (!access_ok(type, guest_addr, len))
51
+ if (!access_ok_untagged(type, guest_addr, len)) {
52
return NULL;
53
+ }
54
#ifdef DEBUG_REMAP
55
{
56
void *addr;
57
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
58
index XXXXXXX..XXXXXXX 100644
70
index XXXXXXX..XXXXXXX 100644
59
--- a/linux-user/elfload.c
71
--- a/target/arm/translate-sve.c
60
+++ b/linux-user/elfload.c
72
+++ b/target/arm/translate-sve.c
61
@@ -XXX,XX +XXX,XX @@ static int vma_get_mapping_count(const struct mm_struct *mm)
73
@@ -XXX,XX +XXX,XX @@ TRANS_FEAT(REVH, aa64_sve, gen_gvec_ool_arg_zpz, revh_fns[a->esz], a, 0)
62
static abi_ulong vma_dump_size(const struct vm_area_struct *vma)
74
TRANS_FEAT(REVW, aa64_sve, gen_gvec_ool_arg_zpz,
63
{
75
a->esz == 3 ? gen_helper_sve_revw_d : NULL, a, 0)
64
/* if we cannot even read the first page, skip it */
76
65
- if (!access_ok(VERIFY_READ, vma->vma_start, TARGET_PAGE_SIZE))
77
+TRANS_FEAT(REVD, aa64_sme, gen_gvec_ool_arg_zpz, gen_helper_sme_revd_q, a, 0)
66
+ if (!access_ok_untagged(VERIFY_READ, vma->vma_start, TARGET_PAGE_SIZE))
78
+
67
return (0);
79
TRANS_FEAT(SPLICE, aa64_sve, gen_gvec_ool_arg_zpzz,
68
80
gen_helper_sve_splice, a, a->esz)
69
/*
70
diff --git a/linux-user/hppa/cpu_loop.c b/linux-user/hppa/cpu_loop.c
71
index XXXXXXX..XXXXXXX 100644
72
--- a/linux-user/hppa/cpu_loop.c
73
+++ b/linux-user/hppa/cpu_loop.c
74
@@ -XXX,XX +XXX,XX @@ static abi_ulong hppa_lws(CPUHPPAState *env)
75
return -TARGET_ENOSYS;
76
77
case 0: /* elf32 atomic 32bit cmpxchg */
78
- if ((addr & 3) || !access_ok(VERIFY_WRITE, addr, 4)) {
79
+ if ((addr & 3) || !access_ok(cs, VERIFY_WRITE, addr, 4)) {
80
return -TARGET_EFAULT;
81
}
82
old = tswap32(old);
83
@@ -XXX,XX +XXX,XX @@ static abi_ulong hppa_lws(CPUHPPAState *env)
84
return -TARGET_ENOSYS;
85
}
86
if (((addr | old | new) & ((1 << size) - 1))
87
- || !access_ok(VERIFY_WRITE, addr, 1 << size)
88
- || !access_ok(VERIFY_READ, old, 1 << size)
89
- || !access_ok(VERIFY_READ, new, 1 << size)) {
90
+ || !access_ok(cs, VERIFY_WRITE, addr, 1 << size)
91
+ || !access_ok(cs, VERIFY_READ, old, 1 << size)
92
+ || !access_ok(cs, VERIFY_READ, new, 1 << size)) {
93
return -TARGET_EFAULT;
94
}
95
/* Note that below we use host-endian loads so that the cmpxchg
96
diff --git a/linux-user/i386/cpu_loop.c b/linux-user/i386/cpu_loop.c
97
index XXXXXXX..XXXXXXX 100644
98
--- a/linux-user/i386/cpu_loop.c
99
+++ b/linux-user/i386/cpu_loop.c
100
@@ -XXX,XX +XXX,XX @@ static bool write_ok_or_segv(CPUX86State *env, abi_ptr addr, size_t len)
101
* For all the vsyscalls, NULL means "don't write anything" not
102
* "write it at address 0".
103
*/
104
- if (addr == 0 || access_ok(VERIFY_WRITE, addr, len)) {
105
+ if (addr == 0 || access_ok(env_cpu(env), VERIFY_WRITE, addr, len)) {
106
return true;
107
}
108
109
diff --git a/linux-user/i386/signal.c b/linux-user/i386/signal.c
110
index XXXXXXX..XXXXXXX 100644
111
--- a/linux-user/i386/signal.c
112
+++ b/linux-user/i386/signal.c
113
@@ -XXX,XX +XXX,XX @@ restore_sigcontext(CPUX86State *env, struct target_sigcontext *sc)
114
115
fpstate_addr = tswapl(sc->fpstate);
116
if (fpstate_addr != 0) {
117
- if (!access_ok(VERIFY_READ, fpstate_addr,
118
- sizeof(struct target_fpstate)))
119
+ if (!access_ok(env_cpu(env), VERIFY_READ, fpstate_addr,
120
+ sizeof(struct target_fpstate))) {
121
goto badframe;
122
+ }
123
#ifndef TARGET_X86_64
124
cpu_x86_frstor(env, fpstate_addr, 1);
125
#else
126
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
127
index XXXXXXX..XXXXXXX 100644
128
--- a/linux-user/syscall.c
129
+++ b/linux-user/syscall.c
130
@@ -XXX,XX +XXX,XX @@ static abi_long do_accept4(int fd, abi_ulong target_addr,
131
return -TARGET_EINVAL;
132
}
133
134
- if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
135
+ if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
136
return -TARGET_EFAULT;
137
+ }
138
139
addr = alloca(addrlen);
140
141
@@ -XXX,XX +XXX,XX @@ static abi_long do_getpeername(int fd, abi_ulong target_addr,
142
return -TARGET_EINVAL;
143
}
144
145
- if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
146
+ if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
147
return -TARGET_EFAULT;
148
+ }
149
150
addr = alloca(addrlen);
151
152
@@ -XXX,XX +XXX,XX @@ static abi_long do_getsockname(int fd, abi_ulong target_addr,
153
return -TARGET_EINVAL;
154
}
155
156
- if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
157
+ if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
158
return -TARGET_EFAULT;
159
+ }
160
161
addr = alloca(addrlen);
162
81
163
--
82
--
164
2.20.1
83
2.25.1
165
166
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Record whether the backing page is anonymous, or if it has file
3
This is an SVE instruction that operates using the SVE vector
4
backing. This will allow us to get close to the Linux AArch64
4
length but that it is present only if SME is implemented.
5
ABI for MTE, which allows tag memory only on ram-backed VMAs.
6
7
The real ABI allows tag memory on files, when those files are
8
on ram-backed filesystems, such as tmpfs. We will not be able
9
to implement that in QEMU linux-user.
10
11
Thankfully, anonymous memory for malloc arenas is the primary
12
consumer of this feature, so this restricted version should
13
still be of use.
14
5
15
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
16
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
17
Message-id: 20210212184902.1251044-3-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-31-richard.henderson@linaro.org
18
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
19
---
10
---
20
include/exec/cpu-all.h | 2 ++
11
target/arm/helper.h | 18 +++++++
21
linux-user/mmap.c | 3 +++
12
target/arm/sve.decode | 5 ++
22
2 files changed, 5 insertions(+)
13
target/arm/translate-sve.c | 102 +++++++++++++++++++++++++++++++++++++
14
target/arm/vec_helper.c | 24 +++++++++
15
4 files changed, 149 insertions(+)
23
16
24
diff --git a/include/exec/cpu-all.h b/include/exec/cpu-all.h
17
diff --git a/target/arm/helper.h b/target/arm/helper.h
25
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
26
--- a/include/exec/cpu-all.h
19
--- a/target/arm/helper.h
27
+++ b/include/exec/cpu-all.h
20
+++ b/target/arm/helper.h
28
@@ -XXX,XX +XXX,XX @@ extern intptr_t qemu_host_page_mask;
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(gvec_bfmlal, TCG_CALL_NO_RWG,
29
#define PAGE_WRITE_INV 0x0020
22
DEF_HELPER_FLAGS_6(gvec_bfmlal_idx, TCG_CALL_NO_RWG,
30
/* For use with page_set_flags: page is being replaced; target_data cleared. */
23
void, ptr, ptr, ptr, ptr, ptr, i32)
31
#define PAGE_RESET 0x0040
24
32
+/* For linux-user, indicates that the page is MAP_ANON. */
25
+DEF_HELPER_FLAGS_5(gvec_sclamp_b, TCG_CALL_NO_RWG,
33
+#define PAGE_ANON 0x0080
26
+ void, ptr, ptr, ptr, ptr, i32)
34
27
+DEF_HELPER_FLAGS_5(gvec_sclamp_h, TCG_CALL_NO_RWG,
35
#if defined(CONFIG_BSD) && defined(CONFIG_USER_ONLY)
28
+ void, ptr, ptr, ptr, ptr, i32)
36
/* FIXME: Code that sets/uses this is broken and needs to go away. */
29
+DEF_HELPER_FLAGS_5(gvec_sclamp_s, TCG_CALL_NO_RWG,
37
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
30
+ void, ptr, ptr, ptr, ptr, i32)
38
index XXXXXXX..XXXXXXX 100644
31
+DEF_HELPER_FLAGS_5(gvec_sclamp_d, TCG_CALL_NO_RWG,
39
--- a/linux-user/mmap.c
32
+ void, ptr, ptr, ptr, ptr, i32)
40
+++ b/linux-user/mmap.c
33
+
41
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
34
+DEF_HELPER_FLAGS_5(gvec_uclamp_b, TCG_CALL_NO_RWG,
42
}
35
+ void, ptr, ptr, ptr, ptr, i32)
36
+DEF_HELPER_FLAGS_5(gvec_uclamp_h, TCG_CALL_NO_RWG,
37
+ void, ptr, ptr, ptr, ptr, i32)
38
+DEF_HELPER_FLAGS_5(gvec_uclamp_s, TCG_CALL_NO_RWG,
39
+ void, ptr, ptr, ptr, ptr, i32)
40
+DEF_HELPER_FLAGS_5(gvec_uclamp_d, TCG_CALL_NO_RWG,
41
+ void, ptr, ptr, ptr, ptr, i32)
42
+
43
#ifdef TARGET_AARCH64
44
#include "helper-a64.h"
45
#include "helper-sve.h"
46
diff --git a/target/arm/sve.decode b/target/arm/sve.decode
47
index XXXXXXX..XXXXXXX 100644
48
--- a/target/arm/sve.decode
49
+++ b/target/arm/sve.decode
50
@@ -XXX,XX +XXX,XX @@ PSEL 00100101 .. 1 100 .. 01 .... 0 .... 0 .... \
51
@psel esz=2 imm=%psel_imm_s
52
PSEL 00100101 .1 1 000 .. 01 .... 0 .... 0 .... \
53
@psel esz=3 imm=%psel_imm_d
54
+
55
+### SVE clamp
56
+
57
+SCLAMP 01000100 .. 0 ..... 110000 ..... ..... @rda_rn_rm
58
+UCLAMP 01000100 .. 0 ..... 110001 ..... ..... @rda_rn_rm
59
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/target/arm/translate-sve.c
62
+++ b/target/arm/translate-sve.c
63
@@ -XXX,XX +XXX,XX @@ static bool trans_PSEL(DisasContext *s, arg_psel *a)
64
tcg_temp_free_ptr(ptr);
65
return true;
66
}
67
+
68
+static void gen_sclamp_i32(TCGv_i32 d, TCGv_i32 n, TCGv_i32 m, TCGv_i32 a)
69
+{
70
+ tcg_gen_smax_i32(d, a, n);
71
+ tcg_gen_smin_i32(d, d, m);
72
+}
73
+
74
+static void gen_sclamp_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 a)
75
+{
76
+ tcg_gen_smax_i64(d, a, n);
77
+ tcg_gen_smin_i64(d, d, m);
78
+}
79
+
80
+static void gen_sclamp_vec(unsigned vece, TCGv_vec d, TCGv_vec n,
81
+ TCGv_vec m, TCGv_vec a)
82
+{
83
+ tcg_gen_smax_vec(vece, d, a, n);
84
+ tcg_gen_smin_vec(vece, d, d, m);
85
+}
86
+
87
+static void gen_sclamp(unsigned vece, uint32_t d, uint32_t n, uint32_t m,
88
+ uint32_t a, uint32_t oprsz, uint32_t maxsz)
89
+{
90
+ static const TCGOpcode vecop[] = {
91
+ INDEX_op_smin_vec, INDEX_op_smax_vec, 0
92
+ };
93
+ static const GVecGen4 ops[4] = {
94
+ { .fniv = gen_sclamp_vec,
95
+ .fno = gen_helper_gvec_sclamp_b,
96
+ .opt_opc = vecop,
97
+ .vece = MO_8 },
98
+ { .fniv = gen_sclamp_vec,
99
+ .fno = gen_helper_gvec_sclamp_h,
100
+ .opt_opc = vecop,
101
+ .vece = MO_16 },
102
+ { .fni4 = gen_sclamp_i32,
103
+ .fniv = gen_sclamp_vec,
104
+ .fno = gen_helper_gvec_sclamp_s,
105
+ .opt_opc = vecop,
106
+ .vece = MO_32 },
107
+ { .fni8 = gen_sclamp_i64,
108
+ .fniv = gen_sclamp_vec,
109
+ .fno = gen_helper_gvec_sclamp_d,
110
+ .opt_opc = vecop,
111
+ .vece = MO_64,
112
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64 }
113
+ };
114
+ tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &ops[vece]);
115
+}
116
+
117
+TRANS_FEAT(SCLAMP, aa64_sme, gen_gvec_fn_arg_zzzz, gen_sclamp, a)
118
+
119
+static void gen_uclamp_i32(TCGv_i32 d, TCGv_i32 n, TCGv_i32 m, TCGv_i32 a)
120
+{
121
+ tcg_gen_umax_i32(d, a, n);
122
+ tcg_gen_umin_i32(d, d, m);
123
+}
124
+
125
+static void gen_uclamp_i64(TCGv_i64 d, TCGv_i64 n, TCGv_i64 m, TCGv_i64 a)
126
+{
127
+ tcg_gen_umax_i64(d, a, n);
128
+ tcg_gen_umin_i64(d, d, m);
129
+}
130
+
131
+static void gen_uclamp_vec(unsigned vece, TCGv_vec d, TCGv_vec n,
132
+ TCGv_vec m, TCGv_vec a)
133
+{
134
+ tcg_gen_umax_vec(vece, d, a, n);
135
+ tcg_gen_umin_vec(vece, d, d, m);
136
+}
137
+
138
+static void gen_uclamp(unsigned vece, uint32_t d, uint32_t n, uint32_t m,
139
+ uint32_t a, uint32_t oprsz, uint32_t maxsz)
140
+{
141
+ static const TCGOpcode vecop[] = {
142
+ INDEX_op_umin_vec, INDEX_op_umax_vec, 0
143
+ };
144
+ static const GVecGen4 ops[4] = {
145
+ { .fniv = gen_uclamp_vec,
146
+ .fno = gen_helper_gvec_uclamp_b,
147
+ .opt_opc = vecop,
148
+ .vece = MO_8 },
149
+ { .fniv = gen_uclamp_vec,
150
+ .fno = gen_helper_gvec_uclamp_h,
151
+ .opt_opc = vecop,
152
+ .vece = MO_16 },
153
+ { .fni4 = gen_uclamp_i32,
154
+ .fniv = gen_uclamp_vec,
155
+ .fno = gen_helper_gvec_uclamp_s,
156
+ .opt_opc = vecop,
157
+ .vece = MO_32 },
158
+ { .fni8 = gen_uclamp_i64,
159
+ .fniv = gen_uclamp_vec,
160
+ .fno = gen_helper_gvec_uclamp_d,
161
+ .opt_opc = vecop,
162
+ .vece = MO_64,
163
+ .prefer_i64 = TCG_TARGET_REG_BITS == 64 }
164
+ };
165
+ tcg_gen_gvec_4(d, n, m, a, oprsz, maxsz, &ops[vece]);
166
+}
167
+
168
+TRANS_FEAT(UCLAMP, aa64_sme, gen_gvec_fn_arg_zzzz, gen_uclamp, a)
169
diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c
170
index XXXXXXX..XXXXXXX 100644
171
--- a/target/arm/vec_helper.c
172
+++ b/target/arm/vec_helper.c
173
@@ -XXX,XX +XXX,XX @@ void HELPER(gvec_bfmlal_idx)(void *vd, void *vn, void *vm,
43
}
174
}
44
the_end1:
175
clear_tail(d, opr_sz, simd_maxsz(desc));
45
+ if (flags & MAP_ANONYMOUS) {
176
}
46
+ page_flags |= PAGE_ANON;
177
+
47
+ }
178
+#define DO_CLAMP(NAME, TYPE) \
48
page_flags |= PAGE_RESET;
179
+void HELPER(NAME)(void *d, void *n, void *m, void *a, uint32_t desc) \
49
page_set_flags(start, start + len, page_flags);
180
+{ \
50
the_end:
181
+ intptr_t i, opr_sz = simd_oprsz(desc); \
182
+ for (i = 0; i < opr_sz; i += sizeof(TYPE)) { \
183
+ TYPE aa = *(TYPE *)(a + i); \
184
+ TYPE nn = *(TYPE *)(n + i); \
185
+ TYPE mm = *(TYPE *)(m + i); \
186
+ TYPE dd = MIN(MAX(aa, nn), mm); \
187
+ *(TYPE *)(d + i) = dd; \
188
+ } \
189
+ clear_tail(d, opr_sz, simd_maxsz(desc)); \
190
+}
191
+
192
+DO_CLAMP(gvec_sclamp_b, int8_t)
193
+DO_CLAMP(gvec_sclamp_h, int16_t)
194
+DO_CLAMP(gvec_sclamp_s, int32_t)
195
+DO_CLAMP(gvec_sclamp_d, int64_t)
196
+
197
+DO_CLAMP(gvec_uclamp_b, uint8_t)
198
+DO_CLAMP(gvec_uclamp_h, uint16_t)
199
+DO_CLAMP(gvec_uclamp_s, uint32_t)
200
+DO_CLAMP(gvec_uclamp_d, uint64_t)
51
--
201
--
52
2.20.1
202
2.25.1
53
54
diff view generated by jsdifflib
1
From: Doug Evans <dje@google.com>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is a 10/100 ethernet device that has several features.
3
We can handle both exception entry and exception return by
4
Only the ones needed by the Linux driver have been implemented.
4
hooking into aarch64_sve_change_el.
5
See npcm7xx_emc.c for a list of unimplemented features.
6
5
7
Reviewed-by: Hao Wu <wuhaotsh@google.com>
8
Reviewed-by: Avi Fishman <avi.fishman@nuvoton.com>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Doug Evans <dje@google.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Message-id: 20210213002520.1374134-3-dje@google.com
8
Message-id: 20220708151540.18136-32-richard.henderson@linaro.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
10
---
14
docs/system/arm/nuvoton.rst | 3 ++-
11
target/arm/helper.c | 15 +++++++++++++--
15
include/hw/arm/npcm7xx.h | 2 ++
12
1 file changed, 13 insertions(+), 2 deletions(-)
16
hw/arm/npcm7xx.c | 50 +++++++++++++++++++++++++++++++++++--
17
3 files changed, 52 insertions(+), 3 deletions(-)
18
13
19
diff --git a/docs/system/arm/nuvoton.rst b/docs/system/arm/nuvoton.rst
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
20
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
21
--- a/docs/system/arm/nuvoton.rst
16
--- a/target/arm/helper.c
22
+++ b/docs/system/arm/nuvoton.rst
17
+++ b/target/arm/helper.c
23
@@ -XXX,XX +XXX,XX @@ Supported devices
18
@@ -XXX,XX +XXX,XX @@ void aarch64_sve_change_el(CPUARMState *env, int old_el,
24
* Analog to Digital Converter (ADC)
19
return;
25
* Pulse Width Modulation (PWM)
20
}
26
* SMBus controller (SMBF)
21
27
+ * Ethernet controller (EMC)
22
+ old_a64 = old_el ? arm_el_is_aa64(env, old_el) : el0_a64;
28
23
+ new_a64 = new_el ? arm_el_is_aa64(env, new_el) : el0_a64;
29
Missing devices
30
---------------
31
@@ -XXX,XX +XXX,XX @@ Missing devices
32
* Shared memory (SHM)
33
* eSPI slave interface
34
35
- * Ethernet controllers (GMAC and EMC)
36
+ * Ethernet controller (GMAC)
37
* USB device (USBD)
38
* Peripheral SPI controller (PSPI)
39
* SD/MMC host
40
diff --git a/include/hw/arm/npcm7xx.h b/include/hw/arm/npcm7xx.h
41
index XXXXXXX..XXXXXXX 100644
42
--- a/include/hw/arm/npcm7xx.h
43
+++ b/include/hw/arm/npcm7xx.h
44
@@ -XXX,XX +XXX,XX @@
45
#include "hw/misc/npcm7xx_gcr.h"
46
#include "hw/misc/npcm7xx_pwm.h"
47
#include "hw/misc/npcm7xx_rng.h"
48
+#include "hw/net/npcm7xx_emc.h"
49
#include "hw/nvram/npcm7xx_otp.h"
50
#include "hw/timer/npcm7xx_timer.h"
51
#include "hw/ssi/npcm7xx_fiu.h"
52
@@ -XXX,XX +XXX,XX @@ typedef struct NPCM7xxState {
53
EHCISysBusState ehci;
54
OHCISysBusState ohci;
55
NPCM7xxFIUState fiu[2];
56
+ NPCM7xxEMCState emc[2];
57
} NPCM7xxState;
58
59
#define TYPE_NPCM7XX "npcm7xx"
60
diff --git a/hw/arm/npcm7xx.c b/hw/arm/npcm7xx.c
61
index XXXXXXX..XXXXXXX 100644
62
--- a/hw/arm/npcm7xx.c
63
+++ b/hw/arm/npcm7xx.c
64
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxInterrupt {
65
NPCM7XX_UART1_IRQ,
66
NPCM7XX_UART2_IRQ,
67
NPCM7XX_UART3_IRQ,
68
+ NPCM7XX_EMC1RX_IRQ = 15,
69
+ NPCM7XX_EMC1TX_IRQ,
70
NPCM7XX_TIMER0_IRQ = 32, /* Timer Module 0 */
71
NPCM7XX_TIMER1_IRQ,
72
NPCM7XX_TIMER2_IRQ,
73
@@ -XXX,XX +XXX,XX @@ enum NPCM7xxInterrupt {
74
NPCM7XX_SMBUS15_IRQ,
75
NPCM7XX_PWM0_IRQ = 93, /* PWM module 0 */
76
NPCM7XX_PWM1_IRQ, /* PWM module 1 */
77
+ NPCM7XX_EMC2RX_IRQ = 114,
78
+ NPCM7XX_EMC2TX_IRQ,
79
NPCM7XX_GPIO0_IRQ = 116,
80
NPCM7XX_GPIO1_IRQ,
81
NPCM7XX_GPIO2_IRQ,
82
@@ -XXX,XX +XXX,XX @@ static const hwaddr npcm7xx_smbus_addr[] = {
83
0xf008f000,
84
};
85
86
+/* Register base address for each EMC Module */
87
+static const hwaddr npcm7xx_emc_addr[] = {
88
+ 0xf0825000,
89
+ 0xf0826000,
90
+};
91
+
24
+
92
static const struct {
93
hwaddr regs_addr;
94
uint32_t unconnected_pins;
95
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_init(Object *obj)
96
for (i = 0; i < ARRAY_SIZE(s->pwm); i++) {
97
object_initialize_child(obj, "pwm[*]", &s->pwm[i], TYPE_NPCM7XX_PWM);
98
}
99
+
100
+ for (i = 0; i < ARRAY_SIZE(s->emc); i++) {
101
+ object_initialize_child(obj, "emc[*]", &s->emc[i], TYPE_NPCM7XX_EMC);
102
+ }
103
}
104
105
static void npcm7xx_realize(DeviceState *dev, Error **errp)
106
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_realize(DeviceState *dev, Error **errp)
107
sysbus_connect_irq(sbd, i, npcm7xx_irq(s, NPCM7XX_PWM0_IRQ + i));
108
}
109
110
+ /*
25
+ /*
111
+ * EMC Modules. Cannot fail.
26
+ * Both AArch64.TakeException and AArch64.ExceptionReturn
112
+ * The mapping of the device to its netdev backend works as follows:
27
+ * invoke ResetSVEState when taking an exception from, or
113
+ * emc[i] = nd_table[i]
28
+ * returning to, AArch32 state when PSTATE.SM is enabled.
114
+ * This works around the inability to specify the netdev property for the
115
+ * emc device: it's not pluggable and thus the -device option can't be
116
+ * used.
117
+ */
29
+ */
118
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(npcm7xx_emc_addr) != ARRAY_SIZE(s->emc));
30
+ if (old_a64 != new_a64 && FIELD_EX64(env->svcr, SVCR, SM)) {
119
+ QEMU_BUILD_BUG_ON(ARRAY_SIZE(s->emc) != 2);
31
+ arm_reset_sve_state(env);
120
+ for (i = 0; i < ARRAY_SIZE(s->emc); i++) {
32
+ return;
121
+ s->emc[i].emc_num = i;
122
+ SysBusDevice *sbd = SYS_BUS_DEVICE(&s->emc[i]);
123
+ if (nd_table[i].used) {
124
+ qemu_check_nic_model(&nd_table[i], TYPE_NPCM7XX_EMC);
125
+ qdev_set_nic_properties(DEVICE(sbd), &nd_table[i]);
126
+ }
127
+ /*
128
+ * The device exists regardless of whether it's connected to a QEMU
129
+ * netdev backend. So always instantiate it even if there is no
130
+ * backend.
131
+ */
132
+ sysbus_realize(sbd, &error_abort);
133
+ sysbus_mmio_map(sbd, 0, npcm7xx_emc_addr[i]);
134
+ int tx_irq = i == 0 ? NPCM7XX_EMC1TX_IRQ : NPCM7XX_EMC2TX_IRQ;
135
+ int rx_irq = i == 0 ? NPCM7XX_EMC1RX_IRQ : NPCM7XX_EMC2RX_IRQ;
136
+ /*
137
+ * N.B. The values for the second argument sysbus_connect_irq are
138
+ * chosen to match the registration order in npcm7xx_emc_realize.
139
+ */
140
+ sysbus_connect_irq(sbd, 0, npcm7xx_irq(s, tx_irq));
141
+ sysbus_connect_irq(sbd, 1, npcm7xx_irq(s, rx_irq));
142
+ }
33
+ }
143
+
34
+
144
/*
35
/*
145
* Flash Interface Unit (FIU). Can fail if incorrect number of chip selects
36
* DDI0584A.d sec 3.2: "If SVE instructions are disabled or trapped
146
* specified, but this is a programming error.
37
* at ELx, or not available because the EL is in AArch32 state, then
147
@@ -XXX,XX +XXX,XX @@ static void npcm7xx_realize(DeviceState *dev, Error **errp)
38
@@ -XXX,XX +XXX,XX @@ void aarch64_sve_change_el(CPUARMState *env, int old_el,
148
create_unimplemented_device("npcm7xx.vcd", 0xf0810000, 64 * KiB);
39
* we already have the correct register contents when encountering the
149
create_unimplemented_device("npcm7xx.ece", 0xf0820000, 8 * KiB);
40
* vq0->vq0 transition between EL0->EL1.
150
create_unimplemented_device("npcm7xx.vdma", 0xf0822000, 8 * KiB);
41
*/
151
- create_unimplemented_device("npcm7xx.emc1", 0xf0825000, 4 * KiB);
42
- old_a64 = old_el ? arm_el_is_aa64(env, old_el) : el0_a64;
152
- create_unimplemented_device("npcm7xx.emc2", 0xf0826000, 4 * KiB);
43
old_len = (old_a64 && !sve_exception_el(env, old_el)
153
create_unimplemented_device("npcm7xx.usbd[0]", 0xf0830000, 4 * KiB);
44
? sve_vqm1_for_el(env, old_el) : 0);
154
create_unimplemented_device("npcm7xx.usbd[1]", 0xf0831000, 4 * KiB);
45
- new_a64 = new_el ? arm_el_is_aa64(env, new_el) : el0_a64;
155
create_unimplemented_device("npcm7xx.usbd[2]", 0xf0832000, 4 * KiB);
46
new_len = (new_a64 && !sve_exception_el(env, new_el)
47
? sve_vqm1_for_el(env, new_el) : 0);
48
156
--
49
--
157
2.20.1
50
2.25.1
158
159
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Provide an identity fallback for target that do not
3
Note that SME remains effectively disabled for user-only,
4
use tagged addresses.
4
because we do not yet set CPACR_EL1.SMEN. This needs to
5
wait until the kernel ABI is implemented.
5
6
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20210212184902.1251044-12-richard.henderson@linaro.org
9
Message-id: 20220708151540.18136-33-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
---
11
include/exec/cpu_ldst.h | 7 +++++++
12
docs/system/arm/emulation.rst | 4 ++++
12
1 file changed, 7 insertions(+)
13
target/arm/cpu64.c | 11 +++++++++++
14
2 files changed, 15 insertions(+)
13
15
14
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
16
diff --git a/docs/system/arm/emulation.rst b/docs/system/arm/emulation.rst
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
16
--- a/include/exec/cpu_ldst.h
18
--- a/docs/system/arm/emulation.rst
17
+++ b/include/exec/cpu_ldst.h
19
+++ b/docs/system/arm/emulation.rst
18
@@ -XXX,XX +XXX,XX @@ typedef uint64_t abi_ptr;
20
@@ -XXX,XX +XXX,XX @@ the following architecture extensions:
19
#define TARGET_ABI_FMT_ptr "%"PRIx64
21
- FEAT_SHA512 (Advanced SIMD SHA512 instructions)
20
#endif
22
- FEAT_SM3 (Advanced SIMD SM3 instructions)
21
23
- FEAT_SM4 (Advanced SIMD SM4 instructions)
22
+#ifndef TARGET_TAGGED_ADDRESSES
24
+- FEAT_SME (Scalable Matrix Extension)
23
+static inline abi_ptr cpu_untagged_addr(CPUState *cs, abi_ptr x)
25
+- FEAT_SME_FA64 (Full A64 instruction set in Streaming SVE mode)
24
+{
26
+- FEAT_SME_F64F64 (Double-precision floating-point outer product instructions)
25
+ return x;
27
+- FEAT_SME_I16I64 (16-bit to 64-bit integer widening outer product instructions)
26
+}
28
- FEAT_SPECRES (Speculation restriction instructions)
27
+#endif
29
- FEAT_SSBS (Speculative Store Bypass Safe)
30
- FEAT_TLBIOS (TLB invalidate instructions in Outer Shareable domain)
31
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/cpu64.c
34
+++ b/target/arm/cpu64.c
35
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
36
*/
37
t = FIELD_DP64(t, ID_AA64PFR1, MTE, 3); /* FEAT_MTE3 */
38
t = FIELD_DP64(t, ID_AA64PFR1, RAS_FRAC, 0); /* FEAT_RASv1p1 + FEAT_DoubleFault */
39
+ t = FIELD_DP64(t, ID_AA64PFR1, SME, 1); /* FEAT_SME */
40
t = FIELD_DP64(t, ID_AA64PFR1, CSV2_FRAC, 0); /* FEAT_CSV2_2 */
41
cpu->isar.id_aa64pfr1 = t;
42
43
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
44
t = FIELD_DP64(t, ID_AA64DFR0, PMUVER, 5); /* FEAT_PMUv3p4 */
45
cpu->isar.id_aa64dfr0 = t;
46
47
+ t = cpu->isar.id_aa64smfr0;
48
+ t = FIELD_DP64(t, ID_AA64SMFR0, F32F32, 1); /* FEAT_SME */
49
+ t = FIELD_DP64(t, ID_AA64SMFR0, B16F32, 1); /* FEAT_SME */
50
+ t = FIELD_DP64(t, ID_AA64SMFR0, F16F32, 1); /* FEAT_SME */
51
+ t = FIELD_DP64(t, ID_AA64SMFR0, I8I32, 0xf); /* FEAT_SME */
52
+ t = FIELD_DP64(t, ID_AA64SMFR0, F64F64, 1); /* FEAT_SME_F64F64 */
53
+ t = FIELD_DP64(t, ID_AA64SMFR0, I16I64, 0xf); /* FEAT_SME_I16I64 */
54
+ t = FIELD_DP64(t, ID_AA64SMFR0, FA64, 1); /* FEAT_SME_FA64 */
55
+ cpu->isar.id_aa64smfr0 = t;
28
+
56
+
29
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
57
/* Replicate the same data to the 32-bit id registers. */
30
#define g2h(x) ((void *)((uintptr_t)(abi_ptr)(x) + guest_base))
58
aa32_max_features(cpu);
31
59
32
--
60
--
33
2.20.1
61
2.25.1
34
35
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
These constants are only ever used with access_ok, and friends.
4
Rather than translating them to PAGE_* bits, let them equal
5
the PAGE_* bits to begin.
6
7
Reviewed-by: Warner Losh <imp@bsdimp.com>
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20210212184902.1251044-9-richard.henderson@linaro.org
5
Message-id: 20220708151540.18136-34-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
7
---
13
bsd-user/qemu.h | 9 ++++-----
8
linux-user/aarch64/target_cpu.h | 5 ++++-
14
1 file changed, 4 insertions(+), 5 deletions(-)
9
1 file changed, 4 insertions(+), 1 deletion(-)
15
10
16
diff --git a/bsd-user/qemu.h b/bsd-user/qemu.h
11
diff --git a/linux-user/aarch64/target_cpu.h b/linux-user/aarch64/target_cpu.h
17
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
18
--- a/bsd-user/qemu.h
13
--- a/linux-user/aarch64/target_cpu.h
19
+++ b/bsd-user/qemu.h
14
+++ b/linux-user/aarch64/target_cpu.h
20
@@ -XXX,XX +XXX,XX @@ extern unsigned long x86_stack_size;
15
@@ -XXX,XX +XXX,XX @@ static inline void cpu_clone_regs_parent(CPUARMState *env, unsigned flags)
21
16
22
/* user access */
17
static inline void cpu_set_tls(CPUARMState *env, target_ulong newtls)
23
24
-#define VERIFY_READ 0
25
-#define VERIFY_WRITE 1 /* implies read access */
26
+#define VERIFY_READ PAGE_READ
27
+#define VERIFY_WRITE (PAGE_READ | PAGE_WRITE)
28
29
-static inline int access_ok(int type, abi_ulong addr, abi_ulong size)
30
+static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
31
{
18
{
32
- return page_check_range((target_ulong)addr, size,
19
- /* Note that AArch64 Linux keeps the TLS pointer in TPIDR; this is
33
- (type == VERIFY_READ) ? PAGE_READ : (PAGE_READ | PAGE_WRITE)) == 0;
20
+ /*
34
+ return page_check_range((target_ulong)addr, size, type) == 0;
21
+ * Note that AArch64 Linux keeps the TLS pointer in TPIDR; this is
22
* different from AArch32 Linux, which uses TPIDRRO.
23
*/
24
env->cp15.tpidr_el[0] = newtls;
25
+ /* TPIDR2_EL0 is cleared with CLONE_SETTLS. */
26
+ env->cp15.tpidr2_el0 = 0;
35
}
27
}
36
28
37
/* NOTE __get_user and __put_user use host pointers and don't check access. */
29
static inline abi_ulong get_sp_from_cpustate(CPUARMState *state)
38
--
30
--
39
2.20.1
31
2.25.1
40
41
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20210212184902.1251044-28-richard.henderson@linaro.org
5
Message-id: 20220708151540.18136-35-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
7
---
8
linux-user/aarch64/target_signal.h | 2 ++
8
linux-user/aarch64/cpu_loop.c | 9 +++++++++
9
linux-user/aarch64/cpu_loop.c | 3 +++
9
1 file changed, 9 insertions(+)
10
2 files changed, 5 insertions(+)
11
10
12
diff --git a/linux-user/aarch64/target_signal.h b/linux-user/aarch64/target_signal.h
13
index XXXXXXX..XXXXXXX 100644
14
--- a/linux-user/aarch64/target_signal.h
15
+++ b/linux-user/aarch64/target_signal.h
16
@@ -XXX,XX +XXX,XX @@ typedef struct target_sigaltstack {
17
18
#include "../generic/signal.h"
19
20
+#define TARGET_SEGV_MTESERR 9 /* Synchronous ARM MTE exception */
21
+
22
#define TARGET_ARCH_HAS_SETUP_FRAME
23
#endif /* AARCH64_TARGET_SIGNAL_H */
24
diff --git a/linux-user/aarch64/cpu_loop.c b/linux-user/aarch64/cpu_loop.c
11
diff --git a/linux-user/aarch64/cpu_loop.c b/linux-user/aarch64/cpu_loop.c
25
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
26
--- a/linux-user/aarch64/cpu_loop.c
13
--- a/linux-user/aarch64/cpu_loop.c
27
+++ b/linux-user/aarch64/cpu_loop.c
14
+++ b/linux-user/aarch64/cpu_loop.c
28
@@ -XXX,XX +XXX,XX @@ void cpu_loop(CPUARMState *env)
15
@@ -XXX,XX +XXX,XX @@ void cpu_loop(CPUARMState *env)
29
case 0x0d ... 0x0f: /* Permission fault, level {1-3} */
16
30
info.si_code = TARGET_SEGV_ACCERR;
17
switch (trapnr) {
31
break;
18
case EXCP_SWI:
32
+ case 0x11: /* Synchronous Tag Check Fault */
19
+ /*
33
+ info.si_code = TARGET_SEGV_MTESERR;
20
+ * On syscall, PSTATE.ZA is preserved, along with the ZA matrix.
34
+ break;
21
+ * PSTATE.SM is cleared, per SMSTOP, which does ResetSVEState.
35
default:
22
+ */
36
g_assert_not_reached();
23
+ if (FIELD_EX64(env->svcr, SVCR, SM)) {
37
}
24
+ env->svcr = FIELD_DP64(env->svcr, SVCR, SM, 0);
25
+ arm_rebuild_hflags(env);
26
+ arm_reset_sve_state(env);
27
+ }
28
ret = do_syscall(env,
29
env->xregs[8],
30
env->xregs[0],
38
--
31
--
39
2.20.1
32
2.25.1
40
41
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Return bool not int; pass abi_ulong not 'unsigned long'.
3
Make sure to zero the currently reserved fields.
4
All callers use abi_ulong already, so the change in type
5
has no effect.
6
4
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Message-id: 20220708151540.18136-36-richard.henderson@linaro.org
10
Message-id: 20210212184902.1251044-6-richard.henderson@linaro.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
9
---
13
include/exec/cpu_ldst.h | 2 +-
10
linux-user/aarch64/signal.c | 9 ++++++++-
14
1 file changed, 1 insertion(+), 1 deletion(-)
11
1 file changed, 8 insertions(+), 1 deletion(-)
15
12
16
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
13
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
17
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
18
--- a/include/exec/cpu_ldst.h
15
--- a/linux-user/aarch64/signal.c
19
+++ b/include/exec/cpu_ldst.h
16
+++ b/linux-user/aarch64/signal.c
20
@@ -XXX,XX +XXX,XX @@ typedef uint64_t abi_ptr;
17
@@ -XXX,XX +XXX,XX @@ struct target_extra_context {
21
#endif
18
struct target_sve_context {
22
#define h2g_valid(x) guest_addr_valid((uintptr_t)(x) - guest_base)
19
struct target_aarch64_ctx head;
23
20
uint16_t vl;
24
-static inline int guest_range_valid(unsigned long start, unsigned long len)
21
- uint16_t reserved[3];
25
+static inline bool guest_range_valid(abi_ulong start, abi_ulong len)
22
+ uint16_t flags;
23
+ uint16_t reserved[2];
24
/* The actual SVE data immediately follows. It is laid out
25
* according to TARGET_SVE_SIG_{Z,P}REG_OFFSET, based off of
26
* the original struct pointer.
27
@@ -XXX,XX +XXX,XX @@ struct target_sve_context {
28
#define TARGET_SVE_SIG_CONTEXT_SIZE(VQ) \
29
(TARGET_SVE_SIG_PREG_OFFSET(VQ, 17))
30
31
+#define TARGET_SVE_SIG_FLAG_SM 1
32
+
33
struct target_rt_sigframe {
34
struct target_siginfo info;
35
struct target_ucontext uc;
36
@@ -XXX,XX +XXX,XX @@ static void target_setup_sve_record(struct target_sve_context *sve,
26
{
37
{
27
return len - 1 <= GUEST_ADDR_MAX && start <= GUEST_ADDR_MAX - len + 1;
38
int i, j;
28
}
39
40
+ memset(sve, 0, sizeof(*sve));
41
__put_user(TARGET_SVE_MAGIC, &sve->head.magic);
42
__put_user(size, &sve->head.size);
43
__put_user(vq * TARGET_SVE_VQ_BYTES, &sve->vl);
44
+ if (FIELD_EX64(env->svcr, SVCR, SM)) {
45
+ __put_user(TARGET_SVE_SIG_FLAG_SM, &sve->flags);
46
+ }
47
48
/* Note that SVE regs are stored as a byte stream, with each byte element
49
* at a subsequent address. This corresponds to a little-endian store
29
--
50
--
30
2.20.1
51
2.25.1
31
32
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is more descriptive than 'unsigned long'.
3
Fold the return value setting into the goto, so each
4
No functional change, since these match on all linux+bsd hosts.
4
point of failure need not do both.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20220708151540.18136-37-richard.henderson@linaro.org
9
Message-id: 20210212184902.1251044-5-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
10
---
12
include/exec/cpu_ldst.h | 6 +++---
11
linux-user/aarch64/signal.c | 26 +++++++++++---------------
13
1 file changed, 3 insertions(+), 3 deletions(-)
12
1 file changed, 11 insertions(+), 15 deletions(-)
14
13
15
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
14
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
16
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
17
--- a/include/exec/cpu_ldst.h
16
--- a/linux-user/aarch64/signal.c
18
+++ b/include/exec/cpu_ldst.h
17
+++ b/linux-user/aarch64/signal.c
19
@@ -XXX,XX +XXX,XX @@ typedef uint64_t abi_ptr;
18
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
20
#endif
19
struct target_sve_context *sve = NULL;
21
20
uint64_t extra_datap = 0;
22
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
21
bool used_extra = false;
23
-#define g2h(x) ((void *)((unsigned long)(abi_ptr)(x) + guest_base))
22
- bool err = false;
24
+#define g2h(x) ((void *)((uintptr_t)(abi_ptr)(x) + guest_base))
23
int vq = 0, sve_size = 0;
25
24
26
#if HOST_LONG_BITS <= TARGET_VIRT_ADDR_SPACE_BITS
25
target_restore_general_frame(env, sf);
27
#define guest_addr_valid(x) (1)
26
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
28
#else
27
switch (magic) {
29
#define guest_addr_valid(x) ((x) <= GUEST_ADDR_MAX)
28
case 0:
30
#endif
29
if (size != 0) {
31
-#define h2g_valid(x) guest_addr_valid((unsigned long)(x) - guest_base)
30
- err = true;
32
+#define h2g_valid(x) guest_addr_valid((uintptr_t)(x) - guest_base)
31
- goto exit;
33
32
+ goto err;
34
static inline int guest_range_valid(unsigned long start, unsigned long len)
33
}
35
{
34
if (used_extra) {
36
@@ -XXX,XX +XXX,XX @@ static inline int guest_range_valid(unsigned long start, unsigned long len)
35
ctx = NULL;
36
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
37
38
case TARGET_FPSIMD_MAGIC:
39
if (fpsimd || size != sizeof(struct target_fpsimd_context)) {
40
- err = true;
41
- goto exit;
42
+ goto err;
43
}
44
fpsimd = (struct target_fpsimd_context *)ctx;
45
break;
46
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
47
break;
48
}
49
}
50
- err = true;
51
- goto exit;
52
+ goto err;
53
54
case TARGET_EXTRA_MAGIC:
55
if (extra || size != sizeof(struct target_extra_context)) {
56
- err = true;
57
- goto exit;
58
+ goto err;
59
}
60
__get_user(extra_datap,
61
&((struct target_extra_context *)ctx)->datap);
62
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
63
/* Unknown record -- we certainly didn't generate it.
64
* Did we in fact get out of sync?
65
*/
66
- err = true;
67
- goto exit;
68
+ goto err;
69
}
70
ctx = (void *)ctx + size;
71
}
72
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
73
if (fpsimd) {
74
target_restore_fpsimd_record(env, fpsimd);
75
} else {
76
- err = true;
77
+ goto err;
78
}
79
80
/* SVE data, if present, overwrites FPSIMD data. */
81
if (sve) {
82
target_restore_sve_record(env, sve, vq);
83
}
84
-
85
- exit:
86
unlock_user(extra, extra_datap, 0);
87
- return err;
88
+ return 0;
89
+
90
+ err:
91
+ unlock_user(extra, extra_datap, 0);
92
+ return 1;
37
}
93
}
38
94
39
#define h2g_nocheck(x) ({ \
95
static abi_ulong get_sigframe(struct target_sigaction *ka,
40
- unsigned long __ret = (unsigned long)(x) - guest_base; \
41
+ uintptr_t __ret = (uintptr_t)(x) - guest_base; \
42
(abi_ptr)__ret; \
43
})
44
45
--
96
--
46
2.20.1
97
2.25.1
47
48
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
This is the only use of guest_addr_valid that does not begin
3
In parse_user_sigframe, the kernel rejects duplicate sve records,
4
with a guest address, but a host address being transformed to
4
or records that are smaller than the header. We were silently
5
a guest address.
5
allowing these cases to pass, dropping the record.
6
7
We will shortly adjust guest_addr_valid to handle guest memory
8
tags, and the host address should not be subjected to that.
9
10
Move h2g_valid adjacent to the other h2g macros.
11
6
12
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20210212184902.1251044-10-richard.henderson@linaro.org
9
Message-id: 20220708151540.18136-38-richard.henderson@linaro.org
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
---
11
---
17
include/exec/cpu_ldst.h | 5 ++++-
12
linux-user/aarch64/signal.c | 5 ++++-
18
1 file changed, 4 insertions(+), 1 deletion(-)
13
1 file changed, 4 insertions(+), 1 deletion(-)
19
14
20
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
15
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
21
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
22
--- a/include/exec/cpu_ldst.h
17
--- a/linux-user/aarch64/signal.c
23
+++ b/include/exec/cpu_ldst.h
18
+++ b/linux-user/aarch64/signal.c
24
@@ -XXX,XX +XXX,XX @@ typedef uint64_t abi_ptr;
19
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
25
#else
20
break;
26
#define guest_addr_valid(x) ((x) <= GUEST_ADDR_MAX)
21
27
#endif
22
case TARGET_SVE_MAGIC:
28
-#define h2g_valid(x) guest_addr_valid((uintptr_t)(x) - guest_base)
23
+ if (sve || size < sizeof(struct target_sve_context)) {
29
24
+ goto err;
30
static inline bool guest_range_valid(abi_ulong start, abi_ulong len)
25
+ }
31
{
26
if (cpu_isar_feature(aa64_sve, env_archcpu(env))) {
32
return len - 1 <= GUEST_ADDR_MAX && start <= GUEST_ADDR_MAX - len + 1;
27
vq = sve_vq(env);
33
}
28
sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(vq), 16);
34
29
- if (!sve && size == sve_size) {
35
+#define h2g_valid(x) \
30
+ if (size == sve_size) {
36
+ (HOST_LONG_BITS <= TARGET_VIRT_ADDR_SPACE_BITS || \
31
sve = (struct target_sve_context *)ctx;
37
+ (uintptr_t)(x) - guest_base <= GUEST_ADDR_MAX)
32
break;
38
+
33
}
39
#define h2g_nocheck(x) ({ \
40
uintptr_t __ret = (uintptr_t)(x) - guest_base; \
41
(abi_ptr)__ret; \
42
--
34
--
43
2.20.1
35
2.25.1
44
45
diff view generated by jsdifflib
1
From: Luc Michel <luc@lmichel.fr>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Also add Damien as a reviewer.
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Signed-off-by: Luc Michel <luc@lmichel.fr>
5
Message-id: 20220708151540.18136-39-richard.henderson@linaro.org
6
Acked-by: Damien Hedde <damien.hedde@greensocs.com>
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20210211085318.2507-1-luc@lmichel.fr
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
7
---
11
MAINTAINERS | 11 +++++++++++
8
linux-user/aarch64/signal.c | 3 +++
12
1 file changed, 11 insertions(+)
9
1 file changed, 3 insertions(+)
13
10
14
diff --git a/MAINTAINERS b/MAINTAINERS
11
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
15
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
16
--- a/MAINTAINERS
13
--- a/linux-user/aarch64/signal.c
17
+++ b/MAINTAINERS
14
+++ b/linux-user/aarch64/signal.c
18
@@ -XXX,XX +XXX,XX @@ F: pc-bios/opensbi-*
15
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
19
F: .gitlab-ci.d/opensbi.yml
16
__get_user(extra_size,
20
F: .gitlab-ci.d/opensbi/
17
&((struct target_extra_context *)ctx)->size);
21
18
extra = lock_user(VERIFY_READ, extra_datap, extra_size, 0);
22
+Clock framework
19
+ if (!extra) {
23
+M: Luc Michel <luc@lmichel.fr>
20
+ return 1;
24
+R: Damien Hedde <damien.hedde@greensocs.com>
21
+ }
25
+S: Maintained
22
break;
26
+F: include/hw/clock.h
23
27
+F: include/hw/qdev-clock.h
24
default:
28
+F: hw/core/clock.c
29
+F: hw/core/clock-vmstate.c
30
+F: hw/core/qdev-clock.c
31
+F: docs/devel/clocks.rst
32
+
33
Usermode Emulation
34
------------------
35
Overall usermode emulation
36
--
25
--
37
2.20.1
26
2.25.1
38
39
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
Verify that addr + size - 1 does not wrap around.
3
Move the checks out of the parsing loop and into the
4
restore function. This more closely mirrors the code
5
structure in the kernel, and is slightly clearer.
6
7
Reject rather than silently skip incorrect VL and SVE record sizes,
8
bringing our checks in to line with those the kernel does.
4
9
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20210212184902.1251044-7-richard.henderson@linaro.org
12
Message-id: 20220708151540.18136-40-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
14
---
10
linux-user/qemu.h | 17 ++++++++++++-----
15
linux-user/aarch64/signal.c | 51 +++++++++++++++++++++++++------------
11
1 file changed, 12 insertions(+), 5 deletions(-)
16
1 file changed, 35 insertions(+), 16 deletions(-)
12
17
13
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
18
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
14
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
15
--- a/linux-user/qemu.h
20
--- a/linux-user/aarch64/signal.c
16
+++ b/linux-user/qemu.h
21
+++ b/linux-user/aarch64/signal.c
17
@@ -XXX,XX +XXX,XX @@ extern unsigned long guest_stack_size;
22
@@ -XXX,XX +XXX,XX @@ static void target_restore_fpsimd_record(CPUARMState *env,
18
#define VERIFY_READ 0
23
}
19
#define VERIFY_WRITE 1 /* implies read access */
24
}
20
25
21
-static inline int access_ok(int type, abi_ulong addr, abi_ulong size)
26
-static void target_restore_sve_record(CPUARMState *env,
22
+static inline bool access_ok(int type, abi_ulong addr, abi_ulong size)
27
- struct target_sve_context *sve, int vq)
28
+static bool target_restore_sve_record(CPUARMState *env,
29
+ struct target_sve_context *sve,
30
+ int size)
23
{
31
{
24
- return guest_addr_valid(addr) &&
32
- int i, j;
25
- (size == 0 || guest_addr_valid(addr + size - 1)) &&
33
+ int i, j, vl, vq;
26
- page_check_range((target_ulong)addr, size,
34
27
- (type == VERIFY_READ) ? PAGE_READ : (PAGE_READ | PAGE_WRITE)) == 0;
35
- /* Note that SVE regs are stored as a byte stream, with each byte element
28
+ if (!guest_addr_valid(addr)) {
36
+ if (!cpu_isar_feature(aa64_sve, env_archcpu(env))) {
29
+ return false;
37
+ return false;
30
+ }
38
+ }
31
+ if (size != 0 &&
39
+
32
+ (addr + size - 1 < addr ||
40
+ __get_user(vl, &sve->vl);
33
+ !guest_addr_valid(addr + size - 1))) {
41
+ vq = sve_vq(env);
42
+
43
+ /* Reject mismatched VL. */
44
+ if (vl != vq * TARGET_SVE_VQ_BYTES) {
34
+ return false;
45
+ return false;
35
+ }
46
+ }
36
+ return page_check_range((target_ulong)addr, size,
47
+
37
+ (type == VERIFY_READ) ? PAGE_READ :
48
+ /* Accept empty record -- used to clear PSTATE.SM. */
38
+ (PAGE_READ | PAGE_WRITE)) == 0;
49
+ if (size <= sizeof(*sve)) {
50
+ return true;
51
+ }
52
+
53
+ /* Reject non-empty but incomplete record. */
54
+ if (size < TARGET_SVE_SIG_CONTEXT_SIZE(vq)) {
55
+ return false;
56
+ }
57
+
58
+ /*
59
+ * Note that SVE regs are stored as a byte stream, with each byte element
60
* at a subsequent address. This corresponds to a little-endian load
61
* of our 64-bit hunks.
62
*/
63
@@ -XXX,XX +XXX,XX @@ static void target_restore_sve_record(CPUARMState *env,
64
}
65
}
66
}
67
+ return true;
39
}
68
}
40
69
41
/* NOTE __get_user and __put_user use host pointers and don't check access.
70
static int target_restore_sigframe(CPUARMState *env,
71
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
72
struct target_sve_context *sve = NULL;
73
uint64_t extra_datap = 0;
74
bool used_extra = false;
75
- int vq = 0, sve_size = 0;
76
+ int sve_size = 0;
77
78
target_restore_general_frame(env, sf);
79
80
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
81
if (sve || size < sizeof(struct target_sve_context)) {
82
goto err;
83
}
84
- if (cpu_isar_feature(aa64_sve, env_archcpu(env))) {
85
- vq = sve_vq(env);
86
- sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(vq), 16);
87
- if (size == sve_size) {
88
- sve = (struct target_sve_context *)ctx;
89
- break;
90
- }
91
- }
92
- goto err;
93
+ sve = (struct target_sve_context *)ctx;
94
+ sve_size = size;
95
+ break;
96
97
case TARGET_EXTRA_MAGIC:
98
if (extra || size != sizeof(struct target_extra_context)) {
99
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
100
}
101
102
/* SVE data, if present, overwrites FPSIMD data. */
103
- if (sve) {
104
- target_restore_sve_record(env, sve, vq);
105
+ if (sve && !target_restore_sve_record(env, sve, sve_size)) {
106
+ goto err;
107
}
108
unlock_user(extra, extra_datap, 0);
109
return 0;
42
--
110
--
43
2.20.1
111
2.25.1
44
45
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
For copy_*_user, only 0 and -TARGET_EFAULT are returned; no need
3
Set the SM bit in the SVE record on signal delivery, create the ZA record.
4
to involve abi_long. Use size_t for lengths. Use bool for the
4
Restore SM and ZA state according to the records present on return.
5
lock_user copy argument. Use ssize_t for target_strlen, because
6
we can't overflow the host memory space.
7
5
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20220708151540.18136-41-richard.henderson@linaro.org
11
Message-id: 20210212184902.1251044-19-richard.henderson@linaro.org
12
[PMM: moved fix for ifdef error to previous commit]
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
10
---
15
linux-user/qemu.h | 12 +++++-------
11
linux-user/aarch64/signal.c | 167 +++++++++++++++++++++++++++++++++---
16
linux-user/uaccess.c | 45 ++++++++++++++++++++++----------------------
12
1 file changed, 154 insertions(+), 13 deletions(-)
17
2 files changed, 28 insertions(+), 29 deletions(-)
18
13
19
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
14
diff --git a/linux-user/aarch64/signal.c b/linux-user/aarch64/signal.c
20
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
21
--- a/linux-user/qemu.h
16
--- a/linux-user/aarch64/signal.c
22
+++ b/linux-user/qemu.h
17
+++ b/linux-user/aarch64/signal.c
23
@@ -XXX,XX +XXX,XX @@
18
@@ -XXX,XX +XXX,XX @@ struct target_sve_context {
24
#include "exec/cpu_ldst.h"
19
25
20
#define TARGET_SVE_SIG_FLAG_SM 1
26
#undef DEBUG_REMAP
21
27
-#ifdef DEBUG_REMAP
22
+#define TARGET_ZA_MAGIC 0x54366345
28
-#endif /* DEBUG_REMAP */
23
+
29
24
+struct target_za_context {
30
#include "exec/user/abitypes.h"
25
+ struct target_aarch64_ctx head;
31
26
+ uint16_t vl;
32
@@ -XXX,XX +XXX,XX @@ static inline bool access_ok(CPUState *cpu, int type,
27
+ uint16_t reserved[3];
33
* buffers between the target and host. These internally perform
28
+ /* The actual ZA data immediately follows. */
34
* locking/unlocking of the memory.
29
+};
35
*/
30
+
36
-abi_long copy_from_user(void *hptr, abi_ulong gaddr, size_t len);
31
+#define TARGET_ZA_SIG_REGS_OFFSET \
37
-abi_long copy_to_user(abi_ulong gaddr, void *hptr, size_t len);
32
+ QEMU_ALIGN_UP(sizeof(struct target_za_context), TARGET_SVE_VQ_BYTES)
38
+int copy_from_user(void *hptr, abi_ulong gaddr, size_t len);
33
+#define TARGET_ZA_SIG_ZAV_OFFSET(VQ, N) \
39
+int copy_to_user(abi_ulong gaddr, void *hptr, size_t len);
34
+ (TARGET_ZA_SIG_REGS_OFFSET + (VQ) * TARGET_SVE_VQ_BYTES * (N))
40
35
+#define TARGET_ZA_SIG_CONTEXT_SIZE(VQ) \
41
/* Functions for accessing guest memory. The tget and tput functions
36
+ TARGET_ZA_SIG_ZAV_OFFSET(VQ, VQ * TARGET_SVE_VQ_BYTES)
42
read/write single values, byteswapping as necessary. The lock_user function
37
+
43
@@ -XXX,XX +XXX,XX @@ abi_long copy_to_user(abi_ulong gaddr, void *hptr, size_t len);
38
struct target_rt_sigframe {
44
39
struct target_siginfo info;
45
/* Lock an area of guest memory into the host. If copy is true then the
40
struct target_ucontext uc;
46
host area will have the same contents as the guest. */
41
@@ -XXX,XX +XXX,XX @@ static void target_setup_end_record(struct target_aarch64_ctx *end)
47
-void *lock_user(int type, abi_ulong guest_addr, long len, int copy);
48
+void *lock_user(int type, abi_ulong guest_addr, size_t len, bool copy);
49
50
/* Unlock an area of guest memory. The first LEN bytes must be
51
flushed back to guest memory. host_ptr = NULL is explicitly
52
allowed and does nothing. */
53
#ifndef DEBUG_REMAP
54
-static inline void unlock_user(void *host_ptr, abi_ulong guest_addr, long len)
55
+static inline void unlock_user(void *host_ptr, abi_ulong guest_addr, size_t len)
56
{ }
57
#else
58
void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
59
@@ -XXX,XX +XXX,XX @@ void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
60
61
/* Return the length of a string in target memory or -TARGET_EFAULT if
62
access error. */
63
-abi_long target_strlen(abi_ulong gaddr);
64
+ssize_t target_strlen(abi_ulong gaddr);
65
66
/* Like lock_user but for null terminated strings. */
67
void *lock_user_string(abi_ulong guest_addr);
68
diff --git a/linux-user/uaccess.c b/linux-user/uaccess.c
69
index XXXXXXX..XXXXXXX 100644
70
--- a/linux-user/uaccess.c
71
+++ b/linux-user/uaccess.c
72
@@ -XXX,XX +XXX,XX @@
73
74
#include "qemu.h"
75
76
-void *lock_user(int type, abi_ulong guest_addr, long len, int copy)
77
+void *lock_user(int type, abi_ulong guest_addr, size_t len, bool copy)
78
{
79
if (!access_ok_untagged(type, guest_addr, len)) {
80
return NULL;
81
@@ -XXX,XX +XXX,XX @@ void *lock_user(int type, abi_ulong guest_addr, long len, int copy)
82
}
42
}
83
43
84
#ifdef DEBUG_REMAP
44
static void target_setup_sve_record(struct target_sve_context *sve,
85
-void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
45
- CPUARMState *env, int vq, int size)
86
+void unlock_user(void *host_ptr, abi_ulong guest_addr, size_t len);
46
+ CPUARMState *env, int size)
87
{
47
{
88
if (!host_ptr) {
48
- int i, j;
89
return;
49
+ int i, j, vq = sve_vq(env);
90
@@ -XXX,XX +XXX,XX @@ void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
50
91
if (host_ptr == g2h_untagged(guest_addr)) {
51
memset(sve, 0, sizeof(*sve));
92
return;
52
__put_user(TARGET_SVE_MAGIC, &sve->head.magic);
93
}
53
@@ -XXX,XX +XXX,XX @@ static void target_setup_sve_record(struct target_sve_context *sve,
94
- if (len > 0) {
54
}
95
+ if (len != 0) {
96
memcpy(g2h_untagged(guest_addr), host_ptr, len);
97
}
98
g_free(host_ptr);
99
@@ -XXX,XX +XXX,XX @@ void unlock_user(void *host_ptr, abi_ulong guest_addr, long len);
100
101
void *lock_user_string(abi_ulong guest_addr)
102
{
103
- abi_long len = target_strlen(guest_addr);
104
+ ssize_t len = target_strlen(guest_addr);
105
if (len < 0) {
106
return NULL;
107
}
108
- return lock_user(VERIFY_READ, guest_addr, (long)(len + 1), 1);
109
+ return lock_user(VERIFY_READ, guest_addr, (size_t)len + 1, 1);
110
}
55
}
111
56
112
/* copy_from_user() and copy_to_user() are usually used to copy data
57
+static void target_setup_za_record(struct target_za_context *za,
113
* buffers between the target and host. These internally perform
58
+ CPUARMState *env, int size)
114
* locking/unlocking of the memory.
59
+{
115
*/
60
+ int vq = sme_vq(env);
116
-abi_long copy_from_user(void *hptr, abi_ulong gaddr, size_t len)
61
+ int vl = vq * TARGET_SVE_VQ_BYTES;
117
+int copy_from_user(void *hptr, abi_ulong gaddr, size_t len)
62
+ int i, j;
118
{
63
+
119
- abi_long ret = 0;
64
+ memset(za, 0, sizeof(*za));
120
- void *ghptr;
65
+ __put_user(TARGET_ZA_MAGIC, &za->head.magic);
121
+ int ret = 0;
66
+ __put_user(size, &za->head.size);
122
+ void *ghptr = lock_user(VERIFY_READ, gaddr, len, 1);
67
+ __put_user(vl, &za->vl);
123
68
+
124
- if ((ghptr = lock_user(VERIFY_READ, gaddr, len, 1))) {
69
+ if (size == TARGET_ZA_SIG_CONTEXT_SIZE(0)) {
125
+ if (ghptr) {
70
+ return;
126
memcpy(hptr, ghptr, len);
71
+ }
127
unlock_user(ghptr, gaddr, 0);
72
+ assert(size == TARGET_ZA_SIG_CONTEXT_SIZE(vq));
128
- } else
73
+
129
+ } else {
74
+ /*
130
ret = -TARGET_EFAULT;
75
+ * Note that ZA vectors are stored as a byte stream,
131
-
76
+ * with each byte element at a subsequent address.
132
+ }
77
+ */
133
return ret;
78
+ for (i = 0; i < vl; ++i) {
79
+ uint64_t *z = (void *)za + TARGET_ZA_SIG_ZAV_OFFSET(vq, i);
80
+ for (j = 0; j < vq * 2; ++j) {
81
+ __put_user_e(env->zarray[i].d[j], z + j, le);
82
+ }
83
+ }
84
+}
85
+
86
static void target_restore_general_frame(CPUARMState *env,
87
struct target_rt_sigframe *sf)
88
{
89
@@ -XXX,XX +XXX,XX @@ static void target_restore_fpsimd_record(CPUARMState *env,
90
91
static bool target_restore_sve_record(CPUARMState *env,
92
struct target_sve_context *sve,
93
- int size)
94
+ int size, int *svcr)
95
{
96
- int i, j, vl, vq;
97
+ int i, j, vl, vq, flags;
98
+ bool sm;
99
100
- if (!cpu_isar_feature(aa64_sve, env_archcpu(env))) {
101
+ __get_user(vl, &sve->vl);
102
+ __get_user(flags, &sve->flags);
103
+
104
+ sm = flags & TARGET_SVE_SIG_FLAG_SM;
105
+
106
+ /* The cpu must support Streaming or Non-streaming SVE. */
107
+ if (sm
108
+ ? !cpu_isar_feature(aa64_sme, env_archcpu(env))
109
+ : !cpu_isar_feature(aa64_sve, env_archcpu(env))) {
110
return false;
111
}
112
113
- __get_user(vl, &sve->vl);
114
- vq = sve_vq(env);
115
+ /*
116
+ * Note that we cannot use sve_vq() because that depends on the
117
+ * current setting of PSTATE.SM, not the state to be restored.
118
+ */
119
+ vq = sve_vqm1_for_el_sm(env, 0, sm) + 1;
120
121
/* Reject mismatched VL. */
122
if (vl != vq * TARGET_SVE_VQ_BYTES) {
123
@@ -XXX,XX +XXX,XX @@ static bool target_restore_sve_record(CPUARMState *env,
124
return false;
125
}
126
127
+ *svcr = FIELD_DP64(*svcr, SVCR, SM, sm);
128
+
129
/*
130
* Note that SVE regs are stored as a byte stream, with each byte element
131
* at a subsequent address. This corresponds to a little-endian load
132
@@ -XXX,XX +XXX,XX @@ static bool target_restore_sve_record(CPUARMState *env,
133
return true;
134
}
134
}
135
135
136
-
136
+static bool target_restore_za_record(CPUARMState *env,
137
-abi_long copy_to_user(abi_ulong gaddr, void *hptr, size_t len)
137
+ struct target_za_context *za,
138
+int copy_to_user(abi_ulong gaddr, void *hptr, size_t len)
138
+ int size, int *svcr)
139
{
139
+{
140
- abi_long ret = 0;
140
+ int i, j, vl, vq;
141
- void *ghptr;
141
+
142
+ int ret = 0;
142
+ if (!cpu_isar_feature(aa64_sme, env_archcpu(env))) {
143
+ void *ghptr = lock_user(VERIFY_WRITE, gaddr, len, 0);
143
+ return false;
144
144
+ }
145
- if ((ghptr = lock_user(VERIFY_WRITE, gaddr, len, 0))) {
145
+
146
+ if (ghptr) {
146
+ __get_user(vl, &za->vl);
147
memcpy(ghptr, hptr, len);
147
+ vq = sme_vq(env);
148
unlock_user(ghptr, gaddr, len);
148
+
149
- } else
149
+ /* Reject mismatched VL. */
150
+ } else {
150
+ if (vl != vq * TARGET_SVE_VQ_BYTES) {
151
ret = -TARGET_EFAULT;
151
+ return false;
152
+ }
152
+ }
153
153
+
154
return ret;
154
+ /* Accept empty record -- used to clear PSTATE.ZA. */
155
}
155
+ if (size <= TARGET_ZA_SIG_CONTEXT_SIZE(0)) {
156
156
+ return true;
157
/* Return the length of a string in target memory or -TARGET_EFAULT if
157
+ }
158
access error */
158
+
159
-abi_long target_strlen(abi_ulong guest_addr1)
159
+ /* Reject non-empty but incomplete record. */
160
+ssize_t target_strlen(abi_ulong guest_addr1)
160
+ if (size < TARGET_ZA_SIG_CONTEXT_SIZE(vq)) {
161
{
161
+ return false;
162
uint8_t *ptr;
162
+ }
163
abi_ulong guest_addr;
163
+
164
- int max_len, len;
164
+ *svcr = FIELD_DP64(*svcr, SVCR, ZA, 1);
165
+ size_t max_len, len;
165
+
166
166
+ for (i = 0; i < vl; ++i) {
167
guest_addr = guest_addr1;
167
+ uint64_t *z = (void *)za + TARGET_ZA_SIG_ZAV_OFFSET(vq, i);
168
for(;;) {
168
+ for (j = 0; j < vq * 2; ++j) {
169
@@ -XXX,XX +XXX,XX @@ abi_long target_strlen(abi_ulong guest_addr1)
169
+ __get_user_e(env->zarray[i].d[j], z + j, le);
170
unlock_user(ptr, guest_addr, 0);
171
guest_addr += len;
172
/* we don't allow wrapping or integer overflow */
173
- if (guest_addr == 0 ||
174
- (guest_addr - guest_addr1) > 0x7fffffff)
175
+ if (guest_addr == 0 || (guest_addr - guest_addr1) > 0x7fffffff) {
176
return -TARGET_EFAULT;
177
- if (len != max_len)
178
+ }
170
+ }
179
+ if (len != max_len) {
171
+ }
172
+ return true;
173
+}
174
+
175
static int target_restore_sigframe(CPUARMState *env,
176
struct target_rt_sigframe *sf)
177
{
178
struct target_aarch64_ctx *ctx, *extra = NULL;
179
struct target_fpsimd_context *fpsimd = NULL;
180
struct target_sve_context *sve = NULL;
181
+ struct target_za_context *za = NULL;
182
uint64_t extra_datap = 0;
183
bool used_extra = false;
184
int sve_size = 0;
185
+ int za_size = 0;
186
+ int svcr = 0;
187
188
target_restore_general_frame(env, sf);
189
190
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
191
sve_size = size;
180
break;
192
break;
193
194
+ case TARGET_ZA_MAGIC:
195
+ if (za || size < sizeof(struct target_za_context)) {
196
+ goto err;
197
+ }
198
+ za = (struct target_za_context *)ctx;
199
+ za_size = size;
200
+ break;
201
+
202
case TARGET_EXTRA_MAGIC:
203
if (extra || size != sizeof(struct target_extra_context)) {
204
goto err;
205
@@ -XXX,XX +XXX,XX @@ static int target_restore_sigframe(CPUARMState *env,
206
}
207
208
/* SVE data, if present, overwrites FPSIMD data. */
209
- if (sve && !target_restore_sve_record(env, sve, sve_size)) {
210
+ if (sve && !target_restore_sve_record(env, sve, sve_size, &svcr)) {
211
goto err;
212
}
213
+ if (za && !target_restore_za_record(env, za, za_size, &svcr)) {
214
+ goto err;
215
+ }
216
+ if (env->svcr != svcr) {
217
+ env->svcr = svcr;
218
+ arm_rebuild_hflags(env);
219
+ }
220
unlock_user(extra, extra_datap, 0);
221
return 0;
222
223
@@ -XXX,XX +XXX,XX @@ static void target_setup_frame(int usig, struct target_sigaction *ka,
224
.total_size = offsetof(struct target_rt_sigframe,
225
uc.tuc_mcontext.__reserved),
226
};
227
- int fpsimd_ofs, fr_ofs, sve_ofs = 0, vq = 0, sve_size = 0;
228
+ int fpsimd_ofs, fr_ofs, sve_ofs = 0, za_ofs = 0;
229
+ int sve_size = 0, za_size = 0;
230
struct target_rt_sigframe *frame;
231
struct target_rt_frame_record *fr;
232
abi_ulong frame_addr, return_addr;
233
@@ -XXX,XX +XXX,XX @@ static void target_setup_frame(int usig, struct target_sigaction *ka,
234
&layout);
235
236
/* SVE state needs saving only if it exists. */
237
- if (cpu_isar_feature(aa64_sve, env_archcpu(env))) {
238
- vq = sve_vq(env);
239
- sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(vq), 16);
240
+ if (cpu_isar_feature(aa64_sve, env_archcpu(env)) ||
241
+ cpu_isar_feature(aa64_sme, env_archcpu(env))) {
242
+ sve_size = QEMU_ALIGN_UP(TARGET_SVE_SIG_CONTEXT_SIZE(sve_vq(env)), 16);
243
sve_ofs = alloc_sigframe_space(sve_size, &layout);
244
}
245
+ if (cpu_isar_feature(aa64_sme, env_archcpu(env))) {
246
+ /* ZA state needs saving only if it is enabled. */
247
+ if (FIELD_EX64(env->svcr, SVCR, ZA)) {
248
+ za_size = TARGET_ZA_SIG_CONTEXT_SIZE(sme_vq(env));
249
+ } else {
250
+ za_size = TARGET_ZA_SIG_CONTEXT_SIZE(0);
181
+ }
251
+ }
182
}
252
+ za_ofs = alloc_sigframe_space(za_size, &layout);
183
return guest_addr - guest_addr1;
253
+ }
184
}
254
255
if (layout.extra_ofs) {
256
/* Reserve space for the extra end marker. The standard end marker
257
@@ -XXX,XX +XXX,XX @@ static void target_setup_frame(int usig, struct target_sigaction *ka,
258
target_setup_end_record((void *)frame + layout.extra_end_ofs);
259
}
260
if (sve_ofs) {
261
- target_setup_sve_record((void *)frame + sve_ofs, env, vq, sve_size);
262
+ target_setup_sve_record((void *)frame + sve_ofs, env, sve_size);
263
+ }
264
+ if (za_ofs) {
265
+ target_setup_za_record((void *)frame + za_ofs, env, za_size);
266
}
267
268
/* Set up the stack frame for unwinding. */
269
@@ -XXX,XX +XXX,XX @@ static void target_setup_frame(int usig, struct target_sigaction *ka,
270
env->btype = 2;
271
}
272
273
+ /*
274
+ * Invoke the signal handler with both SM and ZA disabled.
275
+ * When clearing SM, ResetSVEState, per SMSTOP.
276
+ */
277
+ if (FIELD_EX64(env->svcr, SVCR, SM)) {
278
+ arm_reset_sve_state(env);
279
+ }
280
+ if (env->svcr) {
281
+ env->svcr = 0;
282
+ arm_rebuild_hflags(env);
283
+ }
284
+
285
if (info) {
286
tswap_siginfo(&frame->info, info);
287
env->xregs[1] = frame_addr + offsetof(struct target_rt_sigframe, info);
185
--
288
--
186
2.20.1
289
2.25.1
187
188
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
These prctl fields are required for the function of MTE.
3
Add "sve" to the sve prctl functions, to distinguish
4
them from the coming "sme" prctls with similar names.
4
5
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20210212184902.1251044-24-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-42-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
---
10
linux-user/aarch64/target_syscall.h | 9 ++++++
11
linux-user/aarch64/target_prctl.h | 8 ++++----
11
linux-user/syscall.c | 43 +++++++++++++++++++++++++++++
12
linux-user/syscall.c | 12 ++++++------
12
2 files changed, 52 insertions(+)
13
2 files changed, 10 insertions(+), 10 deletions(-)
13
14
14
diff --git a/linux-user/aarch64/target_syscall.h b/linux-user/aarch64/target_syscall.h
15
diff --git a/linux-user/aarch64/target_prctl.h b/linux-user/aarch64/target_prctl.h
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/linux-user/aarch64/target_syscall.h
17
--- a/linux-user/aarch64/target_prctl.h
17
+++ b/linux-user/aarch64/target_syscall.h
18
+++ b/linux-user/aarch64/target_prctl.h
18
@@ -XXX,XX +XXX,XX @@ struct target_pt_regs {
19
@@ -XXX,XX +XXX,XX @@
19
#define TARGET_PR_SET_TAGGED_ADDR_CTRL 55
20
#ifndef AARCH64_TARGET_PRCTL_H
20
#define TARGET_PR_GET_TAGGED_ADDR_CTRL 56
21
#define AARCH64_TARGET_PRCTL_H
21
# define TARGET_PR_TAGGED_ADDR_ENABLE (1UL << 0)
22
22
+/* MTE tag check fault modes */
23
-static abi_long do_prctl_get_vl(CPUArchState *env)
23
+# define TARGET_PR_MTE_TCF_SHIFT 1
24
+static abi_long do_prctl_sve_get_vl(CPUArchState *env)
24
+# define TARGET_PR_MTE_TCF_NONE (0UL << TARGET_PR_MTE_TCF_SHIFT)
25
{
25
+# define TARGET_PR_MTE_TCF_SYNC (1UL << TARGET_PR_MTE_TCF_SHIFT)
26
ARMCPU *cpu = env_archcpu(env);
26
+# define TARGET_PR_MTE_TCF_ASYNC (2UL << TARGET_PR_MTE_TCF_SHIFT)
27
if (cpu_isar_feature(aa64_sve, cpu)) {
27
+# define TARGET_PR_MTE_TCF_MASK (3UL << TARGET_PR_MTE_TCF_SHIFT)
28
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_get_vl(CPUArchState *env)
28
+/* MTE tag inclusion mask */
29
}
29
+# define TARGET_PR_MTE_TAG_SHIFT 3
30
return -TARGET_EINVAL;
30
+# define TARGET_PR_MTE_TAG_MASK (0xffffUL << TARGET_PR_MTE_TAG_SHIFT)
31
}
31
32
-#define do_prctl_get_vl do_prctl_get_vl
32
#endif /* AARCH64_TARGET_SYSCALL_H */
33
+#define do_prctl_sve_get_vl do_prctl_sve_get_vl
34
35
-static abi_long do_prctl_set_vl(CPUArchState *env, abi_long arg2)
36
+static abi_long do_prctl_sve_set_vl(CPUArchState *env, abi_long arg2)
37
{
38
/*
39
* We cannot support either PR_SVE_SET_VL_ONEXEC or PR_SVE_VL_INHERIT.
40
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_set_vl(CPUArchState *env, abi_long arg2)
41
}
42
return -TARGET_EINVAL;
43
}
44
-#define do_prctl_set_vl do_prctl_set_vl
45
+#define do_prctl_sve_set_vl do_prctl_sve_set_vl
46
47
static abi_long do_prctl_reset_keys(CPUArchState *env, abi_long arg2)
48
{
33
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
49
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
34
index XXXXXXX..XXXXXXX 100644
50
index XXXXXXX..XXXXXXX 100644
35
--- a/linux-user/syscall.c
51
--- a/linux-user/syscall.c
36
+++ b/linux-user/syscall.c
52
+++ b/linux-user/syscall.c
37
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
53
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_inval1(CPUArchState *env, abi_long arg2)
38
{
54
#ifndef do_prctl_set_fp_mode
39
abi_ulong valid_mask = TARGET_PR_TAGGED_ADDR_ENABLE;
55
#define do_prctl_set_fp_mode do_prctl_inval1
40
CPUARMState *env = cpu_env;
56
#endif
41
+ ARMCPU *cpu = env_archcpu(env);
57
-#ifndef do_prctl_get_vl
42
+
58
-#define do_prctl_get_vl do_prctl_inval0
43
+ if (cpu_isar_feature(aa64_mte, cpu)) {
59
+#ifndef do_prctl_sve_get_vl
44
+ valid_mask |= TARGET_PR_MTE_TCF_MASK;
60
+#define do_prctl_sve_get_vl do_prctl_inval0
45
+ valid_mask |= TARGET_PR_MTE_TAG_MASK;
61
#endif
46
+ }
62
-#ifndef do_prctl_set_vl
47
63
-#define do_prctl_set_vl do_prctl_inval1
48
if ((arg2 & ~valid_mask) || arg3 || arg4 || arg5) {
64
+#ifndef do_prctl_sve_set_vl
49
return -TARGET_EINVAL;
65
+#define do_prctl_sve_set_vl do_prctl_inval1
50
}
66
#endif
51
env->tagged_addr_enable = arg2 & TARGET_PR_TAGGED_ADDR_ENABLE;
67
#ifndef do_prctl_reset_keys
52
+
68
#define do_prctl_reset_keys do_prctl_inval1
53
+ if (cpu_isar_feature(aa64_mte, cpu)) {
69
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl(CPUArchState *env, abi_long option, abi_long arg2,
54
+ switch (arg2 & TARGET_PR_MTE_TCF_MASK) {
70
case PR_SET_FP_MODE:
55
+ case TARGET_PR_MTE_TCF_NONE:
71
return do_prctl_set_fp_mode(env, arg2);
56
+ case TARGET_PR_MTE_TCF_SYNC:
72
case PR_SVE_GET_VL:
57
+ case TARGET_PR_MTE_TCF_ASYNC:
73
- return do_prctl_get_vl(env);
58
+ break;
74
+ return do_prctl_sve_get_vl(env);
59
+ default:
75
case PR_SVE_SET_VL:
60
+ return -EINVAL;
76
- return do_prctl_set_vl(env, arg2);
61
+ }
77
+ return do_prctl_sve_set_vl(env, arg2);
62
+
78
case PR_PAC_RESET_KEYS:
63
+ /*
79
if (arg3 || arg4 || arg5) {
64
+ * Write PR_MTE_TCF to SCTLR_EL1[TCF0].
80
return -TARGET_EINVAL;
65
+ * Note that the syscall values are consistent with hw.
66
+ */
67
+ env->cp15.sctlr_el[1] =
68
+ deposit64(env->cp15.sctlr_el[1], 38, 2,
69
+ arg2 >> TARGET_PR_MTE_TCF_SHIFT);
70
+
71
+ /*
72
+ * Write PR_MTE_TAG to GCR_EL1[Exclude].
73
+ * Note that the syscall uses an include mask,
74
+ * and hardware uses an exclude mask -- invert.
75
+ */
76
+ env->cp15.gcr_el1 =
77
+ deposit64(env->cp15.gcr_el1, 0, 16,
78
+ ~arg2 >> TARGET_PR_MTE_TAG_SHIFT);
79
+ arm_rebuild_hflags(env);
80
+ }
81
return 0;
82
}
83
case TARGET_PR_GET_TAGGED_ADDR_CTRL:
84
{
85
abi_long ret = 0;
86
CPUARMState *env = cpu_env;
87
+ ARMCPU *cpu = env_archcpu(env);
88
89
if (arg2 || arg3 || arg4 || arg5) {
90
return -TARGET_EINVAL;
91
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
92
if (env->tagged_addr_enable) {
93
ret |= TARGET_PR_TAGGED_ADDR_ENABLE;
94
}
95
+ if (cpu_isar_feature(aa64_mte, cpu)) {
96
+ /* See above. */
97
+ ret |= (extract64(env->cp15.sctlr_el[1], 38, 2)
98
+ << TARGET_PR_MTE_TCF_SHIFT);
99
+ ret = deposit64(ret, TARGET_PR_MTE_TAG_SHIFT, 16,
100
+ ~env->cp15.gcr_el1);
101
+ }
102
return ret;
103
}
104
#endif /* AARCH64 */
105
--
81
--
106
2.20.1
82
2.25.1
107
108
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We define target_mmap et al as untagged, so that they can be
3
These prctl set the Streaming SVE vector length, which may
4
used from the binary loaders. Explicitly call cpu_untagged_addr
4
be completely different from the Normal SVE vector length.
5
for munmap, mprotect, mremap syscall entry points.
6
7
Add a few comments for the syscalls that are exempted by the
8
kernel's tagged-address-abi.rst.
9
5
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20210212184902.1251044-14-richard.henderson@linaro.org
8
Message-id: 20220708151540.18136-43-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
10
---
15
linux-user/syscall.c | 11 +++++++++++
11
linux-user/aarch64/target_prctl.h | 54 +++++++++++++++++++++++++++++++
16
1 file changed, 11 insertions(+)
12
linux-user/syscall.c | 16 +++++++++
13
2 files changed, 70 insertions(+)
17
14
15
diff --git a/linux-user/aarch64/target_prctl.h b/linux-user/aarch64/target_prctl.h
16
index XXXXXXX..XXXXXXX 100644
17
--- a/linux-user/aarch64/target_prctl.h
18
+++ b/linux-user/aarch64/target_prctl.h
19
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_sve_get_vl(CPUArchState *env)
20
{
21
ARMCPU *cpu = env_archcpu(env);
22
if (cpu_isar_feature(aa64_sve, cpu)) {
23
+ /* PSTATE.SM is always unset on syscall entry. */
24
return sve_vq(env) * 16;
25
}
26
return -TARGET_EINVAL;
27
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_sve_set_vl(CPUArchState *env, abi_long arg2)
28
&& arg2 >= 0 && arg2 <= 512 * 16 && !(arg2 & 15)) {
29
uint32_t vq, old_vq;
30
31
+ /* PSTATE.SM is always unset on syscall entry. */
32
old_vq = sve_vq(env);
33
34
/*
35
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_sve_set_vl(CPUArchState *env, abi_long arg2)
36
}
37
#define do_prctl_sve_set_vl do_prctl_sve_set_vl
38
39
+static abi_long do_prctl_sme_get_vl(CPUArchState *env)
40
+{
41
+ ARMCPU *cpu = env_archcpu(env);
42
+ if (cpu_isar_feature(aa64_sme, cpu)) {
43
+ return sme_vq(env) * 16;
44
+ }
45
+ return -TARGET_EINVAL;
46
+}
47
+#define do_prctl_sme_get_vl do_prctl_sme_get_vl
48
+
49
+static abi_long do_prctl_sme_set_vl(CPUArchState *env, abi_long arg2)
50
+{
51
+ /*
52
+ * We cannot support either PR_SME_SET_VL_ONEXEC or PR_SME_VL_INHERIT.
53
+ * Note the kernel definition of sve_vl_valid allows for VQ=512,
54
+ * i.e. VL=8192, even though the architectural maximum is VQ=16.
55
+ */
56
+ if (cpu_isar_feature(aa64_sme, env_archcpu(env))
57
+ && arg2 >= 0 && arg2 <= 512 * 16 && !(arg2 & 15)) {
58
+ int vq, old_vq;
59
+
60
+ old_vq = sme_vq(env);
61
+
62
+ /*
63
+ * Bound the value of vq, so that we know that it fits into
64
+ * the 4-bit field in SMCR_EL1. Because PSTATE.SM is cleared
65
+ * on syscall entry, we are not modifying the current SVE
66
+ * vector length.
67
+ */
68
+ vq = MAX(arg2 / 16, 1);
69
+ vq = MIN(vq, 16);
70
+ env->vfp.smcr_el[1] =
71
+ FIELD_DP64(env->vfp.smcr_el[1], SMCR, LEN, vq - 1);
72
+
73
+ /* Delay rebuilding hflags until we know if ZA must change. */
74
+ vq = sve_vqm1_for_el_sm(env, 0, true) + 1;
75
+
76
+ if (vq != old_vq) {
77
+ /*
78
+ * PSTATE.ZA state is cleared on any change to SVL.
79
+ * We need not call arm_rebuild_hflags because PSTATE.SM was
80
+ * cleared on syscall entry, so this hasn't changed VL.
81
+ */
82
+ env->svcr = FIELD_DP64(env->svcr, SVCR, ZA, 0);
83
+ arm_rebuild_hflags(env);
84
+ }
85
+ return vq * 16;
86
+ }
87
+ return -TARGET_EINVAL;
88
+}
89
+#define do_prctl_sme_set_vl do_prctl_sme_set_vl
90
+
91
static abi_long do_prctl_reset_keys(CPUArchState *env, abi_long arg2)
92
{
93
ARMCPU *cpu = env_archcpu(env);
18
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
94
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
19
index XXXXXXX..XXXXXXX 100644
95
index XXXXXXX..XXXXXXX 100644
20
--- a/linux-user/syscall.c
96
--- a/linux-user/syscall.c
21
+++ b/linux-user/syscall.c
97
+++ b/linux-user/syscall.c
22
@@ -XXX,XX +XXX,XX @@ abi_long do_brk(abi_ulong new_brk)
98
@@ -XXX,XX +XXX,XX @@ abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr)
23
abi_long mapped_addr;
99
#ifndef PR_SET_SYSCALL_USER_DISPATCH
24
abi_ulong new_alloc_size;
100
# define PR_SET_SYSCALL_USER_DISPATCH 59
25
26
+ /* brk pointers are always untagged */
27
+
28
DEBUGF_BRK("do_brk(" TARGET_ABI_FMT_lx ") -> ", new_brk);
29
30
if (!new_brk) {
31
@@ -XXX,XX +XXX,XX @@ static inline abi_ulong do_shmat(CPUArchState *cpu_env,
32
int i,ret;
33
abi_ulong shmlba;
34
35
+ /* shmat pointers are always untagged */
36
+
37
/* find out the length of the shared memory segment */
38
ret = get_errno(shmctl(shmid, IPC_STAT, &shm_info));
39
if (is_error(ret)) {
40
@@ -XXX,XX +XXX,XX @@ static inline abi_long do_shmdt(abi_ulong shmaddr)
41
int i;
42
abi_long rv;
43
44
+ /* shmdt pointers are always untagged */
45
+
46
mmap_lock();
47
48
for (i = 0; i < N_SHM_REGIONS; ++i) {
49
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
50
v5, v6));
51
}
52
#else
53
+ /* mmap pointers are always untagged */
54
ret = get_errno(target_mmap(arg1, arg2, arg3,
55
target_to_host_bitmask(arg4, mmap_flags_tbl),
56
arg5,
57
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
58
return get_errno(ret);
59
#endif
101
#endif
60
case TARGET_NR_munmap:
102
+#ifndef PR_SME_SET_VL
61
+ arg1 = cpu_untagged_addr(cpu, arg1);
103
+# define PR_SME_SET_VL 63
62
return get_errno(target_munmap(arg1, arg2));
104
+# define PR_SME_GET_VL 64
63
case TARGET_NR_mprotect:
105
+# define PR_SME_VL_LEN_MASK 0xffff
64
+ arg1 = cpu_untagged_addr(cpu, arg1);
106
+# define PR_SME_VL_INHERIT (1 << 17)
65
{
107
+#endif
66
TaskState *ts = cpu->opaque;
108
67
/* Special hack to detect libc making the stack executable. */
109
#include "target_prctl.h"
68
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
110
69
return get_errno(target_mprotect(arg1, arg2, arg3));
111
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl_inval1(CPUArchState *env, abi_long arg2)
70
#ifdef TARGET_NR_mremap
112
#ifndef do_prctl_set_unalign
71
case TARGET_NR_mremap:
113
#define do_prctl_set_unalign do_prctl_inval1
72
+ arg1 = cpu_untagged_addr(cpu, arg1);
73
+ /* mremap new_addr (arg5) is always untagged */
74
return get_errno(target_mremap(arg1, arg2, arg3, arg4, arg5));
75
#endif
114
#endif
76
/* ??? msync/mlock/munlock are broken for softmmu. */
115
+#ifndef do_prctl_sme_get_vl
116
+#define do_prctl_sme_get_vl do_prctl_inval0
117
+#endif
118
+#ifndef do_prctl_sme_set_vl
119
+#define do_prctl_sme_set_vl do_prctl_inval1
120
+#endif
121
122
static abi_long do_prctl(CPUArchState *env, abi_long option, abi_long arg2,
123
abi_long arg3, abi_long arg4, abi_long arg5)
124
@@ -XXX,XX +XXX,XX @@ static abi_long do_prctl(CPUArchState *env, abi_long option, abi_long arg2,
125
return do_prctl_sve_get_vl(env);
126
case PR_SVE_SET_VL:
127
return do_prctl_sve_set_vl(env, arg2);
128
+ case PR_SME_GET_VL:
129
+ return do_prctl_sme_get_vl(env);
130
+ case PR_SME_SET_VL:
131
+ return do_prctl_sme_set_vl(env, arg2);
132
case PR_PAC_RESET_KEYS:
133
if (arg3 || arg4 || arg5) {
134
return -TARGET_EINVAL;
77
--
135
--
78
2.20.1
136
2.25.1
79
80
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
2
3
We were fudging TBI1 enabled to speed up the generated code.
3
There's no reason to set CPACR_EL1.ZEN if SVE disabled.
4
Now that we've improved the code generation, remove this.
5
Also, tidy the comment to reflect the current code.
6
7
The pauth test was testing a kernel address (-1) and making
8
incorrect assumptions about TBI1; stick to userland addresses.
9
4
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20210212184902.1251044-23-richard.henderson@linaro.org
7
Message-id: 20220708151540.18136-44-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
9
---
15
target/arm/internals.h | 4 ++--
10
target/arm/cpu.c | 7 +++----
16
target/arm/cpu.c | 10 +++-------
11
1 file changed, 3 insertions(+), 4 deletions(-)
17
tests/tcg/aarch64/pauth-2.c | 1 -
18
3 files changed, 5 insertions(+), 10 deletions(-)
19
12
20
diff --git a/target/arm/internals.h b/target/arm/internals.h
21
index XXXXXXX..XXXXXXX 100644
22
--- a/target/arm/internals.h
23
+++ b/target/arm/internals.h
24
@@ -XXX,XX +XXX,XX @@ static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
25
*/
26
static inline uint64_t useronly_clean_ptr(uint64_t ptr)
27
{
28
- /* TBI is known to be enabled. */
29
#ifdef CONFIG_USER_ONLY
30
- ptr = sextract64(ptr, 0, 56);
31
+ /* TBI0 is known to be enabled, while TBI1 is disabled. */
32
+ ptr &= sextract64(ptr, 0, 56);
33
#endif
34
return ptr;
35
}
36
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
13
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
37
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/cpu.c
15
--- a/target/arm/cpu.c
39
+++ b/target/arm/cpu.c
16
+++ b/target/arm/cpu.c
40
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
41
env->vfp.zcr_el[1] = MIN(cpu->sve_max_vq - 1, 3);
18
/* and to the FP/Neon instructions */
19
env->cp15.cpacr_el1 = FIELD_DP64(env->cp15.cpacr_el1,
20
CPACR_EL1, FPEN, 3);
21
- /* and to the SVE instructions */
22
- env->cp15.cpacr_el1 = FIELD_DP64(env->cp15.cpacr_el1,
23
- CPACR_EL1, ZEN, 3);
24
- /* with reasonable vector length */
25
+ /* and to the SVE instructions, with default vector length */
26
if (cpu_isar_feature(aa64_sve, cpu)) {
27
+ env->cp15.cpacr_el1 = FIELD_DP64(env->cp15.cpacr_el1,
28
+ CPACR_EL1, ZEN, 3);
29
env->vfp.zcr_el[1] = cpu->sve_default_vq - 1;
42
}
30
}
43
/*
31
/*
44
- * Enable TBI0 and TBI1. While the real kernel only enables TBI0,
45
- * turning on both here will produce smaller code and otherwise
46
- * make no difference to the user-level emulation.
47
- *
48
- * In sve_probe_page, we assume that this is set.
49
- * Do not modify this without other changes.
50
+ * Enable TBI0 but not TBI1.
51
+ * Note that this must match useronly_clean_ptr.
52
*/
53
- env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
54
+ env->cp15.tcr_el[1].raw_tcr = (1ULL << 37);
55
#else
56
/* Reset into the highest available EL */
57
if (arm_feature(env, ARM_FEATURE_EL3)) {
58
diff --git a/tests/tcg/aarch64/pauth-2.c b/tests/tcg/aarch64/pauth-2.c
59
index XXXXXXX..XXXXXXX 100644
60
--- a/tests/tcg/aarch64/pauth-2.c
61
+++ b/tests/tcg/aarch64/pauth-2.c
62
@@ -XXX,XX +XXX,XX @@ void do_test(uint64_t value)
63
int main()
64
{
65
do_test(0);
66
- do_test(-1);
67
do_test(0xda004acedeadbeefull);
68
return 0;
69
}
70
--
32
--
71
2.20.1
33
2.25.1
72
73
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Enable SME, TPIDR2_EL0, and FA64 if supported by the cpu.
2
4
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20210212184902.1251044-31-richard.henderson@linaro.org
7
Message-id: 20220708151540.18136-45-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
9
---
8
target/arm/cpu.c | 15 +++++++++++++++
10
target/arm/cpu.c | 11 +++++++++++
9
1 file changed, 15 insertions(+)
11
1 file changed, 11 insertions(+)
10
12
11
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
13
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
12
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/cpu.c
15
--- a/target/arm/cpu.c
14
+++ b/target/arm/cpu.c
16
+++ b/target/arm/cpu.c
15
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
16
* Note that this must match useronly_clean_ptr.
18
CPACR_EL1, ZEN, 3);
17
*/
19
env->vfp.zcr_el[1] = cpu->sve_default_vq - 1;
18
env->cp15.tcr_el[1].raw_tcr = (1ULL << 37);
20
}
19
+
21
+ /* and for SME instructions, with default vector length, and TPIDR2 */
20
+ /* Enable MTE */
22
+ if (cpu_isar_feature(aa64_sme, cpu)) {
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
23
+ env->cp15.sctlr_el[1] |= SCTLR_EnTP2;
22
+ /* Enable tag access, but leave TCF0 as No Effect (0). */
24
+ env->cp15.cpacr_el1 = FIELD_DP64(env->cp15.cpacr_el1,
23
+ env->cp15.sctlr_el[1] |= SCTLR_ATA0;
25
+ CPACR_EL1, SMEN, 3);
24
+ /*
26
+ env->vfp.smcr_el[1] = cpu->sme_default_vq - 1;
25
+ * Exclude all tags, so that tag 0 is always used.
27
+ if (cpu_isar_feature(aa64_sme_fa64, cpu)) {
26
+ * This corresponds to Linux current->thread.gcr_incl = 0.
28
+ env->vfp.smcr_el[1] = FIELD_DP64(env->vfp.smcr_el[1],
27
+ *
29
+ SMCR, FA64, 1);
28
+ * Set RRND, so that helper_irg() will generate a seed later.
30
+ }
29
+ * Here in cpu_reset(), the crypto subsystem has not yet been
30
+ * initialized.
31
+ */
32
+ env->cp15.gcr_el1 = 0x1ffff;
33
+ }
31
+ }
34
#else
32
/*
35
/* Reset into the highest available EL */
33
* Enable 48-bit address space (TODO: take reserved_va into account).
36
if (arm_feature(env, ARM_FEATURE_EL3)) {
34
* Enable TBI0 but not TBI1.
37
--
35
--
38
2.20.1
36
2.25.1
39
40
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Richard Henderson <richard.henderson@linaro.org>
2
3
Use g2h_untagged in contexts that have no cpu, e.g. the binary
4
loaders that operate before the primary cpu is created. As a
5
colollary, target_mmap and friends must use untagged addresses,
6
since they are used by the loaders.
7
8
Use g2h_untagged on values returned from target_mmap, as the
9
kernel never applies a tag itself.
10
11
Use g2h_untagged on all pc values. The only current user of
12
tags, aarch64, removes tags from code addresses upon branch,
13
so "pc" is always untagged.
14
15
Use g2h with the cpu context on hand wherever possible.
16
17
Use g2h_untagged in lock_user, which will be updated soon.
18
2
19
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
20
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
21
Message-id: 20210212184902.1251044-13-richard.henderson@linaro.org
5
Message-id: 20220708151540.18136-46-richard.henderson@linaro.org
22
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
23
---
7
---
24
bsd-user/qemu.h | 8 ++--
8
linux-user/elfload.c | 20 ++++++++++++++++++++
25
include/exec/cpu_ldst.h | 12 +++++-
9
1 file changed, 20 insertions(+)
26
include/exec/exec-all.h | 2 +-
27
linux-user/qemu.h | 6 +--
28
accel/tcg/translate-all.c | 4 +-
29
accel/tcg/user-exec.c | 48 ++++++++++++------------
30
bsd-user/elfload.c | 2 +-
31
bsd-user/main.c | 4 +-
32
bsd-user/mmap.c | 23 ++++++------
33
linux-user/elfload.c | 12 +++---
34
linux-user/flatload.c | 2 +-
35
linux-user/hppa/cpu_loop.c | 31 ++++++++--------
36
linux-user/i386/cpu_loop.c | 4 +-
37
linux-user/mmap.c | 45 +++++++++++-----------
38
linux-user/ppc/signal.c | 4 +-
39
linux-user/syscall.c | 72 +++++++++++++++++++-----------------
40
target/arm/helper-a64.c | 4 +-
41
target/hppa/op_helper.c | 2 +-
42
target/i386/tcg/mem_helper.c | 2 +-
43
target/s390x/mem_helper.c | 4 +-
44
20 files changed, 154 insertions(+), 137 deletions(-)
45
10
46
diff --git a/bsd-user/qemu.h b/bsd-user/qemu.h
47
index XXXXXXX..XXXXXXX 100644
48
--- a/bsd-user/qemu.h
49
+++ b/bsd-user/qemu.h
50
@@ -XXX,XX +XXX,XX @@ static inline void *lock_user(int type, abi_ulong guest_addr, long len, int copy
51
void *addr;
52
addr = g_malloc(len);
53
if (copy)
54
- memcpy(addr, g2h(guest_addr), len);
55
+ memcpy(addr, g2h_untagged(guest_addr), len);
56
else
57
memset(addr, 0, len);
58
return addr;
59
}
60
#else
61
- return g2h(guest_addr);
62
+ return g2h_untagged(guest_addr);
63
#endif
64
}
65
66
@@ -XXX,XX +XXX,XX @@ static inline void unlock_user(void *host_ptr, abi_ulong guest_addr,
67
#ifdef DEBUG_REMAP
68
if (!host_ptr)
69
return;
70
- if (host_ptr == g2h(guest_addr))
71
+ if (host_ptr == g2h_untagged(guest_addr))
72
return;
73
if (len > 0)
74
- memcpy(g2h(guest_addr), host_ptr, len);
75
+ memcpy(g2h_untagged(guest_addr), host_ptr, len);
76
g_free(host_ptr);
77
#endif
78
}
79
diff --git a/include/exec/cpu_ldst.h b/include/exec/cpu_ldst.h
80
index XXXXXXX..XXXXXXX 100644
81
--- a/include/exec/cpu_ldst.h
82
+++ b/include/exec/cpu_ldst.h
83
@@ -XXX,XX +XXX,XX @@ static inline abi_ptr cpu_untagged_addr(CPUState *cs, abi_ptr x)
84
#endif
85
86
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
87
-#define g2h(x) ((void *)((uintptr_t)(abi_ptr)(x) + guest_base))
88
+static inline void *g2h_untagged(abi_ptr x)
89
+{
90
+ return (void *)((uintptr_t)(x) + guest_base);
91
+}
92
+
93
+static inline void *g2h(CPUState *cs, abi_ptr x)
94
+{
95
+ return g2h_untagged(cpu_untagged_addr(cs, x));
96
+}
97
98
static inline bool guest_addr_valid(abi_ulong x)
99
{
100
@@ -XXX,XX +XXX,XX @@ static inline int cpu_ldsw_code(CPUArchState *env, abi_ptr addr)
101
static inline void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
102
MMUAccessType access_type, int mmu_idx)
103
{
104
- return g2h(addr);
105
+ return g2h(env_cpu(env), addr);
106
}
107
#else
108
void *tlb_vaddr_to_host(CPUArchState *env, abi_ptr addr,
109
diff --git a/include/exec/exec-all.h b/include/exec/exec-all.h
110
index XXXXXXX..XXXXXXX 100644
111
--- a/include/exec/exec-all.h
112
+++ b/include/exec/exec-all.h
113
@@ -XXX,XX +XXX,XX @@ static inline tb_page_addr_t get_page_addr_code_hostp(CPUArchState *env,
114
void **hostp)
115
{
116
if (hostp) {
117
- *hostp = g2h(addr);
118
+ *hostp = g2h_untagged(addr);
119
}
120
return addr;
121
}
122
diff --git a/linux-user/qemu.h b/linux-user/qemu.h
123
index XXXXXXX..XXXXXXX 100644
124
--- a/linux-user/qemu.h
125
+++ b/linux-user/qemu.h
126
@@ -XXX,XX +XXX,XX @@ static inline void *lock_user(int type, abi_ulong guest_addr, long len, int copy
127
return addr;
128
}
129
#else
130
- return g2h(guest_addr);
131
+ return g2h_untagged(guest_addr);
132
#endif
133
}
134
135
@@ -XXX,XX +XXX,XX @@ static inline void unlock_user(void *host_ptr, abi_ulong guest_addr,
136
#ifdef DEBUG_REMAP
137
if (!host_ptr)
138
return;
139
- if (host_ptr == g2h(guest_addr))
140
+ if (host_ptr == g2h_untagged(guest_addr))
141
return;
142
if (len > 0)
143
- memcpy(g2h(guest_addr), host_ptr, len);
144
+ memcpy(g2h_untagged(guest_addr), host_ptr, len);
145
g_free(host_ptr);
146
#endif
147
}
148
diff --git a/accel/tcg/translate-all.c b/accel/tcg/translate-all.c
149
index XXXXXXX..XXXXXXX 100644
150
--- a/accel/tcg/translate-all.c
151
+++ b/accel/tcg/translate-all.c
152
@@ -XXX,XX +XXX,XX @@ static inline void tb_page_add(PageDesc *p, TranslationBlock *tb,
153
prot |= p2->flags;
154
p2->flags &= ~PAGE_WRITE;
155
}
156
- mprotect(g2h(page_addr), qemu_host_page_size,
157
+ mprotect(g2h_untagged(page_addr), qemu_host_page_size,
158
(prot & PAGE_BITS) & ~PAGE_WRITE);
159
if (DEBUG_TB_INVALIDATE_GATE) {
160
printf("protecting code page: 0x" TB_PAGE_ADDR_FMT "\n", page_addr);
161
@@ -XXX,XX +XXX,XX @@ int page_unprotect(target_ulong address, uintptr_t pc)
162
}
163
#endif
164
}
165
- mprotect((void *)g2h(host_start), qemu_host_page_size,
166
+ mprotect((void *)g2h_untagged(host_start), qemu_host_page_size,
167
prot & PAGE_BITS);
168
}
169
mmap_unlock();
170
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
171
index XXXXXXX..XXXXXXX 100644
172
--- a/accel/tcg/user-exec.c
173
+++ b/accel/tcg/user-exec.c
174
@@ -XXX,XX +XXX,XX @@ int probe_access_flags(CPUArchState *env, target_ulong addr,
175
int flags;
176
177
flags = probe_access_internal(env, addr, 0, access_type, nonfault, ra);
178
- *phost = flags ? NULL : g2h(addr);
179
+ *phost = flags ? NULL : g2h(env_cpu(env), addr);
180
return flags;
181
}
182
183
@@ -XXX,XX +XXX,XX @@ void *probe_access(CPUArchState *env, target_ulong addr, int size,
184
flags = probe_access_internal(env, addr, size, access_type, false, ra);
185
g_assert(flags == 0);
186
187
- return size ? g2h(addr) : NULL;
188
+ return size ? g2h(env_cpu(env), addr) : NULL;
189
}
190
191
#if defined(__i386__)
192
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_ldub_data(CPUArchState *env, abi_ptr ptr)
193
uint16_t meminfo = trace_mem_get_info(MO_UB, MMU_USER_IDX, false);
194
195
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
196
- ret = ldub_p(g2h(ptr));
197
+ ret = ldub_p(g2h(env_cpu(env), ptr));
198
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
199
return ret;
200
}
201
@@ -XXX,XX +XXX,XX @@ int cpu_ldsb_data(CPUArchState *env, abi_ptr ptr)
202
uint16_t meminfo = trace_mem_get_info(MO_SB, MMU_USER_IDX, false);
203
204
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
205
- ret = ldsb_p(g2h(ptr));
206
+ ret = ldsb_p(g2h(env_cpu(env), ptr));
207
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
208
return ret;
209
}
210
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_lduw_be_data(CPUArchState *env, abi_ptr ptr)
211
uint16_t meminfo = trace_mem_get_info(MO_BEUW, MMU_USER_IDX, false);
212
213
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
214
- ret = lduw_be_p(g2h(ptr));
215
+ ret = lduw_be_p(g2h(env_cpu(env), ptr));
216
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
217
return ret;
218
}
219
@@ -XXX,XX +XXX,XX @@ int cpu_ldsw_be_data(CPUArchState *env, abi_ptr ptr)
220
uint16_t meminfo = trace_mem_get_info(MO_BESW, MMU_USER_IDX, false);
221
222
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
223
- ret = ldsw_be_p(g2h(ptr));
224
+ ret = ldsw_be_p(g2h(env_cpu(env), ptr));
225
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
226
return ret;
227
}
228
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_ldl_be_data(CPUArchState *env, abi_ptr ptr)
229
uint16_t meminfo = trace_mem_get_info(MO_BEUL, MMU_USER_IDX, false);
230
231
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
232
- ret = ldl_be_p(g2h(ptr));
233
+ ret = ldl_be_p(g2h(env_cpu(env), ptr));
234
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
235
return ret;
236
}
237
@@ -XXX,XX +XXX,XX @@ uint64_t cpu_ldq_be_data(CPUArchState *env, abi_ptr ptr)
238
uint16_t meminfo = trace_mem_get_info(MO_BEQ, MMU_USER_IDX, false);
239
240
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
241
- ret = ldq_be_p(g2h(ptr));
242
+ ret = ldq_be_p(g2h(env_cpu(env), ptr));
243
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
244
return ret;
245
}
246
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_lduw_le_data(CPUArchState *env, abi_ptr ptr)
247
uint16_t meminfo = trace_mem_get_info(MO_LEUW, MMU_USER_IDX, false);
248
249
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
250
- ret = lduw_le_p(g2h(ptr));
251
+ ret = lduw_le_p(g2h(env_cpu(env), ptr));
252
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
253
return ret;
254
}
255
@@ -XXX,XX +XXX,XX @@ int cpu_ldsw_le_data(CPUArchState *env, abi_ptr ptr)
256
uint16_t meminfo = trace_mem_get_info(MO_LESW, MMU_USER_IDX, false);
257
258
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
259
- ret = ldsw_le_p(g2h(ptr));
260
+ ret = ldsw_le_p(g2h(env_cpu(env), ptr));
261
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
262
return ret;
263
}
264
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_ldl_le_data(CPUArchState *env, abi_ptr ptr)
265
uint16_t meminfo = trace_mem_get_info(MO_LEUL, MMU_USER_IDX, false);
266
267
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
268
- ret = ldl_le_p(g2h(ptr));
269
+ ret = ldl_le_p(g2h(env_cpu(env), ptr));
270
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
271
return ret;
272
}
273
@@ -XXX,XX +XXX,XX @@ uint64_t cpu_ldq_le_data(CPUArchState *env, abi_ptr ptr)
274
uint16_t meminfo = trace_mem_get_info(MO_LEQ, MMU_USER_IDX, false);
275
276
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
277
- ret = ldq_le_p(g2h(ptr));
278
+ ret = ldq_le_p(g2h(env_cpu(env), ptr));
279
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
280
return ret;
281
}
282
@@ -XXX,XX +XXX,XX @@ void cpu_stb_data(CPUArchState *env, abi_ptr ptr, uint32_t val)
283
uint16_t meminfo = trace_mem_get_info(MO_UB, MMU_USER_IDX, true);
284
285
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
286
- stb_p(g2h(ptr), val);
287
+ stb_p(g2h(env_cpu(env), ptr), val);
288
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
289
}
290
291
@@ -XXX,XX +XXX,XX @@ void cpu_stw_be_data(CPUArchState *env, abi_ptr ptr, uint32_t val)
292
uint16_t meminfo = trace_mem_get_info(MO_BEUW, MMU_USER_IDX, true);
293
294
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
295
- stw_be_p(g2h(ptr), val);
296
+ stw_be_p(g2h(env_cpu(env), ptr), val);
297
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
298
}
299
300
@@ -XXX,XX +XXX,XX @@ void cpu_stl_be_data(CPUArchState *env, abi_ptr ptr, uint32_t val)
301
uint16_t meminfo = trace_mem_get_info(MO_BEUL, MMU_USER_IDX, true);
302
303
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
304
- stl_be_p(g2h(ptr), val);
305
+ stl_be_p(g2h(env_cpu(env), ptr), val);
306
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
307
}
308
309
@@ -XXX,XX +XXX,XX @@ void cpu_stq_be_data(CPUArchState *env, abi_ptr ptr, uint64_t val)
310
uint16_t meminfo = trace_mem_get_info(MO_BEQ, MMU_USER_IDX, true);
311
312
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
313
- stq_be_p(g2h(ptr), val);
314
+ stq_be_p(g2h(env_cpu(env), ptr), val);
315
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
316
}
317
318
@@ -XXX,XX +XXX,XX @@ void cpu_stw_le_data(CPUArchState *env, abi_ptr ptr, uint32_t val)
319
uint16_t meminfo = trace_mem_get_info(MO_LEUW, MMU_USER_IDX, true);
320
321
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
322
- stw_le_p(g2h(ptr), val);
323
+ stw_le_p(g2h(env_cpu(env), ptr), val);
324
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
325
}
326
327
@@ -XXX,XX +XXX,XX @@ void cpu_stl_le_data(CPUArchState *env, abi_ptr ptr, uint32_t val)
328
uint16_t meminfo = trace_mem_get_info(MO_LEUL, MMU_USER_IDX, true);
329
330
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
331
- stl_le_p(g2h(ptr), val);
332
+ stl_le_p(g2h(env_cpu(env), ptr), val);
333
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
334
}
335
336
@@ -XXX,XX +XXX,XX @@ void cpu_stq_le_data(CPUArchState *env, abi_ptr ptr, uint64_t val)
337
uint16_t meminfo = trace_mem_get_info(MO_LEQ, MMU_USER_IDX, true);
338
339
trace_guest_mem_before_exec(env_cpu(env), ptr, meminfo);
340
- stq_le_p(g2h(ptr), val);
341
+ stq_le_p(g2h(env_cpu(env), ptr), val);
342
qemu_plugin_vcpu_mem_cb(env_cpu(env), ptr, meminfo);
343
}
344
345
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_ldub_code(CPUArchState *env, abi_ptr ptr)
346
uint32_t ret;
347
348
set_helper_retaddr(1);
349
- ret = ldub_p(g2h(ptr));
350
+ ret = ldub_p(g2h_untagged(ptr));
351
clear_helper_retaddr();
352
return ret;
353
}
354
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_lduw_code(CPUArchState *env, abi_ptr ptr)
355
uint32_t ret;
356
357
set_helper_retaddr(1);
358
- ret = lduw_p(g2h(ptr));
359
+ ret = lduw_p(g2h_untagged(ptr));
360
clear_helper_retaddr();
361
return ret;
362
}
363
@@ -XXX,XX +XXX,XX @@ uint32_t cpu_ldl_code(CPUArchState *env, abi_ptr ptr)
364
uint32_t ret;
365
366
set_helper_retaddr(1);
367
- ret = ldl_p(g2h(ptr));
368
+ ret = ldl_p(g2h_untagged(ptr));
369
clear_helper_retaddr();
370
return ret;
371
}
372
@@ -XXX,XX +XXX,XX @@ uint64_t cpu_ldq_code(CPUArchState *env, abi_ptr ptr)
373
uint64_t ret;
374
375
set_helper_retaddr(1);
376
- ret = ldq_p(g2h(ptr));
377
+ ret = ldq_p(g2h_untagged(ptr));
378
clear_helper_retaddr();
379
return ret;
380
}
381
@@ -XXX,XX +XXX,XX @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr,
382
if (unlikely(addr & (size - 1))) {
383
cpu_loop_exit_atomic(env_cpu(env), retaddr);
384
}
385
- void *ret = g2h(addr);
386
+ void *ret = g2h(env_cpu(env), addr);
387
set_helper_retaddr(retaddr);
388
return ret;
389
}
390
diff --git a/bsd-user/elfload.c b/bsd-user/elfload.c
391
index XXXXXXX..XXXXXXX 100644
392
--- a/bsd-user/elfload.c
393
+++ b/bsd-user/elfload.c
394
@@ -XXX,XX +XXX,XX @@ static void padzero(abi_ulong elf_bss, abi_ulong last_bss)
395
end_addr1 = REAL_HOST_PAGE_ALIGN(elf_bss);
396
end_addr = HOST_PAGE_ALIGN(elf_bss);
397
if (end_addr1 < end_addr) {
398
- mmap((void *)g2h(end_addr1), end_addr - end_addr1,
399
+ mmap((void *)g2h_untagged(end_addr1), end_addr - end_addr1,
400
PROT_READ|PROT_WRITE|PROT_EXEC,
401
MAP_FIXED|MAP_PRIVATE|MAP_ANON, -1, 0);
402
}
403
diff --git a/bsd-user/main.c b/bsd-user/main.c
404
index XXXXXXX..XXXXXXX 100644
405
--- a/bsd-user/main.c
406
+++ b/bsd-user/main.c
407
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
408
env->idt.base = target_mmap(0, sizeof(uint64_t) * (env->idt.limit + 1),
409
PROT_READ|PROT_WRITE,
410
MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
411
- idt_table = g2h(env->idt.base);
412
+ idt_table = g2h_untagged(env->idt.base);
413
set_idt(0, 0);
414
set_idt(1, 0);
415
set_idt(2, 0);
416
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
417
PROT_READ|PROT_WRITE,
418
MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
419
env->gdt.limit = sizeof(uint64_t) * TARGET_GDT_ENTRIES - 1;
420
- gdt_table = g2h(env->gdt.base);
421
+ gdt_table = g2h_untagged(env->gdt.base);
422
#ifdef TARGET_ABI32
423
write_dt(&gdt_table[__USER_CS >> 3], 0, 0xfffff,
424
DESC_G_MASK | DESC_B_MASK | DESC_P_MASK | DESC_S_MASK |
425
diff --git a/bsd-user/mmap.c b/bsd-user/mmap.c
426
index XXXXXXX..XXXXXXX 100644
427
--- a/bsd-user/mmap.c
428
+++ b/bsd-user/mmap.c
429
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int prot)
430
}
431
end = host_end;
432
}
433
- ret = mprotect(g2h(host_start), qemu_host_page_size, prot1 & PAGE_BITS);
434
+ ret = mprotect(g2h_untagged(host_start),
435
+ qemu_host_page_size, prot1 & PAGE_BITS);
436
if (ret != 0)
437
goto error;
438
host_start += qemu_host_page_size;
439
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int prot)
440
for(addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) {
441
prot1 |= page_get_flags(addr);
442
}
443
- ret = mprotect(g2h(host_end - qemu_host_page_size), qemu_host_page_size,
444
- prot1 & PAGE_BITS);
445
+ ret = mprotect(g2h_untagged(host_end - qemu_host_page_size),
446
+ qemu_host_page_size, prot1 & PAGE_BITS);
447
if (ret != 0)
448
goto error;
449
host_end -= qemu_host_page_size;
450
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int prot)
451
452
/* handle the pages in the middle */
453
if (host_start < host_end) {
454
- ret = mprotect(g2h(host_start), host_end - host_start, prot);
455
+ ret = mprotect(g2h_untagged(host_start), host_end - host_start, prot);
456
if (ret != 0)
457
goto error;
458
}
459
@@ -XXX,XX +XXX,XX @@ static int mmap_frag(abi_ulong real_start,
460
int prot1, prot_new;
461
462
real_end = real_start + qemu_host_page_size;
463
- host_start = g2h(real_start);
464
+ host_start = g2h_untagged(real_start);
465
466
/* get the protection of the target pages outside the mapping */
467
prot1 = 0;
468
@@ -XXX,XX +XXX,XX @@ static int mmap_frag(abi_ulong real_start,
469
mprotect(host_start, qemu_host_page_size, prot1 | PROT_WRITE);
470
471
/* read the corresponding file data */
472
- pread(fd, g2h(start), end - start, offset);
473
+ pread(fd, g2h_untagged(start), end - start, offset);
474
475
/* put final protection */
476
if (prot_new != (prot1 | PROT_WRITE))
477
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int prot,
478
/* Note: we prefer to control the mapping address. It is
479
especially important if qemu_host_page_size >
480
qemu_real_host_page_size */
481
- p = mmap(g2h(mmap_start),
482
+ p = mmap(g2h_untagged(mmap_start),
483
host_len, prot, flags | MAP_FIXED, fd, host_offset);
484
if (p == MAP_FAILED)
485
goto fail;
486
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int prot,
487
-1, 0);
488
if (retaddr == -1)
489
goto fail;
490
- pread(fd, g2h(start), len, offset);
491
+ pread(fd, g2h_untagged(start), len, offset);
492
if (!(prot & PROT_WRITE)) {
493
ret = target_mprotect(start, len, prot);
494
if (ret != 0) {
495
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int prot,
496
offset1 = 0;
497
else
498
offset1 = offset + real_start - start;
499
- p = mmap(g2h(real_start), real_end - real_start,
500
+ p = mmap(g2h_untagged(real_start), real_end - real_start,
501
prot, flags, fd, offset1);
502
if (p == MAP_FAILED)
503
goto fail;
504
@@ -XXX,XX +XXX,XX @@ int target_munmap(abi_ulong start, abi_ulong len)
505
ret = 0;
506
/* unmap what we can */
507
if (real_start < real_end) {
508
- ret = munmap(g2h(real_start), real_end - real_start);
509
+ ret = munmap(g2h_untagged(real_start), real_end - real_start);
510
}
511
512
if (ret == 0)
513
@@ -XXX,XX +XXX,XX @@ int target_msync(abi_ulong start, abi_ulong len, int flags)
514
return 0;
515
516
start &= qemu_host_page_mask;
517
- return msync(g2h(start), end - start, flags);
518
+ return msync(g2h_untagged(start), end - start, flags);
519
}
520
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
11
diff --git a/linux-user/elfload.c b/linux-user/elfload.c
521
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
522
--- a/linux-user/elfload.c
13
--- a/linux-user/elfload.c
523
+++ b/linux-user/elfload.c
14
+++ b/linux-user/elfload.c
524
@@ -XXX,XX +XXX,XX @@ enum {
15
@@ -XXX,XX +XXX,XX @@ enum {
525
16
ARM_HWCAP2_A64_RNG = 1 << 16,
526
static bool init_guest_commpage(void)
17
ARM_HWCAP2_A64_BTI = 1 << 17,
527
{
18
ARM_HWCAP2_A64_MTE = 1 << 18,
528
- void *want = g2h(ARM_COMMPAGE & -qemu_host_page_size);
19
+ ARM_HWCAP2_A64_ECV = 1 << 19,
529
+ void *want = g2h_untagged(ARM_COMMPAGE & -qemu_host_page_size);
20
+ ARM_HWCAP2_A64_AFP = 1 << 20,
530
void *addr = mmap(want, qemu_host_page_size, PROT_READ | PROT_WRITE,
21
+ ARM_HWCAP2_A64_RPRES = 1 << 21,
531
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
22
+ ARM_HWCAP2_A64_MTE3 = 1 << 22,
532
23
+ ARM_HWCAP2_A64_SME = 1 << 23,
533
@@ -XXX,XX +XXX,XX @@ static bool init_guest_commpage(void)
24
+ ARM_HWCAP2_A64_SME_I16I64 = 1 << 24,
534
}
25
+ ARM_HWCAP2_A64_SME_F64F64 = 1 << 25,
535
26
+ ARM_HWCAP2_A64_SME_I8I32 = 1 << 26,
536
/* Set kernel helper versions; rest of page is 0. */
27
+ ARM_HWCAP2_A64_SME_F16F32 = 1 << 27,
537
- __put_user(5, (uint32_t *)g2h(0xffff0ffcu));
28
+ ARM_HWCAP2_A64_SME_B16F32 = 1 << 28,
538
+ __put_user(5, (uint32_t *)g2h_untagged(0xffff0ffcu));
29
+ ARM_HWCAP2_A64_SME_F32F32 = 1 << 29,
539
30
+ ARM_HWCAP2_A64_SME_FA64 = 1 << 30,
540
if (mprotect(addr, qemu_host_page_size, PROT_READ)) {
31
};
541
perror("Protecting guest commpage");
32
542
@@ -XXX,XX +XXX,XX @@ static void zero_bss(abi_ulong elf_bss, abi_ulong last_bss, int prot)
33
#define ELF_HWCAP get_elf_hwcap()
543
here is still actually needed. For now, continue with it,
34
@@ -XXX,XX +XXX,XX @@ static uint32_t get_elf_hwcap2(void)
544
but merge it with the "normal" mmap that would allocate the bss. */
35
GET_FEATURE_ID(aa64_rndr, ARM_HWCAP2_A64_RNG);
545
36
GET_FEATURE_ID(aa64_bti, ARM_HWCAP2_A64_BTI);
546
- host_start = (uintptr_t) g2h(elf_bss);
37
GET_FEATURE_ID(aa64_mte, ARM_HWCAP2_A64_MTE);
547
- host_end = (uintptr_t) g2h(last_bss);
38
+ GET_FEATURE_ID(aa64_sme, (ARM_HWCAP2_A64_SME |
548
+ host_start = (uintptr_t) g2h_untagged(elf_bss);
39
+ ARM_HWCAP2_A64_SME_F32F32 |
549
+ host_end = (uintptr_t) g2h_untagged(last_bss);
40
+ ARM_HWCAP2_A64_SME_B16F32 |
550
host_map_start = REAL_HOST_PAGE_ALIGN(host_start);
41
+ ARM_HWCAP2_A64_SME_F16F32 |
551
42
+ ARM_HWCAP2_A64_SME_I8I32));
552
if (host_map_start < host_end) {
43
+ GET_FEATURE_ID(aa64_sme_f64f64, ARM_HWCAP2_A64_SME_F64F64);
553
@@ -XXX,XX +XXX,XX @@ static void pgb_have_guest_base(const char *image_name, abi_ulong guest_loaddr,
44
+ GET_FEATURE_ID(aa64_sme_i16i64, ARM_HWCAP2_A64_SME_I16I64);
554
}
45
+ GET_FEATURE_ID(aa64_sme_fa64, ARM_HWCAP2_A64_SME_FA64);
555
46
556
/* Reserve the address space for the binary, or reserved_va. */
47
return hwcaps;
557
- test = g2h(guest_loaddr);
558
+ test = g2h_untagged(guest_loaddr);
559
addr = mmap(test, guest_hiaddr - guest_loaddr, PROT_NONE, flags, -1, 0);
560
if (test != addr) {
561
pgb_fail_in_use(image_name);
562
@@ -XXX,XX +XXX,XX @@ static void pgb_reserved_va(const char *image_name, abi_ulong guest_loaddr,
563
564
/* Reserve the memory on the host. */
565
assert(guest_base != 0);
566
- test = g2h(0);
567
+ test = g2h_untagged(0);
568
addr = mmap(test, reserved_va, PROT_NONE, flags, -1, 0);
569
if (addr == MAP_FAILED || addr != test) {
570
error_report("Unable to reserve 0x%lx bytes of virtual address "
571
diff --git a/linux-user/flatload.c b/linux-user/flatload.c
572
index XXXXXXX..XXXXXXX 100644
573
--- a/linux-user/flatload.c
574
+++ b/linux-user/flatload.c
575
@@ -XXX,XX +XXX,XX @@ static int load_flat_file(struct linux_binprm * bprm,
576
}
577
578
/* zero the BSS. */
579
- memset(g2h(datapos + data_len), 0, bss_len);
580
+ memset(g2h_untagged(datapos + data_len), 0, bss_len);
581
582
return 0;
583
}
48
}
584
diff --git a/linux-user/hppa/cpu_loop.c b/linux-user/hppa/cpu_loop.c
585
index XXXXXXX..XXXXXXX 100644
586
--- a/linux-user/hppa/cpu_loop.c
587
+++ b/linux-user/hppa/cpu_loop.c
588
@@ -XXX,XX +XXX,XX @@
589
590
static abi_ulong hppa_lws(CPUHPPAState *env)
591
{
592
+ CPUState *cs = env_cpu(env);
593
uint32_t which = env->gr[20];
594
abi_ulong addr = env->gr[26];
595
abi_ulong old = env->gr[25];
596
@@ -XXX,XX +XXX,XX @@ static abi_ulong hppa_lws(CPUHPPAState *env)
597
}
598
old = tswap32(old);
599
new = tswap32(new);
600
- ret = qatomic_cmpxchg((uint32_t *)g2h(addr), old, new);
601
+ ret = qatomic_cmpxchg((uint32_t *)g2h(cs, addr), old, new);
602
ret = tswap32(ret);
603
break;
604
605
@@ -XXX,XX +XXX,XX @@ static abi_ulong hppa_lws(CPUHPPAState *env)
606
can be host-endian as well. */
607
switch (size) {
608
case 0:
609
- old = *(uint8_t *)g2h(old);
610
- new = *(uint8_t *)g2h(new);
611
- ret = qatomic_cmpxchg((uint8_t *)g2h(addr), old, new);
612
+ old = *(uint8_t *)g2h(cs, old);
613
+ new = *(uint8_t *)g2h(cs, new);
614
+ ret = qatomic_cmpxchg((uint8_t *)g2h(cs, addr), old, new);
615
ret = ret != old;
616
break;
617
case 1:
618
- old = *(uint16_t *)g2h(old);
619
- new = *(uint16_t *)g2h(new);
620
- ret = qatomic_cmpxchg((uint16_t *)g2h(addr), old, new);
621
+ old = *(uint16_t *)g2h(cs, old);
622
+ new = *(uint16_t *)g2h(cs, new);
623
+ ret = qatomic_cmpxchg((uint16_t *)g2h(cs, addr), old, new);
624
ret = ret != old;
625
break;
626
case 2:
627
- old = *(uint32_t *)g2h(old);
628
- new = *(uint32_t *)g2h(new);
629
- ret = qatomic_cmpxchg((uint32_t *)g2h(addr), old, new);
630
+ old = *(uint32_t *)g2h(cs, old);
631
+ new = *(uint32_t *)g2h(cs, new);
632
+ ret = qatomic_cmpxchg((uint32_t *)g2h(cs, addr), old, new);
633
ret = ret != old;
634
break;
635
case 3:
636
{
637
uint64_t o64, n64, r64;
638
- o64 = *(uint64_t *)g2h(old);
639
- n64 = *(uint64_t *)g2h(new);
640
+ o64 = *(uint64_t *)g2h(cs, old);
641
+ n64 = *(uint64_t *)g2h(cs, new);
642
#ifdef CONFIG_ATOMIC64
643
- r64 = qatomic_cmpxchg__nocheck((uint64_t *)g2h(addr),
644
+ r64 = qatomic_cmpxchg__nocheck((uint64_t *)g2h(cs, addr),
645
o64, n64);
646
ret = r64 != o64;
647
#else
648
start_exclusive();
649
- r64 = *(uint64_t *)g2h(addr);
650
+ r64 = *(uint64_t *)g2h(cs, addr);
651
ret = 1;
652
if (r64 == o64) {
653
- *(uint64_t *)g2h(addr) = n64;
654
+ *(uint64_t *)g2h(cs, addr) = n64;
655
ret = 0;
656
}
657
end_exclusive();
658
diff --git a/linux-user/i386/cpu_loop.c b/linux-user/i386/cpu_loop.c
659
index XXXXXXX..XXXXXXX 100644
660
--- a/linux-user/i386/cpu_loop.c
661
+++ b/linux-user/i386/cpu_loop.c
662
@@ -XXX,XX +XXX,XX @@ void target_cpu_copy_regs(CPUArchState *env, struct target_pt_regs *regs)
663
env->idt.base = target_mmap(0, sizeof(uint64_t) * (env->idt.limit + 1),
664
PROT_READ|PROT_WRITE,
665
MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
666
- idt_table = g2h(env->idt.base);
667
+ idt_table = g2h_untagged(env->idt.base);
668
set_idt(0, 0);
669
set_idt(1, 0);
670
set_idt(2, 0);
671
@@ -XXX,XX +XXX,XX @@ void target_cpu_copy_regs(CPUArchState *env, struct target_pt_regs *regs)
672
PROT_READ|PROT_WRITE,
673
MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
674
env->gdt.limit = sizeof(uint64_t) * TARGET_GDT_ENTRIES - 1;
675
- gdt_table = g2h(env->gdt.base);
676
+ gdt_table = g2h_untagged(env->gdt.base);
677
#ifdef TARGET_ABI32
678
write_dt(&gdt_table[__USER_CS >> 3], 0, 0xfffff,
679
DESC_G_MASK | DESC_B_MASK | DESC_P_MASK | DESC_S_MASK |
680
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
681
index XXXXXXX..XXXXXXX 100644
682
--- a/linux-user/mmap.c
683
+++ b/linux-user/mmap.c
684
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
685
}
686
end = host_end;
687
}
688
- ret = mprotect(g2h(host_start), qemu_host_page_size,
689
+ ret = mprotect(g2h_untagged(host_start), qemu_host_page_size,
690
prot1 & PAGE_BITS);
691
if (ret != 0) {
692
goto error;
693
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
694
for (addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) {
695
prot1 |= page_get_flags(addr);
696
}
697
- ret = mprotect(g2h(host_end - qemu_host_page_size),
698
+ ret = mprotect(g2h_untagged(host_end - qemu_host_page_size),
699
qemu_host_page_size, prot1 & PAGE_BITS);
700
if (ret != 0) {
701
goto error;
702
@@ -XXX,XX +XXX,XX @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
703
704
/* handle the pages in the middle */
705
if (host_start < host_end) {
706
- ret = mprotect(g2h(host_start), host_end - host_start, host_prot);
707
+ ret = mprotect(g2h_untagged(host_start),
708
+ host_end - host_start, host_prot);
709
if (ret != 0) {
710
goto error;
711
}
712
@@ -XXX,XX +XXX,XX @@ static int mmap_frag(abi_ulong real_start,
713
int prot1, prot_new;
714
715
real_end = real_start + qemu_host_page_size;
716
- host_start = g2h(real_start);
717
+ host_start = g2h_untagged(real_start);
718
719
/* get the protection of the target pages outside the mapping */
720
prot1 = 0;
721
@@ -XXX,XX +XXX,XX @@ static int mmap_frag(abi_ulong real_start,
722
mprotect(host_start, qemu_host_page_size, prot1 | PROT_WRITE);
723
724
/* read the corresponding file data */
725
- if (pread(fd, g2h(start), end - start, offset) == -1)
726
+ if (pread(fd, g2h_untagged(start), end - start, offset) == -1)
727
return -1;
728
729
/* put final protection */
730
@@ -XXX,XX +XXX,XX @@ static int mmap_frag(abi_ulong real_start,
731
mprotect(host_start, qemu_host_page_size, prot_new);
732
}
733
if (prot_new & PROT_WRITE) {
734
- memset(g2h(start), 0, end - start);
735
+ memset(g2h_untagged(start), 0, end - start);
736
}
737
}
738
return 0;
739
@@ -XXX,XX +XXX,XX @@ abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size, abi_ulong align)
740
* - mremap() with MREMAP_FIXED flag
741
* - shmat() with SHM_REMAP flag
742
*/
743
- ptr = mmap(g2h(addr), size, PROT_NONE,
744
+ ptr = mmap(g2h_untagged(addr), size, PROT_NONE,
745
MAP_ANONYMOUS|MAP_PRIVATE|MAP_NORESERVE, -1, 0);
746
747
/* ENOMEM, if host address space has no memory */
748
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
749
/* Note: we prefer to control the mapping address. It is
750
especially important if qemu_host_page_size >
751
qemu_real_host_page_size */
752
- p = mmap(g2h(start), host_len, host_prot,
753
+ p = mmap(g2h_untagged(start), host_len, host_prot,
754
flags | MAP_FIXED | MAP_ANONYMOUS, -1, 0);
755
if (p == MAP_FAILED) {
756
goto fail;
757
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
758
/* update start so that it points to the file position at 'offset' */
759
host_start = (unsigned long)p;
760
if (!(flags & MAP_ANONYMOUS)) {
761
- p = mmap(g2h(start), len, host_prot,
762
+ p = mmap(g2h_untagged(start), len, host_prot,
763
flags | MAP_FIXED, fd, host_offset);
764
if (p == MAP_FAILED) {
765
- munmap(g2h(start), host_len);
766
+ munmap(g2h_untagged(start), host_len);
767
goto fail;
768
}
769
host_start += offset - host_offset;
770
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
771
-1, 0);
772
if (retaddr == -1)
773
goto fail;
774
- if (pread(fd, g2h(start), len, offset) == -1)
775
+ if (pread(fd, g2h_untagged(start), len, offset) == -1)
776
goto fail;
777
if (!(host_prot & PROT_WRITE)) {
778
ret = target_mprotect(start, len, target_prot);
779
@@ -XXX,XX +XXX,XX @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
780
offset1 = 0;
781
else
782
offset1 = offset + real_start - start;
783
- p = mmap(g2h(real_start), real_end - real_start,
784
+ p = mmap(g2h_untagged(real_start), real_end - real_start,
785
host_prot, flags, fd, offset1);
786
if (p == MAP_FAILED)
787
goto fail;
788
@@ -XXX,XX +XXX,XX @@ static void mmap_reserve(abi_ulong start, abi_ulong size)
789
real_end -= qemu_host_page_size;
790
}
791
if (real_start != real_end) {
792
- mmap(g2h(real_start), real_end - real_start, PROT_NONE,
793
+ mmap(g2h_untagged(real_start), real_end - real_start, PROT_NONE,
794
MAP_FIXED | MAP_ANONYMOUS | MAP_PRIVATE | MAP_NORESERVE,
795
-1, 0);
796
}
797
@@ -XXX,XX +XXX,XX @@ int target_munmap(abi_ulong start, abi_ulong len)
798
if (reserved_va) {
799
mmap_reserve(real_start, real_end - real_start);
800
} else {
801
- ret = munmap(g2h(real_start), real_end - real_start);
802
+ ret = munmap(g2h_untagged(real_start), real_end - real_start);
803
}
804
}
805
806
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
807
mmap_lock();
808
809
if (flags & MREMAP_FIXED) {
810
- host_addr = mremap(g2h(old_addr), old_size, new_size,
811
- flags, g2h(new_addr));
812
+ host_addr = mremap(g2h_untagged(old_addr), old_size, new_size,
813
+ flags, g2h_untagged(new_addr));
814
815
if (reserved_va && host_addr != MAP_FAILED) {
816
/* If new and old addresses overlap then the above mremap will
817
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
818
errno = ENOMEM;
819
host_addr = MAP_FAILED;
820
} else {
821
- host_addr = mremap(g2h(old_addr), old_size, new_size,
822
- flags | MREMAP_FIXED, g2h(mmap_start));
823
+ host_addr = mremap(g2h_untagged(old_addr), old_size, new_size,
824
+ flags | MREMAP_FIXED,
825
+ g2h_untagged(mmap_start));
826
if (reserved_va) {
827
mmap_reserve(old_addr, old_size);
828
}
829
@@ -XXX,XX +XXX,XX @@ abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
830
}
831
}
832
if (prot == 0) {
833
- host_addr = mremap(g2h(old_addr), old_size, new_size, flags);
834
+ host_addr = mremap(g2h_untagged(old_addr),
835
+ old_size, new_size, flags);
836
837
if (host_addr != MAP_FAILED) {
838
/* Check if address fits target address space */
839
if (!guest_range_valid(h2g(host_addr), new_size)) {
840
/* Revert mremap() changes */
841
- host_addr = mremap(g2h(old_addr), new_size, old_size,
842
- flags);
843
+ host_addr = mremap(g2h_untagged(old_addr),
844
+ new_size, old_size, flags);
845
errno = ENOMEM;
846
host_addr = MAP_FAILED;
847
} else if (reserved_va && old_size > new_size) {
848
diff --git a/linux-user/ppc/signal.c b/linux-user/ppc/signal.c
849
index XXXXXXX..XXXXXXX 100644
850
--- a/linux-user/ppc/signal.c
851
+++ b/linux-user/ppc/signal.c
852
@@ -XXX,XX +XXX,XX @@ static void restore_user_regs(CPUPPCState *env,
853
uint64_t v_addr;
854
/* 64-bit needs to recover the pointer to the vectors from the frame */
855
__get_user(v_addr, &frame->v_regs);
856
- v_regs = g2h(v_addr);
857
+ v_regs = g2h(env_cpu(env), v_addr);
858
#else
859
v_regs = (ppc_avr_t *)frame->mc_vregs.altivec;
860
#endif
861
@@ -XXX,XX +XXX,XX @@ void setup_rt_frame(int sig, struct target_sigaction *ka,
862
if (get_ppc64_abi(image) < 2) {
863
/* ELFv1 PPC64 function pointers are pointers to OPD entries. */
864
struct target_func_ptr *handler =
865
- (struct target_func_ptr *)g2h(ka->_sa_handler);
866
+ (struct target_func_ptr *)g2h(env_cpu(env), ka->_sa_handler);
867
env->nip = tswapl(handler->entry);
868
env->gpr[2] = tswapl(handler->toc);
869
} else {
870
diff --git a/linux-user/syscall.c b/linux-user/syscall.c
871
index XXXXXXX..XXXXXXX 100644
872
--- a/linux-user/syscall.c
873
+++ b/linux-user/syscall.c
874
@@ -XXX,XX +XXX,XX @@ abi_long do_brk(abi_ulong new_brk)
875
/* Heap contents are initialized to zero, as for anonymous
876
* mapped pages. */
877
if (new_brk > target_brk) {
878
- memset(g2h(target_brk), 0, new_brk - target_brk);
879
+ memset(g2h_untagged(target_brk), 0, new_brk - target_brk);
880
}
881
    target_brk = new_brk;
882
DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk <= brk_page)\n", target_brk);
883
@@ -XXX,XX +XXX,XX @@ abi_long do_brk(abi_ulong new_brk)
884
* come from the remaining part of the previous page: it may
885
* contains garbage data due to a previous heap usage (grown
886
* then shrunken). */
887
- memset(g2h(target_brk), 0, brk_page - target_brk);
888
+ memset(g2h_untagged(target_brk), 0, brk_page - target_brk);
889
890
target_brk = new_brk;
891
brk_page = HOST_PAGE_ALIGN(target_brk);
892
@@ -XXX,XX +XXX,XX @@ static inline abi_ulong do_shmat(CPUArchState *cpu_env,
893
mmap_lock();
894
895
if (shmaddr)
896
- host_raddr = shmat(shmid, (void *)g2h(shmaddr), shmflg);
897
+ host_raddr = shmat(shmid, (void *)g2h_untagged(shmaddr), shmflg);
898
else {
899
abi_ulong mmap_start;
900
901
@@ -XXX,XX +XXX,XX @@ static inline abi_ulong do_shmat(CPUArchState *cpu_env,
902
errno = ENOMEM;
903
host_raddr = (void *)-1;
904
} else
905
- host_raddr = shmat(shmid, g2h(mmap_start), shmflg | SHM_REMAP);
906
+ host_raddr = shmat(shmid, g2h_untagged(mmap_start),
907
+ shmflg | SHM_REMAP);
908
}
909
910
if (host_raddr == (void *)-1) {
911
@@ -XXX,XX +XXX,XX @@ static inline abi_long do_shmdt(abi_ulong shmaddr)
912
break;
913
}
914
}
915
- rv = get_errno(shmdt(g2h(shmaddr)));
916
+ rv = get_errno(shmdt(g2h_untagged(shmaddr)));
917
918
mmap_unlock();
919
920
@@ -XXX,XX +XXX,XX @@ static abi_long write_ldt(CPUX86State *env,
921
MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
922
if (env->ldt.base == -1)
923
return -TARGET_ENOMEM;
924
- memset(g2h(env->ldt.base), 0,
925
+ memset(g2h_untagged(env->ldt.base), 0,
926
TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE);
927
env->ldt.limit = 0xffff;
928
- ldt_table = g2h(env->ldt.base);
929
+ ldt_table = g2h_untagged(env->ldt.base);
930
}
931
932
/* NOTE: same code as Linux kernel */
933
@@ -XXX,XX +XXX,XX @@ static abi_long do_modify_ldt(CPUX86State *env, int func, abi_ulong ptr,
934
#if defined(TARGET_ABI32)
935
abi_long do_set_thread_area(CPUX86State *env, abi_ulong ptr)
936
{
937
- uint64_t *gdt_table = g2h(env->gdt.base);
938
+ uint64_t *gdt_table = g2h_untagged(env->gdt.base);
939
struct target_modify_ldt_ldt_s ldt_info;
940
struct target_modify_ldt_ldt_s *target_ldt_info;
941
int seg_32bit, contents, read_exec_only, limit_in_pages;
942
@@ -XXX,XX +XXX,XX @@ install:
943
static abi_long do_get_thread_area(CPUX86State *env, abi_ulong ptr)
944
{
945
struct target_modify_ldt_ldt_s *target_ldt_info;
946
- uint64_t *gdt_table = g2h(env->gdt.base);
947
+ uint64_t *gdt_table = g2h_untagged(env->gdt.base);
948
uint32_t base_addr, limit, flags;
949
int seg_32bit, contents, read_exec_only, limit_in_pages, idx;
950
int seg_not_present, useable, lm;
951
@@ -XXX,XX +XXX,XX @@ static int do_safe_futex(int *uaddr, int op, int val,
952
tricky. However they're probably useless because guest atomic
953
operations won't work either. */
954
#if defined(TARGET_NR_futex)
955
-static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout,
956
- target_ulong uaddr2, int val3)
957
+static int do_futex(CPUState *cpu, target_ulong uaddr, int op, int val,
958
+ target_ulong timeout, target_ulong uaddr2, int val3)
959
{
960
struct timespec ts, *pts;
961
int base_op;
962
@@ -XXX,XX +XXX,XX @@ static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout,
963
} else {
964
pts = NULL;
965
}
966
- return do_safe_futex(g2h(uaddr), op, tswap32(val), pts, NULL, val3);
967
+ return do_safe_futex(g2h(cpu, uaddr),
968
+ op, tswap32(val), pts, NULL, val3);
969
case FUTEX_WAKE:
970
- return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
971
+ return do_safe_futex(g2h(cpu, uaddr),
972
+ op, val, NULL, NULL, 0);
973
case FUTEX_FD:
974
- return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
975
+ return do_safe_futex(g2h(cpu, uaddr),
976
+ op, val, NULL, NULL, 0);
977
case FUTEX_REQUEUE:
978
case FUTEX_CMP_REQUEUE:
979
case FUTEX_WAKE_OP:
980
@@ -XXX,XX +XXX,XX @@ static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout,
981
to satisfy the compiler. We do not need to tswap TIMEOUT
982
since it's not compared to guest memory. */
983
pts = (struct timespec *)(uintptr_t) timeout;
984
- return do_safe_futex(g2h(uaddr), op, val, pts, g2h(uaddr2),
985
+ return do_safe_futex(g2h(cpu, uaddr), op, val, pts, g2h(cpu, uaddr2),
986
(base_op == FUTEX_CMP_REQUEUE
987
- ? tswap32(val3)
988
- : val3));
989
+ ? tswap32(val3) : val3));
990
default:
991
return -TARGET_ENOSYS;
992
}
993
@@ -XXX,XX +XXX,XX @@ static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout,
994
#endif
995
996
#if defined(TARGET_NR_futex_time64)
997
-static int do_futex_time64(target_ulong uaddr, int op, int val, target_ulong timeout,
998
+static int do_futex_time64(CPUState *cpu, target_ulong uaddr, int op,
999
+ int val, target_ulong timeout,
1000
target_ulong uaddr2, int val3)
1001
{
1002
struct timespec ts, *pts;
1003
@@ -XXX,XX +XXX,XX @@ static int do_futex_time64(target_ulong uaddr, int op, int val, target_ulong tim
1004
} else {
1005
pts = NULL;
1006
}
1007
- return do_safe_futex(g2h(uaddr), op, tswap32(val), pts, NULL, val3);
1008
+ return do_safe_futex(g2h(cpu, uaddr), op,
1009
+ tswap32(val), pts, NULL, val3);
1010
case FUTEX_WAKE:
1011
- return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
1012
+ return do_safe_futex(g2h(cpu, uaddr), op, val, NULL, NULL, 0);
1013
case FUTEX_FD:
1014
- return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
1015
+ return do_safe_futex(g2h(cpu, uaddr), op, val, NULL, NULL, 0);
1016
case FUTEX_REQUEUE:
1017
case FUTEX_CMP_REQUEUE:
1018
case FUTEX_WAKE_OP:
1019
@@ -XXX,XX +XXX,XX @@ static int do_futex_time64(target_ulong uaddr, int op, int val, target_ulong tim
1020
to satisfy the compiler. We do not need to tswap TIMEOUT
1021
since it's not compared to guest memory. */
1022
pts = (struct timespec *)(uintptr_t) timeout;
1023
- return do_safe_futex(g2h(uaddr), op, val, pts, g2h(uaddr2),
1024
+ return do_safe_futex(g2h(cpu, uaddr), op, val, pts, g2h(cpu, uaddr2),
1025
(base_op == FUTEX_CMP_REQUEUE
1026
- ? tswap32(val3)
1027
- : val3));
1028
+ ? tswap32(val3) : val3));
1029
default:
1030
return -TARGET_ENOSYS;
1031
}
1032
@@ -XXX,XX +XXX,XX @@ static int open_self_maps(void *cpu_env, int fd)
1033
const char *path;
1034
1035
max = h2g_valid(max - 1) ?
1036
- max : (uintptr_t) g2h(GUEST_ADDR_MAX) + 1;
1037
+ max : (uintptr_t) g2h_untagged(GUEST_ADDR_MAX) + 1;
1038
1039
if (page_check_range(h2g(min), max - min, flags) == -1) {
1040
continue;
1041
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
1042
1043
if (ts->child_tidptr) {
1044
put_user_u32(0, ts->child_tidptr);
1045
- do_sys_futex(g2h(ts->child_tidptr), FUTEX_WAKE, INT_MAX,
1046
- NULL, NULL, 0);
1047
+ do_sys_futex(g2h(cpu, ts->child_tidptr),
1048
+ FUTEX_WAKE, INT_MAX, NULL, NULL, 0);
1049
}
1050
thread_cpu = NULL;
1051
g_free(ts);
1052
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
1053
if (!arg5) {
1054
ret = mount(p, p2, p3, (unsigned long)arg4, NULL);
1055
} else {
1056
- ret = mount(p, p2, p3, (unsigned long)arg4, g2h(arg5));
1057
+ ret = mount(p, p2, p3, (unsigned long)arg4, g2h(cpu, arg5));
1058
}
1059
ret = get_errno(ret);
1060
1061
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
1062
/* ??? msync/mlock/munlock are broken for softmmu. */
1063
#ifdef TARGET_NR_msync
1064
case TARGET_NR_msync:
1065
- return get_errno(msync(g2h(arg1), arg2, arg3));
1066
+ return get_errno(msync(g2h(cpu, arg1), arg2, arg3));
1067
#endif
1068
#ifdef TARGET_NR_mlock
1069
case TARGET_NR_mlock:
1070
- return get_errno(mlock(g2h(arg1), arg2));
1071
+ return get_errno(mlock(g2h(cpu, arg1), arg2));
1072
#endif
1073
#ifdef TARGET_NR_munlock
1074
case TARGET_NR_munlock:
1075
- return get_errno(munlock(g2h(arg1), arg2));
1076
+ return get_errno(munlock(g2h(cpu, arg1), arg2));
1077
#endif
1078
#ifdef TARGET_NR_mlockall
1079
case TARGET_NR_mlockall:
1080
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
1081
1082
#if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address)
1083
case TARGET_NR_set_tid_address:
1084
- return get_errno(set_tid_address((int *)g2h(arg1)));
1085
+ return get_errno(set_tid_address((int *)g2h(cpu, arg1)));
1086
#endif
1087
1088
case TARGET_NR_tkill:
1089
@@ -XXX,XX +XXX,XX @@ static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
1090
#endif
1091
#ifdef TARGET_NR_futex
1092
case TARGET_NR_futex:
1093
- return do_futex(arg1, arg2, arg3, arg4, arg5, arg6);
1094
+ return do_futex(cpu, arg1, arg2, arg3, arg4, arg5, arg6);
1095
#endif
1096
#ifdef TARGET_NR_futex_time64
1097
case TARGET_NR_futex_time64:
1098
- return do_futex_time64(arg1, arg2, arg3, arg4, arg5, arg6);
1099
+ return do_futex_time64(cpu, arg1, arg2, arg3, arg4, arg5, arg6);
1100
#endif
1101
#if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init)
1102
case TARGET_NR_inotify_init:
1103
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
1104
index XXXXXXX..XXXXXXX 100644
1105
--- a/target/arm/helper-a64.c
1106
+++ b/target/arm/helper-a64.c
1107
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(paired_cmpxchg64_le)(CPUARMState *env, uint64_t addr,
1108
1109
#ifdef CONFIG_USER_ONLY
1110
/* ??? Enforce alignment. */
1111
- uint64_t *haddr = g2h(addr);
1112
+ uint64_t *haddr = g2h(env_cpu(env), addr);
1113
1114
set_helper_retaddr(ra);
1115
o0 = ldq_le_p(haddr + 0);
1116
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(paired_cmpxchg64_be)(CPUARMState *env, uint64_t addr,
1117
1118
#ifdef CONFIG_USER_ONLY
1119
/* ??? Enforce alignment. */
1120
- uint64_t *haddr = g2h(addr);
1121
+ uint64_t *haddr = g2h(env_cpu(env), addr);
1122
1123
set_helper_retaddr(ra);
1124
o1 = ldq_be_p(haddr + 0);
1125
diff --git a/target/hppa/op_helper.c b/target/hppa/op_helper.c
1126
index XXXXXXX..XXXXXXX 100644
1127
--- a/target/hppa/op_helper.c
1128
+++ b/target/hppa/op_helper.c
1129
@@ -XXX,XX +XXX,XX @@ static void atomic_store_3(CPUHPPAState *env, target_ulong addr, uint32_t val,
1130
#ifdef CONFIG_USER_ONLY
1131
uint32_t old, new, cmp;
1132
1133
- uint32_t *haddr = g2h(addr - 1);
1134
+ uint32_t *haddr = g2h(env_cpu(env), addr - 1);
1135
old = *haddr;
1136
while (1) {
1137
new = (old & ~mask) | (val & mask);
1138
diff --git a/target/i386/tcg/mem_helper.c b/target/i386/tcg/mem_helper.c
1139
index XXXXXXX..XXXXXXX 100644
1140
--- a/target/i386/tcg/mem_helper.c
1141
+++ b/target/i386/tcg/mem_helper.c
1142
@@ -XXX,XX +XXX,XX @@ void helper_cmpxchg8b(CPUX86State *env, target_ulong a0)
1143
1144
#ifdef CONFIG_USER_ONLY
1145
{
1146
- uint64_t *haddr = g2h(a0);
1147
+ uint64_t *haddr = g2h(env_cpu(env), a0);
1148
cmpv = cpu_to_le64(cmpv);
1149
newv = cpu_to_le64(newv);
1150
oldv = qatomic_cmpxchg__nocheck(haddr, cmpv, newv);
1151
diff --git a/target/s390x/mem_helper.c b/target/s390x/mem_helper.c
1152
index XXXXXXX..XXXXXXX 100644
1153
--- a/target/s390x/mem_helper.c
1154
+++ b/target/s390x/mem_helper.c
1155
@@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1,
1156
1157
if (parallel) {
1158
#ifdef CONFIG_USER_ONLY
1159
- uint32_t *haddr = g2h(a1);
1160
+ uint32_t *haddr = g2h(env_cpu(env), a1);
1161
ov = qatomic_cmpxchg__nocheck(haddr, cv, nv);
1162
#else
1163
TCGMemOpIdx oi = make_memop_idx(MO_TEUL | MO_ALIGN, mem_idx);
1164
@@ -XXX,XX +XXX,XX @@ static uint32_t do_csst(CPUS390XState *env, uint32_t r3, uint64_t a1,
1165
if (parallel) {
1166
#ifdef CONFIG_ATOMIC64
1167
# ifdef CONFIG_USER_ONLY
1168
- uint64_t *haddr = g2h(a1);
1169
+ uint64_t *haddr = g2h(env_cpu(env), a1);
1170
ov = qatomic_cmpxchg__nocheck(haddr, cv, nv);
1171
# else
1172
TCGMemOpIdx oi = make_memop_idx(MO_TEQ | MO_ALIGN, mem_idx);
1173
--
49
--
1174
2.20.1
50
2.25.1
1175
1176
diff view generated by jsdifflib