1
Mostly this is RTH's memtag series, but there are also some cleanups
1
First pullreq for 6.0: mostly my v8.1M work, plus some other
2
from Philippe.
2
bits and pieces. (I still have a lot of stuff in my to-review
3
folder, which I may or may not get to before the Christmas break...)
3
4
4
thanks
5
thanks
5
-- PMM
6
-- PMM
6
7
7
The following changes since commit 10f7ffabf9c507fc02382b89912003b1c43c3231:
8
The following changes since commit 5e7b204dbfae9a562fc73684986f936b97f63877:
8
9
9
Merge remote-tracking branch 'remotes/mcayland/tags/qemu-macppc-20200626' into staging (2020-06-26 12:14:18 +0100)
10
Merge remote-tracking branch 'remotes/mst/tags/for_upstream' into staging (2020-12-09 20:08:54 +0000)
10
11
11
are available in the Git repository at:
12
are available in the Git repository at:
12
13
13
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20200626
14
https://git.linaro.org/people/pmaydell/qemu-arm.git tags/pull-target-arm-20201210
14
15
15
for you to fetch changes up to c7459633baa71d1781fde4a245d6ec9ce2f008cf:
16
for you to fetch changes up to 71f916be1c7e9ede0e37d9cabc781b5a9e8638ff:
16
17
17
target/arm: Enable MTE (2020-06-26 14:32:24 +0100)
18
hw/arm/armv7m: Correct typo in QOM object name (2020-12-10 11:44:56 +0000)
18
19
19
----------------------------------------------------------------
20
----------------------------------------------------------------
20
target-arm queue:
21
target-arm queue:
21
* hw/arm/aspeed: improve QOM usage
22
* hw/arm/smmuv3: Fix up L1STD_SPAN decoding
22
* hw/misc/pca9552: trace GPIO change events
23
* xlnx-zynqmp: Support Xilinx ZynqMP CAN controllers
23
* target/arm: Implement ARMv8.5-MemTag for system emulation
24
* sbsa-ref: allow to use Cortex-A53/57/72 cpus
25
* Various minor code cleanups
26
* hw/intc/armv7m_nvic: Make all of system PPB range be RAZWI/BusFault
27
* Implement more pieces of ARMv8.1M support
24
28
25
----------------------------------------------------------------
29
----------------------------------------------------------------
26
Philippe Mathieu-Daudé (12):
30
Alex Chen (4):
27
hw/arm/aspeed: Remove extraneous MemoryRegion object owner
31
i.MX25: Fix bad printf format specifiers
28
hw/arm/aspeed: Rename AspeedBoardState as AspeedMachineState
32
i.MX31: Fix bad printf format specifiers
29
hw/arm/aspeed: QOM'ify AspeedMachineState
33
i.MX6: Fix bad printf format specifiers
30
hw/i2c/core: Add i2c_try_create_slave() and i2c_realize_and_unref()
34
i.MX6ul: Fix bad printf format specifiers
31
hw/misc/pca9552: Rename 'nr_leds' as 'pin_count'
32
hw/misc/pca9552: Rename generic code as pca955x
33
hw/misc/pca9552: Add generic PCA955xClass, parent of TYPE_PCA9552
34
hw/misc/pca9552: Add a 'description' property for debugging purpose
35
hw/misc/pca9552: Trace GPIO High/Low events
36
hw/arm/aspeed: Describe each PCA9552 device
37
hw/misc/pca9552: Trace GPIO change events
38
hw/misc/pca9552: Model qdev output GPIOs
39
35
40
Richard Henderson (45):
36
Havard Skinnemoen (1):
41
target/arm: Add isar tests for mte
37
tests/qtest/npcm7xx_rng-test: dump random data on failure
42
target/arm: Improve masking of SCR RES0 bits
43
target/arm: Add support for MTE to SCTLR_ELx
44
target/arm: Add support for MTE to HCR_EL2 and SCR_EL3
45
target/arm: Rename DISAS_UPDATE to DISAS_UPDATE_EXIT
46
target/arm: Add DISAS_UPDATE_NOCHAIN
47
target/arm: Add MTE system registers
48
target/arm: Add MTE bits to tb_flags
49
target/arm: Implement the IRG instruction
50
target/arm: Revise decoding for disas_add_sub_imm
51
target/arm: Implement the ADDG, SUBG instructions
52
target/arm: Implement the GMI instruction
53
target/arm: Implement the SUBP instruction
54
target/arm: Define arm_cpu_do_unaligned_access for user-only
55
target/arm: Implement LDG, STG, ST2G instructions
56
target/arm: Implement the STGP instruction
57
target/arm: Restrict the values of DCZID.BS under TCG
58
target/arm: Simplify DC_ZVA
59
target/arm: Implement the LDGM, STGM, STZGM instructions
60
target/arm: Implement the access tag cache flushes
61
target/arm: Move regime_el to internals.h
62
target/arm: Move regime_tcr to internals.h
63
target/arm: Add gen_mte_check1
64
target/arm: Add gen_mte_checkN
65
target/arm: Implement helper_mte_check1
66
target/arm: Implement helper_mte_checkN
67
target/arm: Add helper_mte_check_zva
68
target/arm: Use mte_checkN for sve unpredicated loads
69
target/arm: Use mte_checkN for sve unpredicated stores
70
target/arm: Use mte_check1 for sve LD1R
71
target/arm: Tidy trans_LD1R_zpri
72
target/arm: Add arm_tlb_bti_gp
73
target/arm: Add mte helpers for sve scalar + int loads
74
target/arm: Add mte helpers for sve scalar + int stores
75
target/arm: Add mte helpers for sve scalar + int ff/nf loads
76
target/arm: Handle TBI for sve scalar + int memory ops
77
target/arm: Add mte helpers for sve scatter/gather memory ops
78
target/arm: Complete TBI clearing for user-only for SVE
79
target/arm: Implement data cache set allocation tags
80
target/arm: Set PSTATE.TCO on exception entry
81
target/arm: Always pass cacheattr to get_phys_addr
82
target/arm: Cache the Tagged bit for a page in MemTxAttrs
83
target/arm: Create tagged ram when MTE is enabled
84
target/arm: Add allocation tag storage for system mode
85
target/arm: Enable MTE
86
38
87
include/hw/arm/aspeed.h | 12 +-
39
Kunkun Jiang (1):
88
include/hw/i2c/i2c.h | 2 +
40
hw/arm/smmuv3: Fix up L1STD_SPAN decoding
89
include/hw/misc/pca9552.h | 16 +-
90
target/arm/cpu.h | 50 +-
91
target/arm/helper-a64.h | 16 +
92
target/arm/helper-sve.h | 488 ++++++++++++++
93
target/arm/helper.h | 2 +
94
target/arm/internals.h | 153 ++++-
95
target/arm/translate-a64.h | 5 +
96
target/arm/translate.h | 23 +-
97
hw/arm/aspeed.c | 46 +-
98
hw/arm/virt.c | 55 +-
99
hw/i2c/core.c | 18 +-
100
hw/misc/pca9552.c | 216 +++++--
101
target/arm/cpu.c | 81 ++-
102
target/arm/cpu64.c | 5 +
103
target/arm/helper-a64.c | 94 +--
104
target/arm/helper.c | 423 ++++++++++---
105
target/arm/m_helper.c | 11 +-
106
target/arm/mte_helper.c | 906 ++++++++++++++++++++++++++
107
target/arm/op_helper.c | 16 +
108
target/arm/sve_helper.c | 616 ++++++++++++++----
109
target/arm/tlb_helper.c | 13 +-
110
target/arm/translate-a64.c | 657 ++++++++++++++++---
111
target/arm/translate-sve.c | 1366 ++++++++++++++++++++++++++--------------
112
target/arm/translate-vfp.inc.c | 4 +-
113
target/arm/translate.c | 16 +-
114
hw/misc/trace-events | 4 +
115
target/arm/Makefile.objs | 1 +
116
29 files changed, 4391 insertions(+), 924 deletions(-)
117
create mode 100644 target/arm/mte_helper.c
118
41
42
Marcin Juszkiewicz (1):
43
sbsa-ref: allow to use Cortex-A53/57/72 cpus
44
45
Peter Maydell (25):
46
hw/intc/armv7m_nvic: Make all of system PPB range be RAZWI/BusFault
47
target/arm: Implement v8.1M PXN extension
48
target/arm: Don't clobber ID_PFR1.Security on M-profile cores
49
target/arm: Implement VSCCLRM insn
50
target/arm: Implement CLRM instruction
51
target/arm: Enforce M-profile VMRS/VMSR register restrictions
52
target/arm: Refactor M-profile VMSR/VMRS handling
53
target/arm: Move general-use constant expanders up in translate.c
54
target/arm: Implement VLDR/VSTR system register
55
target/arm: Implement M-profile FPSCR_nzcvqc
56
target/arm: Use new FPCR_NZCV_MASK constant
57
target/arm: Factor out preserve-fp-state from full_vfp_access_check()
58
target/arm: Implement FPCXT_S fp system register
59
hw/intc/armv7m_nvic: Update FPDSCR masking for v8.1M
60
target/arm: For v8.1M, always clear R0-R3, R12, APSR, EPSR on exception entry
61
target/arm: In v8.1M, don't set HFSR.FORCED on vector table fetch failures
62
target/arm: Implement v8.1M REVIDR register
63
target/arm: Implement new v8.1M NOCP check for exception return
64
target/arm: Implement new v8.1M VLLDM and VLSTM encodings
65
hw/intc/armv7m_nvic: Support v8.1M CCR.TRD bit
66
target/arm: Implement CCR_S.TRD behaviour for SG insns
67
hw/intc/armv7m_nvic: Fix "return from inactive handler" check
68
target/arm: Implement M-profile "minimal RAS implementation"
69
hw/intc/armv7m_nvic: Implement read/write for RAS register block
70
hw/arm/armv7m: Correct typo in QOM object name
71
72
Vikram Garhwal (4):
73
hw/net/can: Introduce Xilinx ZynqMP CAN controller
74
xlnx-zynqmp: Connect Xilinx ZynqMP CAN controllers
75
tests/qtest: Introduce tests for Xilinx ZynqMP CAN controller
76
MAINTAINERS: Add maintainer entry for Xilinx ZynqMP CAN controller
77
78
meson.build | 1 +
79
hw/arm/smmuv3-internal.h | 2 +-
80
hw/net/can/trace.h | 1 +
81
include/hw/arm/xlnx-zynqmp.h | 8 +
82
include/hw/intc/armv7m_nvic.h | 2 +
83
include/hw/net/xlnx-zynqmp-can.h | 78 +++
84
target/arm/cpu.h | 46 ++
85
target/arm/m-nocp.decode | 10 +-
86
target/arm/t32.decode | 10 +-
87
target/arm/vfp.decode | 14 +
88
hw/arm/armv7m.c | 4 +-
89
hw/arm/sbsa-ref.c | 23 +-
90
hw/arm/xlnx-zcu102.c | 20 +
91
hw/arm/xlnx-zynqmp.c | 34 ++
92
hw/intc/armv7m_nvic.c | 246 ++++++--
93
hw/misc/imx25_ccm.c | 12 +-
94
hw/misc/imx31_ccm.c | 14 +-
95
hw/misc/imx6_ccm.c | 20 +-
96
hw/misc/imx6_src.c | 2 +-
97
hw/misc/imx6ul_ccm.c | 4 +-
98
hw/misc/imx_ccm.c | 4 +-
99
hw/net/can/xlnx-zynqmp-can.c | 1161 ++++++++++++++++++++++++++++++++++++++
100
target/arm/cpu.c | 5 +-
101
target/arm/helper.c | 7 +-
102
target/arm/m_helper.c | 130 ++++-
103
target/arm/translate.c | 105 +++-
104
tests/qtest/npcm7xx_rng-test.c | 12 +
105
tests/qtest/xlnx-can-test.c | 360 ++++++++++++
106
MAINTAINERS | 8 +
107
hw/Kconfig | 1 +
108
hw/net/can/meson.build | 1 +
109
hw/net/can/trace-events | 9 +
110
target/arm/translate-vfp.c.inc | 511 ++++++++++++++++-
111
tests/qtest/meson.build | 1 +
112
34 files changed, 2713 insertions(+), 153 deletions(-)
113
create mode 100644 hw/net/can/trace.h
114
create mode 100644 include/hw/net/xlnx-zynqmp-can.h
115
create mode 100644 hw/net/can/xlnx-zynqmp-can.c
116
create mode 100644 tests/qtest/xlnx-can-test.c
117
create mode 100644 hw/net/can/trace-events
118
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
I'm confused by this code, 'bmc' is created as:
4
5
bmc = g_new0(AspeedBoardState, 1);
6
7
Then we use it as QOM owner for different MemoryRegion objects.
8
But looking at memory_region_init_ram (similarly for ROM):
9
10
void memory_region_init_ram(MemoryRegion *mr,
11
struct Object *owner,
12
const char *name,
13
uint64_t size,
14
Error **errp)
15
{
16
DeviceState *owner_dev;
17
Error *err = NULL;
18
19
memory_region_init_ram_nomigrate(mr, owner, name, size, &err);
20
if (err) {
21
error_propagate(errp, err);
22
return;
23
}
24
/* This will assert if owner is neither NULL nor a DeviceState.
25
* We only want the owner here for the purposes of defining a
26
* unique name for migration. TODO: Ideally we should implement
27
* a naming scheme for Objects which are not DeviceStates, in
28
* which case we can relax this restriction.
29
*/
30
owner_dev = DEVICE(owner);
31
vmstate_register_ram(mr, owner_dev);
32
}
33
34
The expected assertion is not triggered ('bmc' is not NULL neither
35
a DeviceState).
36
37
'bmc' structure is defined as:
38
39
struct AspeedBoardState {
40
AspeedSoCState soc;
41
MemoryRegion ram_container;
42
MemoryRegion max_ram;
43
};
44
45
What happens is when using 'OBJECT(bmc)', the QOM macros cast the
46
memory pointed by bmc, which first member is 'soc', which is
47
initialized ...:
48
49
object_initialize_child(OBJECT(machine), "soc",
50
&bmc->soc, amc->soc_name);
51
52
The 'soc' object is indeed a DeviceState, so the assertion passes.
53
54
Since this is fragile and only happens to work by luck, remove the
55
dangerous OBJECT(bmc) owner argument.
56
57
Note, this probably breaks migration for this machine.
58
59
Reviewed-by: Cédric Le Goater <clg@kaod.org>
60
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
61
Message-id: 20200623072132.2868-2-f4bug@amsat.org
62
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
63
---
64
hw/arm/aspeed.c | 6 +++---
65
1 file changed, 3 insertions(+), 3 deletions(-)
66
67
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
68
index XXXXXXX..XXXXXXX 100644
69
--- a/hw/arm/aspeed.c
70
+++ b/hw/arm/aspeed.c
71
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
72
* needed by the flash modules of the Aspeed machines.
73
*/
74
if (ASPEED_MACHINE(machine)->mmio_exec) {
75
- memory_region_init_alias(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
76
+ memory_region_init_alias(boot_rom, NULL, "aspeed.boot_rom",
77
&fl->mmio, 0, fl->size);
78
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
79
boot_rom);
80
} else {
81
- memory_region_init_rom(boot_rom, OBJECT(bmc), "aspeed.boot_rom",
82
+ memory_region_init_rom(boot_rom, NULL, "aspeed.boot_rom",
83
fl->size, &error_abort);
84
memory_region_add_subregion(get_system_memory(), FIRMWARE_ADDR,
85
boot_rom);
86
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
87
if (machine->kernel_filename && sc->num_cpus > 1) {
88
/* With no u-boot we must set up a boot stub for the secondary CPU */
89
MemoryRegion *smpboot = g_new(MemoryRegion, 1);
90
- memory_region_init_ram(smpboot, OBJECT(bmc), "aspeed.smpboot",
91
+ memory_region_init_ram(smpboot, NULL, "aspeed.smpboot",
92
0x80, &error_abort);
93
memory_region_add_subregion(get_system_memory(),
94
AST_SMP_MAILBOX_BASE, smpboot);
95
--
96
2.20.1
97
98
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Kunkun Jiang <jiangkunkun@huawei.com>
2
2
3
This "bit" is a particular value of the page's MemAttr.
3
Accroding to the SMMUv3 spec, the SPAN field of Level1 Stream Table
4
Descriptor is 5 bits([4:0]).
4
5
6
Fixes: 9bde7f0674f(hw/arm/smmuv3: Implement translate callback)
7
Signed-off-by: Kunkun Jiang <jiangkunkun@huawei.com>
8
Message-id: 20201124023711.1184-1-jiangkunkun@huawei.com
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Acked-by: Eric Auger <eric.auger@redhat.com>
7
Message-id: 20200626033144.790098-43-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
12
---
10
target/arm/helper.c | 48 ++++++++++++++++++++++++++++++++++++++---
13
hw/arm/smmuv3-internal.h | 2 +-
11
target/arm/tlb_helper.c | 5 +++++
14
1 file changed, 1 insertion(+), 1 deletion(-)
12
2 files changed, 50 insertions(+), 3 deletions(-)
13
15
14
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
diff --git a/hw/arm/smmuv3-internal.h b/hw/arm/smmuv3-internal.h
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper.c
18
--- a/hw/arm/smmuv3-internal.h
17
+++ b/target/arm/helper.c
19
+++ b/hw/arm/smmuv3-internal.h
18
@@ -XXX,XX +XXX,XX @@ static uint8_t combine_cacheattr_nibble(uint8_t s1, uint8_t s2)
20
@@ -XXX,XX +XXX,XX @@ static inline uint64_t l1std_l2ptr(STEDesc *desc)
19
*/
21
return hi << 32 | lo;
20
static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
21
{
22
- uint8_t s1lo = extract32(s1.attrs, 0, 4), s2lo = extract32(s2.attrs, 0, 4);
23
- uint8_t s1hi = extract32(s1.attrs, 4, 4), s2hi = extract32(s2.attrs, 4, 4);
24
+ uint8_t s1lo, s2lo, s1hi, s2hi;
25
ARMCacheAttrs ret;
26
+ bool tagged = false;
27
+
28
+ if (s1.attrs == 0xf0) {
29
+ tagged = true;
30
+ s1.attrs = 0xff;
31
+ }
32
+
33
+ s1lo = extract32(s1.attrs, 0, 4);
34
+ s2lo = extract32(s2.attrs, 0, 4);
35
+ s1hi = extract32(s1.attrs, 4, 4);
36
+ s2hi = extract32(s2.attrs, 4, 4);
37
38
/* Combine shareability attributes (table D4-43) */
39
if (s1.shareability == 2 || s2.shareability == 2) {
40
@@ -XXX,XX +XXX,XX @@ static ARMCacheAttrs combine_cacheattrs(ARMCacheAttrs s1, ARMCacheAttrs s2)
41
}
42
}
43
44
+ /* TODO: CombineS1S2Desc does not consider transient, only WB, RWA. */
45
+ if (tagged && ret.attrs == 0xff) {
46
+ ret.attrs = 0xf0;
47
+ }
48
+
49
return ret;
50
}
22
}
51
23
52
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
24
-#define L1STD_SPAN(stm) (extract32((stm)->word[0], 0, 4))
53
* Normal Non-Shareable,
25
+#define L1STD_SPAN(stm) (extract32((stm)->word[0], 0, 5))
54
* Inner Write-Back Read-Allocate Write-Allocate,
26
55
* Outer Write-Back Read-Allocate Write-Allocate.
27
#endif
56
+ * Do not overwrite Tagged within attrs.
57
*/
58
- cacheattrs->attrs = 0xff;
59
+ if (cacheattrs->attrs != 0xf0) {
60
+ cacheattrs->attrs = 0xff;
61
+ }
62
cacheattrs->shareability = 0;
63
}
64
*cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
65
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
66
/* Definitely a real MMU, not an MPU */
67
68
if (regime_translation_disabled(env, mmu_idx)) {
69
+ uint64_t hcr;
70
+ uint8_t memattr;
71
+
72
/*
73
* MMU disabled. S1 addresses within aa64 translation regimes are
74
* still checked for bounds -- see AArch64.TranslateAddressS1Off.
75
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
76
*phys_ptr = address;
77
*prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
78
*page_size = TARGET_PAGE_SIZE;
79
+
80
+ /* Fill in cacheattr a-la AArch64.TranslateAddressS1Off. */
81
+ hcr = arm_hcr_el2_eff(env);
82
+ cacheattrs->shareability = 0;
83
+ if (hcr & HCR_DC) {
84
+ if (hcr & HCR_DCT) {
85
+ memattr = 0xf0; /* Tagged, Normal, WB, RWA */
86
+ } else {
87
+ memattr = 0xff; /* Normal, WB, RWA */
88
+ }
89
+ } else if (access_type == MMU_INST_FETCH) {
90
+ if (regime_sctlr(env, mmu_idx) & SCTLR_I) {
91
+ memattr = 0xee; /* Normal, WT, RA, NT */
92
+ } else {
93
+ memattr = 0x44; /* Normal, NC, No */
94
+ }
95
+ cacheattrs->shareability = 2; /* outer sharable */
96
+ } else {
97
+ memattr = 0x00; /* Device, nGnRnE */
98
+ }
99
+ cacheattrs->attrs = memattr;
100
return 0;
101
}
102
103
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
104
index XXXXXXX..XXXXXXX 100644
105
--- a/target/arm/tlb_helper.c
106
+++ b/target/arm/tlb_helper.c
107
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
108
phys_addr &= TARGET_PAGE_MASK;
109
address &= TARGET_PAGE_MASK;
110
}
111
+ /* Notice and record tagged memory. */
112
+ if (cpu_isar_feature(aa64_mte, cpu) && cacheattrs.attrs == 0xf0) {
113
+ arm_tlb_mte_tagged(&attrs) = true;
114
+ }
115
+
116
tlb_set_page_with_attrs(cs, address, phys_addr, attrs,
117
prot, mmu_idx, page_size);
118
return true;
119
--
28
--
120
2.20.1
29
2.20.1
121
30
122
31
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Vikram Garhwal <fnu.vikram@xilinx.com>
2
2
3
Look up the physical address for the given virtual address,
3
The Xilinx ZynqMP CAN controller is developed based on SocketCAN, QEMU CAN bus
4
convert that to a tag physical address, and finally return
4
implementation. Bus connection and socketCAN connection for each CAN module
5
the host address that backs it.
5
can be set through command lines.
6
6
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Example for using single CAN:
8
Message-id: 20200626033144.790098-45-richard.henderson@linaro.org
8
-object can-bus,id=canbus0 \
9
-machine xlnx-zcu102.canbus0=canbus0 \
10
-object can-host-socketcan,id=socketcan0,if=vcan0,canbus=canbus0
11
12
Example for connecting both CAN to same virtual CAN on host machine:
13
-object can-bus,id=canbus0 -object can-bus,id=canbus1 \
14
-machine xlnx-zcu102.canbus0=canbus0 \
15
-machine xlnx-zcu102.canbus1=canbus1 \
16
-object can-host-socketcan,id=socketcan0,if=vcan0,canbus=canbus0 \
17
-object can-host-socketcan,id=socketcan1,if=vcan0,canbus=canbus1
18
19
To create virtual CAN on the host machine, please check the QEMU CAN docs:
20
https://github.com/qemu/qemu/blob/master/docs/can.txt
21
22
Signed-off-by: Vikram Garhwal <fnu.vikram@xilinx.com>
23
Message-id: 1605728926-352690-2-git-send-email-fnu.vikram@xilinx.com
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
24
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
25
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
26
---
12
target/arm/mte_helper.c | 131 ++++++++++++++++++++++++++++++++++++++++
27
meson.build | 1 +
13
1 file changed, 131 insertions(+)
28
hw/net/can/trace.h | 1 +
29
include/hw/net/xlnx-zynqmp-can.h | 78 ++
30
hw/net/can/xlnx-zynqmp-can.c | 1161 ++++++++++++++++++++++++++++++
31
hw/Kconfig | 1 +
32
hw/net/can/meson.build | 1 +
33
hw/net/can/trace-events | 9 +
34
7 files changed, 1252 insertions(+)
35
create mode 100644 hw/net/can/trace.h
36
create mode 100644 include/hw/net/xlnx-zynqmp-can.h
37
create mode 100644 hw/net/can/xlnx-zynqmp-can.c
38
create mode 100644 hw/net/can/trace-events
14
39
15
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
40
diff --git a/meson.build b/meson.build
16
index XXXXXXX..XXXXXXX 100644
41
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/mte_helper.c
42
--- a/meson.build
18
+++ b/target/arm/mte_helper.c
43
+++ b/meson.build
44
@@ -XXX,XX +XXX,XX @@ if have_system
45
'hw/misc',
46
'hw/misc/macio',
47
'hw/net',
48
+ 'hw/net/can',
49
'hw/nvram',
50
'hw/pci',
51
'hw/pci-host',
52
diff --git a/hw/net/can/trace.h b/hw/net/can/trace.h
53
new file mode 100644
54
index XXXXXXX..XXXXXXX
55
--- /dev/null
56
+++ b/hw/net/can/trace.h
57
@@ -0,0 +1 @@
58
+#include "trace/trace-hw_net_can.h"
59
diff --git a/include/hw/net/xlnx-zynqmp-can.h b/include/hw/net/xlnx-zynqmp-can.h
60
new file mode 100644
61
index XXXXXXX..XXXXXXX
62
--- /dev/null
63
+++ b/include/hw/net/xlnx-zynqmp-can.h
19
@@ -XXX,XX +XXX,XX @@
64
@@ -XXX,XX +XXX,XX @@
20
#include "cpu.h"
65
+/*
21
#include "internals.h"
66
+ * QEMU model of the Xilinx ZynqMP CAN controller.
22
#include "exec/exec-all.h"
67
+ *
23
+#include "exec/ram_addr.h"
68
+ * Copyright (c) 2020 Xilinx Inc.
24
#include "exec/cpu_ldst.h"
69
+ *
25
#include "exec/helper-proto.h"
70
+ * Written-by: Vikram Garhwal<fnu.vikram@xilinx.com>
26
71
+ *
27
@@ -XXX,XX +XXX,XX @@ static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
72
+ * Based on QEMU CAN Device emulation implemented by Jin Yang, Deniz Eren and
28
int ptr_size, MMUAccessType tag_access,
73
+ * Pavel Pisa.
29
int tag_size, uintptr_t ra)
74
+ *
30
{
75
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
31
+#ifdef CONFIG_USER_ONLY
76
+ * of this software and associated documentation files (the "Software"), to deal
32
/* Tag storage not implemented. */
77
+ * in the Software without restriction, including without limitation the rights
33
return NULL;
78
+ * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
34
+#else
79
+ * copies of the Software, and to permit persons to whom the Software is
35
+ uintptr_t index;
80
+ * furnished to do so, subject to the following conditions:
36
+ CPUIOTLBEntry *iotlbentry;
81
+ *
37
+ int in_page, flags;
82
+ * The above copyright notice and this permission notice shall be included in
38
+ ram_addr_t ptr_ra;
83
+ * all copies or substantial portions of the Software.
39
+ hwaddr ptr_paddr, tag_paddr, xlat;
84
+ *
40
+ MemoryRegion *mr;
85
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
41
+ ARMASIdx tag_asi;
86
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
42
+ AddressSpace *tag_as;
87
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
43
+ void *host;
88
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
89
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
90
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
91
+ * THE SOFTWARE.
92
+ */
93
+
94
+#ifndef XLNX_ZYNQMP_CAN_H
95
+#define XLNX_ZYNQMP_CAN_H
96
+
97
+#include "hw/register.h"
98
+#include "net/can_emu.h"
99
+#include "net/can_host.h"
100
+#include "qemu/fifo32.h"
101
+#include "hw/ptimer.h"
102
+#include "hw/qdev-clock.h"
103
+
104
+#define TYPE_XLNX_ZYNQMP_CAN "xlnx.zynqmp-can"
105
+
106
+#define XLNX_ZYNQMP_CAN(obj) \
107
+ OBJECT_CHECK(XlnxZynqMPCANState, (obj), TYPE_XLNX_ZYNQMP_CAN)
108
+
109
+#define MAX_CAN_CTRLS 2
110
+#define XLNX_ZYNQMP_CAN_R_MAX (0x84 / 4)
111
+#define MAILBOX_CAPACITY 64
112
+#define CAN_TIMER_MAX 0XFFFFUL
113
+#define CAN_DEFAULT_CLOCK (24 * 1000 * 1000)
114
+
115
+/* Each CAN_FRAME will have 4 * 32bit size. */
116
+#define CAN_FRAME_SIZE 4
117
+#define RXFIFO_SIZE (MAILBOX_CAPACITY * CAN_FRAME_SIZE)
118
+
119
+typedef struct XlnxZynqMPCANState {
120
+ SysBusDevice parent_obj;
121
+ MemoryRegion iomem;
122
+
123
+ qemu_irq irq;
124
+
125
+ CanBusClientState bus_client;
126
+ CanBusState *canbus;
127
+
128
+ struct {
129
+ uint32_t ext_clk_freq;
130
+ } cfg;
131
+
132
+ RegisterInfo reg_info[XLNX_ZYNQMP_CAN_R_MAX];
133
+ uint32_t regs[XLNX_ZYNQMP_CAN_R_MAX];
134
+
135
+ Fifo32 rx_fifo;
136
+ Fifo32 tx_fifo;
137
+ Fifo32 txhpb_fifo;
138
+
139
+ ptimer_state *can_timer;
140
+} XlnxZynqMPCANState;
141
+
142
+#endif
143
diff --git a/hw/net/can/xlnx-zynqmp-can.c b/hw/net/can/xlnx-zynqmp-can.c
144
new file mode 100644
145
index XXXXXXX..XXXXXXX
146
--- /dev/null
147
+++ b/hw/net/can/xlnx-zynqmp-can.c
148
@@ -XXX,XX +XXX,XX @@
149
+/*
150
+ * QEMU model of the Xilinx ZynqMP CAN controller.
151
+ * This implementation is based on the following datasheet:
152
+ * https://www.xilinx.com/support/documentation/user_guides/ug1085-zynq-ultrascale-trm.pdf
153
+ *
154
+ * Copyright (c) 2020 Xilinx Inc.
155
+ *
156
+ * Written-by: Vikram Garhwal<fnu.vikram@xilinx.com>
157
+ *
158
+ * Based on QEMU CAN Device emulation implemented by Jin Yang, Deniz Eren and
159
+ * Pavel Pisa
160
+ *
161
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
162
+ * of this software and associated documentation files (the "Software"), to deal
163
+ * in the Software without restriction, including without limitation the rights
164
+ * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
165
+ * copies of the Software, and to permit persons to whom the Software is
166
+ * furnished to do so, subject to the following conditions:
167
+ *
168
+ * The above copyright notice and this permission notice shall be included in
169
+ * all copies or substantial portions of the Software.
170
+ *
171
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
172
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
173
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
174
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
175
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
176
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
177
+ * THE SOFTWARE.
178
+ */
179
+
180
+#include "qemu/osdep.h"
181
+#include "hw/sysbus.h"
182
+#include "hw/register.h"
183
+#include "hw/irq.h"
184
+#include "qapi/error.h"
185
+#include "qemu/bitops.h"
186
+#include "qemu/log.h"
187
+#include "qemu/cutils.h"
188
+#include "sysemu/sysemu.h"
189
+#include "migration/vmstate.h"
190
+#include "hw/qdev-properties.h"
191
+#include "net/can_emu.h"
192
+#include "net/can_host.h"
193
+#include "qemu/event_notifier.h"
194
+#include "qom/object_interfaces.h"
195
+#include "hw/net/xlnx-zynqmp-can.h"
196
+#include "trace.h"
197
+
198
+#ifndef XLNX_ZYNQMP_CAN_ERR_DEBUG
199
+#define XLNX_ZYNQMP_CAN_ERR_DEBUG 0
200
+#endif
201
+
202
+#define MAX_DLC 8
203
+#undef ERROR
204
+
205
+REG32(SOFTWARE_RESET_REGISTER, 0x0)
206
+ FIELD(SOFTWARE_RESET_REGISTER, CEN, 1, 1)
207
+ FIELD(SOFTWARE_RESET_REGISTER, SRST, 0, 1)
208
+REG32(MODE_SELECT_REGISTER, 0x4)
209
+ FIELD(MODE_SELECT_REGISTER, SNOOP, 2, 1)
210
+ FIELD(MODE_SELECT_REGISTER, LBACK, 1, 1)
211
+ FIELD(MODE_SELECT_REGISTER, SLEEP, 0, 1)
212
+REG32(ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER, 0x8)
213
+ FIELD(ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER, BRP, 0, 8)
214
+REG32(ARBITRATION_PHASE_BIT_TIMING_REGISTER, 0xc)
215
+ FIELD(ARBITRATION_PHASE_BIT_TIMING_REGISTER, SJW, 7, 2)
216
+ FIELD(ARBITRATION_PHASE_BIT_TIMING_REGISTER, TS2, 4, 3)
217
+ FIELD(ARBITRATION_PHASE_BIT_TIMING_REGISTER, TS1, 0, 4)
218
+REG32(ERROR_COUNTER_REGISTER, 0x10)
219
+ FIELD(ERROR_COUNTER_REGISTER, REC, 8, 8)
220
+ FIELD(ERROR_COUNTER_REGISTER, TEC, 0, 8)
221
+REG32(ERROR_STATUS_REGISTER, 0x14)
222
+ FIELD(ERROR_STATUS_REGISTER, ACKER, 4, 1)
223
+ FIELD(ERROR_STATUS_REGISTER, BERR, 3, 1)
224
+ FIELD(ERROR_STATUS_REGISTER, STER, 2, 1)
225
+ FIELD(ERROR_STATUS_REGISTER, FMER, 1, 1)
226
+ FIELD(ERROR_STATUS_REGISTER, CRCER, 0, 1)
227
+REG32(STATUS_REGISTER, 0x18)
228
+ FIELD(STATUS_REGISTER, SNOOP, 12, 1)
229
+ FIELD(STATUS_REGISTER, ACFBSY, 11, 1)
230
+ FIELD(STATUS_REGISTER, TXFLL, 10, 1)
231
+ FIELD(STATUS_REGISTER, TXBFLL, 9, 1)
232
+ FIELD(STATUS_REGISTER, ESTAT, 7, 2)
233
+ FIELD(STATUS_REGISTER, ERRWRN, 6, 1)
234
+ FIELD(STATUS_REGISTER, BBSY, 5, 1)
235
+ FIELD(STATUS_REGISTER, BIDLE, 4, 1)
236
+ FIELD(STATUS_REGISTER, NORMAL, 3, 1)
237
+ FIELD(STATUS_REGISTER, SLEEP, 2, 1)
238
+ FIELD(STATUS_REGISTER, LBACK, 1, 1)
239
+ FIELD(STATUS_REGISTER, CONFIG, 0, 1)
240
+REG32(INTERRUPT_STATUS_REGISTER, 0x1c)
241
+ FIELD(INTERRUPT_STATUS_REGISTER, TXFEMP, 14, 1)
242
+ FIELD(INTERRUPT_STATUS_REGISTER, TXFWMEMP, 13, 1)
243
+ FIELD(INTERRUPT_STATUS_REGISTER, RXFWMFLL, 12, 1)
244
+ FIELD(INTERRUPT_STATUS_REGISTER, WKUP, 11, 1)
245
+ FIELD(INTERRUPT_STATUS_REGISTER, SLP, 10, 1)
246
+ FIELD(INTERRUPT_STATUS_REGISTER, BSOFF, 9, 1)
247
+ FIELD(INTERRUPT_STATUS_REGISTER, ERROR, 8, 1)
248
+ FIELD(INTERRUPT_STATUS_REGISTER, RXNEMP, 7, 1)
249
+ FIELD(INTERRUPT_STATUS_REGISTER, RXOFLW, 6, 1)
250
+ FIELD(INTERRUPT_STATUS_REGISTER, RXUFLW, 5, 1)
251
+ FIELD(INTERRUPT_STATUS_REGISTER, RXOK, 4, 1)
252
+ FIELD(INTERRUPT_STATUS_REGISTER, TXBFLL, 3, 1)
253
+ FIELD(INTERRUPT_STATUS_REGISTER, TXFLL, 2, 1)
254
+ FIELD(INTERRUPT_STATUS_REGISTER, TXOK, 1, 1)
255
+ FIELD(INTERRUPT_STATUS_REGISTER, ARBLST, 0, 1)
256
+REG32(INTERRUPT_ENABLE_REGISTER, 0x20)
257
+ FIELD(INTERRUPT_ENABLE_REGISTER, ETXFEMP, 14, 1)
258
+ FIELD(INTERRUPT_ENABLE_REGISTER, ETXFWMEMP, 13, 1)
259
+ FIELD(INTERRUPT_ENABLE_REGISTER, ERXFWMFLL, 12, 1)
260
+ FIELD(INTERRUPT_ENABLE_REGISTER, EWKUP, 11, 1)
261
+ FIELD(INTERRUPT_ENABLE_REGISTER, ESLP, 10, 1)
262
+ FIELD(INTERRUPT_ENABLE_REGISTER, EBSOFF, 9, 1)
263
+ FIELD(INTERRUPT_ENABLE_REGISTER, EERROR, 8, 1)
264
+ FIELD(INTERRUPT_ENABLE_REGISTER, ERXNEMP, 7, 1)
265
+ FIELD(INTERRUPT_ENABLE_REGISTER, ERXOFLW, 6, 1)
266
+ FIELD(INTERRUPT_ENABLE_REGISTER, ERXUFLW, 5, 1)
267
+ FIELD(INTERRUPT_ENABLE_REGISTER, ERXOK, 4, 1)
268
+ FIELD(INTERRUPT_ENABLE_REGISTER, ETXBFLL, 3, 1)
269
+ FIELD(INTERRUPT_ENABLE_REGISTER, ETXFLL, 2, 1)
270
+ FIELD(INTERRUPT_ENABLE_REGISTER, ETXOK, 1, 1)
271
+ FIELD(INTERRUPT_ENABLE_REGISTER, EARBLST, 0, 1)
272
+REG32(INTERRUPT_CLEAR_REGISTER, 0x24)
273
+ FIELD(INTERRUPT_CLEAR_REGISTER, CTXFEMP, 14, 1)
274
+ FIELD(INTERRUPT_CLEAR_REGISTER, CTXFWMEMP, 13, 1)
275
+ FIELD(INTERRUPT_CLEAR_REGISTER, CRXFWMFLL, 12, 1)
276
+ FIELD(INTERRUPT_CLEAR_REGISTER, CWKUP, 11, 1)
277
+ FIELD(INTERRUPT_CLEAR_REGISTER, CSLP, 10, 1)
278
+ FIELD(INTERRUPT_CLEAR_REGISTER, CBSOFF, 9, 1)
279
+ FIELD(INTERRUPT_CLEAR_REGISTER, CERROR, 8, 1)
280
+ FIELD(INTERRUPT_CLEAR_REGISTER, CRXNEMP, 7, 1)
281
+ FIELD(INTERRUPT_CLEAR_REGISTER, CRXOFLW, 6, 1)
282
+ FIELD(INTERRUPT_CLEAR_REGISTER, CRXUFLW, 5, 1)
283
+ FIELD(INTERRUPT_CLEAR_REGISTER, CRXOK, 4, 1)
284
+ FIELD(INTERRUPT_CLEAR_REGISTER, CTXBFLL, 3, 1)
285
+ FIELD(INTERRUPT_CLEAR_REGISTER, CTXFLL, 2, 1)
286
+ FIELD(INTERRUPT_CLEAR_REGISTER, CTXOK, 1, 1)
287
+ FIELD(INTERRUPT_CLEAR_REGISTER, CARBLST, 0, 1)
288
+REG32(TIMESTAMP_REGISTER, 0x28)
289
+ FIELD(TIMESTAMP_REGISTER, CTS, 0, 1)
290
+REG32(WIR, 0x2c)
291
+ FIELD(WIR, EW, 8, 8)
292
+ FIELD(WIR, FW, 0, 8)
293
+REG32(TXFIFO_ID, 0x30)
294
+ FIELD(TXFIFO_ID, IDH, 21, 11)
295
+ FIELD(TXFIFO_ID, SRRRTR, 20, 1)
296
+ FIELD(TXFIFO_ID, IDE, 19, 1)
297
+ FIELD(TXFIFO_ID, IDL, 1, 18)
298
+ FIELD(TXFIFO_ID, RTR, 0, 1)
299
+REG32(TXFIFO_DLC, 0x34)
300
+ FIELD(TXFIFO_DLC, DLC, 28, 4)
301
+REG32(TXFIFO_DATA1, 0x38)
302
+ FIELD(TXFIFO_DATA1, DB0, 24, 8)
303
+ FIELD(TXFIFO_DATA1, DB1, 16, 8)
304
+ FIELD(TXFIFO_DATA1, DB2, 8, 8)
305
+ FIELD(TXFIFO_DATA1, DB3, 0, 8)
306
+REG32(TXFIFO_DATA2, 0x3c)
307
+ FIELD(TXFIFO_DATA2, DB4, 24, 8)
308
+ FIELD(TXFIFO_DATA2, DB5, 16, 8)
309
+ FIELD(TXFIFO_DATA2, DB6, 8, 8)
310
+ FIELD(TXFIFO_DATA2, DB7, 0, 8)
311
+REG32(TXHPB_ID, 0x40)
312
+ FIELD(TXHPB_ID, IDH, 21, 11)
313
+ FIELD(TXHPB_ID, SRRRTR, 20, 1)
314
+ FIELD(TXHPB_ID, IDE, 19, 1)
315
+ FIELD(TXHPB_ID, IDL, 1, 18)
316
+ FIELD(TXHPB_ID, RTR, 0, 1)
317
+REG32(TXHPB_DLC, 0x44)
318
+ FIELD(TXHPB_DLC, DLC, 28, 4)
319
+REG32(TXHPB_DATA1, 0x48)
320
+ FIELD(TXHPB_DATA1, DB0, 24, 8)
321
+ FIELD(TXHPB_DATA1, DB1, 16, 8)
322
+ FIELD(TXHPB_DATA1, DB2, 8, 8)
323
+ FIELD(TXHPB_DATA1, DB3, 0, 8)
324
+REG32(TXHPB_DATA2, 0x4c)
325
+ FIELD(TXHPB_DATA2, DB4, 24, 8)
326
+ FIELD(TXHPB_DATA2, DB5, 16, 8)
327
+ FIELD(TXHPB_DATA2, DB6, 8, 8)
328
+ FIELD(TXHPB_DATA2, DB7, 0, 8)
329
+REG32(RXFIFO_ID, 0x50)
330
+ FIELD(RXFIFO_ID, IDH, 21, 11)
331
+ FIELD(RXFIFO_ID, SRRRTR, 20, 1)
332
+ FIELD(RXFIFO_ID, IDE, 19, 1)
333
+ FIELD(RXFIFO_ID, IDL, 1, 18)
334
+ FIELD(RXFIFO_ID, RTR, 0, 1)
335
+REG32(RXFIFO_DLC, 0x54)
336
+ FIELD(RXFIFO_DLC, DLC, 28, 4)
337
+ FIELD(RXFIFO_DLC, RXT, 0, 16)
338
+REG32(RXFIFO_DATA1, 0x58)
339
+ FIELD(RXFIFO_DATA1, DB0, 24, 8)
340
+ FIELD(RXFIFO_DATA1, DB1, 16, 8)
341
+ FIELD(RXFIFO_DATA1, DB2, 8, 8)
342
+ FIELD(RXFIFO_DATA1, DB3, 0, 8)
343
+REG32(RXFIFO_DATA2, 0x5c)
344
+ FIELD(RXFIFO_DATA2, DB4, 24, 8)
345
+ FIELD(RXFIFO_DATA2, DB5, 16, 8)
346
+ FIELD(RXFIFO_DATA2, DB6, 8, 8)
347
+ FIELD(RXFIFO_DATA2, DB7, 0, 8)
348
+REG32(AFR, 0x60)
349
+ FIELD(AFR, UAF4, 3, 1)
350
+ FIELD(AFR, UAF3, 2, 1)
351
+ FIELD(AFR, UAF2, 1, 1)
352
+ FIELD(AFR, UAF1, 0, 1)
353
+REG32(AFMR1, 0x64)
354
+ FIELD(AFMR1, AMIDH, 21, 11)
355
+ FIELD(AFMR1, AMSRR, 20, 1)
356
+ FIELD(AFMR1, AMIDE, 19, 1)
357
+ FIELD(AFMR1, AMIDL, 1, 18)
358
+ FIELD(AFMR1, AMRTR, 0, 1)
359
+REG32(AFIR1, 0x68)
360
+ FIELD(AFIR1, AIIDH, 21, 11)
361
+ FIELD(AFIR1, AISRR, 20, 1)
362
+ FIELD(AFIR1, AIIDE, 19, 1)
363
+ FIELD(AFIR1, AIIDL, 1, 18)
364
+ FIELD(AFIR1, AIRTR, 0, 1)
365
+REG32(AFMR2, 0x6c)
366
+ FIELD(AFMR2, AMIDH, 21, 11)
367
+ FIELD(AFMR2, AMSRR, 20, 1)
368
+ FIELD(AFMR2, AMIDE, 19, 1)
369
+ FIELD(AFMR2, AMIDL, 1, 18)
370
+ FIELD(AFMR2, AMRTR, 0, 1)
371
+REG32(AFIR2, 0x70)
372
+ FIELD(AFIR2, AIIDH, 21, 11)
373
+ FIELD(AFIR2, AISRR, 20, 1)
374
+ FIELD(AFIR2, AIIDE, 19, 1)
375
+ FIELD(AFIR2, AIIDL, 1, 18)
376
+ FIELD(AFIR2, AIRTR, 0, 1)
377
+REG32(AFMR3, 0x74)
378
+ FIELD(AFMR3, AMIDH, 21, 11)
379
+ FIELD(AFMR3, AMSRR, 20, 1)
380
+ FIELD(AFMR3, AMIDE, 19, 1)
381
+ FIELD(AFMR3, AMIDL, 1, 18)
382
+ FIELD(AFMR3, AMRTR, 0, 1)
383
+REG32(AFIR3, 0x78)
384
+ FIELD(AFIR3, AIIDH, 21, 11)
385
+ FIELD(AFIR3, AISRR, 20, 1)
386
+ FIELD(AFIR3, AIIDE, 19, 1)
387
+ FIELD(AFIR3, AIIDL, 1, 18)
388
+ FIELD(AFIR3, AIRTR, 0, 1)
389
+REG32(AFMR4, 0x7c)
390
+ FIELD(AFMR4, AMIDH, 21, 11)
391
+ FIELD(AFMR4, AMSRR, 20, 1)
392
+ FIELD(AFMR4, AMIDE, 19, 1)
393
+ FIELD(AFMR4, AMIDL, 1, 18)
394
+ FIELD(AFMR4, AMRTR, 0, 1)
395
+REG32(AFIR4, 0x80)
396
+ FIELD(AFIR4, AIIDH, 21, 11)
397
+ FIELD(AFIR4, AISRR, 20, 1)
398
+ FIELD(AFIR4, AIIDE, 19, 1)
399
+ FIELD(AFIR4, AIIDL, 1, 18)
400
+ FIELD(AFIR4, AIRTR, 0, 1)
401
+
402
+static void can_update_irq(XlnxZynqMPCANState *s)
403
+{
404
+ uint32_t irq;
405
+
406
+ /* Watermark register interrupts. */
407
+ if ((fifo32_num_free(&s->tx_fifo) / CAN_FRAME_SIZE) >
408
+ ARRAY_FIELD_EX32(s->regs, WIR, EW)) {
409
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXFWMEMP, 1);
410
+ }
411
+
412
+ if ((fifo32_num_used(&s->rx_fifo) / CAN_FRAME_SIZE) >
413
+ ARRAY_FIELD_EX32(s->regs, WIR, FW)) {
414
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXFWMFLL, 1);
415
+ }
416
+
417
+ /* RX Interrupts. */
418
+ if (fifo32_num_used(&s->rx_fifo) >= CAN_FRAME_SIZE) {
419
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXNEMP, 1);
420
+ }
421
+
422
+ /* TX interrupts. */
423
+ if (fifo32_is_empty(&s->tx_fifo)) {
424
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXFEMP, 1);
425
+ }
426
+
427
+ if (fifo32_is_full(&s->tx_fifo)) {
428
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXFLL, 1);
429
+ }
430
+
431
+ if (fifo32_is_full(&s->txhpb_fifo)) {
432
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXBFLL, 1);
433
+ }
434
+
435
+ irq = s->regs[R_INTERRUPT_STATUS_REGISTER];
436
+ irq &= s->regs[R_INTERRUPT_ENABLE_REGISTER];
437
+
438
+ trace_xlnx_can_update_irq(s->regs[R_INTERRUPT_STATUS_REGISTER],
439
+ s->regs[R_INTERRUPT_ENABLE_REGISTER], irq);
440
+ qemu_set_irq(s->irq, irq);
441
+}
442
+
443
+static void can_ier_post_write(RegisterInfo *reg, uint64_t val)
444
+{
445
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
446
+
447
+ can_update_irq(s);
448
+}
449
+
450
+static uint64_t can_icr_pre_write(RegisterInfo *reg, uint64_t val)
451
+{
452
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
453
+
454
+ s->regs[R_INTERRUPT_STATUS_REGISTER] &= ~val;
455
+ can_update_irq(s);
456
+
457
+ return 0;
458
+}
459
+
460
+static void can_config_reset(XlnxZynqMPCANState *s)
461
+{
462
+ /* Reset all the configuration registers. */
463
+ register_reset(&s->reg_info[R_SOFTWARE_RESET_REGISTER]);
464
+ register_reset(&s->reg_info[R_MODE_SELECT_REGISTER]);
465
+ register_reset(
466
+ &s->reg_info[R_ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER]);
467
+ register_reset(&s->reg_info[R_ARBITRATION_PHASE_BIT_TIMING_REGISTER]);
468
+ register_reset(&s->reg_info[R_STATUS_REGISTER]);
469
+ register_reset(&s->reg_info[R_INTERRUPT_STATUS_REGISTER]);
470
+ register_reset(&s->reg_info[R_INTERRUPT_ENABLE_REGISTER]);
471
+ register_reset(&s->reg_info[R_INTERRUPT_CLEAR_REGISTER]);
472
+ register_reset(&s->reg_info[R_WIR]);
473
+}
474
+
475
+static void can_config_mode(XlnxZynqMPCANState *s)
476
+{
477
+ register_reset(&s->reg_info[R_ERROR_COUNTER_REGISTER]);
478
+ register_reset(&s->reg_info[R_ERROR_STATUS_REGISTER]);
479
+
480
+ /* Put XlnxZynqMPCAN in configuration mode. */
481
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, CONFIG, 1);
482
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, WKUP, 0);
483
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, SLP, 0);
484
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, BSOFF, 0);
485
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, ERROR, 0);
486
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOFLW, 0);
487
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOK, 0);
488
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXOK, 0);
489
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, ARBLST, 0);
490
+
491
+ can_update_irq(s);
492
+}
493
+
494
+static void update_status_register_mode_bits(XlnxZynqMPCANState *s)
495
+{
496
+ bool sleep_status = ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SLEEP);
497
+ bool sleep_mode = ARRAY_FIELD_EX32(s->regs, MODE_SELECT_REGISTER, SLEEP);
498
+ /* Wake up interrupt bit. */
499
+ bool wakeup_irq_val = sleep_status && (sleep_mode == 0);
500
+ /* Sleep interrupt bit. */
501
+ bool sleep_irq_val = sleep_mode && (sleep_status == 0);
502
+
503
+ /* Clear previous core mode status bits. */
504
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, LBACK, 0);
505
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, SLEEP, 0);
506
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, SNOOP, 0);
507
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, NORMAL, 0);
508
+
509
+ /* set current mode bit and generate irqs accordingly. */
510
+ if (ARRAY_FIELD_EX32(s->regs, MODE_SELECT_REGISTER, LBACK)) {
511
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, LBACK, 1);
512
+ } else if (ARRAY_FIELD_EX32(s->regs, MODE_SELECT_REGISTER, SLEEP)) {
513
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, SLEEP, 1);
514
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, SLP,
515
+ sleep_irq_val);
516
+ } else if (ARRAY_FIELD_EX32(s->regs, MODE_SELECT_REGISTER, SNOOP)) {
517
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, SNOOP, 1);
518
+ } else {
519
+ /*
520
+ * If all bits are zero then XlnxZynqMPCAN is set in normal mode.
521
+ */
522
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, NORMAL, 1);
523
+ /* Set wakeup interrupt bit. */
524
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, WKUP,
525
+ wakeup_irq_val);
526
+ }
527
+
528
+ can_update_irq(s);
529
+}
530
+
531
+static void can_exit_sleep_mode(XlnxZynqMPCANState *s)
532
+{
533
+ ARRAY_FIELD_DP32(s->regs, MODE_SELECT_REGISTER, SLEEP, 0);
534
+ update_status_register_mode_bits(s);
535
+}
536
+
537
+static void generate_frame(qemu_can_frame *frame, uint32_t *data)
538
+{
539
+ frame->can_id = data[0];
540
+ frame->can_dlc = FIELD_EX32(data[1], TXFIFO_DLC, DLC);
541
+
542
+ frame->data[0] = FIELD_EX32(data[2], TXFIFO_DATA1, DB3);
543
+ frame->data[1] = FIELD_EX32(data[2], TXFIFO_DATA1, DB2);
544
+ frame->data[2] = FIELD_EX32(data[2], TXFIFO_DATA1, DB1);
545
+ frame->data[3] = FIELD_EX32(data[2], TXFIFO_DATA1, DB0);
546
+
547
+ frame->data[4] = FIELD_EX32(data[3], TXFIFO_DATA2, DB7);
548
+ frame->data[5] = FIELD_EX32(data[3], TXFIFO_DATA2, DB6);
549
+ frame->data[6] = FIELD_EX32(data[3], TXFIFO_DATA2, DB5);
550
+ frame->data[7] = FIELD_EX32(data[3], TXFIFO_DATA2, DB4);
551
+}
552
+
553
+static bool tx_ready_check(XlnxZynqMPCANState *s)
554
+{
555
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, SRST)) {
556
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
557
+
558
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to transfer data while"
559
+ " data while controller is in reset mode.\n",
560
+ path);
561
+ return false;
562
+ }
563
+
564
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN) == 0) {
565
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
566
+
567
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to transfer"
568
+ " data while controller is in configuration mode. Reset"
569
+ " the core so operations can start fresh.\n",
570
+ path);
571
+ return false;
572
+ }
573
+
574
+ if (ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SNOOP)) {
575
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
576
+
577
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to transfer"
578
+ " data while controller is in SNOOP MODE.\n",
579
+ path);
580
+ return false;
581
+ }
582
+
583
+ return true;
584
+}
585
+
586
+static void transfer_fifo(XlnxZynqMPCANState *s, Fifo32 *fifo)
587
+{
588
+ qemu_can_frame frame;
589
+ uint32_t data[CAN_FRAME_SIZE];
590
+ int i;
591
+ bool can_tx = tx_ready_check(s);
592
+
593
+ if (!can_tx) {
594
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
595
+
596
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Controller is not enabled for data"
597
+ " transfer.\n", path);
598
+ can_update_irq(s);
599
+ return;
600
+ }
601
+
602
+ while (!fifo32_is_empty(fifo)) {
603
+ for (i = 0; i < CAN_FRAME_SIZE; i++) {
604
+ data[i] = fifo32_pop(fifo);
605
+ }
606
+
607
+ if (ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, LBACK)) {
608
+ /*
609
+ * Controller is in loopback. In Loopback mode, the CAN core
610
+ * transmits a recessive bitstream on to the XlnxZynqMPCAN Bus.
611
+ * Any message transmitted is looped back to the RX line and
612
+ * acknowledged. The XlnxZynqMPCAN core receives any message
613
+ * that it transmits.
614
+ */
615
+ if (fifo32_is_full(&s->rx_fifo)) {
616
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOFLW, 1);
617
+ } else {
618
+ for (i = 0; i < CAN_FRAME_SIZE; i++) {
619
+ fifo32_push(&s->rx_fifo, data[i]);
620
+ }
621
+
622
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOK, 1);
623
+ }
624
+ } else {
625
+ /* Normal mode Tx. */
626
+ generate_frame(&frame, data);
627
+
628
+ trace_xlnx_can_tx_data(frame.can_id, frame.can_dlc,
629
+ frame.data[0], frame.data[1],
630
+ frame.data[2], frame.data[3],
631
+ frame.data[4], frame.data[5],
632
+ frame.data[6], frame.data[7]);
633
+ can_bus_client_send(&s->bus_client, &frame, 1);
634
+ }
635
+ }
636
+
637
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, TXOK, 1);
638
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, TXBFLL, 0);
639
+
640
+ if (ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SLEEP)) {
641
+ can_exit_sleep_mode(s);
642
+ }
643
+
644
+ can_update_irq(s);
645
+}
646
+
647
+static uint64_t can_srr_pre_write(RegisterInfo *reg, uint64_t val)
648
+{
649
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
650
+
651
+ ARRAY_FIELD_DP32(s->regs, SOFTWARE_RESET_REGISTER, CEN,
652
+ FIELD_EX32(val, SOFTWARE_RESET_REGISTER, CEN));
653
+
654
+ if (FIELD_EX32(val, SOFTWARE_RESET_REGISTER, SRST)) {
655
+ trace_xlnx_can_reset(val);
656
+
657
+ /* First, core will do software reset then will enter in config mode. */
658
+ can_config_reset(s);
659
+ }
660
+
661
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN) == 0) {
662
+ can_config_mode(s);
663
+ } else {
664
+ /*
665
+ * Leave config mode. Now XlnxZynqMPCAN core will enter normal,
666
+ * sleep, snoop or loopback mode depending upon LBACK, SLEEP, SNOOP
667
+ * register states.
668
+ */
669
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, CONFIG, 0);
670
+
671
+ ptimer_transaction_begin(s->can_timer);
672
+ ptimer_set_count(s->can_timer, 0);
673
+ ptimer_transaction_commit(s->can_timer);
674
+
675
+ /* XlnxZynqMPCAN is out of config mode. It will send pending data. */
676
+ transfer_fifo(s, &s->txhpb_fifo);
677
+ transfer_fifo(s, &s->tx_fifo);
678
+ }
679
+
680
+ update_status_register_mode_bits(s);
681
+
682
+ return s->regs[R_SOFTWARE_RESET_REGISTER];
683
+}
684
+
685
+static uint64_t can_msr_pre_write(RegisterInfo *reg, uint64_t val)
686
+{
687
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
688
+ uint8_t multi_mode;
44
+
689
+
45
+ /*
690
+ /*
46
+ * Probe the first byte of the virtual address. This raises an
691
+ * Multiple mode set check. This is done to make sure user doesn't set
47
+ * exception for inaccessible pages, and resolves the virtual address
692
+ * multiple modes.
48
+ * into the softmmu tlb.
49
+ *
50
+ * When RA == 0, this is for mte_probe1. The page is expected to be
51
+ * valid. Indicate to probe_access_flags no-fault, then assert that
52
+ * we received a valid page.
53
+ */
693
+ */
54
+ flags = probe_access_flags(env, ptr, ptr_access, ptr_mmu_idx,
694
+ multi_mode = FIELD_EX32(val, MODE_SELECT_REGISTER, LBACK) +
55
+ ra == 0, &host, ra);
695
+ FIELD_EX32(val, MODE_SELECT_REGISTER, SLEEP) +
56
+ assert(!(flags & TLB_INVALID_MASK));
696
+ FIELD_EX32(val, MODE_SELECT_REGISTER, SNOOP);
697
+
698
+ if (multi_mode > 1) {
699
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
700
+
701
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to config"
702
+ " several modes simultaneously. One mode will be selected"
703
+ " according to their priority: LBACK > SLEEP > SNOOP.\n",
704
+ path);
705
+ }
706
+
707
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN) == 0) {
708
+ /* We are in configuration mode, any mode can be selected. */
709
+ s->regs[R_MODE_SELECT_REGISTER] = val;
710
+ } else {
711
+ bool sleep_mode_bit = FIELD_EX32(val, MODE_SELECT_REGISTER, SLEEP);
712
+
713
+ ARRAY_FIELD_DP32(s->regs, MODE_SELECT_REGISTER, SLEEP, sleep_mode_bit);
714
+
715
+ if (FIELD_EX32(val, MODE_SELECT_REGISTER, LBACK)) {
716
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
717
+
718
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to set"
719
+ " LBACK mode without setting CEN bit as 0.\n",
720
+ path);
721
+ } else if (FIELD_EX32(val, MODE_SELECT_REGISTER, SNOOP)) {
722
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
723
+
724
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Attempting to set"
725
+ " SNOOP mode without setting CEN bit as 0.\n",
726
+ path);
727
+ }
728
+
729
+ update_status_register_mode_bits(s);
730
+ }
731
+
732
+ return s->regs[R_MODE_SELECT_REGISTER];
733
+}
734
+
735
+static uint64_t can_brpr_pre_write(RegisterInfo *reg, uint64_t val)
736
+{
737
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
738
+
739
+ /* Only allow writes when in config mode. */
740
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN)) {
741
+ return s->regs[R_ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER];
742
+ }
743
+
744
+ return val;
745
+}
746
+
747
+static uint64_t can_btr_pre_write(RegisterInfo *reg, uint64_t val)
748
+{
749
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
750
+
751
+ /* Only allow writes when in config mode. */
752
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN)) {
753
+ return s->regs[R_ARBITRATION_PHASE_BIT_TIMING_REGISTER];
754
+ }
755
+
756
+ return val;
757
+}
758
+
759
+static uint64_t can_tcr_pre_write(RegisterInfo *reg, uint64_t val)
760
+{
761
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
762
+
763
+ if (FIELD_EX32(val, TIMESTAMP_REGISTER, CTS)) {
764
+ ptimer_transaction_begin(s->can_timer);
765
+ ptimer_set_count(s->can_timer, 0);
766
+ ptimer_transaction_commit(s->can_timer);
767
+ }
768
+
769
+ return 0;
770
+}
771
+
772
+static void update_rx_fifo(XlnxZynqMPCANState *s, const qemu_can_frame *frame)
773
+{
774
+ bool filter_pass = false;
775
+ uint16_t timestamp = 0;
776
+
777
+ /* If no filter is enabled. Message will be stored in FIFO. */
778
+ if (!((ARRAY_FIELD_EX32(s->regs, AFR, UAF1)) |
779
+ (ARRAY_FIELD_EX32(s->regs, AFR, UAF2)) |
780
+ (ARRAY_FIELD_EX32(s->regs, AFR, UAF3)) |
781
+ (ARRAY_FIELD_EX32(s->regs, AFR, UAF4)))) {
782
+ filter_pass = true;
783
+ }
57
+
784
+
58
+ /*
785
+ /*
59
+ * Find the iotlbentry for ptr. This *must* be present in the TLB
786
+ * Messages that pass any of the acceptance filters will be stored in
60
+ * because we just found the mapping.
787
+ * the RX FIFO.
61
+ * TODO: Perhaps there should be a cputlb helper that returns a
62
+ * matching tlb entry + iotlb entry.
63
+ */
788
+ */
64
+ index = tlb_index(env, ptr_mmu_idx, ptr);
789
+ if (ARRAY_FIELD_EX32(s->regs, AFR, UAF1)) {
65
+# ifdef CONFIG_DEBUG_TCG
790
+ uint32_t id_masked = s->regs[R_AFMR1] & frame->can_id;
66
+ {
791
+ uint32_t filter_id_masked = s->regs[R_AFMR1] & s->regs[R_AFIR1];
67
+ CPUTLBEntry *entry = tlb_entry(env, ptr_mmu_idx, ptr);
792
+
68
+ target_ulong comparator = (ptr_access == MMU_DATA_LOAD
793
+ if (filter_id_masked == id_masked) {
69
+ ? entry->addr_read
794
+ filter_pass = true;
70
+ : tlb_addr_write(entry));
795
+ }
71
+ g_assert(tlb_hit(comparator, ptr));
796
+ }
72
+ }
797
+
73
+# endif
798
+ if (ARRAY_FIELD_EX32(s->regs, AFR, UAF2)) {
74
+ iotlbentry = &env_tlb(env)->d[ptr_mmu_idx].iotlb[index];
799
+ uint32_t id_masked = s->regs[R_AFMR2] & frame->can_id;
75
+
800
+ uint32_t filter_id_masked = s->regs[R_AFMR2] & s->regs[R_AFIR2];
76
+ /* If the virtual page MemAttr != Tagged, access unchecked. */
801
+
77
+ if (!arm_tlb_mte_tagged(&iotlbentry->attrs)) {
802
+ if (filter_id_masked == id_masked) {
78
+ return NULL;
803
+ filter_pass = true;
804
+ }
805
+ }
806
+
807
+ if (ARRAY_FIELD_EX32(s->regs, AFR, UAF3)) {
808
+ uint32_t id_masked = s->regs[R_AFMR3] & frame->can_id;
809
+ uint32_t filter_id_masked = s->regs[R_AFMR3] & s->regs[R_AFIR3];
810
+
811
+ if (filter_id_masked == id_masked) {
812
+ filter_pass = true;
813
+ }
814
+ }
815
+
816
+ if (ARRAY_FIELD_EX32(s->regs, AFR, UAF4)) {
817
+ uint32_t id_masked = s->regs[R_AFMR4] & frame->can_id;
818
+ uint32_t filter_id_masked = s->regs[R_AFMR4] & s->regs[R_AFIR4];
819
+
820
+ if (filter_id_masked == id_masked) {
821
+ filter_pass = true;
822
+ }
823
+ }
824
+
825
+ if (!filter_pass) {
826
+ trace_xlnx_can_rx_fifo_filter_reject(frame->can_id, frame->can_dlc);
827
+ return;
828
+ }
829
+
830
+ /* Store the message in fifo if it passed through any of the filters. */
831
+ if (filter_pass && frame->can_dlc <= MAX_DLC) {
832
+
833
+ if (fifo32_is_full(&s->rx_fifo)) {
834
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOFLW, 1);
835
+ } else {
836
+ timestamp = CAN_TIMER_MAX - ptimer_get_count(s->can_timer);
837
+
838
+ fifo32_push(&s->rx_fifo, frame->can_id);
839
+
840
+ fifo32_push(&s->rx_fifo, deposit32(0, R_RXFIFO_DLC_DLC_SHIFT,
841
+ R_RXFIFO_DLC_DLC_LENGTH,
842
+ frame->can_dlc) |
843
+ deposit32(0, R_RXFIFO_DLC_RXT_SHIFT,
844
+ R_RXFIFO_DLC_RXT_LENGTH,
845
+ timestamp));
846
+
847
+ /* First 32 bit of the data. */
848
+ fifo32_push(&s->rx_fifo, deposit32(0, R_TXFIFO_DATA1_DB3_SHIFT,
849
+ R_TXFIFO_DATA1_DB3_LENGTH,
850
+ frame->data[0]) |
851
+ deposit32(0, R_TXFIFO_DATA1_DB2_SHIFT,
852
+ R_TXFIFO_DATA1_DB2_LENGTH,
853
+ frame->data[1]) |
854
+ deposit32(0, R_TXFIFO_DATA1_DB1_SHIFT,
855
+ R_TXFIFO_DATA1_DB1_LENGTH,
856
+ frame->data[2]) |
857
+ deposit32(0, R_TXFIFO_DATA1_DB0_SHIFT,
858
+ R_TXFIFO_DATA1_DB0_LENGTH,
859
+ frame->data[3]));
860
+ /* Last 32 bit of the data. */
861
+ fifo32_push(&s->rx_fifo, deposit32(0, R_TXFIFO_DATA2_DB7_SHIFT,
862
+ R_TXFIFO_DATA2_DB7_LENGTH,
863
+ frame->data[4]) |
864
+ deposit32(0, R_TXFIFO_DATA2_DB6_SHIFT,
865
+ R_TXFIFO_DATA2_DB6_LENGTH,
866
+ frame->data[5]) |
867
+ deposit32(0, R_TXFIFO_DATA2_DB5_SHIFT,
868
+ R_TXFIFO_DATA2_DB5_LENGTH,
869
+ frame->data[6]) |
870
+ deposit32(0, R_TXFIFO_DATA2_DB4_SHIFT,
871
+ R_TXFIFO_DATA2_DB4_LENGTH,
872
+ frame->data[7]));
873
+
874
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXOK, 1);
875
+ trace_xlnx_can_rx_data(frame->can_id, frame->can_dlc,
876
+ frame->data[0], frame->data[1],
877
+ frame->data[2], frame->data[3],
878
+ frame->data[4], frame->data[5],
879
+ frame->data[6], frame->data[7]);
880
+ }
881
+
882
+ can_update_irq(s);
883
+ }
884
+}
885
+
886
+static uint64_t can_rxfifo_pre_read(RegisterInfo *reg, uint64_t val)
887
+{
888
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
889
+
890
+ if (!fifo32_is_empty(&s->rx_fifo)) {
891
+ val = fifo32_pop(&s->rx_fifo);
892
+ } else {
893
+ ARRAY_FIELD_DP32(s->regs, INTERRUPT_STATUS_REGISTER, RXUFLW, 1);
894
+ }
895
+
896
+ can_update_irq(s);
897
+ return val;
898
+}
899
+
900
+static void can_filter_enable_post_write(RegisterInfo *reg, uint64_t val)
901
+{
902
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
903
+
904
+ if (ARRAY_FIELD_EX32(s->regs, AFR, UAF1) &&
905
+ ARRAY_FIELD_EX32(s->regs, AFR, UAF2) &&
906
+ ARRAY_FIELD_EX32(s->regs, AFR, UAF3) &&
907
+ ARRAY_FIELD_EX32(s->regs, AFR, UAF4)) {
908
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, ACFBSY, 1);
909
+ } else {
910
+ ARRAY_FIELD_DP32(s->regs, STATUS_REGISTER, ACFBSY, 0);
911
+ }
912
+}
913
+
914
+static uint64_t can_filter_mask_pre_write(RegisterInfo *reg, uint64_t val)
915
+{
916
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
917
+ uint32_t reg_idx = (reg->access->addr) / 4;
918
+ uint32_t filter_number = (reg_idx - R_AFMR1) / 2;
919
+
920
+ /* modify an acceptance filter, the corresponding UAF bit should be '0'. */
921
+ if (!(s->regs[R_AFR] & (1 << filter_number))) {
922
+ s->regs[reg_idx] = val;
923
+
924
+ trace_xlnx_can_filter_mask_pre_write(filter_number, s->regs[reg_idx]);
925
+ } else {
926
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
927
+
928
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Acceptance filter %d"
929
+ " mask is not set as corresponding UAF bit is not 0.\n",
930
+ path, filter_number + 1);
931
+ }
932
+
933
+ return s->regs[reg_idx];
934
+}
935
+
936
+static uint64_t can_filter_id_pre_write(RegisterInfo *reg, uint64_t val)
937
+{
938
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
939
+ uint32_t reg_idx = (reg->access->addr) / 4;
940
+ uint32_t filter_number = (reg_idx - R_AFIR1) / 2;
941
+
942
+ if (!(s->regs[R_AFR] & (1 << filter_number))) {
943
+ s->regs[reg_idx] = val;
944
+
945
+ trace_xlnx_can_filter_id_pre_write(filter_number, s->regs[reg_idx]);
946
+ } else {
947
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
948
+
949
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Acceptance filter %d"
950
+ " id is not set as corresponding UAF bit is not 0.\n",
951
+ path, filter_number + 1);
952
+ }
953
+
954
+ return s->regs[reg_idx];
955
+}
956
+
957
+static void can_tx_post_write(RegisterInfo *reg, uint64_t val)
958
+{
959
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(reg->opaque);
960
+
961
+ bool is_txhpb = reg->access->addr > A_TXFIFO_DATA2;
962
+
963
+ bool initiate_transfer = (reg->access->addr == A_TXFIFO_DATA2) ||
964
+ (reg->access->addr == A_TXHPB_DATA2);
965
+
966
+ Fifo32 *f = is_txhpb ? &s->txhpb_fifo : &s->tx_fifo;
967
+
968
+ if (!fifo32_is_full(f)) {
969
+ fifo32_push(f, val);
970
+ } else {
971
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
972
+
973
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: TX FIFO is full.\n", path);
974
+ }
975
+
976
+ /* Initiate the message send if TX register is written. */
977
+ if (initiate_transfer &&
978
+ ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN)) {
979
+ transfer_fifo(s, f);
980
+ }
981
+
982
+ can_update_irq(s);
983
+}
984
+
985
+static const RegisterAccessInfo can_regs_info[] = {
986
+ { .name = "SOFTWARE_RESET_REGISTER",
987
+ .addr = A_SOFTWARE_RESET_REGISTER,
988
+ .rsvd = 0xfffffffc,
989
+ .pre_write = can_srr_pre_write,
990
+ },{ .name = "MODE_SELECT_REGISTER",
991
+ .addr = A_MODE_SELECT_REGISTER,
992
+ .rsvd = 0xfffffff8,
993
+ .pre_write = can_msr_pre_write,
994
+ },{ .name = "ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER",
995
+ .addr = A_ARBITRATION_PHASE_BAUD_RATE_PRESCALER_REGISTER,
996
+ .rsvd = 0xffffff00,
997
+ .pre_write = can_brpr_pre_write,
998
+ },{ .name = "ARBITRATION_PHASE_BIT_TIMING_REGISTER",
999
+ .addr = A_ARBITRATION_PHASE_BIT_TIMING_REGISTER,
1000
+ .rsvd = 0xfffffe00,
1001
+ .pre_write = can_btr_pre_write,
1002
+ },{ .name = "ERROR_COUNTER_REGISTER",
1003
+ .addr = A_ERROR_COUNTER_REGISTER,
1004
+ .rsvd = 0xffff0000,
1005
+ .ro = 0xffffffff,
1006
+ },{ .name = "ERROR_STATUS_REGISTER",
1007
+ .addr = A_ERROR_STATUS_REGISTER,
1008
+ .rsvd = 0xffffffe0,
1009
+ .w1c = 0x1f,
1010
+ },{ .name = "STATUS_REGISTER", .addr = A_STATUS_REGISTER,
1011
+ .reset = 0x1,
1012
+ .rsvd = 0xffffe000,
1013
+ .ro = 0x1fff,
1014
+ },{ .name = "INTERRUPT_STATUS_REGISTER",
1015
+ .addr = A_INTERRUPT_STATUS_REGISTER,
1016
+ .reset = 0x6000,
1017
+ .rsvd = 0xffff8000,
1018
+ .ro = 0x7fff,
1019
+ },{ .name = "INTERRUPT_ENABLE_REGISTER",
1020
+ .addr = A_INTERRUPT_ENABLE_REGISTER,
1021
+ .rsvd = 0xffff8000,
1022
+ .post_write = can_ier_post_write,
1023
+ },{ .name = "INTERRUPT_CLEAR_REGISTER",
1024
+ .addr = A_INTERRUPT_CLEAR_REGISTER,
1025
+ .rsvd = 0xffff8000,
1026
+ .pre_write = can_icr_pre_write,
1027
+ },{ .name = "TIMESTAMP_REGISTER",
1028
+ .addr = A_TIMESTAMP_REGISTER,
1029
+ .rsvd = 0xfffffffe,
1030
+ .pre_write = can_tcr_pre_write,
1031
+ },{ .name = "WIR", .addr = A_WIR,
1032
+ .reset = 0x3f3f,
1033
+ .rsvd = 0xffff0000,
1034
+ },{ .name = "TXFIFO_ID", .addr = A_TXFIFO_ID,
1035
+ .post_write = can_tx_post_write,
1036
+ },{ .name = "TXFIFO_DLC", .addr = A_TXFIFO_DLC,
1037
+ .rsvd = 0xfffffff,
1038
+ .post_write = can_tx_post_write,
1039
+ },{ .name = "TXFIFO_DATA1", .addr = A_TXFIFO_DATA1,
1040
+ .post_write = can_tx_post_write,
1041
+ },{ .name = "TXFIFO_DATA2", .addr = A_TXFIFO_DATA2,
1042
+ .post_write = can_tx_post_write,
1043
+ },{ .name = "TXHPB_ID", .addr = A_TXHPB_ID,
1044
+ .post_write = can_tx_post_write,
1045
+ },{ .name = "TXHPB_DLC", .addr = A_TXHPB_DLC,
1046
+ .rsvd = 0xfffffff,
1047
+ .post_write = can_tx_post_write,
1048
+ },{ .name = "TXHPB_DATA1", .addr = A_TXHPB_DATA1,
1049
+ .post_write = can_tx_post_write,
1050
+ },{ .name = "TXHPB_DATA2", .addr = A_TXHPB_DATA2,
1051
+ .post_write = can_tx_post_write,
1052
+ },{ .name = "RXFIFO_ID", .addr = A_RXFIFO_ID,
1053
+ .ro = 0xffffffff,
1054
+ .post_read = can_rxfifo_pre_read,
1055
+ },{ .name = "RXFIFO_DLC", .addr = A_RXFIFO_DLC,
1056
+ .rsvd = 0xfff0000,
1057
+ .post_read = can_rxfifo_pre_read,
1058
+ },{ .name = "RXFIFO_DATA1", .addr = A_RXFIFO_DATA1,
1059
+ .post_read = can_rxfifo_pre_read,
1060
+ },{ .name = "RXFIFO_DATA2", .addr = A_RXFIFO_DATA2,
1061
+ .post_read = can_rxfifo_pre_read,
1062
+ },{ .name = "AFR", .addr = A_AFR,
1063
+ .rsvd = 0xfffffff0,
1064
+ .post_write = can_filter_enable_post_write,
1065
+ },{ .name = "AFMR1", .addr = A_AFMR1,
1066
+ .pre_write = can_filter_mask_pre_write,
1067
+ },{ .name = "AFIR1", .addr = A_AFIR1,
1068
+ .pre_write = can_filter_id_pre_write,
1069
+ },{ .name = "AFMR2", .addr = A_AFMR2,
1070
+ .pre_write = can_filter_mask_pre_write,
1071
+ },{ .name = "AFIR2", .addr = A_AFIR2,
1072
+ .pre_write = can_filter_id_pre_write,
1073
+ },{ .name = "AFMR3", .addr = A_AFMR3,
1074
+ .pre_write = can_filter_mask_pre_write,
1075
+ },{ .name = "AFIR3", .addr = A_AFIR3,
1076
+ .pre_write = can_filter_id_pre_write,
1077
+ },{ .name = "AFMR4", .addr = A_AFMR4,
1078
+ .pre_write = can_filter_mask_pre_write,
1079
+ },{ .name = "AFIR4", .addr = A_AFIR4,
1080
+ .pre_write = can_filter_id_pre_write,
1081
+ }
1082
+};
1083
+
1084
+static void xlnx_zynqmp_can_ptimer_cb(void *opaque)
1085
+{
1086
+ /* No action required on the timer rollover. */
1087
+}
1088
+
1089
+static const MemoryRegionOps can_ops = {
1090
+ .read = register_read_memory,
1091
+ .write = register_write_memory,
1092
+ .endianness = DEVICE_LITTLE_ENDIAN,
1093
+ .valid = {
1094
+ .min_access_size = 4,
1095
+ .max_access_size = 4,
1096
+ },
1097
+};
1098
+
1099
+static void xlnx_zynqmp_can_reset_init(Object *obj, ResetType type)
1100
+{
1101
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(obj);
1102
+ unsigned int i;
1103
+
1104
+ for (i = R_RXFIFO_ID; i < ARRAY_SIZE(s->reg_info); ++i) {
1105
+ register_reset(&s->reg_info[i]);
1106
+ }
1107
+
1108
+ ptimer_transaction_begin(s->can_timer);
1109
+ ptimer_set_count(s->can_timer, 0);
1110
+ ptimer_transaction_commit(s->can_timer);
1111
+}
1112
+
1113
+static void xlnx_zynqmp_can_reset_hold(Object *obj)
1114
+{
1115
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(obj);
1116
+ unsigned int i;
1117
+
1118
+ for (i = 0; i < R_RXFIFO_ID; ++i) {
1119
+ register_reset(&s->reg_info[i]);
79
+ }
1120
+ }
80
+
1121
+
81
+ /*
1122
+ /*
82
+ * If not backed by host ram, there is no tag storage: access unchecked.
1123
+ * Reset FIFOs when CAN model is reset. This will clear the fifo writes
83
+ * This is probably a guest os bug though, so log it.
1124
+ * done by post_write which gets called from register_reset function,
1125
+ * post_write handle will not be able to trigger tx because CAN will be
1126
+ * disabled when software_reset_register is cleared first.
84
+ */
1127
+ */
85
+ if (unlikely(flags & TLB_MMIO)) {
1128
+ fifo32_reset(&s->rx_fifo);
86
+ qemu_log_mask(LOG_GUEST_ERROR,
1129
+ fifo32_reset(&s->tx_fifo);
87
+ "Page @ 0x%" PRIx64 " indicates Tagged Normal memory "
1130
+ fifo32_reset(&s->txhpb_fifo);
88
+ "but is not backed by host ram\n", ptr);
1131
+}
89
+ return NULL;
1132
+
90
+ }
1133
+static bool xlnx_zynqmp_can_can_receive(CanBusClientState *client)
91
+
1134
+{
92
+ /*
1135
+ XlnxZynqMPCANState *s = container_of(client, XlnxZynqMPCANState,
93
+ * The Normal memory access can extend to the next page. E.g. a single
1136
+ bus_client);
94
+ * 8-byte access to the last byte of a page will check only the last
1137
+
95
+ * tag on the first page.
1138
+ if (ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, SRST)) {
96
+ * Any page access exception has priority over tag check exception.
1139
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
97
+ */
1140
+
98
+ in_page = -(ptr | TARGET_PAGE_MASK);
1141
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Controller is in reset state.\n",
99
+ if (unlikely(ptr_size > in_page)) {
1142
+ path);
100
+ void *ignore;
1143
+ return false;
101
+ flags |= probe_access_flags(env, ptr + in_page, ptr_access,
1144
+ }
102
+ ptr_mmu_idx, ra == 0, &ignore, ra);
1145
+
103
+ assert(!(flags & TLB_INVALID_MASK));
1146
+ if ((ARRAY_FIELD_EX32(s->regs, SOFTWARE_RESET_REGISTER, CEN)) == 0) {
104
+ }
1147
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
105
+
1148
+
106
+ /* Any debug exception has priority over a tag check exception. */
1149
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Controller is disabled. Incoming"
107
+ if (unlikely(flags & TLB_WATCHPOINT)) {
1150
+ " messages will be discarded.\n", path);
108
+ int wp = ptr_access == MMU_DATA_LOAD ? BP_MEM_READ : BP_MEM_WRITE;
1151
+ return false;
109
+ assert(ra != 0);
1152
+ }
110
+ cpu_check_watchpoint(env_cpu(env), ptr, ptr_size,
1153
+
111
+ iotlbentry->attrs, wp, ra);
1154
+ return true;
112
+ }
1155
+}
113
+
1156
+
114
+ /*
1157
+static ssize_t xlnx_zynqmp_can_receive(CanBusClientState *client,
115
+ * Find the physical address within the normal mem space.
1158
+ const qemu_can_frame *buf, size_t buf_size) {
116
+ * The memory region lookup must succeed because TLB_MMIO was
1159
+ XlnxZynqMPCANState *s = container_of(client, XlnxZynqMPCANState,
117
+ * not set in the cputlb lookup above.
1160
+ bus_client);
118
+ */
1161
+ const qemu_can_frame *frame = buf;
119
+ mr = memory_region_from_host(host, &ptr_ra);
1162
+
120
+ tcg_debug_assert(mr != NULL);
1163
+ if (buf_size <= 0) {
121
+ tcg_debug_assert(memory_region_is_ram(mr));
1164
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
122
+ ptr_paddr = ptr_ra;
1165
+
123
+ do {
1166
+ qemu_log_mask(LOG_GUEST_ERROR, "%s: Error in the data received.\n",
124
+ ptr_paddr += mr->addr;
1167
+ path);
125
+ mr = mr->container;
1168
+ return 0;
126
+ } while (mr);
1169
+ }
127
+
1170
+
128
+ /* Convert to the physical address in tag space. */
1171
+ if (ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SNOOP)) {
129
+ tag_paddr = ptr_paddr >> (LOG2_TAG_GRANULE + 1);
1172
+ /* Snoop Mode: Just keep the data. no response back. */
130
+
1173
+ update_rx_fifo(s, frame);
131
+ /* Look up the address in tag space. */
1174
+ } else if ((ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SLEEP))) {
132
+ tag_asi = iotlbentry->attrs.secure ? ARMASIdx_TagS : ARMASIdx_TagNS;
1175
+ /*
133
+ tag_as = cpu_get_address_space(env_cpu(env), tag_asi);
1176
+ * XlnxZynqMPCAN is in sleep mode. Any data on bus will bring it to wake
134
+ mr = address_space_translate(tag_as, tag_paddr, &xlat, NULL,
1177
+ * up state.
135
+ tag_access == MMU_DATA_STORE,
1178
+ */
136
+ iotlbentry->attrs);
1179
+ can_exit_sleep_mode(s);
137
+
1180
+ update_rx_fifo(s, frame);
138
+ /*
1181
+ } else if ((ARRAY_FIELD_EX32(s->regs, STATUS_REGISTER, SLEEP)) == 0) {
139
+ * Note that @mr will never be NULL. If there is nothing in the address
1182
+ update_rx_fifo(s, frame);
140
+ * space at @tag_paddr, the translation will return the unallocated memory
1183
+ } else {
141
+ * region. For our purposes, the result must be ram.
1184
+ /*
142
+ */
1185
+ * XlnxZynqMPCAN will not participate in normal bus communication
143
+ if (unlikely(!memory_region_is_ram(mr))) {
1186
+ * and will not receive any messages transmitted by other CAN nodes.
144
+ /* ??? Failure is a board configuration error. */
1187
+ */
145
+ qemu_log_mask(LOG_UNIMP,
1188
+ trace_xlnx_can_rx_discard(s->regs[R_STATUS_REGISTER]);
146
+ "Tag Memory @ 0x%" HWADDR_PRIx " not found for "
1189
+ }
147
+ "Normal Memory @ 0x%" HWADDR_PRIx "\n",
1190
+
148
+ tag_paddr, ptr_paddr);
1191
+ return 1;
149
+ return NULL;
1192
+}
150
+ }
1193
+
151
+
1194
+static CanBusClientInfo can_xilinx_bus_client_info = {
152
+ /*
1195
+ .can_receive = xlnx_zynqmp_can_can_receive,
153
+ * Ensure the tag memory is dirty on write, for migration.
1196
+ .receive = xlnx_zynqmp_can_receive,
154
+ * Tag memory can never contain code or display memory (vga).
1197
+};
155
+ */
1198
+
156
+ if (tag_access == MMU_DATA_STORE) {
1199
+static int xlnx_zynqmp_can_connect_to_bus(XlnxZynqMPCANState *s,
157
+ ram_addr_t tag_ra = memory_region_get_ram_addr(mr) + xlat;
1200
+ CanBusState *bus)
158
+ cpu_physical_memory_set_dirty_flag(tag_ra, DIRTY_MEMORY_MIGRATION);
1201
+{
159
+ }
1202
+ s->bus_client.info = &can_xilinx_bus_client_info;
160
+
1203
+
161
+ return memory_region_get_ram_ptr(mr) + xlat;
1204
+ if (can_bus_insert_client(bus, &s->bus_client) < 0) {
162
+#endif
1205
+ return -1;
163
}
1206
+ }
164
1207
+ return 0;
165
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
1208
+}
1209
+
1210
+static void xlnx_zynqmp_can_realize(DeviceState *dev, Error **errp)
1211
+{
1212
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(dev);
1213
+
1214
+ if (s->canbus) {
1215
+ if (xlnx_zynqmp_can_connect_to_bus(s, s->canbus) < 0) {
1216
+ g_autofree char *path = object_get_canonical_path(OBJECT(s));
1217
+
1218
+ error_setg(errp, "%s: xlnx_zynqmp_can_connect_to_bus"
1219
+ " failed.", path);
1220
+ return;
1221
+ }
1222
+ }
1223
+
1224
+ /* Create RX FIFO, TXFIFO, TXHPB storage. */
1225
+ fifo32_create(&s->rx_fifo, RXFIFO_SIZE);
1226
+ fifo32_create(&s->tx_fifo, RXFIFO_SIZE);
1227
+ fifo32_create(&s->txhpb_fifo, CAN_FRAME_SIZE);
1228
+
1229
+ /* Allocate a new timer. */
1230
+ s->can_timer = ptimer_init(xlnx_zynqmp_can_ptimer_cb, s,
1231
+ PTIMER_POLICY_DEFAULT);
1232
+
1233
+ ptimer_transaction_begin(s->can_timer);
1234
+
1235
+ ptimer_set_freq(s->can_timer, s->cfg.ext_clk_freq);
1236
+ ptimer_set_limit(s->can_timer, CAN_TIMER_MAX, 1);
1237
+ ptimer_run(s->can_timer, 0);
1238
+ ptimer_transaction_commit(s->can_timer);
1239
+}
1240
+
1241
+static void xlnx_zynqmp_can_init(Object *obj)
1242
+{
1243
+ XlnxZynqMPCANState *s = XLNX_ZYNQMP_CAN(obj);
1244
+ SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
1245
+
1246
+ RegisterInfoArray *reg_array;
1247
+
1248
+ memory_region_init(&s->iomem, obj, TYPE_XLNX_ZYNQMP_CAN,
1249
+ XLNX_ZYNQMP_CAN_R_MAX * 4);
1250
+ reg_array = register_init_block32(DEVICE(obj), can_regs_info,
1251
+ ARRAY_SIZE(can_regs_info),
1252
+ s->reg_info, s->regs,
1253
+ &can_ops,
1254
+ XLNX_ZYNQMP_CAN_ERR_DEBUG,
1255
+ XLNX_ZYNQMP_CAN_R_MAX * 4);
1256
+
1257
+ memory_region_add_subregion(&s->iomem, 0x00, &reg_array->mem);
1258
+ sysbus_init_mmio(sbd, &s->iomem);
1259
+ sysbus_init_irq(SYS_BUS_DEVICE(obj), &s->irq);
1260
+}
1261
+
1262
+static const VMStateDescription vmstate_can = {
1263
+ .name = TYPE_XLNX_ZYNQMP_CAN,
1264
+ .version_id = 1,
1265
+ .minimum_version_id = 1,
1266
+ .fields = (VMStateField[]) {
1267
+ VMSTATE_FIFO32(rx_fifo, XlnxZynqMPCANState),
1268
+ VMSTATE_FIFO32(tx_fifo, XlnxZynqMPCANState),
1269
+ VMSTATE_FIFO32(txhpb_fifo, XlnxZynqMPCANState),
1270
+ VMSTATE_UINT32_ARRAY(regs, XlnxZynqMPCANState, XLNX_ZYNQMP_CAN_R_MAX),
1271
+ VMSTATE_PTIMER(can_timer, XlnxZynqMPCANState),
1272
+ VMSTATE_END_OF_LIST(),
1273
+ }
1274
+};
1275
+
1276
+static Property xlnx_zynqmp_can_properties[] = {
1277
+ DEFINE_PROP_UINT32("ext_clk_freq", XlnxZynqMPCANState, cfg.ext_clk_freq,
1278
+ CAN_DEFAULT_CLOCK),
1279
+ DEFINE_PROP_LINK("canbus", XlnxZynqMPCANState, canbus, TYPE_CAN_BUS,
1280
+ CanBusState *),
1281
+ DEFINE_PROP_END_OF_LIST(),
1282
+};
1283
+
1284
+static void xlnx_zynqmp_can_class_init(ObjectClass *klass, void *data)
1285
+{
1286
+ DeviceClass *dc = DEVICE_CLASS(klass);
1287
+ ResettableClass *rc = RESETTABLE_CLASS(klass);
1288
+
1289
+ rc->phases.enter = xlnx_zynqmp_can_reset_init;
1290
+ rc->phases.hold = xlnx_zynqmp_can_reset_hold;
1291
+ dc->realize = xlnx_zynqmp_can_realize;
1292
+ device_class_set_props(dc, xlnx_zynqmp_can_properties);
1293
+ dc->vmsd = &vmstate_can;
1294
+}
1295
+
1296
+static const TypeInfo can_info = {
1297
+ .name = TYPE_XLNX_ZYNQMP_CAN,
1298
+ .parent = TYPE_SYS_BUS_DEVICE,
1299
+ .instance_size = sizeof(XlnxZynqMPCANState),
1300
+ .class_init = xlnx_zynqmp_can_class_init,
1301
+ .instance_init = xlnx_zynqmp_can_init,
1302
+};
1303
+
1304
+static void can_register_types(void)
1305
+{
1306
+ type_register_static(&can_info);
1307
+}
1308
+
1309
+type_init(can_register_types)
1310
diff --git a/hw/Kconfig b/hw/Kconfig
1311
index XXXXXXX..XXXXXXX 100644
1312
--- a/hw/Kconfig
1313
+++ b/hw/Kconfig
1314
@@ -XXX,XX +XXX,XX @@ config XILINX_AXI
1315
config XLNX_ZYNQMP
1316
bool
1317
select REGISTER
1318
+ select CAN_BUS
1319
diff --git a/hw/net/can/meson.build b/hw/net/can/meson.build
1320
index XXXXXXX..XXXXXXX 100644
1321
--- a/hw/net/can/meson.build
1322
+++ b/hw/net/can/meson.build
1323
@@ -XXX,XX +XXX,XX @@ softmmu_ss.add(when: 'CONFIG_CAN_PCI', if_true: files('can_pcm3680_pci.c'))
1324
softmmu_ss.add(when: 'CONFIG_CAN_PCI', if_true: files('can_mioe3680_pci.c'))
1325
softmmu_ss.add(when: 'CONFIG_CAN_CTUCANFD', if_true: files('ctucan_core.c'))
1326
softmmu_ss.add(when: 'CONFIG_CAN_CTUCANFD_PCI', if_true: files('ctucan_pci.c'))
1327
+softmmu_ss.add(when: 'CONFIG_XLNX_ZYNQMP', if_true: files('xlnx-zynqmp-can.c'))
1328
diff --git a/hw/net/can/trace-events b/hw/net/can/trace-events
1329
new file mode 100644
1330
index XXXXXXX..XXXXXXX
1331
--- /dev/null
1332
+++ b/hw/net/can/trace-events
1333
@@ -XXX,XX +XXX,XX @@
1334
+# xlnx-zynqmp-can.c
1335
+xlnx_can_update_irq(uint32_t isr, uint32_t ier, uint32_t irq) "ISR: 0x%08x IER: 0x%08x IRQ: 0x%08x"
1336
+xlnx_can_reset(uint32_t val) "Resetting controller with value = 0x%08x"
1337
+xlnx_can_rx_fifo_filter_reject(uint32_t id, uint8_t dlc) "Frame: ID: 0x%08x DLC: 0x%02x"
1338
+xlnx_can_filter_id_pre_write(uint8_t filter_num, uint32_t value) "Filter%d ID: 0x%08x"
1339
+xlnx_can_filter_mask_pre_write(uint8_t filter_num, uint32_t value) "Filter%d MASK: 0x%08x"
1340
+xlnx_can_tx_data(uint32_t id, uint8_t dlc, uint8_t db0, uint8_t db1, uint8_t db2, uint8_t db3, uint8_t db4, uint8_t db5, uint8_t db6, uint8_t db7) "Frame: ID: 0x%08x DLC: 0x%02x DATA: 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x"
1341
+xlnx_can_rx_data(uint32_t id, uint32_t dlc, uint8_t db0, uint8_t db1, uint8_t db2, uint8_t db3, uint8_t db4, uint8_t db5, uint8_t db6, uint8_t db7) "Frame: ID: 0x%08x DLC: 0x%02x DATA: 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x 0x%02x"
1342
+xlnx_can_rx_discard(uint32_t status) "Controller is not enabled for bus communication. Status Register: 0x%08x"
166
--
1343
--
167
2.20.1
1344
2.20.1
168
1345
169
1346
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Vikram Garhwal <fnu.vikram@xilinx.com>
2
2
3
Because the elements are sequential, we can eliminate many tests all
3
Connect CAN0 and CAN1 on the ZynqMP.
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
4
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Francisco Iglesias <francisco.iglesias@xilinx.com>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
8
Message-id: 20200626033144.790098-36-richard.henderson@linaro.org
7
Signed-off-by: Vikram Garhwal <fnu.vikram@xilinx.com>
8
Message-id: 1605728926-352690-3-git-send-email-fnu.vikram@xilinx.com
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
10
---
11
target/arm/helper-sve.h | 98 ++++++++++++++++
11
include/hw/arm/xlnx-zynqmp.h | 8 ++++++++
12
target/arm/sve_helper.c | 99 ++++++++++++++--
12
hw/arm/xlnx-zcu102.c | 20 ++++++++++++++++++++
13
target/arm/translate-sve.c | 232 +++++++++++++++++++++++++------------
13
hw/arm/xlnx-zynqmp.c | 34 ++++++++++++++++++++++++++++++++++
14
3 files changed, 343 insertions(+), 86 deletions(-)
14
3 files changed, 62 insertions(+)
15
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
16
diff --git a/include/hw/arm/xlnx-zynqmp.h b/include/hw/arm/xlnx-zynqmp.h
17
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-sve.h
18
--- a/include/hw/arm/xlnx-zynqmp.h
19
+++ b/target/arm/helper-sve.h
19
+++ b/include/hw/arm/xlnx-zynqmp.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
20
@@ -XXX,XX +XXX,XX @@
21
DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
#include "hw/intc/arm_gic.h"
22
DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
#include "hw/net/cadence_gem.h"
23
23
#include "hw/char/cadence_uart.h"
24
+DEF_HELPER_FLAGS_4(sve_ldff1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
24
+#include "hw/net/xlnx-zynqmp-can.h"
25
+DEF_HELPER_FLAGS_4(sve_ldff1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
#include "hw/ide/ahci.h"
26
+DEF_HELPER_FLAGS_4(sve_ldff1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
#include "hw/sd/sdhci.h"
27
+DEF_HELPER_FLAGS_4(sve_ldff1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
#include "hw/ssi/xilinx_spips.h"
28
+DEF_HELPER_FLAGS_4(sve_ldff1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
@@ -XXX,XX +XXX,XX @@
29
+DEF_HELPER_FLAGS_4(sve_ldff1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
29
#include "hw/cpu/cluster.h"
30
+DEF_HELPER_FLAGS_4(sve_ldff1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
#include "target/arm/cpu.h"
31
#include "qom/object.h"
32
+#include "net/can_emu.h"
33
34
#define TYPE_XLNX_ZYNQMP "xlnx,zynqmp"
35
OBJECT_DECLARE_SIMPLE_TYPE(XlnxZynqMPState, XLNX_ZYNQMP)
36
@@ -XXX,XX +XXX,XX @@ OBJECT_DECLARE_SIMPLE_TYPE(XlnxZynqMPState, XLNX_ZYNQMP)
37
#define XLNX_ZYNQMP_NUM_RPU_CPUS 2
38
#define XLNX_ZYNQMP_NUM_GEMS 4
39
#define XLNX_ZYNQMP_NUM_UARTS 2
40
+#define XLNX_ZYNQMP_NUM_CAN 2
41
+#define XLNX_ZYNQMP_CAN_REF_CLK (24 * 1000 * 1000)
42
#define XLNX_ZYNQMP_NUM_SDHCI 2
43
#define XLNX_ZYNQMP_NUM_SPIS 2
44
#define XLNX_ZYNQMP_NUM_GDMA_CH 8
45
@@ -XXX,XX +XXX,XX @@ struct XlnxZynqMPState {
46
47
CadenceGEMState gem[XLNX_ZYNQMP_NUM_GEMS];
48
CadenceUARTState uart[XLNX_ZYNQMP_NUM_UARTS];
49
+ XlnxZynqMPCANState can[XLNX_ZYNQMP_NUM_CAN];
50
SysbusAHCIState sata;
51
SDHCIState sdhci[XLNX_ZYNQMP_NUM_SDHCI];
52
XilinxSPIPS spi[XLNX_ZYNQMP_NUM_SPIS];
53
@@ -XXX,XX +XXX,XX @@ struct XlnxZynqMPState {
54
bool virt;
55
/* Has the RPU subsystem? */
56
bool has_rpu;
31
+
57
+
32
+DEF_HELPER_FLAGS_4(sve_ldff1hh_le_r_mte, TCG_CALL_NO_WG,
58
+ /* CAN bus. */
33
+ void, env, ptr, tl, i32)
59
+ CanBusState *canbus[XLNX_ZYNQMP_NUM_CAN];
34
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_le_r_mte, TCG_CALL_NO_WG,
60
};
35
+ void, env, ptr, tl, i32)
61
36
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_le_r_mte, TCG_CALL_NO_WG,
62
#endif
37
+ void, env, ptr, tl, i32)
63
diff --git a/hw/arm/xlnx-zcu102.c b/hw/arm/xlnx-zcu102.c
38
+DEF_HELPER_FLAGS_4(sve_ldff1hss_le_r_mte, TCG_CALL_NO_WG,
64
index XXXXXXX..XXXXXXX 100644
39
+ void, env, ptr, tl, i32)
65
--- a/hw/arm/xlnx-zcu102.c
40
+DEF_HELPER_FLAGS_4(sve_ldff1hds_le_r_mte, TCG_CALL_NO_WG,
66
+++ b/hw/arm/xlnx-zcu102.c
41
+ void, env, ptr, tl, i32)
67
@@ -XXX,XX +XXX,XX @@
68
#include "sysemu/qtest.h"
69
#include "sysemu/device_tree.h"
70
#include "qom/object.h"
71
+#include "net/can_emu.h"
72
73
struct XlnxZCU102 {
74
MachineState parent_obj;
75
@@ -XXX,XX +XXX,XX @@ struct XlnxZCU102 {
76
bool secure;
77
bool virt;
78
79
+ CanBusState *canbus[XLNX_ZYNQMP_NUM_CAN];
42
+
80
+
43
+DEF_HELPER_FLAGS_4(sve_ldff1hh_be_r_mte, TCG_CALL_NO_WG,
81
struct arm_boot_info binfo;
44
+ void, env, ptr, tl, i32)
82
};
45
+DEF_HELPER_FLAGS_4(sve_ldff1hsu_be_r_mte, TCG_CALL_NO_WG,
83
46
+ void, env, ptr, tl, i32)
84
@@ -XXX,XX +XXX,XX @@ static void xlnx_zcu102_init(MachineState *machine)
47
+DEF_HELPER_FLAGS_4(sve_ldff1hdu_be_r_mte, TCG_CALL_NO_WG,
85
object_property_set_bool(OBJECT(&s->soc), "virtualization", s->virt,
48
+ void, env, ptr, tl, i32)
86
&error_fatal);
49
+DEF_HELPER_FLAGS_4(sve_ldff1hss_be_r_mte, TCG_CALL_NO_WG,
87
50
+ void, env, ptr, tl, i32)
88
+ for (i = 0; i < XLNX_ZYNQMP_NUM_CAN; i++) {
51
+DEF_HELPER_FLAGS_4(sve_ldff1hds_be_r_mte, TCG_CALL_NO_WG,
89
+ gchar *bus_name = g_strdup_printf("canbus%d", i);
52
+ void, env, ptr, tl, i32)
53
+
90
+
54
+DEF_HELPER_FLAGS_4(sve_ldff1ss_le_r_mte, TCG_CALL_NO_WG,
91
+ object_property_set_link(OBJECT(&s->soc), bus_name,
55
+ void, env, ptr, tl, i32)
92
+ OBJECT(s->canbus[i]), &error_fatal);
56
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_le_r_mte, TCG_CALL_NO_WG,
93
+ g_free(bus_name);
57
+ void, env, ptr, tl, i32)
58
+DEF_HELPER_FLAGS_4(sve_ldff1sds_le_r_mte, TCG_CALL_NO_WG,
59
+ void, env, ptr, tl, i32)
60
+
61
+DEF_HELPER_FLAGS_4(sve_ldff1ss_be_r_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_4(sve_ldff1sdu_be_r_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ldff1sds_be_r_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ldff1dd_le_r_mte, TCG_CALL_NO_WG,
69
+ void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ldff1dd_be_r_mte, TCG_CALL_NO_WG,
71
+ void, env, ptr, tl, i32)
72
+
73
DEF_HELPER_FLAGS_4(sve_ldnf1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
80
+DEF_HELPER_FLAGS_4(sve_ldnf1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ldnf1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ldnf1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ldnf1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+DEF_HELPER_FLAGS_4(sve_ldnf1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
85
+DEF_HELPER_FLAGS_4(sve_ldnf1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ldnf1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+
88
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_le_r_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_le_r_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_le_r_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_le_r_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_le_r_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, tl, i32)
98
+
99
+DEF_HELPER_FLAGS_4(sve_ldnf1hh_be_r_mte, TCG_CALL_NO_WG,
100
+ void, env, ptr, tl, i32)
101
+DEF_HELPER_FLAGS_4(sve_ldnf1hsu_be_r_mte, TCG_CALL_NO_WG,
102
+ void, env, ptr, tl, i32)
103
+DEF_HELPER_FLAGS_4(sve_ldnf1hdu_be_r_mte, TCG_CALL_NO_WG,
104
+ void, env, ptr, tl, i32)
105
+DEF_HELPER_FLAGS_4(sve_ldnf1hss_be_r_mte, TCG_CALL_NO_WG,
106
+ void, env, ptr, tl, i32)
107
+DEF_HELPER_FLAGS_4(sve_ldnf1hds_be_r_mte, TCG_CALL_NO_WG,
108
+ void, env, ptr, tl, i32)
109
+
110
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_le_r_mte, TCG_CALL_NO_WG,
111
+ void, env, ptr, tl, i32)
112
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_le_r_mte, TCG_CALL_NO_WG,
113
+ void, env, ptr, tl, i32)
114
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_le_r_mte, TCG_CALL_NO_WG,
115
+ void, env, ptr, tl, i32)
116
+
117
+DEF_HELPER_FLAGS_4(sve_ldnf1ss_be_r_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_4(sve_ldnf1sdu_be_r_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_4(sve_ldnf1sds_be_r_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, tl, i32)
123
+
124
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_le_r_mte, TCG_CALL_NO_WG,
125
+ void, env, ptr, tl, i32)
126
+DEF_HELPER_FLAGS_4(sve_ldnf1dd_be_r_mte, TCG_CALL_NO_WG,
127
+ void, env, ptr, tl, i32)
128
+
129
DEF_HELPER_FLAGS_4(sve_st1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
130
DEF_HELPER_FLAGS_4(sve_st2bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
131
DEF_HELPER_FLAGS_4(sve_st3bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
132
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
133
index XXXXXXX..XXXXXXX 100644
134
--- a/target/arm/sve_helper.c
135
+++ b/target/arm/sve_helper.c
136
@@ -XXX,XX +XXX,XX @@ static void record_fault(CPUARMState *env, uintptr_t i, uintptr_t oprsz)
137
*/
138
static inline QEMU_ALWAYS_INLINE
139
void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
140
- uint32_t desc, const uintptr_t retaddr,
141
+ uint32_t desc, const uintptr_t retaddr, uint32_t mtedesc,
142
const int esz, const int msz, const SVEContFault fault,
143
sve_ldst1_host_fn *host_fn,
144
sve_ldst1_tlb_fn *tlb_fn)
145
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
146
mem_off = info.mem_off_first[0];
147
flags = info.page[0].flags;
148
149
+ /*
150
+ * Disable MTE checking if the Tagged bit is not set. Since TBI must
151
+ * be set within MTEDESC for MTE, !mtedesc => !mte_active.
152
+ */
153
+ if (arm_tlb_mte_tagged(&info.page[0].attrs)) {
154
+ mtedesc = 0;
155
+ }
94
+ }
156
+
95
+
157
if (fault == FAULT_FIRST) {
96
qdev_realize(DEVICE(&s->soc), NULL, &error_fatal);
158
+ /* Trapping mte check for the first-fault element. */
97
159
+ if (mtedesc) {
98
/* Create and plug in the SD cards */
160
+ mte_check1(env, mtedesc, addr + mem_off, retaddr);
99
@@ -XXX,XX +XXX,XX @@ static void xlnx_zcu102_machine_instance_init(Object *obj)
161
+ }
100
s->secure = false;
101
/* Default to virt (EL2) being disabled */
102
s->virt = false;
103
+ object_property_add_link(obj, "xlnx-zcu102.canbus0", TYPE_CAN_BUS,
104
+ (Object **)&s->canbus[0],
105
+ object_property_allow_set_link,
106
+ 0);
162
+
107
+
163
/*
108
+ object_property_add_link(obj, "xlnx-zcu102.canbus1", TYPE_CAN_BUS,
164
* Special handling of the first active element,
109
+ (Object **)&s->canbus[1],
165
* if it crosses a page boundary or is MMIO.
110
+ object_property_allow_set_link,
166
*/
111
+ 0);
167
bool is_split = mem_off == info.mem_off_split;
168
- /* TODO: MTE check. */
169
if (unlikely(flags != 0) || unlikely(is_split)) {
170
/*
171
* Use the slow path for cross-page handling.
172
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
173
/* Watchpoint hit, see below. */
174
goto do_fault;
175
}
176
- /* TODO: MTE check. */
177
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
178
+ goto do_fault;
179
+ }
180
/*
181
* Use the slow path for cross-page handling.
182
* This is RAM, without a watchpoint, and will not trap.
183
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
184
& BP_MEM_READ)) {
185
goto do_fault;
186
}
187
- /* TODO: MTE check. */
188
+ if (mtedesc && !mte_probe1(env, mtedesc, addr + mem_off)) {
189
+ goto do_fault;
190
+ }
191
host_fn(vd, reg_off, host + mem_off);
192
}
193
reg_off += 1 << esz;
194
@@ -XXX,XX +XXX,XX @@ void sve_ldnfff1_r(CPUARMState *env, void *vg, const target_ulong addr,
195
record_fault(env, reg_off, reg_max);
196
}
112
}
197
113
198
-#define DO_LDFF1_LDNF1_1(PART, ESZ) \
114
static void xlnx_zcu102_machine_class_init(ObjectClass *oc, void *data)
199
+static inline QEMU_ALWAYS_INLINE
115
diff --git a/hw/arm/xlnx-zynqmp.c b/hw/arm/xlnx-zynqmp.c
200
+void sve_ldnfff1_r_mte(CPUARMState *env, void *vg, target_ulong addr,
116
index XXXXXXX..XXXXXXX 100644
201
+ uint32_t desc, const uintptr_t retaddr,
117
--- a/hw/arm/xlnx-zynqmp.c
202
+ const int esz, const int msz, const SVEContFault fault,
118
+++ b/hw/arm/xlnx-zynqmp.c
203
+ sve_ldst1_host_fn *host_fn,
119
@@ -XXX,XX +XXX,XX @@ static const int uart_intr[XLNX_ZYNQMP_NUM_UARTS] = {
204
+ sve_ldst1_tlb_fn *tlb_fn)
120
21, 22,
205
+{
121
};
206
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
122
207
+ int bit55 = extract64(addr, 55, 1);
123
+static const uint64_t can_addr[XLNX_ZYNQMP_NUM_CAN] = {
124
+ 0xFF060000, 0xFF070000,
125
+};
208
+
126
+
209
+ /* Remove mtedesc from the normal sve descriptor. */
127
+static const int can_intr[XLNX_ZYNQMP_NUM_CAN] = {
210
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
128
+ 23, 24,
129
+};
211
+
130
+
212
+ /* Perform gross MTE suppression early. */
131
static const uint64_t sdhci_addr[XLNX_ZYNQMP_NUM_SDHCI] = {
213
+ if (!tbi_check(desc, bit55) ||
132
0xFF160000, 0xFF170000,
214
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
133
};
215
+ mtedesc = 0;
134
@@ -XXX,XX +XXX,XX @@ static void xlnx_zynqmp_init(Object *obj)
135
TYPE_CADENCE_UART);
136
}
137
138
+ for (i = 0; i < XLNX_ZYNQMP_NUM_CAN; i++) {
139
+ object_initialize_child(obj, "can[*]", &s->can[i],
140
+ TYPE_XLNX_ZYNQMP_CAN);
216
+ }
141
+ }
217
+
142
+
218
+ sve_ldnfff1_r(env, vg, addr, desc, retaddr, mtedesc,
143
object_initialize_child(obj, "sata", &s->sata, TYPE_SYSBUS_AHCI);
219
+ esz, msz, fault, host_fn, tlb_fn);
144
220
+}
145
for (i = 0; i < XLNX_ZYNQMP_NUM_SDHCI; i++) {
146
@@ -XXX,XX +XXX,XX @@ static void xlnx_zynqmp_realize(DeviceState *dev, Error **errp)
147
gic_spi[uart_intr[i]]);
148
}
149
150
+ for (i = 0; i < XLNX_ZYNQMP_NUM_CAN; i++) {
151
+ object_property_set_int(OBJECT(&s->can[i]), "ext_clk_freq",
152
+ XLNX_ZYNQMP_CAN_REF_CLK, &error_abort);
221
+
153
+
222
+#define DO_LDFF1_LDNF1_1(PART, ESZ) \
154
+ object_property_set_link(OBJECT(&s->can[i]), "canbus",
223
void HELPER(sve_ldff1##PART##_r)(CPUARMState *env, void *vg, \
155
+ OBJECT(s->canbus[i]), &error_fatal);
224
target_ulong addr, uint32_t desc) \
225
{ \
226
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
227
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_FIRST, \
228
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
229
} \
230
void HELPER(sve_ldnf1##PART##_r)(CPUARMState *env, void *vg, \
231
target_ulong addr, uint32_t desc) \
232
{ \
233
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
234
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MO_8, FAULT_NO, \
235
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
236
+} \
237
+void HELPER(sve_ldff1##PART##_r_mte)(CPUARMState *env, void *vg, \
238
+ target_ulong addr, uint32_t desc) \
239
+{ \
240
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_FIRST, \
241
+ sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
242
+} \
243
+void HELPER(sve_ldnf1##PART##_r_mte)(CPUARMState *env, void *vg, \
244
+ target_ulong addr, uint32_t desc) \
245
+{ \
246
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, FAULT_NO, \
247
sve_ld1##PART##_host, sve_ld1##PART##_tlb); \
248
}
249
250
-#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
251
+#define DO_LDFF1_LDNF1_2(PART, ESZ, MSZ) \
252
void HELPER(sve_ldff1##PART##_le_r)(CPUARMState *env, void *vg, \
253
target_ulong addr, uint32_t desc) \
254
{ \
255
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
256
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
257
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
258
} \
259
void HELPER(sve_ldnf1##PART##_le_r)(CPUARMState *env, void *vg, \
260
target_ulong addr, uint32_t desc) \
261
{ \
262
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
263
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
264
sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
265
} \
266
void HELPER(sve_ldff1##PART##_be_r)(CPUARMState *env, void *vg, \
267
target_ulong addr, uint32_t desc) \
268
{ \
269
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
270
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_FIRST, \
271
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
272
} \
273
void HELPER(sve_ldnf1##PART##_be_r)(CPUARMState *env, void *vg, \
274
target_ulong addr, uint32_t desc) \
275
{ \
276
- sve_ldnfff1_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
277
+ sve_ldnfff1_r(env, vg, addr, desc, GETPC(), 0, ESZ, MSZ, FAULT_NO, \
278
sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
279
+} \
280
+void HELPER(sve_ldff1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
284
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
285
+} \
286
+void HELPER(sve_ldnf1##PART##_le_r_mte)(CPUARMState *env, void *vg, \
287
+ target_ulong addr, uint32_t desc) \
288
+{ \
289
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
290
+ sve_ld1##PART##_le_host, sve_ld1##PART##_le_tlb); \
291
+} \
292
+void HELPER(sve_ldff1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
293
+ target_ulong addr, uint32_t desc) \
294
+{ \
295
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_FIRST, \
296
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
297
+} \
298
+void HELPER(sve_ldnf1##PART##_be_r_mte)(CPUARMState *env, void *vg, \
299
+ target_ulong addr, uint32_t desc) \
300
+{ \
301
+ sve_ldnfff1_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, FAULT_NO, \
302
+ sve_ld1##PART##_be_host, sve_ld1##PART##_be_tlb); \
303
}
304
305
DO_LDFF1_LDNF1_1(bb, MO_8)
306
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
307
index XXXXXXX..XXXXXXX 100644
308
--- a/target/arm/translate-sve.c
309
+++ b/target/arm/translate-sve.c
310
@@ -XXX,XX +XXX,XX @@ static bool trans_LD_zpri(DisasContext *s, arg_rpri_load *a)
311
312
static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
313
{
314
- static gen_helper_gvec_mem * const fns[2][16] = {
315
- /* Little-endian */
316
- { gen_helper_sve_ldff1bb_r,
317
- gen_helper_sve_ldff1bhu_r,
318
- gen_helper_sve_ldff1bsu_r,
319
- gen_helper_sve_ldff1bdu_r,
320
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
321
+ { /* mte inactive, little-endian */
322
+ { gen_helper_sve_ldff1bb_r,
323
+ gen_helper_sve_ldff1bhu_r,
324
+ gen_helper_sve_ldff1bsu_r,
325
+ gen_helper_sve_ldff1bdu_r,
326
327
- gen_helper_sve_ldff1sds_le_r,
328
- gen_helper_sve_ldff1hh_le_r,
329
- gen_helper_sve_ldff1hsu_le_r,
330
- gen_helper_sve_ldff1hdu_le_r,
331
+ gen_helper_sve_ldff1sds_le_r,
332
+ gen_helper_sve_ldff1hh_le_r,
333
+ gen_helper_sve_ldff1hsu_le_r,
334
+ gen_helper_sve_ldff1hdu_le_r,
335
336
- gen_helper_sve_ldff1hds_le_r,
337
- gen_helper_sve_ldff1hss_le_r,
338
- gen_helper_sve_ldff1ss_le_r,
339
- gen_helper_sve_ldff1sdu_le_r,
340
+ gen_helper_sve_ldff1hds_le_r,
341
+ gen_helper_sve_ldff1hss_le_r,
342
+ gen_helper_sve_ldff1ss_le_r,
343
+ gen_helper_sve_ldff1sdu_le_r,
344
345
- gen_helper_sve_ldff1bds_r,
346
- gen_helper_sve_ldff1bss_r,
347
- gen_helper_sve_ldff1bhs_r,
348
- gen_helper_sve_ldff1dd_le_r },
349
+ gen_helper_sve_ldff1bds_r,
350
+ gen_helper_sve_ldff1bss_r,
351
+ gen_helper_sve_ldff1bhs_r,
352
+ gen_helper_sve_ldff1dd_le_r },
353
354
- /* Big-endian */
355
- { gen_helper_sve_ldff1bb_r,
356
- gen_helper_sve_ldff1bhu_r,
357
- gen_helper_sve_ldff1bsu_r,
358
- gen_helper_sve_ldff1bdu_r,
359
+ /* mte inactive, big-endian */
360
+ { gen_helper_sve_ldff1bb_r,
361
+ gen_helper_sve_ldff1bhu_r,
362
+ gen_helper_sve_ldff1bsu_r,
363
+ gen_helper_sve_ldff1bdu_r,
364
365
- gen_helper_sve_ldff1sds_be_r,
366
- gen_helper_sve_ldff1hh_be_r,
367
- gen_helper_sve_ldff1hsu_be_r,
368
- gen_helper_sve_ldff1hdu_be_r,
369
+ gen_helper_sve_ldff1sds_be_r,
370
+ gen_helper_sve_ldff1hh_be_r,
371
+ gen_helper_sve_ldff1hsu_be_r,
372
+ gen_helper_sve_ldff1hdu_be_r,
373
374
- gen_helper_sve_ldff1hds_be_r,
375
- gen_helper_sve_ldff1hss_be_r,
376
- gen_helper_sve_ldff1ss_be_r,
377
- gen_helper_sve_ldff1sdu_be_r,
378
+ gen_helper_sve_ldff1hds_be_r,
379
+ gen_helper_sve_ldff1hss_be_r,
380
+ gen_helper_sve_ldff1ss_be_r,
381
+ gen_helper_sve_ldff1sdu_be_r,
382
383
- gen_helper_sve_ldff1bds_r,
384
- gen_helper_sve_ldff1bss_r,
385
- gen_helper_sve_ldff1bhs_r,
386
- gen_helper_sve_ldff1dd_be_r },
387
+ gen_helper_sve_ldff1bds_r,
388
+ gen_helper_sve_ldff1bss_r,
389
+ gen_helper_sve_ldff1bhs_r,
390
+ gen_helper_sve_ldff1dd_be_r } },
391
+
156
+
392
+ { /* mte active, little-endian */
157
+ sysbus_realize(SYS_BUS_DEVICE(&s->can[i]), &err);
393
+ { gen_helper_sve_ldff1bb_r_mte,
158
+ if (err) {
394
+ gen_helper_sve_ldff1bhu_r_mte,
159
+ error_propagate(errp, err);
395
+ gen_helper_sve_ldff1bsu_r_mte,
160
+ return;
396
+ gen_helper_sve_ldff1bdu_r_mte,
161
+ }
162
+ sysbus_mmio_map(SYS_BUS_DEVICE(&s->can[i]), 0, can_addr[i]);
163
+ sysbus_connect_irq(SYS_BUS_DEVICE(&s->can[i]), 0,
164
+ gic_spi[can_intr[i]]);
165
+ }
397
+
166
+
398
+ gen_helper_sve_ldff1sds_le_r_mte,
167
object_property_set_int(OBJECT(&s->sata), "num-ports", SATA_NUM_PORTS,
399
+ gen_helper_sve_ldff1hh_le_r_mte,
168
&error_abort);
400
+ gen_helper_sve_ldff1hsu_le_r_mte,
169
if (!sysbus_realize(SYS_BUS_DEVICE(&s->sata), errp)) {
401
+ gen_helper_sve_ldff1hdu_le_r_mte,
170
@@ -XXX,XX +XXX,XX @@ static Property xlnx_zynqmp_props[] = {
402
+
171
DEFINE_PROP_BOOL("has_rpu", XlnxZynqMPState, has_rpu, false),
403
+ gen_helper_sve_ldff1hds_le_r_mte,
172
DEFINE_PROP_LINK("ddr-ram", XlnxZynqMPState, ddr_ram, TYPE_MEMORY_REGION,
404
+ gen_helper_sve_ldff1hss_le_r_mte,
173
MemoryRegion *),
405
+ gen_helper_sve_ldff1ss_le_r_mte,
174
+ DEFINE_PROP_LINK("canbus0", XlnxZynqMPState, canbus[0], TYPE_CAN_BUS,
406
+ gen_helper_sve_ldff1sdu_le_r_mte,
175
+ CanBusState *),
407
+
176
+ DEFINE_PROP_LINK("canbus1", XlnxZynqMPState, canbus[1], TYPE_CAN_BUS,
408
+ gen_helper_sve_ldff1bds_r_mte,
177
+ CanBusState *),
409
+ gen_helper_sve_ldff1bss_r_mte,
178
DEFINE_PROP_END_OF_LIST()
410
+ gen_helper_sve_ldff1bhs_r_mte,
179
};
411
+ gen_helper_sve_ldff1dd_le_r_mte },
180
412
+
413
+ /* mte active, big-endian */
414
+ { gen_helper_sve_ldff1bb_r_mte,
415
+ gen_helper_sve_ldff1bhu_r_mte,
416
+ gen_helper_sve_ldff1bsu_r_mte,
417
+ gen_helper_sve_ldff1bdu_r_mte,
418
+
419
+ gen_helper_sve_ldff1sds_be_r_mte,
420
+ gen_helper_sve_ldff1hh_be_r_mte,
421
+ gen_helper_sve_ldff1hsu_be_r_mte,
422
+ gen_helper_sve_ldff1hdu_be_r_mte,
423
+
424
+ gen_helper_sve_ldff1hds_be_r_mte,
425
+ gen_helper_sve_ldff1hss_be_r_mte,
426
+ gen_helper_sve_ldff1ss_be_r_mte,
427
+ gen_helper_sve_ldff1sdu_be_r_mte,
428
+
429
+ gen_helper_sve_ldff1bds_r_mte,
430
+ gen_helper_sve_ldff1bss_r_mte,
431
+ gen_helper_sve_ldff1bhs_r_mte,
432
+ gen_helper_sve_ldff1dd_be_r_mte } },
433
};
434
435
if (sve_access_check(s)) {
436
TCGv_i64 addr = new_tmp_a64(s);
437
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
438
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
439
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
440
- fns[s->be_data == MO_BE][a->dtype]);
441
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
442
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
443
}
444
return true;
445
}
446
447
static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
448
{
449
- static gen_helper_gvec_mem * const fns[2][16] = {
450
- /* Little-endian */
451
- { gen_helper_sve_ldnf1bb_r,
452
- gen_helper_sve_ldnf1bhu_r,
453
- gen_helper_sve_ldnf1bsu_r,
454
- gen_helper_sve_ldnf1bdu_r,
455
+ static gen_helper_gvec_mem * const fns[2][2][16] = {
456
+ { /* mte inactive, little-endian */
457
+ { gen_helper_sve_ldnf1bb_r,
458
+ gen_helper_sve_ldnf1bhu_r,
459
+ gen_helper_sve_ldnf1bsu_r,
460
+ gen_helper_sve_ldnf1bdu_r,
461
462
- gen_helper_sve_ldnf1sds_le_r,
463
- gen_helper_sve_ldnf1hh_le_r,
464
- gen_helper_sve_ldnf1hsu_le_r,
465
- gen_helper_sve_ldnf1hdu_le_r,
466
+ gen_helper_sve_ldnf1sds_le_r,
467
+ gen_helper_sve_ldnf1hh_le_r,
468
+ gen_helper_sve_ldnf1hsu_le_r,
469
+ gen_helper_sve_ldnf1hdu_le_r,
470
471
- gen_helper_sve_ldnf1hds_le_r,
472
- gen_helper_sve_ldnf1hss_le_r,
473
- gen_helper_sve_ldnf1ss_le_r,
474
- gen_helper_sve_ldnf1sdu_le_r,
475
+ gen_helper_sve_ldnf1hds_le_r,
476
+ gen_helper_sve_ldnf1hss_le_r,
477
+ gen_helper_sve_ldnf1ss_le_r,
478
+ gen_helper_sve_ldnf1sdu_le_r,
479
480
- gen_helper_sve_ldnf1bds_r,
481
- gen_helper_sve_ldnf1bss_r,
482
- gen_helper_sve_ldnf1bhs_r,
483
- gen_helper_sve_ldnf1dd_le_r },
484
+ gen_helper_sve_ldnf1bds_r,
485
+ gen_helper_sve_ldnf1bss_r,
486
+ gen_helper_sve_ldnf1bhs_r,
487
+ gen_helper_sve_ldnf1dd_le_r },
488
489
- /* Big-endian */
490
- { gen_helper_sve_ldnf1bb_r,
491
- gen_helper_sve_ldnf1bhu_r,
492
- gen_helper_sve_ldnf1bsu_r,
493
- gen_helper_sve_ldnf1bdu_r,
494
+ /* mte inactive, big-endian */
495
+ { gen_helper_sve_ldnf1bb_r,
496
+ gen_helper_sve_ldnf1bhu_r,
497
+ gen_helper_sve_ldnf1bsu_r,
498
+ gen_helper_sve_ldnf1bdu_r,
499
500
- gen_helper_sve_ldnf1sds_be_r,
501
- gen_helper_sve_ldnf1hh_be_r,
502
- gen_helper_sve_ldnf1hsu_be_r,
503
- gen_helper_sve_ldnf1hdu_be_r,
504
+ gen_helper_sve_ldnf1sds_be_r,
505
+ gen_helper_sve_ldnf1hh_be_r,
506
+ gen_helper_sve_ldnf1hsu_be_r,
507
+ gen_helper_sve_ldnf1hdu_be_r,
508
509
- gen_helper_sve_ldnf1hds_be_r,
510
- gen_helper_sve_ldnf1hss_be_r,
511
- gen_helper_sve_ldnf1ss_be_r,
512
- gen_helper_sve_ldnf1sdu_be_r,
513
+ gen_helper_sve_ldnf1hds_be_r,
514
+ gen_helper_sve_ldnf1hss_be_r,
515
+ gen_helper_sve_ldnf1ss_be_r,
516
+ gen_helper_sve_ldnf1sdu_be_r,
517
518
- gen_helper_sve_ldnf1bds_r,
519
- gen_helper_sve_ldnf1bss_r,
520
- gen_helper_sve_ldnf1bhs_r,
521
- gen_helper_sve_ldnf1dd_be_r },
522
+ gen_helper_sve_ldnf1bds_r,
523
+ gen_helper_sve_ldnf1bss_r,
524
+ gen_helper_sve_ldnf1bhs_r,
525
+ gen_helper_sve_ldnf1dd_be_r } },
526
+
527
+ { /* mte inactive, little-endian */
528
+ { gen_helper_sve_ldnf1bb_r_mte,
529
+ gen_helper_sve_ldnf1bhu_r_mte,
530
+ gen_helper_sve_ldnf1bsu_r_mte,
531
+ gen_helper_sve_ldnf1bdu_r_mte,
532
+
533
+ gen_helper_sve_ldnf1sds_le_r_mte,
534
+ gen_helper_sve_ldnf1hh_le_r_mte,
535
+ gen_helper_sve_ldnf1hsu_le_r_mte,
536
+ gen_helper_sve_ldnf1hdu_le_r_mte,
537
+
538
+ gen_helper_sve_ldnf1hds_le_r_mte,
539
+ gen_helper_sve_ldnf1hss_le_r_mte,
540
+ gen_helper_sve_ldnf1ss_le_r_mte,
541
+ gen_helper_sve_ldnf1sdu_le_r_mte,
542
+
543
+ gen_helper_sve_ldnf1bds_r_mte,
544
+ gen_helper_sve_ldnf1bss_r_mte,
545
+ gen_helper_sve_ldnf1bhs_r_mte,
546
+ gen_helper_sve_ldnf1dd_le_r_mte },
547
+
548
+ /* mte inactive, big-endian */
549
+ { gen_helper_sve_ldnf1bb_r_mte,
550
+ gen_helper_sve_ldnf1bhu_r_mte,
551
+ gen_helper_sve_ldnf1bsu_r_mte,
552
+ gen_helper_sve_ldnf1bdu_r_mte,
553
+
554
+ gen_helper_sve_ldnf1sds_be_r_mte,
555
+ gen_helper_sve_ldnf1hh_be_r_mte,
556
+ gen_helper_sve_ldnf1hsu_be_r_mte,
557
+ gen_helper_sve_ldnf1hdu_be_r_mte,
558
+
559
+ gen_helper_sve_ldnf1hds_be_r_mte,
560
+ gen_helper_sve_ldnf1hss_be_r_mte,
561
+ gen_helper_sve_ldnf1ss_be_r_mte,
562
+ gen_helper_sve_ldnf1sdu_be_r_mte,
563
+
564
+ gen_helper_sve_ldnf1bds_r_mte,
565
+ gen_helper_sve_ldnf1bss_r_mte,
566
+ gen_helper_sve_ldnf1bhs_r_mte,
567
+ gen_helper_sve_ldnf1dd_be_r_mte } },
568
};
569
570
if (sve_access_check(s)) {
571
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
572
TCGv_i64 addr = new_tmp_a64(s);
573
574
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
575
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
576
- fns[s->be_data == MO_BE][a->dtype]);
577
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 1, false,
578
+ fns[s->mte_active[0]][s->be_data == MO_BE][a->dtype]);
579
}
580
return true;
581
}
582
--
181
--
583
2.20.1
182
2.20.1
584
183
585
184
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Vikram Garhwal <fnu.vikram@xilinx.com>
2
2
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
3
The QTests perform five tests on the Xilinx ZynqMP CAN controller:
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Tests the CAN controller in loopback, sleep and snoop mode.
5
Message-id: 20200626033144.790098-10-richard.henderson@linaro.org
5
Tests filtering of incoming CAN messages.
6
7
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Reviewed-by: Francisco Iglesias <francisco.iglesias@xilinx.com>
9
Signed-off-by: Vikram Garhwal <fnu.vikram@xilinx.com>
10
Message-id: 1605728926-352690-4-git-send-email-fnu.vikram@xilinx.com
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
12
---
8
target/arm/helper-a64.h | 2 ++
13
tests/qtest/xlnx-can-test.c | 360 ++++++++++++++++++++++++++++++++++++
9
target/arm/internals.h | 5 +++
14
tests/qtest/meson.build | 1 +
10
target/arm/mte_helper.c | 72 ++++++++++++++++++++++++++++++++++++++
15
2 files changed, 361 insertions(+)
11
target/arm/translate-a64.c | 18 ++++++++++
16
create mode 100644 tests/qtest/xlnx-can-test.c
12
target/arm/Makefile.objs | 1 +
17
13
5 files changed, 98 insertions(+)
18
diff --git a/tests/qtest/xlnx-can-test.c b/tests/qtest/xlnx-can-test.c
14
create mode 100644 target/arm/mte_helper.c
15
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-a64.h
19
+++ b/target/arm/helper-a64.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autda, TCG_CALL_NO_WG, i64, env, i64, i64)
21
DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
22
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
+
25
+DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
26
diff --git a/target/arm/internals.h b/target/arm/internals.h
27
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/internals.h
29
+++ b/target/arm/internals.h
30
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
31
*/
32
#define GMID_EL1_BS 6
33
34
+static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
35
+{
36
+ return deposit64(ptr, 56, 4, rtag);
37
+}
38
+
39
#endif
40
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
41
new file mode 100644
19
new file mode 100644
42
index XXXXXXX..XXXXXXX
20
index XXXXXXX..XXXXXXX
43
--- /dev/null
21
--- /dev/null
44
+++ b/target/arm/mte_helper.c
22
+++ b/tests/qtest/xlnx-can-test.c
45
@@ -XXX,XX +XXX,XX @@
23
@@ -XXX,XX +XXX,XX @@
46
+/*
24
+/*
47
+ * ARM v8.5-MemTag Operations
25
+ * QTests for the Xilinx ZynqMP CAN controller.
48
+ *
26
+ *
49
+ * Copyright (c) 2020 Linaro, Ltd.
27
+ * Copyright (c) 2020 Xilinx Inc.
50
+ *
28
+ *
51
+ * This library is free software; you can redistribute it and/or
29
+ * Written-by: Vikram Garhwal<fnu.vikram@xilinx.com>
52
+ * modify it under the terms of the GNU Lesser General Public
30
+ *
53
+ * License as published by the Free Software Foundation; either
31
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
54
+ * version 2.1 of the License, or (at your option) any later version.
32
+ * of this software and associated documentation files (the "Software"), to deal
55
+ *
33
+ * in the Software without restriction, including without limitation the rights
56
+ * This library is distributed in the hope that it will be useful,
34
+ * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
57
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
35
+ * copies of the Software, and to permit persons to whom the Software is
58
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
36
+ * furnished to do so, subject to the following conditions:
59
+ * Lesser General Public License for more details.
37
+ *
60
+ *
38
+ * The above copyright notice and this permission notice shall be included in
61
+ * You should have received a copy of the GNU Lesser General Public
39
+ * all copies or substantial portions of the Software.
62
+ * License along with this library; if not, see <http://www.gnu.org/licenses/>.
40
+ *
41
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
42
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
43
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
44
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
45
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
46
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
47
+ * THE SOFTWARE.
63
+ */
48
+ */
64
+
49
+
65
+#include "qemu/osdep.h"
50
+#include "qemu/osdep.h"
66
+#include "cpu.h"
51
+#include "libqos/libqtest.h"
67
+#include "internals.h"
52
+
68
+#include "exec/exec-all.h"
53
+/* Base address. */
69
+#include "exec/cpu_ldst.h"
54
+#define CAN0_BASE_ADDR 0xFF060000
70
+#include "exec/helper-proto.h"
55
+#define CAN1_BASE_ADDR 0xFF070000
71
+
56
+
72
+
57
+/* Register addresses. */
73
+static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
58
+#define R_SRR_OFFSET 0x00
74
+{
59
+#define R_MSR_OFFSET 0x04
75
+ if (exclude == 0xffff) {
60
+#define R_SR_OFFSET 0x18
76
+ return 0;
61
+#define R_ISR_OFFSET 0x1C
62
+#define R_ICR_OFFSET 0x24
63
+#define R_TXID_OFFSET 0x30
64
+#define R_TXDLC_OFFSET 0x34
65
+#define R_TXDATA1_OFFSET 0x38
66
+#define R_TXDATA2_OFFSET 0x3C
67
+#define R_RXID_OFFSET 0x50
68
+#define R_RXDLC_OFFSET 0x54
69
+#define R_RXDATA1_OFFSET 0x58
70
+#define R_RXDATA2_OFFSET 0x5C
71
+#define R_AFR 0x60
72
+#define R_AFMR1 0x64
73
+#define R_AFIR1 0x68
74
+#define R_AFMR2 0x6C
75
+#define R_AFIR2 0x70
76
+#define R_AFMR3 0x74
77
+#define R_AFIR3 0x78
78
+#define R_AFMR4 0x7C
79
+#define R_AFIR4 0x80
80
+
81
+/* CAN modes. */
82
+#define CONFIG_MODE 0x00
83
+#define NORMAL_MODE 0x00
84
+#define LOOPBACK_MODE 0x02
85
+#define SNOOP_MODE 0x04
86
+#define SLEEP_MODE 0x01
87
+#define ENABLE_CAN (1 << 1)
88
+#define STATUS_NORMAL_MODE (1 << 3)
89
+#define STATUS_LOOPBACK_MODE (1 << 1)
90
+#define STATUS_SNOOP_MODE (1 << 12)
91
+#define STATUS_SLEEP_MODE (1 << 2)
92
+#define ISR_TXOK (1 << 1)
93
+#define ISR_RXOK (1 << 4)
94
+
95
+static void match_rx_tx_data(const uint32_t *buf_tx, const uint32_t *buf_rx,
96
+ uint8_t can_timestamp)
97
+{
98
+ uint16_t size = 0;
99
+ uint8_t len = 4;
100
+
101
+ while (size < len) {
102
+ if (R_RXID_OFFSET + 4 * size == R_RXDLC_OFFSET) {
103
+ g_assert_cmpint(buf_rx[size], ==, buf_tx[size] + can_timestamp);
104
+ } else {
105
+ g_assert_cmpint(buf_rx[size], ==, buf_tx[size]);
106
+ }
107
+
108
+ size++;
77
+ }
109
+ }
78
+ if (offset == 0) {
110
+}
79
+ while (exclude & (1 << tag)) {
111
+
80
+ tag = (tag + 1) & 15;
112
+static void read_data(QTestState *qts, uint64_t can_base_addr, uint32_t *buf_rx)
81
+ }
113
+{
82
+ } else {
114
+ uint32_t int_status;
83
+ do {
115
+
84
+ do {
116
+ /* Read the interrupt on CAN rx. */
85
+ tag = (tag + 1) & 15;
117
+ int_status = qtest_readl(qts, can_base_addr + R_ISR_OFFSET) & ISR_RXOK;
86
+ } while (exclude & (1 << tag));
118
+
87
+ } while (--offset > 0);
119
+ g_assert_cmpint(int_status, ==, ISR_RXOK);
88
+ }
120
+
89
+ return tag;
121
+ /* Read the RX register data for CAN. */
90
+}
122
+ buf_rx[0] = qtest_readl(qts, can_base_addr + R_RXID_OFFSET);
91
+
123
+ buf_rx[1] = qtest_readl(qts, can_base_addr + R_RXDLC_OFFSET);
92
+uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
124
+ buf_rx[2] = qtest_readl(qts, can_base_addr + R_RXDATA1_OFFSET);
93
+{
125
+ buf_rx[3] = qtest_readl(qts, can_base_addr + R_RXDATA2_OFFSET);
94
+ int rtag;
126
+
127
+ /* Clear the RX interrupt. */
128
+ qtest_writel(qts, CAN1_BASE_ADDR + R_ICR_OFFSET, ISR_RXOK);
129
+}
130
+
131
+static void send_data(QTestState *qts, uint64_t can_base_addr,
132
+ const uint32_t *buf_tx)
133
+{
134
+ uint32_t int_status;
135
+
136
+ /* Write the TX register data for CAN. */
137
+ qtest_writel(qts, can_base_addr + R_TXID_OFFSET, buf_tx[0]);
138
+ qtest_writel(qts, can_base_addr + R_TXDLC_OFFSET, buf_tx[1]);
139
+ qtest_writel(qts, can_base_addr + R_TXDATA1_OFFSET, buf_tx[2]);
140
+ qtest_writel(qts, can_base_addr + R_TXDATA2_OFFSET, buf_tx[3]);
141
+
142
+ /* Read the interrupt on CAN for tx. */
143
+ int_status = qtest_readl(qts, can_base_addr + R_ISR_OFFSET) & ISR_TXOK;
144
+
145
+ g_assert_cmpint(int_status, ==, ISR_TXOK);
146
+
147
+ /* Clear the interrupt for tx. */
148
+ qtest_writel(qts, CAN0_BASE_ADDR + R_ICR_OFFSET, ISR_TXOK);
149
+}
150
+
151
+/*
152
+ * This test will be transferring data from CAN0 and CAN1 through canbus. CAN0
153
+ * initiate the data transfer to can-bus, CAN1 receives the data. Test compares
154
+ * the data sent from CAN0 with received on CAN1.
155
+ */
156
+static void test_can_bus(void)
157
+{
158
+ const uint32_t buf_tx[4] = { 0xFF, 0x80000000, 0x12345678, 0x87654321 };
159
+ uint32_t buf_rx[4] = { 0x00, 0x00, 0x00, 0x00 };
160
+ uint32_t status = 0;
161
+ uint8_t can_timestamp = 1;
162
+
163
+ QTestState *qts = qtest_init("-machine xlnx-zcu102"
164
+ " -object can-bus,id=canbus0"
165
+ " -machine xlnx-zcu102.canbus0=canbus0"
166
+ " -machine xlnx-zcu102.canbus1=canbus0"
167
+ );
168
+
169
+ /* Configure the CAN0 and CAN1. */
170
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
171
+ qtest_writel(qts, CAN0_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
172
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
173
+ qtest_writel(qts, CAN1_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
174
+
175
+ /* Check here if CAN0 and CAN1 are in normal mode. */
176
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
177
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
178
+
179
+ status = qtest_readl(qts, CAN1_BASE_ADDR + R_SR_OFFSET);
180
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
181
+
182
+ send_data(qts, CAN0_BASE_ADDR, buf_tx);
183
+
184
+ read_data(qts, CAN1_BASE_ADDR, buf_rx);
185
+ match_rx_tx_data(buf_tx, buf_rx, can_timestamp);
186
+
187
+ qtest_quit(qts);
188
+}
189
+
190
+/*
191
+ * This test is performing loopback mode on CAN0 and CAN1. Data sent from TX of
192
+ * each CAN0 and CAN1 are compared with RX register data for respective CAN.
193
+ */
194
+static void test_can_loopback(void)
195
+{
196
+ uint32_t buf_tx[4] = { 0xFF, 0x80000000, 0x12345678, 0x87654321 };
197
+ uint32_t buf_rx[4] = { 0x00, 0x00, 0x00, 0x00 };
198
+ uint32_t status = 0;
199
+
200
+ QTestState *qts = qtest_init("-machine xlnx-zcu102"
201
+ " -object can-bus,id=canbus0"
202
+ " -machine xlnx-zcu102.canbus0=canbus0"
203
+ " -machine xlnx-zcu102.canbus1=canbus0"
204
+ );
205
+
206
+ /* Configure the CAN0 in loopback mode. */
207
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, CONFIG_MODE);
208
+ qtest_writel(qts, CAN0_BASE_ADDR + R_MSR_OFFSET, LOOPBACK_MODE);
209
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
210
+
211
+ /* Check here if CAN0 is set in loopback mode. */
212
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
213
+
214
+ g_assert_cmpint(status, ==, STATUS_LOOPBACK_MODE);
215
+
216
+ send_data(qts, CAN0_BASE_ADDR, buf_tx);
217
+ read_data(qts, CAN0_BASE_ADDR, buf_rx);
218
+ match_rx_tx_data(buf_tx, buf_rx, 0);
219
+
220
+ /* Configure the CAN1 in loopback mode. */
221
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, CONFIG_MODE);
222
+ qtest_writel(qts, CAN1_BASE_ADDR + R_MSR_OFFSET, LOOPBACK_MODE);
223
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
224
+
225
+ /* Check here if CAN1 is set in loopback mode. */
226
+ status = qtest_readl(qts, CAN1_BASE_ADDR + R_SR_OFFSET);
227
+
228
+ g_assert_cmpint(status, ==, STATUS_LOOPBACK_MODE);
229
+
230
+ send_data(qts, CAN1_BASE_ADDR, buf_tx);
231
+ read_data(qts, CAN1_BASE_ADDR, buf_rx);
232
+ match_rx_tx_data(buf_tx, buf_rx, 0);
233
+
234
+ qtest_quit(qts);
235
+}
236
+
237
+/*
238
+ * Enable filters for CAN1. This will filter incoming messages with ID. In this
239
+ * test message will pass through filter 2.
240
+ */
241
+static void test_can_filter(void)
242
+{
243
+ uint32_t buf_tx[4] = { 0x14, 0x80000000, 0x12345678, 0x87654321 };
244
+ uint32_t buf_rx[4] = { 0x00, 0x00, 0x00, 0x00 };
245
+ uint32_t status = 0;
246
+ uint8_t can_timestamp = 1;
247
+
248
+ QTestState *qts = qtest_init("-machine xlnx-zcu102"
249
+ " -object can-bus,id=canbus0"
250
+ " -machine xlnx-zcu102.canbus0=canbus0"
251
+ " -machine xlnx-zcu102.canbus1=canbus0"
252
+ );
253
+
254
+ /* Configure the CAN0 and CAN1. */
255
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
256
+ qtest_writel(qts, CAN0_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
257
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
258
+ qtest_writel(qts, CAN1_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
259
+
260
+ /* Check here if CAN0 and CAN1 are in normal mode. */
261
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
262
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
263
+
264
+ status = qtest_readl(qts, CAN1_BASE_ADDR + R_SR_OFFSET);
265
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
266
+
267
+ /* Set filter for CAN1 for incoming messages. */
268
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFR, 0x0);
269
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFMR1, 0xF7);
270
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFIR1, 0x121F);
271
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFMR2, 0x5431);
272
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFIR2, 0x14);
273
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFMR3, 0x1234);
274
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFIR3, 0x5431);
275
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFMR4, 0xFFF);
276
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFIR4, 0x1234);
277
+
278
+ qtest_writel(qts, CAN1_BASE_ADDR + R_AFR, 0xF);
279
+
280
+ send_data(qts, CAN0_BASE_ADDR, buf_tx);
281
+
282
+ read_data(qts, CAN1_BASE_ADDR, buf_rx);
283
+ match_rx_tx_data(buf_tx, buf_rx, can_timestamp);
284
+
285
+ qtest_quit(qts);
286
+}
287
+
288
+/* Testing sleep mode on CAN0 while CAN1 is in normal mode. */
289
+static void test_can_sleepmode(void)
290
+{
291
+ uint32_t buf_tx[4] = { 0x14, 0x80000000, 0x12345678, 0x87654321 };
292
+ uint32_t buf_rx[4] = { 0x00, 0x00, 0x00, 0x00 };
293
+ uint32_t status = 0;
294
+ uint8_t can_timestamp = 1;
295
+
296
+ QTestState *qts = qtest_init("-machine xlnx-zcu102"
297
+ " -object can-bus,id=canbus0"
298
+ " -machine xlnx-zcu102.canbus0=canbus0"
299
+ " -machine xlnx-zcu102.canbus1=canbus0"
300
+ );
301
+
302
+ /* Configure the CAN0. */
303
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, CONFIG_MODE);
304
+ qtest_writel(qts, CAN0_BASE_ADDR + R_MSR_OFFSET, SLEEP_MODE);
305
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
306
+
307
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
308
+ qtest_writel(qts, CAN1_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
309
+
310
+ /* Check here if CAN0 is in SLEEP mode and CAN1 in normal mode. */
311
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
312
+ g_assert_cmpint(status, ==, STATUS_SLEEP_MODE);
313
+
314
+ status = qtest_readl(qts, CAN1_BASE_ADDR + R_SR_OFFSET);
315
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
316
+
317
+ send_data(qts, CAN1_BASE_ADDR, buf_tx);
95
+
318
+
96
+ /*
319
+ /*
97
+ * Our IMPDEF choice for GCR_EL1.RRND==1 is to behave as if
320
+ * Once CAN1 sends data on can-bus. CAN0 should exit sleep mode.
98
+ * GCR_EL1.RRND==0, always producing deterministic results.
321
+ * Check the CAN0 status now. It should exit the sleep mode and receive the
322
+ * incoming data.
99
+ */
323
+ */
100
+ uint16_t exclude = extract32(rm | env->cp15.gcr_el1, 0, 16);
324
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
101
+ int start = extract32(env->cp15.rgsr_el1, 0, 4);
325
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
102
+ int seed = extract32(env->cp15.rgsr_el1, 8, 16);
326
+
103
+ int offset, i;
327
+ read_data(qts, CAN0_BASE_ADDR, buf_rx);
104
+
328
+
105
+ /* RandomTag */
329
+ match_rx_tx_data(buf_tx, buf_rx, can_timestamp);
106
+ for (i = offset = 0; i < 4; ++i) {
330
+
107
+ /* NextRandomTagBit */
331
+ qtest_quit(qts);
108
+ int top = (extract32(seed, 5, 1) ^ extract32(seed, 3, 1) ^
332
+}
109
+ extract32(seed, 2, 1) ^ extract32(seed, 0, 1));
333
+
110
+ seed = (top << 15) | (seed >> 1);
334
+/* Testing Snoop mode on CAN0 while CAN1 is in normal mode. */
111
+ offset |= top << i;
335
+static void test_can_snoopmode(void)
112
+ }
336
+{
113
+ rtag = choose_nonexcluded_tag(start, offset, exclude);
337
+ uint32_t buf_tx[4] = { 0x14, 0x80000000, 0x12345678, 0x87654321 };
114
+ env->cp15.rgsr_el1 = rtag | (seed << 8);
338
+ uint32_t buf_rx[4] = { 0x00, 0x00, 0x00, 0x00 };
115
+
339
+ uint32_t status = 0;
116
+ return address_with_allocation_tag(rn, rtag);
340
+ uint8_t can_timestamp = 1;
117
+}
341
+
118
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
342
+ QTestState *qts = qtest_init("-machine xlnx-zcu102"
343
+ " -object can-bus,id=canbus0"
344
+ " -machine xlnx-zcu102.canbus0=canbus0"
345
+ " -machine xlnx-zcu102.canbus1=canbus0"
346
+ );
347
+
348
+ /* Configure the CAN0. */
349
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, CONFIG_MODE);
350
+ qtest_writel(qts, CAN0_BASE_ADDR + R_MSR_OFFSET, SNOOP_MODE);
351
+ qtest_writel(qts, CAN0_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
352
+
353
+ qtest_writel(qts, CAN1_BASE_ADDR + R_SRR_OFFSET, ENABLE_CAN);
354
+ qtest_writel(qts, CAN1_BASE_ADDR + R_MSR_OFFSET, NORMAL_MODE);
355
+
356
+ /* Check here if CAN0 is in SNOOP mode and CAN1 in normal mode. */
357
+ status = qtest_readl(qts, CAN0_BASE_ADDR + R_SR_OFFSET);
358
+ g_assert_cmpint(status, ==, STATUS_SNOOP_MODE);
359
+
360
+ status = qtest_readl(qts, CAN1_BASE_ADDR + R_SR_OFFSET);
361
+ g_assert_cmpint(status, ==, STATUS_NORMAL_MODE);
362
+
363
+ send_data(qts, CAN1_BASE_ADDR, buf_tx);
364
+
365
+ read_data(qts, CAN0_BASE_ADDR, buf_rx);
366
+
367
+ match_rx_tx_data(buf_tx, buf_rx, can_timestamp);
368
+
369
+ qtest_quit(qts);
370
+}
371
+
372
+int main(int argc, char **argv)
373
+{
374
+ g_test_init(&argc, &argv, NULL);
375
+
376
+ qtest_add_func("/net/can/can_bus", test_can_bus);
377
+ qtest_add_func("/net/can/can_loopback", test_can_loopback);
378
+ qtest_add_func("/net/can/can_filter", test_can_filter);
379
+ qtest_add_func("/net/can/can_test_snoopmode", test_can_snoopmode);
380
+ qtest_add_func("/net/can/can_test_sleepmode", test_can_sleepmode);
381
+
382
+ return g_test_run();
383
+}
384
diff --git a/tests/qtest/meson.build b/tests/qtest/meson.build
119
index XXXXXXX..XXXXXXX 100644
385
index XXXXXXX..XXXXXXX 100644
120
--- a/target/arm/translate-a64.c
386
--- a/tests/qtest/meson.build
121
+++ b/target/arm/translate-a64.c
387
+++ b/tests/qtest/meson.build
122
@@ -XXX,XX +XXX,XX @@ static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
388
@@ -XXX,XX +XXX,XX @@ qtests_aarch64 = \
123
return clean;
389
['arm-cpu-features',
124
}
390
'numa-test',
125
391
'boot-serial-test',
126
+/* Insert a zero tag into src, with the result at dst. */
392
+ 'xlnx-can-test',
127
+static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
393
'migration-test']
128
+{
394
129
+ tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
395
qtests_s390x = \
130
+}
131
+
132
typedef struct DisasCompare64 {
133
TCGCond cond;
134
TCGv_i64 value;
135
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
136
case 3: /* SDIV */
137
handle_div(s, true, sf, rm, rn, rd);
138
break;
139
+ case 4: /* IRG */
140
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
141
+ goto do_unallocated;
142
+ }
143
+ if (s->ata) {
144
+ gen_helper_irg(cpu_reg_sp(s, rd), cpu_env,
145
+ cpu_reg_sp(s, rn), cpu_reg(s, rm));
146
+ } else {
147
+ gen_address_with_allocation_tag0(cpu_reg_sp(s, rd),
148
+ cpu_reg_sp(s, rn));
149
+ }
150
+ break;
151
case 8: /* LSLV */
152
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
153
break;
154
diff --git a/target/arm/Makefile.objs b/target/arm/Makefile.objs
155
index XXXXXXX..XXXXXXX 100644
156
--- a/target/arm/Makefile.objs
157
+++ b/target/arm/Makefile.objs
158
@@ -XXX,XX +XXX,XX @@ obj-$(CONFIG_SOFTMMU) += psci.o
159
obj-$(TARGET_AARCH64) += translate-a64.o helper-a64.o
160
obj-$(TARGET_AARCH64) += translate-sve.o sve_helper.o
161
obj-$(TARGET_AARCH64) += pauth_helper.o
162
+obj-$(TARGET_AARCH64) += mte_helper.o
163
--
396
--
164
2.20.1
397
2.20.1
165
398
166
399
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Vikram Garhwal <fnu.vikram@xilinx.com>
2
2
3
D1.10 specifies that exception handlers begin with tag checks overridden.
3
Reviewed-by: Francisco Iglesias <francisco.iglesias@xilinx.com>
4
4
Reviewed-by: Edgar E. Iglesias <edgar.iglesias@xilinx.com>
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Vikram Garhwal <fnu.vikram@xilinx.com>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 1605728926-352690-5-git-send-email-fnu.vikram@xilinx.com
7
Message-id: 20200626033144.790098-41-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
8
---
10
target/arm/helper.c | 3 +++
9
MAINTAINERS | 8 ++++++++
11
1 file changed, 3 insertions(+)
10
1 file changed, 8 insertions(+)
12
11
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
12
diff --git a/MAINTAINERS b/MAINTAINERS
14
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
14
--- a/MAINTAINERS
16
+++ b/target/arm/helper.c
15
+++ b/MAINTAINERS
17
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_do_interrupt_aarch64(CPUState *cs)
16
@@ -XXX,XX +XXX,XX @@ F: hw/net/opencores_eth.c
18
break;
17
19
}
18
Devices
20
}
19
-------
21
+ if (cpu_isar_feature(aa64_mte, cpu)) {
20
+Xilinx CAN
22
+ new_mode |= PSTATE_TCO;
21
+M: Vikram Garhwal <fnu.vikram@xilinx.com>
23
+ }
22
+M: Francisco Iglesias <francisco.iglesias@xilinx.com>
24
23
+S: Maintained
25
pstate_write(env, PSTATE_DAIF | new_mode);
24
+F: hw/net/can/xlnx-*
26
env->aarch64 = 1;
25
+F: include/hw/net/xlnx-*
26
+F: tests/qtest/xlnx-can-test*
27
+
28
EDU
29
M: Jiri Slaby <jslaby@suse.cz>
30
S: Maintained
27
--
31
--
28
2.20.1
32
2.20.1
29
33
30
34
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Marcin Juszkiewicz <marcin.juszkiewicz@linaro.org>
2
2
3
Like the regular data cache flushes, these are nops within qemu.
3
Trusted Firmware now supports A72 on sbsa-ref by default [1] so enable
4
it for QEMU as well. A53 was already enabled there.
4
5
6
1. https://review.trustedfirmware.org/c/TF-A/trusted-firmware-a/+/7117
7
8
Signed-off-by: Marcin Juszkiewicz <marcin.juszkiewicz@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20201120141705.246690-1-marcin.juszkiewicz@linaro.org
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-21-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
13
---
10
target/arm/helper.c | 65 +++++++++++++++++++++++++++++++++++++++++++++
14
hw/arm/sbsa-ref.c | 23 ++++++++++++++++++++---
11
1 file changed, 65 insertions(+)
15
1 file changed, 20 insertions(+), 3 deletions(-)
12
16
13
diff --git a/target/arm/helper.c b/target/arm/helper.c
17
diff --git a/hw/arm/sbsa-ref.c b/hw/arm/sbsa-ref.c
14
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/helper.c
19
--- a/hw/arm/sbsa-ref.c
16
+++ b/target/arm/helper.c
20
+++ b/hw/arm/sbsa-ref.c
17
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_reginfo[] = {
21
@@ -XXX,XX +XXX,XX @@ static const int sbsa_ref_irqmap[] = {
18
.opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
22
[SBSA_GWDT] = 16,
19
.type = ARM_CP_NO_RAW,
20
.access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
21
+ { .name = "DC_IGVAC", .state = ARM_CP_STATE_AA64,
22
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 3,
23
+ .type = ARM_CP_NOP, .access = PL1_W,
24
+ .accessfn = aa64_cacheop_poc_access },
25
+ { .name = "DC_IGSW", .state = ARM_CP_STATE_AA64,
26
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 4,
27
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
28
+ { .name = "DC_IGDVAC", .state = ARM_CP_STATE_AA64,
29
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 5,
30
+ .type = ARM_CP_NOP, .access = PL1_W,
31
+ .accessfn = aa64_cacheop_poc_access },
32
+ { .name = "DC_IGDSW", .state = ARM_CP_STATE_AA64,
33
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 6, .opc2 = 6,
34
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
35
+ { .name = "DC_CGSW", .state = ARM_CP_STATE_AA64,
36
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 4,
37
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
38
+ { .name = "DC_CGDSW", .state = ARM_CP_STATE_AA64,
39
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 10, .opc2 = 6,
40
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
41
+ { .name = "DC_CIGSW", .state = ARM_CP_STATE_AA64,
42
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 4,
43
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
44
+ { .name = "DC_CIGDSW", .state = ARM_CP_STATE_AA64,
45
+ .opc0 = 1, .opc1 = 0, .crn = 7, .crm = 14, .opc2 = 6,
46
+ .type = ARM_CP_NOP, .access = PL1_W, .accessfn = access_tsw },
47
REGINFO_SENTINEL
48
};
23
};
49
24
50
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
25
+static const char * const valid_cpus[] = {
51
.type = ARM_CP_CONST, .access = PL0_RW, },
26
+ ARM_CPU_TYPE_NAME("cortex-a53"),
52
REGINFO_SENTINEL
27
+ ARM_CPU_TYPE_NAME("cortex-a57"),
53
};
28
+ ARM_CPU_TYPE_NAME("cortex-a72"),
54
+
55
+static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
56
+ { .name = "DC_CGVAC", .state = ARM_CP_STATE_AA64,
57
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 3,
58
+ .type = ARM_CP_NOP, .access = PL0_W,
59
+ .accessfn = aa64_cacheop_poc_access },
60
+ { .name = "DC_CGDVAC", .state = ARM_CP_STATE_AA64,
61
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 10, .opc2 = 5,
62
+ .type = ARM_CP_NOP, .access = PL0_W,
63
+ .accessfn = aa64_cacheop_poc_access },
64
+ { .name = "DC_CGVAP", .state = ARM_CP_STATE_AA64,
65
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 3,
66
+ .type = ARM_CP_NOP, .access = PL0_W,
67
+ .accessfn = aa64_cacheop_poc_access },
68
+ { .name = "DC_CGDVAP", .state = ARM_CP_STATE_AA64,
69
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 12, .opc2 = 5,
70
+ .type = ARM_CP_NOP, .access = PL0_W,
71
+ .accessfn = aa64_cacheop_poc_access },
72
+ { .name = "DC_CGVADP", .state = ARM_CP_STATE_AA64,
73
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 3,
74
+ .type = ARM_CP_NOP, .access = PL0_W,
75
+ .accessfn = aa64_cacheop_poc_access },
76
+ { .name = "DC_CGDVADP", .state = ARM_CP_STATE_AA64,
77
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 13, .opc2 = 5,
78
+ .type = ARM_CP_NOP, .access = PL0_W,
79
+ .accessfn = aa64_cacheop_poc_access },
80
+ { .name = "DC_CIGVAC", .state = ARM_CP_STATE_AA64,
81
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 3,
82
+ .type = ARM_CP_NOP, .access = PL0_W,
83
+ .accessfn = aa64_cacheop_poc_access },
84
+ { .name = "DC_CIGDVAC", .state = ARM_CP_STATE_AA64,
85
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
86
+ .type = ARM_CP_NOP, .access = PL0_W,
87
+ .accessfn = aa64_cacheop_poc_access },
88
+ REGINFO_SENTINEL
89
+};
29
+};
90
+
30
+
91
#endif
31
+static bool cpu_type_valid(const char *cpu)
92
32
+{
93
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
33
+ int i;
94
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
34
+
95
*/
35
+ for (i = 0; i < ARRAY_SIZE(valid_cpus); i++) {
96
if (cpu_isar_feature(aa64_mte, cpu)) {
36
+ if (strcmp(cpu, valid_cpus[i]) == 0) {
97
define_arm_cp_regs(cpu, mte_reginfo);
37
+ return true;
98
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
38
+ }
99
} else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
39
+ }
100
define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
40
+ return false;
101
+ define_arm_cp_regs(cpu, mte_el0_cacheop_reginfo);
41
+}
42
+
43
static uint64_t sbsa_ref_cpu_mp_affinity(SBSAMachineState *sms, int idx)
44
{
45
uint8_t clustersz = ARM_DEFAULT_CPUS_PER_CLUSTER;
46
@@ -XXX,XX +XXX,XX @@ static void sbsa_ref_init(MachineState *machine)
47
const CPUArchIdList *possible_cpus;
48
int n, sbsa_max_cpus;
49
50
- if (strcmp(machine->cpu_type, ARM_CPU_TYPE_NAME("cortex-a57"))) {
51
- error_report("sbsa-ref: CPU type other than the built-in "
52
- "cortex-a57 not supported");
53
+ if (!cpu_type_valid(machine->cpu_type)) {
54
+ error_report("mach-virt: CPU type %s not supported", machine->cpu_type);
55
exit(1);
102
}
56
}
103
#endif
104
57
105
--
58
--
106
2.20.1
59
2.20.1
107
60
108
61
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
From: Havard Skinnemoen <hskinnemoen@google.com>
2
2
3
To have a more consistent naming, rename AspeedBoardState
3
Dump the collected random data after a randomness test failure.
4
as AspeedMachineState.
5
4
6
Suggested-by: Cédric Le Goater <clg@kaod.org>
5
Note that this relies on the test having called
7
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
6
g_test_set_nonfatal_assertions() so we don't abort immediately on the
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
7
assertion failure.
9
Message-id: 20200623072132.2868-3-f4bug@amsat.org
8
9
Signed-off-by: Havard Skinnemoen <hskinnemoen@google.com>
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
[PMM: minor commit message tweak]
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
13
---
12
include/hw/arm/aspeed.h | 4 ++--
14
tests/qtest/npcm7xx_rng-test.c | 12 ++++++++++++
13
hw/arm/aspeed.c | 20 ++++++++++----------
15
1 file changed, 12 insertions(+)
14
2 files changed, 12 insertions(+), 12 deletions(-)
15
16
16
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
17
diff --git a/tests/qtest/npcm7xx_rng-test.c b/tests/qtest/npcm7xx_rng-test.c
17
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
18
--- a/include/hw/arm/aspeed.h
19
--- a/tests/qtest/npcm7xx_rng-test.c
19
+++ b/include/hw/arm/aspeed.h
20
+++ b/tests/qtest/npcm7xx_rng-test.c
20
@@ -XXX,XX +XXX,XX @@
21
@@ -XXX,XX +XXX,XX @@
21
22
22
#include "hw/boards.h"
23
#include "libqtest-single.h"
23
24
#include "qemu/bitops.h"
24
-typedef struct AspeedBoardState AspeedBoardState;
25
+#include "qemu-common.h"
25
+typedef struct AspeedMachineState AspeedMachineState;
26
26
27
#define RNG_BASE_ADDR 0xf000b000
27
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
28
28
#define ASPEED_MACHINE(obj) \
29
@@ -XXX,XX +XXX,XX @@
29
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineClass {
30
/* Number of bits to collect for randomness tests. */
30
const char *spi_model;
31
#define TEST_INPUT_BITS (128)
31
uint32_t num_cs;
32
32
uint32_t macs_mask;
33
+static void dump_buf_if_failed(const uint8_t *buf, size_t size)
33
- void (*i2c_init)(AspeedBoardState *bmc);
34
+{
34
+ void (*i2c_init)(AspeedMachineState *bmc);
35
+ if (g_test_failed()) {
35
} AspeedMachineClass;
36
+ qemu_hexdump(stderr, "", buf, size);
36
37
+ }
37
38
+}
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
+
39
index XXXXXXX..XXXXXXX 100644
40
static void rng_writeb(unsigned int offset, uint8_t value)
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
.board_id = -1, /* device-tree-only board */
44
};
45
46
-struct AspeedBoardState {
47
+struct AspeedMachineState {
48
AspeedSoCState soc;
49
MemoryRegion ram_container;
50
MemoryRegion max_ram;
51
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
52
53
static void aspeed_machine_init(MachineState *machine)
54
{
41
{
55
- AspeedBoardState *bmc;
42
writeb(RNG_BASE_ADDR + offset, value);
56
+ AspeedMachineState *bmc;
43
@@ -XXX,XX +XXX,XX @@ static void test_continuous_monobit(void)
57
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
44
}
58
AspeedSoCClass *sc;
45
59
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
46
g_assert_cmpfloat(calc_monobit_p(buf, sizeof(buf)), >, 0.01);
60
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
47
+ dump_buf_if_failed(buf, sizeof(buf));
61
int i;
62
NICInfo *nd = &nd_table[0];
63
64
- bmc = g_new0(AspeedBoardState, 1);
65
+ bmc = g_new0(AspeedMachineState, 1);
66
67
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
68
4 * GiB);
69
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
70
arm_load_kernel(ARM_CPU(first_cpu), machine, &aspeed_board_binfo);
71
}
48
}
72
49
73
-static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
50
/*
74
+static void palmetto_bmc_i2c_init(AspeedMachineState *bmc)
51
@@ -XXX,XX +XXX,XX @@ static void test_continuous_runs(void)
75
{
52
}
76
AspeedSoCState *soc = &bmc->soc;
53
77
DeviceState *dev;
54
g_assert_cmpfloat(calc_runs_p(buf.l, sizeof(buf) * BITS_PER_BYTE), >, 0.01);
78
@@ -XXX,XX +XXX,XX @@ static void palmetto_bmc_i2c_init(AspeedBoardState *bmc)
55
+ dump_buf_if_failed(buf.c, sizeof(buf));
79
object_property_set_int(OBJECT(dev), 110000, "temperature3", &error_abort);
80
}
56
}
81
57
82
-static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
58
/*
83
+static void ast2500_evb_i2c_init(AspeedMachineState *bmc)
59
@@ -XXX,XX +XXX,XX @@ static void test_first_byte_monobit(void)
84
{
60
}
85
AspeedSoCState *soc = &bmc->soc;
61
86
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
62
g_assert_cmpfloat(calc_monobit_p(buf, sizeof(buf)), >, 0.01);
87
@@ -XXX,XX +XXX,XX @@ static void ast2500_evb_i2c_init(AspeedBoardState *bmc)
63
+ dump_buf_if_failed(buf, sizeof(buf));
88
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
89
}
64
}
90
65
91
-static void ast2600_evb_i2c_init(AspeedBoardState *bmc)
66
/*
92
+static void ast2600_evb_i2c_init(AspeedMachineState *bmc)
67
@@ -XXX,XX +XXX,XX @@ static void test_first_byte_runs(void)
93
{
68
}
94
/* Start with some devices on our I2C busses */
69
95
ast2500_evb_i2c_init(bmc);
70
g_assert_cmpfloat(calc_runs_p(buf.l, sizeof(buf) * BITS_PER_BYTE), >, 0.01);
71
+ dump_buf_if_failed(buf.c, sizeof(buf));
96
}
72
}
97
73
98
-static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
74
int main(int argc, char **argv)
99
+static void romulus_bmc_i2c_init(AspeedMachineState *bmc)
100
{
101
AspeedSoCState *soc = &bmc->soc;
102
103
@@ -XXX,XX +XXX,XX @@ static void romulus_bmc_i2c_init(AspeedBoardState *bmc)
104
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), "ds1338", 0x32);
105
}
106
107
-static void swift_bmc_i2c_init(AspeedBoardState *bmc)
108
+static void swift_bmc_i2c_init(AspeedMachineState *bmc)
109
{
110
AspeedSoCState *soc = &bmc->soc;
111
112
@@ -XXX,XX +XXX,XX @@ static void swift_bmc_i2c_init(AspeedBoardState *bmc)
113
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 12), "tmp105", 0x4a);
114
}
115
116
-static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
117
+static void sonorapass_bmc_i2c_init(AspeedMachineState *bmc)
118
{
119
AspeedSoCState *soc = &bmc->soc;
120
121
@@ -XXX,XX +XXX,XX @@ static void sonorapass_bmc_i2c_init(AspeedBoardState *bmc)
122
123
}
124
125
-static void witherspoon_bmc_i2c_init(AspeedBoardState *bmc)
126
+static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
127
{
128
AspeedSoCState *soc = &bmc->soc;
129
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
130
--
75
--
131
2.20.1
76
2.20.1
132
77
133
78
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
AspeedMachineState seems crippled. We use incorrectly 2
4
different structures to do the same thing. Merge them
5
altogether:
6
- Move AspeedMachine fields to AspeedMachineState
7
- AspeedMachineState is now QOM
8
- Remove unused AspeedMachine structure
9
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
12
Message-id: 20200623072132.2868-4-f4bug@amsat.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
include/hw/arm/aspeed.h | 8 +-------
16
hw/arm/aspeed.c | 11 +++++++----
17
2 files changed, 8 insertions(+), 11 deletions(-)
18
19
diff --git a/include/hw/arm/aspeed.h b/include/hw/arm/aspeed.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/include/hw/arm/aspeed.h
22
+++ b/include/hw/arm/aspeed.h
23
@@ -XXX,XX +XXX,XX @@ typedef struct AspeedMachineState AspeedMachineState;
24
25
#define TYPE_ASPEED_MACHINE MACHINE_TYPE_NAME("aspeed")
26
#define ASPEED_MACHINE(obj) \
27
- OBJECT_CHECK(AspeedMachine, (obj), TYPE_ASPEED_MACHINE)
28
-
29
-typedef struct AspeedMachine {
30
- MachineState parent_obj;
31
-
32
- bool mmio_exec;
33
-} AspeedMachine;
34
+ OBJECT_CHECK(AspeedMachineState, (obj), TYPE_ASPEED_MACHINE)
35
36
#define ASPEED_MAC0_ON (1 << 0)
37
#define ASPEED_MAC1_ON (1 << 1)
38
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/hw/arm/aspeed.c
41
+++ b/hw/arm/aspeed.c
42
@@ -XXX,XX +XXX,XX @@ static struct arm_boot_info aspeed_board_binfo = {
43
};
44
45
struct AspeedMachineState {
46
+ /* Private */
47
+ MachineState parent_obj;
48
+ /* Public */
49
+
50
AspeedSoCState soc;
51
MemoryRegion ram_container;
52
MemoryRegion max_ram;
53
+ bool mmio_exec;
54
};
55
56
/* Palmetto hardware value: 0x120CE416 */
57
@@ -XXX,XX +XXX,XX @@ static void sdhci_attach_drive(SDHCIState *sdhci, DriveInfo *dinfo)
58
59
static void aspeed_machine_init(MachineState *machine)
60
{
61
- AspeedMachineState *bmc;
62
+ AspeedMachineState *bmc = ASPEED_MACHINE(machine);
63
AspeedMachineClass *amc = ASPEED_MACHINE_GET_CLASS(machine);
64
AspeedSoCClass *sc;
65
DriveInfo *drive0 = drive_get(IF_MTD, 0, 0);
66
@@ -XXX,XX +XXX,XX @@ static void aspeed_machine_init(MachineState *machine)
67
int i;
68
NICInfo *nd = &nd_table[0];
69
70
- bmc = g_new0(AspeedMachineState, 1);
71
-
72
memory_region_init(&bmc->ram_container, NULL, "aspeed-ram-container",
73
4 * GiB);
74
memory_region_add_subregion(&bmc->ram_container, 0, machine->ram);
75
@@ -XXX,XX +XXX,XX @@ static const TypeInfo aspeed_machine_types[] = {
76
}, {
77
.name = TYPE_ASPEED_MACHINE,
78
.parent = TYPE_MACHINE,
79
- .instance_size = sizeof(AspeedMachine),
80
+ .instance_size = sizeof(AspeedMachineState),
81
.instance_init = aspeed_machine_instance_init,
82
.class_size = sizeof(AspeedMachineClass),
83
.class_init = aspeed_machine_class_init,
84
--
85
2.20.1
86
87
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Alex Chen <alex.chen@huawei.com>
2
2
3
We should use printf format specifier "%u" instead of "%d" for
4
argument of type "unsigned int".
5
6
Reported-by: Euler Robot <euler.robot@huawei.com>
7
Signed-off-by: Alex Chen <alex.chen@huawei.com>
8
Message-id: 20201126111109.112238-2-alex.chen@huawei.com
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-20-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
11
---
8
target/arm/helper-a64.h | 3 ++
12
hw/misc/imx25_ccm.c | 12 ++++++------
9
target/arm/translate.h | 2 +
13
1 file changed, 6 insertions(+), 6 deletions(-)
10
target/arm/mte_helper.c | 84 ++++++++++++++++++++++++++++++++++++++
11
target/arm/translate-a64.c | 72 ++++++++++++++++++++++++++++----
12
4 files changed, 153 insertions(+), 8 deletions(-)
13
14
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
diff --git a/hw/misc/imx25_ccm.c b/hw/misc/imx25_ccm.c
15
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
--- a/hw/misc/imx25_ccm.c
17
+++ b/target/arm/helper-a64.h
18
+++ b/hw/misc/imx25_ccm.c
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
19
@@ -XXX,XX +XXX,XX @@ static const char *imx25_ccm_reg_name(uint32_t reg)
19
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
20
case IMX25_CCM_LPIMR1_REG:
20
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
21
return "lpimr1";
21
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
22
default:
22
+DEF_HELPER_FLAGS_2(ldgm, TCG_CALL_NO_WG, i64, env, i64)
23
- sprintf(unknown, "[%d ?]", reg);
23
+DEF_HELPER_FLAGS_3(stgm, TCG_CALL_NO_WG, void, env, i64, i64)
24
+ sprintf(unknown, "[%u ?]", reg);
24
+DEF_HELPER_FLAGS_3(stzgm_tags, TCG_CALL_NO_WG, void, env, i64, i64)
25
return unknown;
25
diff --git a/target/arm/translate.h b/target/arm/translate.h
26
index XXXXXXX..XXXXXXX 100644
27
--- a/target/arm/translate.h
28
+++ b/target/arm/translate.h
29
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
30
* < 0, set by the current instruction.
31
*/
32
int8_t btype;
33
+ /* A copy of cpu->dcz_blocksize. */
34
+ uint8_t dcz_blocksize;
35
/* True if this page is guarded. */
36
bool guarded_page;
37
/* Bottom two bits of XScale c15_cpar coprocessor access control reg */
38
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
39
index XXXXXXX..XXXXXXX 100644
40
--- a/target/arm/mte_helper.c
41
+++ b/target/arm/mte_helper.c
42
@@ -XXX,XX +XXX,XX @@ void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
43
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
44
}
26
}
45
}
27
}
46
+
28
@@ -XXX,XX +XXX,XX @@ static uint32_t imx25_ccm_get_mpll_clk(IMXCCMState *dev)
47
+#define LDGM_STGM_SIZE (4 << GMID_EL1_BS)
29
freq = imx_ccm_calc_pll(s->reg[IMX25_CCM_MPCTL_REG], CKIH_FREQ);
48
+
49
+uint64_t HELPER(ldgm)(CPUARMState *env, uint64_t ptr)
50
+{
51
+ int mmu_idx = cpu_mmu_index(env, false);
52
+ uintptr_t ra = GETPC();
53
+ void *tag_mem;
54
+
55
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
56
+
57
+ /* Trap if accessing an invalid page. */
58
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD,
59
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
60
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
61
+
62
+ /* The tag is squashed to zero if the page does not support tags. */
63
+ if (!tag_mem) {
64
+ return 0;
65
+ }
66
+
67
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
68
+ /*
69
+ * We are loading 64-bits worth of tags. The ordering of elements
70
+ * within the word corresponds to a 64-bit little-endian operation.
71
+ */
72
+ return ldq_le_p(tag_mem);
73
+}
74
+
75
+void HELPER(stgm)(CPUARMState *env, uint64_t ptr, uint64_t val)
76
+{
77
+ int mmu_idx = cpu_mmu_index(env, false);
78
+ uintptr_t ra = GETPC();
79
+ void *tag_mem;
80
+
81
+ ptr = QEMU_ALIGN_DOWN(ptr, LDGM_STGM_SIZE);
82
+
83
+ /* Trap if accessing an invalid page. */
84
+ tag_mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
85
+ LDGM_STGM_SIZE, MMU_DATA_LOAD,
86
+ LDGM_STGM_SIZE / (2 * TAG_GRANULE), ra);
87
+
88
+ /*
89
+ * Tag store only happens if the page support tags,
90
+ * and if the OS has enabled access to the tags.
91
+ */
92
+ if (!tag_mem) {
93
+ return;
94
+ }
95
+
96
+ QEMU_BUILD_BUG_ON(GMID_EL1_BS != 6);
97
+ /*
98
+ * We are storing 64-bits worth of tags. The ordering of elements
99
+ * within the word corresponds to a 64-bit little-endian operation.
100
+ */
101
+ stq_le_p(tag_mem, val);
102
+}
103
+
104
+void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
105
+{
106
+ uintptr_t ra = GETPC();
107
+ int mmu_idx = cpu_mmu_index(env, false);
108
+ int log2_dcz_bytes, log2_tag_bytes;
109
+ intptr_t dcz_bytes, tag_bytes;
110
+ uint8_t *mem;
111
+
112
+ /*
113
+ * In arm_cpu_realizefn, we assert that dcz > LOG2_TAG_GRANULE+1,
114
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway,
115
+ * to make sure that we can access one complete tag byte here.
116
+ */
117
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
118
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
119
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
120
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
121
+ ptr &= -dcz_bytes;
122
+
123
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, dcz_bytes,
124
+ MMU_DATA_STORE, tag_bytes, ra);
125
+ if (mem) {
126
+ int tag_pair = (val & 0xf) * 0x11;
127
+ memset(mem, tag_pair, tag_bytes);
128
+ }
129
+}
130
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
131
index XXXXXXX..XXXXXXX 100644
132
--- a/target/arm/translate-a64.c
133
+++ b/target/arm/translate-a64.c
134
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
135
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
136
int op2 = extract32(insn, 10, 2);
137
int op1 = extract32(insn, 22, 2);
138
- bool is_load = false, is_pair = false, is_zero = false;
139
+ bool is_load = false, is_pair = false, is_zero = false, is_mult = false;
140
int index = 0;
141
TCGv_i64 addr, clean_addr, tcg_rt;
142
143
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
144
if (op2 != 0) {
145
/* STG */
146
index = op2 - 2;
147
- break;
148
+ } else {
149
+ /* STZGM */
150
+ if (s->current_el == 0 || offset != 0) {
151
+ goto do_unallocated;
152
+ }
153
+ is_mult = is_zero = true;
154
}
155
- goto do_unallocated;
156
+ break;
157
case 1:
158
if (op2 != 0) {
159
/* STZG */
160
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
161
/* ST2G */
162
is_pair = true;
163
index = op2 - 2;
164
- break;
165
+ } else {
166
+ /* STGM */
167
+ if (s->current_el == 0 || offset != 0) {
168
+ goto do_unallocated;
169
+ }
170
+ is_mult = true;
171
}
172
- goto do_unallocated;
173
+ break;
174
case 3:
175
if (op2 != 0) {
176
/* STZ2G */
177
is_pair = is_zero = true;
178
index = op2 - 2;
179
- break;
180
+ } else {
181
+ /* LDGM */
182
+ if (s->current_el == 0 || offset != 0) {
183
+ goto do_unallocated;
184
+ }
185
+ is_mult = is_load = true;
186
}
187
- goto do_unallocated;
188
+ break;
189
190
default:
191
do_unallocated:
192
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
193
return;
194
}
30
}
195
31
196
- if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
32
- DPRINTF("freq = %d\n", freq);
197
+ if (is_mult
33
+ DPRINTF("freq = %u\n", freq);
198
+ ? !dc_isar_feature(aa64_mte, s)
34
199
+ : !dc_isar_feature(aa64_mte_insn_reg, s)) {
35
return freq;
200
goto do_unallocated;
36
}
37
@@ -XXX,XX +XXX,XX @@ static uint32_t imx25_ccm_get_mcu_clk(IMXCCMState *dev)
38
39
freq = freq / (1 + EXTRACT(s->reg[IMX25_CCM_CCTL_REG], ARM_CLK_DIV));
40
41
- DPRINTF("freq = %d\n", freq);
42
+ DPRINTF("freq = %u\n", freq);
43
44
return freq;
45
}
46
@@ -XXX,XX +XXX,XX @@ static uint32_t imx25_ccm_get_ahb_clk(IMXCCMState *dev)
47
freq = imx25_ccm_get_mcu_clk(dev)
48
/ (1 + EXTRACT(s->reg[IMX25_CCM_CCTL_REG], AHB_CLK_DIV));
49
50
- DPRINTF("freq = %d\n", freq);
51
+ DPRINTF("freq = %u\n", freq);
52
53
return freq;
54
}
55
@@ -XXX,XX +XXX,XX @@ static uint32_t imx25_ccm_get_ipg_clk(IMXCCMState *dev)
56
57
freq = imx25_ccm_get_ahb_clk(dev) / 2;
58
59
- DPRINTF("freq = %d\n", freq);
60
+ DPRINTF("freq = %u\n", freq);
61
62
return freq;
63
}
64
@@ -XXX,XX +XXX,XX @@ static uint32_t imx25_ccm_get_clock_frequency(IMXCCMState *dev, IMXClk clock)
65
break;
201
}
66
}
202
67
203
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_tag(DisasContext *s, uint32_t insn)
68
- DPRINTF("Clock = %d) = %d\n", clock, freq);
204
tcg_gen_addi_i64(addr, addr, offset);
69
+ DPRINTF("Clock = %d) = %u\n", clock, freq);
205
}
70
206
71
return freq;
207
+ if (is_mult) {
72
}
208
+ tcg_rt = cpu_reg(s, rt);
209
+
210
+ if (is_zero) {
211
+ int size = 4 << s->dcz_blocksize;
212
+
213
+ if (s->ata) {
214
+ gen_helper_stzgm_tags(cpu_env, addr, tcg_rt);
215
+ }
216
+ /*
217
+ * The non-tags portion of STZGM is mostly like DC_ZVA,
218
+ * except the alignment happens before the access.
219
+ */
220
+ clean_addr = clean_data_tbi(s, addr);
221
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
222
+ gen_helper_dc_zva(cpu_env, clean_addr);
223
+ } else if (s->ata) {
224
+ if (is_load) {
225
+ gen_helper_ldgm(tcg_rt, cpu_env, addr);
226
+ } else {
227
+ gen_helper_stgm(cpu_env, addr, tcg_rt);
228
+ }
229
+ } else {
230
+ MMUAccessType acc = is_load ? MMU_DATA_LOAD : MMU_DATA_STORE;
231
+ int size = 4 << GMID_EL1_BS;
232
+
233
+ clean_addr = clean_data_tbi(s, addr);
234
+ tcg_gen_andi_i64(clean_addr, clean_addr, -size);
235
+ gen_probe_access(s, clean_addr, acc, size);
236
+
237
+ if (is_load) {
238
+ /* The result tags are zeros. */
239
+ tcg_gen_movi_i64(tcg_rt, 0);
240
+ }
241
+ }
242
+ return;
243
+ }
244
+
245
if (is_load) {
246
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
247
tcg_rt = cpu_reg(s, rt);
248
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
249
dc->vec_stride = 0;
250
dc->cp_regs = arm_cpu->cp_regs;
251
dc->features = env->features;
252
+ dc->dcz_blocksize = arm_cpu->dcz_blocksize;
253
254
/* Single step state. The code-generation logic here is:
255
* SS_ACTIVE == 0:
256
--
73
--
257
2.20.1
74
2.20.1
258
75
259
76
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
From: Alex Chen <alex.chen@huawei.com>
2
2
3
We will shortly need this in mte_helper.c as well.
3
We should use printf format specifier "%u" instead of "%d" for
4
argument of type "unsigned int".
4
5
6
Reported-by: Euler Robot <euler.robot@huawei.com>
7
Signed-off-by: Alex Chen <alex.chen@huawei.com>
8
Message-id: 20201126111109.112238-3-alex.chen@huawei.com
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-23-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
11
---
10
target/arm/internals.h | 9 +++++++++
12
hw/misc/imx31_ccm.c | 14 +++++++-------
11
target/arm/helper.c | 9 ---------
13
hw/misc/imx_ccm.c | 4 ++--
12
2 files changed, 9 insertions(+), 9 deletions(-)
14
2 files changed, 9 insertions(+), 9 deletions(-)
13
15
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
16
diff --git a/hw/misc/imx31_ccm.c b/hw/misc/imx31_ccm.c
15
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
18
--- a/hw/misc/imx31_ccm.c
17
+++ b/target/arm/internals.h
19
+++ b/hw/misc/imx31_ccm.c
18
@@ -XXX,XX +XXX,XX @@ static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
20
@@ -XXX,XX +XXX,XX @@ static const char *imx31_ccm_reg_name(uint32_t reg)
21
case IMX31_CCM_PDR2_REG:
22
return "PDR2";
23
default:
24
- sprintf(unknown, "[%d ?]", reg);
25
+ sprintf(unknown, "[%u ?]", reg);
26
return unknown;
19
}
27
}
20
}
28
}
21
29
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_pll_ref_clk(IMXCCMState *dev)
22
+/* Return the TCR controlling this translation regime */
30
freq = CKIH_FREQ;
23
+static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
31
}
24
+{
32
25
+ if (mmu_idx == ARMMMUIdx_Stage2) {
33
- DPRINTF("freq = %d\n", freq);
26
+ return &env->cp15.vtcr_el2;
34
+ DPRINTF("freq = %u\n", freq);
27
+ }
35
28
+ return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
36
return freq;
29
+}
37
}
30
+
38
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_mpll_clk(IMXCCMState *dev)
31
/* Return the FSR value for a debug exception (watchpoint, hardware
39
freq = imx_ccm_calc_pll(s->reg[IMX31_CCM_MPCTL_REG],
32
* breakpoint or BKPT insn) targeting the specified exception level.
40
imx31_ccm_get_pll_ref_clk(dev));
33
*/
41
34
diff --git a/target/arm/helper.c b/target/arm/helper.c
42
- DPRINTF("freq = %d\n", freq);
43
+ DPRINTF("freq = %u\n", freq);
44
45
return freq;
46
}
47
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_mcu_main_clk(IMXCCMState *dev)
48
freq = imx31_ccm_get_mpll_clk(dev);
49
}
50
51
- DPRINTF("freq = %d\n", freq);
52
+ DPRINTF("freq = %u\n", freq);
53
54
return freq;
55
}
56
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_hclk_clk(IMXCCMState *dev)
57
freq = imx31_ccm_get_mcu_main_clk(dev)
58
/ (1 + EXTRACT(s->reg[IMX31_CCM_PDR0_REG], MAX));
59
60
- DPRINTF("freq = %d\n", freq);
61
+ DPRINTF("freq = %u\n", freq);
62
63
return freq;
64
}
65
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_ipg_clk(IMXCCMState *dev)
66
freq = imx31_ccm_get_hclk_clk(dev)
67
/ (1 + EXTRACT(s->reg[IMX31_CCM_PDR0_REG], IPG));
68
69
- DPRINTF("freq = %d\n", freq);
70
+ DPRINTF("freq = %u\n", freq);
71
72
return freq;
73
}
74
@@ -XXX,XX +XXX,XX @@ static uint32_t imx31_ccm_get_clock_frequency(IMXCCMState *dev, IMXClk clock)
75
break;
76
}
77
78
- DPRINTF("Clock = %d) = %d\n", clock, freq);
79
+ DPRINTF("Clock = %d) = %u\n", clock, freq);
80
81
return freq;
82
}
83
diff --git a/hw/misc/imx_ccm.c b/hw/misc/imx_ccm.c
35
index XXXXXXX..XXXXXXX 100644
84
index XXXXXXX..XXXXXXX 100644
36
--- a/target/arm/helper.c
85
--- a/hw/misc/imx_ccm.c
37
+++ b/target/arm/helper.c
86
+++ b/hw/misc/imx_ccm.c
38
@@ -XXX,XX +XXX,XX @@ static inline uint64_t regime_ttbr(CPUARMState *env, ARMMMUIdx mmu_idx,
87
@@ -XXX,XX +XXX,XX @@ uint32_t imx_ccm_get_clock_frequency(IMXCCMState *dev, IMXClk clock)
39
88
freq = klass->get_clock_frequency(dev, clock);
40
#endif /* !CONFIG_USER_ONLY */
89
}
41
90
42
-/* Return the TCR controlling this translation regime */
91
- DPRINTF("(clock = %d) = %d\n", clock, freq);
43
-static inline TCR *regime_tcr(CPUARMState *env, ARMMMUIdx mmu_idx)
92
+ DPRINTF("(clock = %d) = %u\n", clock, freq);
44
-{
93
45
- if (mmu_idx == ARMMMUIdx_Stage2) {
94
return freq;
46
- return &env->cp15.vtcr_el2;
95
}
47
- }
96
@@ -XXX,XX +XXX,XX @@ uint32_t imx_ccm_calc_pll(uint32_t pllreg, uint32_t base_freq)
48
- return &env->cp15.tcr_el[regime_el(env, mmu_idx)];
97
freq = ((2 * (base_freq >> 10) * (mfi * mfd + mfn)) /
49
-}
98
(mfd * pd)) << 10;
50
-
99
51
/* Convert a possible stage1+2 MMU index into the appropriate
100
- DPRINTF("(pllreg = 0x%08x, base_freq = %d) = %d\n", pllreg, base_freq,
52
* stage 1 MMU index
101
+ DPRINTF("(pllreg = 0x%08x, base_freq = %u) = %d\n", pllreg, base_freq,
53
*/
102
freq);
103
104
return freq;
54
--
105
--
55
2.20.1
106
2.20.1
56
107
57
108
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
From: Alex Chen <alex.chen@huawei.com>
2
2
3
Various code from the PCA9552 device model is generic to the
3
We should use printf format specifier "%u" instead of "%d" for
4
PCA955X family. We'll split the generic code in a base class
4
argument of type "unsigned int".
5
in the next commit. To ease review, first do a dumb renaming.
6
5
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Reported-by: Euler Robot <euler.robot@huawei.com>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Signed-off-by: Alex Chen <alex.chen@huawei.com>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20201126111109.112238-4-alex.chen@huawei.com
10
Message-id: 20200623072723.6324-4-f4bug@amsat.org
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
11
---
13
include/hw/misc/pca9552.h | 10 ++---
12
hw/misc/imx6_ccm.c | 20 ++++++++++----------
14
hw/misc/pca9552.c | 80 +++++++++++++++++++--------------------
13
hw/misc/imx6_src.c | 2 +-
15
2 files changed, 45 insertions(+), 45 deletions(-)
14
2 files changed, 11 insertions(+), 11 deletions(-)
16
15
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
16
diff --git a/hw/misc/imx6_ccm.c b/hw/misc/imx6_ccm.c
18
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
18
--- a/hw/misc/imx6_ccm.c
20
+++ b/include/hw/misc/pca9552.h
19
+++ b/hw/misc/imx6_ccm.c
21
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ static const char *imx6_ccm_reg_name(uint32_t reg)
22
#include "hw/i2c/i2c.h"
21
case CCM_CMEOR:
23
22
return "CMEOR";
24
#define TYPE_PCA9552 "pca9552"
23
default:
25
-#define PCA9552(obj) OBJECT_CHECK(PCA9552State, (obj), TYPE_PCA9552)
24
- sprintf(unknown, "%d ?", reg);
26
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
25
+ sprintf(unknown, "%u ?", reg);
27
26
return unknown;
28
-#define PCA9552_NR_REGS 10
29
+#define PCA955X_NR_REGS 10
30
31
-typedef struct PCA9552State {
32
+typedef struct PCA955xState {
33
/*< private >*/
34
I2CSlave i2c;
35
/*< public >*/
36
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
37
uint8_t len;
38
uint8_t pointer;
39
40
- uint8_t regs[PCA9552_NR_REGS];
41
+ uint8_t regs[PCA955X_NR_REGS];
42
uint8_t max_reg;
43
uint8_t pin_count;
44
-} PCA9552State;
45
+} PCA955xState;
46
47
#endif
48
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/hw/misc/pca9552.c
51
+++ b/hw/misc/pca9552.c
52
@@ -XXX,XX +XXX,XX @@
53
54
static const char *led_state[] = {"on", "off", "pwm0", "pwm1"};
55
56
-static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
57
+static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
58
{
59
uint8_t reg = PCA9552_LS0 + (pin / 4);
60
uint8_t shift = (pin % 4) << 1;
61
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_pin_get_config(PCA9552State *s, int pin)
62
return extract32(s->regs[reg], shift, 2);
63
}
64
65
-static void pca9552_update_pin_input(PCA9552State *s)
66
+static void pca955x_update_pin_input(PCA955xState *s)
67
{
68
int i;
69
70
for (i = 0; i < s->pin_count; i++) {
71
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
72
uint8_t input_shift = (i % 8);
73
- uint8_t config = pca9552_pin_get_config(s, i);
74
+ uint8_t config = pca955x_pin_get_config(s, i);
75
76
switch (config) {
77
case PCA9552_LED_ON:
78
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
79
}
27
}
80
}
28
}
81
29
@@ -XXX,XX +XXX,XX @@ static const char *imx6_analog_reg_name(uint32_t reg)
82
-static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
30
case USB_ANALOG_DIGPROG:
83
+static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
31
return "USB_ANALOG_DIGPROG";
84
{
32
default:
85
switch (reg) {
33
- sprintf(unknown, "%d ?", reg);
86
case PCA9552_INPUT0:
34
+ sprintf(unknown, "%u ?", reg);
87
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_read(PCA9552State *s, uint8_t reg)
35
return unknown;
88
}
36
}
89
}
37
}
90
38
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_analog_get_pll2_clk(IMX6CCMState *dev)
91
-static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
39
freq *= 20;
92
+static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
40
}
93
{
41
94
switch (reg) {
42
- DPRINTF("freq = %d\n", (uint32_t)freq);
95
case PCA9552_PSC0:
43
+ DPRINTF("freq = %u\n", (uint32_t)freq);
96
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
44
97
case PCA9552_LS2:
45
return freq;
98
case PCA9552_LS3:
46
}
99
s->regs[reg] = data;
47
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_analog_get_pll2_pfd0_clk(IMX6CCMState *dev)
100
- pca9552_update_pin_input(s);
48
freq = imx6_analog_get_pll2_clk(dev) * 18
101
+ pca955x_update_pin_input(s);
49
/ EXTRACT(dev->analog[CCM_ANALOG_PFD_528], PFD0_FRAC);
50
51
- DPRINTF("freq = %d\n", (uint32_t)freq);
52
+ DPRINTF("freq = %u\n", (uint32_t)freq);
53
54
return freq;
55
}
56
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_analog_get_pll2_pfd2_clk(IMX6CCMState *dev)
57
freq = imx6_analog_get_pll2_clk(dev) * 18
58
/ EXTRACT(dev->analog[CCM_ANALOG_PFD_528], PFD2_FRAC);
59
60
- DPRINTF("freq = %d\n", (uint32_t)freq);
61
+ DPRINTF("freq = %u\n", (uint32_t)freq);
62
63
return freq;
64
}
65
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_analog_get_periph_clk(IMX6CCMState *dev)
102
break;
66
break;
103
67
}
104
case PCA9552_INPUT0:
68
105
@@ -XXX,XX +XXX,XX @@ static void pca9552_write(PCA9552State *s, uint8_t reg, uint8_t data)
69
- DPRINTF("freq = %d\n", (uint32_t)freq);
106
* after each byte is sent to or received by the device. The index
70
+ DPRINTF("freq = %u\n", (uint32_t)freq);
107
* rollovers to 0 when the maximum register address is reached.
71
108
*/
72
return freq;
109
-static void pca9552_autoinc(PCA9552State *s)
73
}
110
+static void pca955x_autoinc(PCA955xState *s)
74
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_ccm_get_ahb_clk(IMX6CCMState *dev)
111
{
75
freq = imx6_analog_get_periph_clk(dev)
112
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
76
/ (1 + EXTRACT(dev->ccm[CCM_CBCDR], AHB_PODF));
113
uint8_t reg = s->pointer & 0xf;
77
114
@@ -XXX,XX +XXX,XX @@ static void pca9552_autoinc(PCA9552State *s)
78
- DPRINTF("freq = %d\n", (uint32_t)freq);
79
+ DPRINTF("freq = %u\n", (uint32_t)freq);
80
81
return freq;
82
}
83
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_ccm_get_ipg_clk(IMX6CCMState *dev)
84
freq = imx6_ccm_get_ahb_clk(dev)
85
/ (1 + EXTRACT(dev->ccm[CCM_CBCDR], IPG_PODF));
86
87
- DPRINTF("freq = %d\n", (uint32_t)freq);
88
+ DPRINTF("freq = %u\n", (uint32_t)freq);
89
90
return freq;
91
}
92
@@ -XXX,XX +XXX,XX @@ static uint64_t imx6_ccm_get_per_clk(IMX6CCMState *dev)
93
freq = imx6_ccm_get_ipg_clk(dev)
94
/ (1 + EXTRACT(dev->ccm[CCM_CSCMR1], PERCLK_PODF));
95
96
- DPRINTF("freq = %d\n", (uint32_t)freq);
97
+ DPRINTF("freq = %u\n", (uint32_t)freq);
98
99
return freq;
100
}
101
@@ -XXX,XX +XXX,XX @@ static uint32_t imx6_ccm_get_clock_frequency(IMXCCMState *dev, IMXClk clock)
102
break;
103
}
104
105
- DPRINTF("Clock = %d) = %d\n", clock, freq);
106
+ DPRINTF("Clock = %d) = %u\n", clock, freq);
107
108
return freq;
109
}
110
diff --git a/hw/misc/imx6_src.c b/hw/misc/imx6_src.c
111
index XXXXXXX..XXXXXXX 100644
112
--- a/hw/misc/imx6_src.c
113
+++ b/hw/misc/imx6_src.c
114
@@ -XXX,XX +XXX,XX @@ static const char *imx6_src_reg_name(uint32_t reg)
115
case SRC_GPR10:
116
return "SRC_GPR10";
117
default:
118
- sprintf(unknown, "%d ?", reg);
119
+ sprintf(unknown, "%u ?", reg);
120
return unknown;
115
}
121
}
116
}
122
}
117
118
-static uint8_t pca9552_recv(I2CSlave *i2c)
119
+static uint8_t pca955x_recv(I2CSlave *i2c)
120
{
121
- PCA9552State *s = PCA9552(i2c);
122
+ PCA955xState *s = PCA955X(i2c);
123
uint8_t ret;
124
125
- ret = pca9552_read(s, s->pointer & 0xf);
126
+ ret = pca955x_read(s, s->pointer & 0xf);
127
128
/*
129
* From the Specs:
130
@@ -XXX,XX +XXX,XX @@ static uint8_t pca9552_recv(I2CSlave *i2c)
131
__func__);
132
}
133
134
- pca9552_autoinc(s);
135
+ pca955x_autoinc(s);
136
137
return ret;
138
}
139
140
-static int pca9552_send(I2CSlave *i2c, uint8_t data)
141
+static int pca955x_send(I2CSlave *i2c, uint8_t data)
142
{
143
- PCA9552State *s = PCA9552(i2c);
144
+ PCA955xState *s = PCA955X(i2c);
145
146
/* First byte sent by is the register address */
147
if (s->len == 0) {
148
s->pointer = data;
149
s->len++;
150
} else {
151
- pca9552_write(s, s->pointer & 0xf, data);
152
+ pca955x_write(s, s->pointer & 0xf, data);
153
154
- pca9552_autoinc(s);
155
+ pca955x_autoinc(s);
156
}
157
158
return 0;
159
}
160
161
-static int pca9552_event(I2CSlave *i2c, enum i2c_event event)
162
+static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
163
{
164
- PCA9552State *s = PCA9552(i2c);
165
+ PCA955xState *s = PCA955X(i2c);
166
167
s->len = 0;
168
return 0;
169
}
170
171
-static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
172
+static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
173
void *opaque, Error **errp)
174
{
175
- PCA9552State *s = PCA9552(obj);
176
+ PCA955xState *s = PCA955X(obj);
177
int led, rc, reg;
178
uint8_t state;
179
180
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
181
* reading the INPUTx reg
182
*/
183
reg = PCA9552_LS0 + led / 4;
184
- state = (pca9552_read(s, reg) >> (led % 8)) & 0x3;
185
+ state = (pca955x_read(s, reg) >> (led % 8)) & 0x3;
186
visit_type_str(v, name, (char **)&led_state[state], errp);
187
}
188
189
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
190
((state & 0x3) << (led_num << 1));
191
}
192
193
-static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
194
+static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
195
void *opaque, Error **errp)
196
{
197
- PCA9552State *s = PCA9552(obj);
198
+ PCA955xState *s = PCA955X(obj);
199
Error *local_err = NULL;
200
int led, rc, reg, val;
201
uint8_t state;
202
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
203
}
204
205
reg = PCA9552_LS0 + led / 4;
206
- val = pca9552_read(s, reg);
207
+ val = pca955x_read(s, reg);
208
val = pca955x_ledsel(val, led % 4, state);
209
- pca9552_write(s, reg, val);
210
+ pca955x_write(s, reg, val);
211
}
212
213
static const VMStateDescription pca9552_vmstate = {
214
@@ -XXX,XX +XXX,XX @@ static const VMStateDescription pca9552_vmstate = {
215
.version_id = 0,
216
.minimum_version_id = 0,
217
.fields = (VMStateField[]) {
218
- VMSTATE_UINT8(len, PCA9552State),
219
- VMSTATE_UINT8(pointer, PCA9552State),
220
- VMSTATE_UINT8_ARRAY(regs, PCA9552State, PCA9552_NR_REGS),
221
- VMSTATE_I2C_SLAVE(i2c, PCA9552State),
222
+ VMSTATE_UINT8(len, PCA955xState),
223
+ VMSTATE_UINT8(pointer, PCA955xState),
224
+ VMSTATE_UINT8_ARRAY(regs, PCA955xState, PCA955X_NR_REGS),
225
+ VMSTATE_I2C_SLAVE(i2c, PCA955xState),
226
VMSTATE_END_OF_LIST()
227
}
228
};
229
230
static void pca9552_reset(DeviceState *dev)
231
{
232
- PCA9552State *s = PCA9552(dev);
233
+ PCA955xState *s = PCA955X(dev);
234
235
s->regs[PCA9552_PSC0] = 0xFF;
236
s->regs[PCA9552_PWM0] = 0x80;
237
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
238
s->regs[PCA9552_LS2] = 0x55;
239
s->regs[PCA9552_LS3] = 0x55;
240
241
- pca9552_update_pin_input(s);
242
+ pca955x_update_pin_input(s);
243
244
s->pointer = 0xFF;
245
s->len = 0;
246
}
247
248
-static void pca9552_initfn(Object *obj)
249
+static void pca955x_initfn(Object *obj)
250
{
251
- PCA9552State *s = PCA9552(obj);
252
+ PCA955xState *s = PCA955X(obj);
253
int led;
254
255
/* If support for the other PCA955X devices are implemented, these
256
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
257
char *name;
258
259
name = g_strdup_printf("led%d", led);
260
- object_property_add(obj, name, "bool", pca9552_get_led, pca9552_set_led,
261
+ object_property_add(obj, name, "bool", pca955x_get_led, pca955x_set_led,
262
NULL, NULL);
263
g_free(name);
264
}
265
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
266
DeviceClass *dc = DEVICE_CLASS(klass);
267
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
268
269
- k->event = pca9552_event;
270
- k->recv = pca9552_recv;
271
- k->send = pca9552_send;
272
+ k->event = pca955x_event;
273
+ k->recv = pca955x_recv;
274
+ k->send = pca955x_send;
275
dc->reset = pca9552_reset;
276
dc->vmsd = &pca9552_vmstate;
277
}
278
@@ -XXX,XX +XXX,XX @@ static void pca9552_class_init(ObjectClass *klass, void *data)
279
static const TypeInfo pca9552_info = {
280
.name = TYPE_PCA9552,
281
.parent = TYPE_I2C_SLAVE,
282
- .instance_init = pca9552_initfn,
283
- .instance_size = sizeof(PCA9552State),
284
+ .instance_init = pca955x_initfn,
285
+ .instance_size = sizeof(PCA955xState),
286
.class_init = pca9552_class_init,
287
};
288
289
-static void pca9552_register_types(void)
290
+static void pca955x_register_types(void)
291
{
292
type_register_static(&pca9552_info);
293
}
294
295
-type_init(pca9552_register_types)
296
+type_init(pca955x_register_types)
297
--
123
--
298
2.20.1
124
2.20.1
299
125
300
126
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
From: Alex Chen <alex.chen@huawei.com>
2
2
3
Extract the code common to the PCA955x family in PCA955xClass,
3
We should use printf format specifier "%u" instead of "%d" for
4
keeping the PCA9552 specific parts into pca9552_class_init().
4
argument of type "unsigned int".
5
Remove the 'TODO' comment added in commit 5141d4158cf.
6
5
7
Suggested-by: Cédric Le Goater <clg@kaod.org>
6
Reported-by: Euler Robot <euler.robot@huawei.com>
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
7
Signed-off-by: Alex Chen <alex.chen@huawei.com>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
8
Message-id: 20201126111109.112238-5-alex.chen@huawei.com
10
Tested-by: Cédric Le Goater <clg@kaod.org>
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Message-id: 20200623072723.6324-5-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
11
---
14
include/hw/misc/pca9552.h | 6 ++--
12
hw/misc/imx6ul_ccm.c | 4 ++--
15
hw/misc/pca9552.c | 66 ++++++++++++++++++++++++++++-----------
13
1 file changed, 2 insertions(+), 2 deletions(-)
16
2 files changed, 51 insertions(+), 21 deletions(-)
17
14
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
15
diff --git a/hw/misc/imx6ul_ccm.c b/hw/misc/imx6ul_ccm.c
19
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
17
--- a/hw/misc/imx6ul_ccm.c
21
+++ b/include/hw/misc/pca9552.h
18
+++ b/hw/misc/imx6ul_ccm.c
22
@@ -XXX,XX +XXX,XX @@
19
@@ -XXX,XX +XXX,XX @@ static const char *imx6ul_ccm_reg_name(uint32_t reg)
23
#include "hw/i2c/i2c.h"
20
case CCM_CMEOR:
24
21
return "CMEOR";
25
#define TYPE_PCA9552 "pca9552"
22
default:
26
-#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA9552)
23
- sprintf(unknown, "%d ?", reg);
27
+#define TYPE_PCA955X "pca955x"
24
+ sprintf(unknown, "%u ?", reg);
28
+#define PCA955X(obj) OBJECT_CHECK(PCA955xState, (obj), TYPE_PCA955X)
25
return unknown;
29
30
#define PCA955X_NR_REGS 10
31
+#define PCA955X_PIN_COUNT_MAX 16
32
33
typedef struct PCA955xState {
34
/*< private >*/
35
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
36
uint8_t pointer;
37
38
uint8_t regs[PCA955X_NR_REGS];
39
- uint8_t max_reg;
40
- uint8_t pin_count;
41
} PCA955xState;
42
43
#endif
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
47
+++ b/hw/misc/pca9552.c
48
@@ -XXX,XX +XXX,XX @@
49
* https://www.nxp.com/docs/en/application-note/AN264.pdf
50
*
51
* Copyright (c) 2017-2018, IBM Corporation.
52
+ * Copyright (c) 2020 Philippe Mathieu-Daudé
53
*
54
* This work is licensed under the terms of the GNU GPL, version 2 or
55
* later. See the COPYING file in the top-level directory.
56
@@ -XXX,XX +XXX,XX @@
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
60
+typedef struct PCA955xClass {
61
+ /*< private >*/
62
+ I2CSlaveClass parent_class;
63
+ /*< public >*/
64
+
65
+ uint8_t pin_count;
66
+ uint8_t max_reg;
67
+} PCA955xClass;
68
+
69
+#define PCA955X_CLASS(klass) \
70
+ OBJECT_CLASS_CHECK(PCA955xClass, (klass), TYPE_PCA955X)
71
+#define PCA955X_GET_CLASS(obj) \
72
+ OBJECT_GET_CLASS(PCA955xClass, (obj), TYPE_PCA955X)
73
+
74
#define PCA9552_LED_ON 0x0
75
#define PCA9552_LED_OFF 0x1
76
#define PCA9552_LED_PWM0 0x2
77
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
78
79
static void pca955x_update_pin_input(PCA955xState *s)
80
{
81
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
82
int i;
83
84
- for (i = 0; i < s->pin_count; i++) {
85
+ for (i = 0; i < k->pin_count; i++) {
86
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
87
uint8_t input_shift = (i % 8);
88
uint8_t config = pca955x_pin_get_config(s, i);
89
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
90
*/
91
static void pca955x_autoinc(PCA955xState *s)
92
{
93
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
94
+
95
if (s->pointer != 0xFF && s->pointer & PCA9552_AUTOINC) {
96
uint8_t reg = s->pointer & 0xf;
97
98
- reg = (reg + 1) % (s->max_reg + 1);
99
+ reg = (reg + 1) % (k->max_reg + 1);
100
s->pointer = reg | PCA9552_AUTOINC;
101
}
26
}
102
}
27
}
103
@@ -XXX,XX +XXX,XX @@ static int pca955x_event(I2CSlave *i2c, enum i2c_event event)
28
@@ -XXX,XX +XXX,XX @@ static const char *imx6ul_analog_reg_name(uint32_t reg)
104
static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
29
case USB_ANALOG_DIGPROG:
105
void *opaque, Error **errp)
30
return "USB_ANALOG_DIGPROG";
106
{
31
default:
107
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
32
- sprintf(unknown, "%d ?", reg);
108
PCA955xState *s = PCA955X(obj);
33
+ sprintf(unknown, "%u ?", reg);
109
int led, rc, reg;
34
return unknown;
110
uint8_t state;
111
@@ -XXX,XX +XXX,XX @@ static void pca955x_get_led(Object *obj, Visitor *v, const char *name,
112
error_setg(errp, "%s: error reading %s", __func__, name);
113
return;
114
}
115
- if (led < 0 || led > s->pin_count) {
116
+ if (led < 0 || led > k->pin_count) {
117
error_setg(errp, "%s invalid led %s", __func__, name);
118
return;
119
}
120
@@ -XXX,XX +XXX,XX @@ static inline uint8_t pca955x_ledsel(uint8_t oldval, int led_num, int state)
121
static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
122
void *opaque, Error **errp)
123
{
124
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
125
PCA955xState *s = PCA955X(obj);
126
Error *local_err = NULL;
127
int led, rc, reg, val;
128
@@ -XXX,XX +XXX,XX @@ static void pca955x_set_led(Object *obj, Visitor *v, const char *name,
129
error_setg(errp, "%s: error reading %s", __func__, name);
130
return;
131
}
132
- if (led < 0 || led > s->pin_count) {
133
+ if (led < 0 || led > k->pin_count) {
134
error_setg(errp, "%s invalid led %s", __func__, name);
135
return;
136
}
137
@@ -XXX,XX +XXX,XX @@ static void pca9552_reset(DeviceState *dev)
138
139
static void pca955x_initfn(Object *obj)
140
{
141
- PCA955xState *s = PCA955X(obj);
142
+ PCA955xClass *k = PCA955X_GET_CLASS(obj);
143
int led;
144
145
- /* If support for the other PCA955X devices are implemented, these
146
- * constant values might be part of class structure describing the
147
- * PCA955X device
148
- */
149
- s->max_reg = PCA9552_LS3;
150
- s->pin_count = 16;
151
-
152
- for (led = 0; led < s->pin_count; led++) {
153
+ assert(k->pin_count <= PCA955X_PIN_COUNT_MAX);
154
+ for (led = 0; led < k->pin_count; led++) {
155
char *name;
156
157
name = g_strdup_printf("led%d", led);
158
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
159
}
35
}
160
}
36
}
161
162
-static void pca9552_class_init(ObjectClass *klass, void *data)
163
+static void pca955x_class_init(ObjectClass *klass, void *data)
164
{
165
- DeviceClass *dc = DEVICE_CLASS(klass);
166
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
167
168
k->event = pca955x_event;
169
k->recv = pca955x_recv;
170
k->send = pca955x_send;
171
+}
172
+
173
+static const TypeInfo pca955x_info = {
174
+ .name = TYPE_PCA955X,
175
+ .parent = TYPE_I2C_SLAVE,
176
+ .instance_init = pca955x_initfn,
177
+ .instance_size = sizeof(PCA955xState),
178
+ .class_init = pca955x_class_init,
179
+ .abstract = true,
180
+};
181
+
182
+static void pca9552_class_init(ObjectClass *oc, void *data)
183
+{
184
+ DeviceClass *dc = DEVICE_CLASS(oc);
185
+ PCA955xClass *pc = PCA955X_CLASS(oc);
186
+
187
dc->reset = pca9552_reset;
188
dc->vmsd = &pca9552_vmstate;
189
+ pc->max_reg = PCA9552_LS3;
190
+ pc->pin_count = 16;
191
}
192
193
static const TypeInfo pca9552_info = {
194
.name = TYPE_PCA9552,
195
- .parent = TYPE_I2C_SLAVE,
196
- .instance_init = pca955x_initfn,
197
- .instance_size = sizeof(PCA955xState),
198
+ .parent = TYPE_PCA955X,
199
.class_init = pca9552_class_init,
200
};
201
202
static void pca955x_register_types(void)
203
{
204
+ type_register_static(&pca955x_info);
205
type_register_static(&pca9552_info);
206
}
207
208
--
37
--
209
2.20.1
38
2.20.1
210
39
211
40
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
For M-profile CPUs, the range from 0xe0000000 to 0xe00fffff is the
2
Private Peripheral Bus range, which includes all of the memory mapped
3
devices and registers that are part of the CPU itself, including the
4
NVIC, systick timer, and debug and trace components like the Data
5
Watchpoint and Trace unit (DWT). Within this large region, the range
6
0xe000e000 to 0xe000efff is the System Control Space (NVIC, system
7
registers, systick) and 0xe002e000 to 0exe002efff is its Non-secure
8
alias.
2
9
3
Extract i2c_try_create_slave() and i2c_realize_and_unref()
10
The architecture is clear that within the SCS unimplemented registers
4
from i2c_create_slave().
11
should be RES0 for privileged accesses and generate BusFault for
5
We can now set properties on a I2CSlave before it is realized.
12
unprivileged accesses, and we currently implement this.
6
13
7
This is in line with the recent qdev/QOM changes merged
14
It is less clear about how to handle accesses to unimplemented
8
in commit 6675a653d2e.
15
regions of the wider PPB. Unprivileged accesses should definitely
16
cause BusFaults (R_DQQS), but the behaviour of privileged accesses is
17
not given as a general rule. However, the register definitions of
18
individual registers for components like the DWT all state that they
19
are RES0 if the relevant component is not implemented, so the
20
simplest way to provide that is to provide RAZ/WI for the whole range
21
for privileged accesses. (The v7M Arm ARM does say that reserved
22
registers should be UNK/SBZP.)
9
23
10
Reviewed-by: Corey Minyard <cminyard@mvista.com>
24
Expand the container MemoryRegion that the NVIC exposes so that
11
Reviewed-by: Cédric Le Goater <clg@kaod.org>
25
it covers the whole PPB space. This means:
12
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
26
* moving the address that the ARMV7M device maps it to down by
13
Reviewed-by: Markus Armbruster <armbru@redhat.com>
27
0xe000 bytes
14
Tested-by: Cédric Le Goater <clg@kaod.org>
28
* moving the off and the offsets within the container of all the
15
Message-id: 20200623072723.6324-2-f4bug@amsat.org
29
subregions forward by 0xe000 bytes
30
* adding a new default MemoryRegion that covers the whole container
31
at a lower priority than anything else and which provides the
32
RAZWI/BusFault behaviour
33
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
34
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
35
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
36
Message-id: 20201119215617.29887-2-peter.maydell@linaro.org
17
---
37
---
18
include/hw/i2c/i2c.h | 2 ++
38
include/hw/intc/armv7m_nvic.h | 1 +
19
hw/i2c/core.c | 18 ++++++++++++++++--
39
hw/arm/armv7m.c | 2 +-
20
2 files changed, 18 insertions(+), 2 deletions(-)
40
hw/intc/armv7m_nvic.c | 78 ++++++++++++++++++++++++++++++-----
41
3 files changed, 69 insertions(+), 12 deletions(-)
21
42
22
diff --git a/include/hw/i2c/i2c.h b/include/hw/i2c/i2c.h
43
diff --git a/include/hw/intc/armv7m_nvic.h b/include/hw/intc/armv7m_nvic.h
23
index XXXXXXX..XXXXXXX 100644
44
index XXXXXXX..XXXXXXX 100644
24
--- a/include/hw/i2c/i2c.h
45
--- a/include/hw/intc/armv7m_nvic.h
25
+++ b/include/hw/i2c/i2c.h
46
+++ b/include/hw/intc/armv7m_nvic.h
26
@@ -XXX,XX +XXX,XX @@ int i2c_send(I2CBus *bus, uint8_t data);
47
@@ -XXX,XX +XXX,XX @@ struct NVICState {
27
uint8_t i2c_recv(I2CBus *bus);
48
MemoryRegion systickmem;
28
49
MemoryRegion systick_ns_mem;
29
DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr);
50
MemoryRegion container;
30
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr);
51
+ MemoryRegion defaultmem;
31
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp);
52
32
53
uint32_t num_irq;
33
/* lm832x.c */
54
qemu_irq excpout;
34
void lm832x_key_event(DeviceState *dev, int key, int state);
55
diff --git a/hw/arm/armv7m.c b/hw/arm/armv7m.c
35
diff --git a/hw/i2c/core.c b/hw/i2c/core.c
36
index XXXXXXX..XXXXXXX 100644
56
index XXXXXXX..XXXXXXX 100644
37
--- a/hw/i2c/core.c
57
--- a/hw/arm/armv7m.c
38
+++ b/hw/i2c/core.c
58
+++ b/hw/arm/armv7m.c
39
@@ -XXX,XX +XXX,XX @@ const VMStateDescription vmstate_i2c_slave = {
59
@@ -XXX,XX +XXX,XX @@ static void armv7m_realize(DeviceState *dev, Error **errp)
40
}
60
sysbus_connect_irq(sbd, 0,
61
qdev_get_gpio_in(DEVICE(s->cpu), ARM_CPU_IRQ));
62
63
- memory_region_add_subregion(&s->container, 0xe000e000,
64
+ memory_region_add_subregion(&s->container, 0xe0000000,
65
sysbus_mmio_get_region(sbd, 0));
66
67
for (i = 0; i < ARRAY_SIZE(s->bitband); i++) {
68
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
69
index XXXXXXX..XXXXXXX 100644
70
--- a/hw/intc/armv7m_nvic.c
71
+++ b/hw/intc/armv7m_nvic.c
72
@@ -XXX,XX +XXX,XX @@ static const MemoryRegionOps nvic_systick_ops = {
73
.endianness = DEVICE_NATIVE_ENDIAN,
41
};
74
};
42
75
43
-DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
76
+/*
44
+DeviceState *i2c_try_create_slave(const char *name, uint8_t addr)
77
+ * Unassigned portions of the PPB space are RAZ/WI for privileged
45
{
78
+ * accesses, and fault for non-privileged accesses.
46
DeviceState *dev;
79
+ */
47
80
+static MemTxResult ppb_default_read(void *opaque, hwaddr addr,
48
dev = qdev_new(name);
81
+ uint64_t *data, unsigned size,
49
qdev_prop_set_uint8(dev, "address", addr);
82
+ MemTxAttrs attrs)
50
- qdev_realize_and_unref(dev, &bus->qbus, &error_fatal);
83
+{
51
+ return dev;
84
+ qemu_log_mask(LOG_UNIMP, "Read of unassigned area of PPB: offset 0x%x\n",
85
+ (uint32_t)addr);
86
+ if (attrs.user) {
87
+ return MEMTX_ERROR;
88
+ }
89
+ *data = 0;
90
+ return MEMTX_OK;
52
+}
91
+}
53
+
92
+
54
+bool i2c_realize_and_unref(DeviceState *dev, I2CBus *bus, Error **errp)
93
+static MemTxResult ppb_default_write(void *opaque, hwaddr addr,
94
+ uint64_t value, unsigned size,
95
+ MemTxAttrs attrs)
55
+{
96
+{
56
+ return qdev_realize_and_unref(dev, &bus->qbus, errp);
97
+ qemu_log_mask(LOG_UNIMP, "Write of unassigned area of PPB: offset 0x%x\n",
98
+ (uint32_t)addr);
99
+ if (attrs.user) {
100
+ return MEMTX_ERROR;
101
+ }
102
+ return MEMTX_OK;
57
+}
103
+}
58
+
104
+
59
+DeviceState *i2c_create_slave(I2CBus *bus, const char *name, uint8_t addr)
105
+static const MemoryRegionOps ppb_default_ops = {
60
+{
106
+ .read_with_attrs = ppb_default_read,
61
+ DeviceState *dev;
107
+ .write_with_attrs = ppb_default_write,
108
+ .endianness = DEVICE_NATIVE_ENDIAN,
109
+ .valid.min_access_size = 1,
110
+ .valid.max_access_size = 8,
111
+};
62
+
112
+
63
+ dev = i2c_try_create_slave(name, addr);
113
static int nvic_post_load(void *opaque, int version_id)
64
+ i2c_realize_and_unref(dev, bus, &error_fatal);
114
{
65
+
115
NVICState *s = opaque;
66
return dev;
116
@@ -XXX,XX +XXX,XX @@ static void nvic_systick_trigger(void *opaque, int n, int level)
67
}
117
static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
118
{
119
NVICState *s = NVIC(dev);
120
- int regionlen;
121
122
/* The armv7m container object will have set our CPU pointer */
123
if (!s->cpu || !arm_feature(&s->cpu->env, ARM_FEATURE_M)) {
124
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
125
M_REG_S));
126
}
127
128
- /* The NVIC and System Control Space (SCS) starts at 0xe000e000
129
+ /*
130
+ * This device provides a single sysbus memory region which
131
+ * represents the whole of the "System PPB" space. This is the
132
+ * range from 0xe0000000 to 0xe00fffff and includes the NVIC,
133
+ * the System Control Space (system registers), the systick timer,
134
+ * and for CPUs with the Security extension an NS banked version
135
+ * of all of these.
136
+ *
137
+ * The default behaviour for unimplemented registers/ranges
138
+ * (for instance the Data Watchpoint and Trace unit at 0xe0001000)
139
+ * is to RAZ/WI for privileged access and BusFault for non-privileged
140
+ * access.
141
+ *
142
+ * The NVIC and System Control Space (SCS) starts at 0xe000e000
143
* and looks like this:
144
* 0x004 - ICTR
145
* 0x010 - 0xff - systick
146
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
147
* generally code determining which banked register to use should
148
* use attrs.secure; code determining actual behaviour of the system
149
* should use env->v7m.secure.
150
+ *
151
+ * The container covers the whole PPB space. Within it the priority
152
+ * of overlapping regions is:
153
+ * - default region (for RAZ/WI and BusFault) : -1
154
+ * - system register regions : 0
155
+ * - systick : 1
156
+ * This is because the systick device is a small block of registers
157
+ * in the middle of the other system control registers.
158
*/
159
- regionlen = arm_feature(&s->cpu->env, ARM_FEATURE_V8) ? 0x21000 : 0x1000;
160
- memory_region_init(&s->container, OBJECT(s), "nvic", regionlen);
161
- /* The system register region goes at the bottom of the priority
162
- * stack as it covers the whole page.
163
- */
164
+ memory_region_init(&s->container, OBJECT(s), "nvic", 0x100000);
165
+ memory_region_init_io(&s->defaultmem, OBJECT(s), &ppb_default_ops, s,
166
+ "nvic-default", 0x100000);
167
+ memory_region_add_subregion_overlap(&s->container, 0, &s->defaultmem, -1);
168
memory_region_init_io(&s->sysregmem, OBJECT(s), &nvic_sysreg_ops, s,
169
"nvic_sysregs", 0x1000);
170
- memory_region_add_subregion(&s->container, 0, &s->sysregmem);
171
+ memory_region_add_subregion(&s->container, 0xe000, &s->sysregmem);
172
173
memory_region_init_io(&s->systickmem, OBJECT(s),
174
&nvic_systick_ops, s,
175
"nvic_systick", 0xe0);
176
177
- memory_region_add_subregion_overlap(&s->container, 0x10,
178
+ memory_region_add_subregion_overlap(&s->container, 0xe010,
179
&s->systickmem, 1);
180
181
if (arm_feature(&s->cpu->env, ARM_FEATURE_V8)) {
182
memory_region_init_io(&s->sysreg_ns_mem, OBJECT(s),
183
&nvic_sysreg_ns_ops, &s->sysregmem,
184
"nvic_sysregs_ns", 0x1000);
185
- memory_region_add_subregion(&s->container, 0x20000, &s->sysreg_ns_mem);
186
+ memory_region_add_subregion(&s->container, 0x2e000, &s->sysreg_ns_mem);
187
memory_region_init_io(&s->systick_ns_mem, OBJECT(s),
188
&nvic_sysreg_ns_ops, &s->systickmem,
189
"nvic_systick_ns", 0xe0);
190
- memory_region_add_subregion_overlap(&s->container, 0x20010,
191
+ memory_region_add_subregion_overlap(&s->container, 0x2e010,
192
&s->systick_ns_mem, 1);
193
}
68
194
69
--
195
--
70
2.20.1
196
2.20.1
71
197
72
198
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
The PCA9552 device does not expose LEDs, but simple pins
4
to connnect LEDs to. To be clearer with the device model,
5
rename 'nr_leds' as 'pin_count'.
6
7
Reviewed-by: Cédric Le Goater <clg@kaod.org>
8
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
9
Tested-by: Cédric Le Goater <clg@kaod.org>
10
Message-id: 20200623072723.6324-3-f4bug@amsat.org
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
---
13
include/hw/misc/pca9552.h | 2 +-
14
hw/misc/pca9552.c | 10 +++++-----
15
2 files changed, 6 insertions(+), 6 deletions(-)
16
17
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/hw/misc/pca9552.h
20
+++ b/include/hw/misc/pca9552.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct PCA9552State {
22
23
uint8_t regs[PCA9552_NR_REGS];
24
uint8_t max_reg;
25
- uint8_t nr_leds;
26
+ uint8_t pin_count;
27
} PCA9552State;
28
29
#endif
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@ static void pca9552_update_pin_input(PCA9552State *s)
35
{
36
int i;
37
38
- for (i = 0; i < s->nr_leds; i++) {
39
+ for (i = 0; i < s->pin_count; i++) {
40
uint8_t input_reg = PCA9552_INPUT0 + (i / 8);
41
uint8_t input_shift = (i % 8);
42
uint8_t config = pca9552_pin_get_config(s, i);
43
@@ -XXX,XX +XXX,XX @@ static void pca9552_get_led(Object *obj, Visitor *v, const char *name,
44
error_setg(errp, "%s: error reading %s", __func__, name);
45
return;
46
}
47
- if (led < 0 || led > s->nr_leds) {
48
+ if (led < 0 || led > s->pin_count) {
49
error_setg(errp, "%s invalid led %s", __func__, name);
50
return;
51
}
52
@@ -XXX,XX +XXX,XX @@ static void pca9552_set_led(Object *obj, Visitor *v, const char *name,
53
error_setg(errp, "%s: error reading %s", __func__, name);
54
return;
55
}
56
- if (led < 0 || led > s->nr_leds) {
57
+ if (led < 0 || led > s->pin_count) {
58
error_setg(errp, "%s invalid led %s", __func__, name);
59
return;
60
}
61
@@ -XXX,XX +XXX,XX @@ static void pca9552_initfn(Object *obj)
62
* PCA955X device
63
*/
64
s->max_reg = PCA9552_LS3;
65
- s->nr_leds = 16;
66
+ s->pin_count = 16;
67
68
- for (led = 0; led < s->nr_leds; led++) {
69
+ for (led = 0; led < s->pin_count; led++) {
70
char *name;
71
72
name = g_strdup_printf("led%d", led);
73
--
74
2.20.1
75
76
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Add a description field to distinguish between multiple devices.
4
5
Reviewed-by: Cédric Le Goater <clg@kaod.org>
6
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
7
Tested-by: Cédric Le Goater <clg@kaod.org>
8
Message-id: 20200623072723.6324-6-f4bug@amsat.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
include/hw/misc/pca9552.h | 1 +
12
hw/misc/pca9552.c | 18 ++++++++++++++++++
13
2 files changed, 19 insertions(+)
14
15
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
16
index XXXXXXX..XXXXXXX 100644
17
--- a/include/hw/misc/pca9552.h
18
+++ b/include/hw/misc/pca9552.h
19
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
20
uint8_t pointer;
21
22
uint8_t regs[PCA955X_NR_REGS];
23
+ char *description; /* For debugging purpose only */
24
} PCA955xState;
25
26
#endif
27
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/misc/pca9552.c
30
+++ b/hw/misc/pca9552.c
31
@@ -XXX,XX +XXX,XX @@
32
#include "qemu/osdep.h"
33
#include "qemu/log.h"
34
#include "qemu/module.h"
35
+#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
#include "migration/vmstate.h"
39
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
40
}
41
}
42
43
+static void pca955x_realize(DeviceState *dev, Error **errp)
44
+{
45
+ PCA955xState *s = PCA955X(dev);
46
+
47
+ if (!s->description) {
48
+ s->description = g_strdup("pca-unspecified");
49
+ }
50
+}
51
+
52
+static Property pca955x_properties[] = {
53
+ DEFINE_PROP_STRING("description", PCA955xState, description),
54
+ DEFINE_PROP_END_OF_LIST(),
55
+};
56
+
57
static void pca955x_class_init(ObjectClass *klass, void *data)
58
{
59
+ DeviceClass *dc = DEVICE_CLASS(klass);
60
I2CSlaveClass *k = I2C_SLAVE_CLASS(klass);
61
62
k->event = pca955x_event;
63
k->recv = pca955x_recv;
64
k->send = pca955x_send;
65
+ dc->realize = pca955x_realize;
66
+ device_class_set_props(dc, pca955x_properties);
67
}
68
69
static const TypeInfo pca955x_info = {
70
--
71
2.20.1
72
73
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.1M the PXN architecture extension adds a new PXN bit to the
2
MPU_RLAR registers, which forbids execution of code in the region
3
from a privileged mode.
2
4
3
Protect reads of aa64 id registers with ARM_CP_STATE_AA64.
5
This is another feature which is just in the generic "in v8.1M" set
4
Use this as a simpler test than arm_el_is_aa64, since EL3
6
and has no ID register field indicating its presence.
5
cannot change mode.
6
7
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-3-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20201119215617.29887-3-peter.maydell@linaro.org
11
---
11
---
12
target/arm/helper.c | 15 ++++++++-------
12
target/arm/helper.c | 7 ++++++-
13
1 file changed, 8 insertions(+), 7 deletions(-)
13
1 file changed, 6 insertions(+), 1 deletion(-)
14
14
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
17
--- a/target/arm/helper.c
18
+++ b/target/arm/helper.c
18
+++ b/target/arm/helper.c
19
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
19
@@ -XXX,XX +XXX,XX @@ bool pmsav8_mpu_lookup(CPUARMState *env, uint32_t address,
20
uint32_t valid_mask = 0x3fff;
20
} else {
21
ARMCPU *cpu = env_archcpu(env);
21
uint32_t ap = extract32(env->pmsav8.rbar[secure][matchregion], 1, 2);
22
22
uint32_t xn = extract32(env->pmsav8.rbar[secure][matchregion], 0, 1);
23
- if (arm_el_is_aa64(env, 3)) {
23
+ bool pxn = false;
24
+ if (ri->state == ARM_CP_STATE_AA64) {
25
value |= SCR_FW | SCR_AW; /* these two bits are RES1. */
26
valid_mask &= ~SCR_NET;
27
+
24
+
28
+ if (cpu_isar_feature(aa64_lor, cpu)) {
25
+ if (arm_feature(env, ARM_FEATURE_V8_1M)) {
29
+ valid_mask |= SCR_TLOR;
26
+ pxn = extract32(env->pmsav8.rlar[secure][matchregion], 4, 1);
30
+ }
27
+ }
31
+ if (cpu_isar_feature(aa64_pauth, cpu)) {
28
32
+ valid_mask |= SCR_API | SCR_APK;
29
if (m_is_system_region(env, address)) {
33
+ }
30
/* System space is always execute never */
34
} else {
31
@@ -XXX,XX +XXX,XX @@ bool pmsav8_mpu_lookup(CPUARMState *env, uint32_t address,
35
valid_mask &= ~(SCR_RW | SCR_ST);
36
}
37
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
38
valid_mask &= ~SCR_SMD;
39
}
32
}
40
}
33
41
- if (cpu_isar_feature(aa64_lor, cpu)) {
34
*prot = simple_ap_to_rw_prot(env, mmu_idx, ap);
42
- valid_mask |= SCR_TLOR;
35
- if (*prot && !xn) {
43
- }
36
+ if (*prot && !xn && !(pxn && !is_user)) {
44
- if (cpu_isar_feature(aa64_pauth, cpu)) {
37
*prot |= PAGE_EXEC;
45
- valid_mask |= SCR_API | SCR_APK;
38
}
46
- }
39
/* We don't need to look the attribute up in the MAIR0/MAIR1
47
48
/* Clear all-context RES0 bits. */
49
value &= valid_mask;
50
--
40
--
51
2.20.1
41
2.20.1
52
42
53
43
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In arm_cpu_realizefn() we check whether the board code disabled EL3
2
via the has_el3 CPU object property, which we create if the CPU
3
starts with the ARM_FEATURE_EL3 feature bit. If it is disabled, then
4
we turn off ARM_FEATURE_EL3 and also zero out the relevant fields in
5
the ID_PFR1 and ID_AA64PFR0 registers.
2
6
3
We can simplify our DC_ZVA if we recognize that the largest BS
7
This codepath was incorrectly being taken for M-profile CPUs, which
4
that we actually use in system mode is 64. Let us just assert
8
do not have an EL3 and don't set ARM_FEATURE_EL3, but which may have
5
that it fits within TARGET_PAGE_SIZE.
9
the M-profile Security extension and so should have non-zero values
10
in the ID_PFR1.Security field.
6
11
7
For DC_GVA and STZGM, we want to be able to write whole bytes
12
Restrict the handling of the feature flag to A/R-profile cores.
8
of tag memory, so assert that BS is >= 2 * TAG_GRANULE, or 32.
9
13
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-18-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
15
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
16
Message-id: 20201119215617.29887-4-peter.maydell@linaro.org
14
---
17
---
15
target/arm/cpu.c | 24 ++++++++++++++++++++++++
18
target/arm/cpu.c | 2 +-
16
1 file changed, 24 insertions(+)
19
1 file changed, 1 insertion(+), 1 deletion(-)
17
20
18
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
21
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
19
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.c
23
--- a/target/arm/cpu.c
21
+++ b/target/arm/cpu.c
24
+++ b/target/arm/cpu.c
22
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
25
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
26
}
23
}
27
}
24
#endif
28
25
29
- if (!cpu->has_el3) {
26
+ if (tcg_enabled()) {
30
+ if (!arm_feature(env, ARM_FEATURE_M) && !cpu->has_el3) {
27
+ int dcz_blocklen = 4 << cpu->dcz_blocksize;
31
/* If the has_el3 CPU property is disabled then we need to disable the
28
+
32
* feature.
29
+ /*
33
*/
30
+ * We only support DCZ blocklen that fits on one page.
31
+ *
32
+ * Architectually this is always true. However TARGET_PAGE_SIZE
33
+ * is variable and, for compatibility with -machine virt-2.7,
34
+ * is only 1KiB, as an artifact of legacy ARMv5 subpage support.
35
+ * But even then, while the largest architectural DCZ blocklen
36
+ * is 2KiB, no cpu actually uses such a large blocklen.
37
+ */
38
+ assert(dcz_blocklen <= TARGET_PAGE_SIZE);
39
+
40
+ /*
41
+ * We only support DCZ blocksize >= 2*TAG_GRANULE, which is to say
42
+ * both nibbles of each byte storing tag data may be written at once.
43
+ * Since TAG_GRANULE is 16, this means that blocklen must be >= 32.
44
+ */
45
+ if (cpu_isar_feature(aa64_mte, cpu)) {
46
+ assert(dcz_blocklen >= 2 * TAG_GRANULE);
47
+ }
48
+ }
49
+
50
qemu_init_vcpu(cs);
51
cpu_reset(cs);
52
53
--
34
--
54
2.20.1
35
2.20.1
55
36
56
37
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
Implement the v8.1M VSCCLRM insn, which zeros floating point
2
2
registers if there is an active floating point context.
3
This is TFSRE0_EL1, TFSR_EL1, TFSR_EL2, TFSR_EL3,
3
This requires support in write_neon_element32() for the MO_32
4
RGSR_EL1, GCR_EL1, GMID_EL1, and PSTATE.TCO.
4
element size, so add it.
5
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Because we want to use arm_gen_condlabel(), we need to move
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
the definition of that function up in translate.c so it is
8
Message-id: 20200626033144.790098-8-richard.henderson@linaro.org
8
before the #include of translate-vfp.c.inc.
9
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20201119215617.29887-5-peter.maydell@linaro.org
10
---
13
---
11
target/arm/cpu.h | 4 ++
14
target/arm/cpu.h | 9 ++++
12
target/arm/internals.h | 9 ++++
15
target/arm/m-nocp.decode | 8 +++-
13
target/arm/helper.c | 94 ++++++++++++++++++++++++++++++++++++++
16
target/arm/translate.c | 21 +++++----
14
target/arm/translate-a64.c | 21 +++++++++
17
target/arm/translate-vfp.c.inc | 84 ++++++++++++++++++++++++++++++++++
15
4 files changed, 128 insertions(+)
18
4 files changed, 111 insertions(+), 11 deletions(-)
16
19
17
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
20
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
18
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/cpu.h
22
--- a/target/arm/cpu.h
20
+++ b/target/arm/cpu.h
23
+++ b/target/arm/cpu.h
21
@@ -XXX,XX +XXX,XX @@ typedef struct CPUARMState {
24
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa32_mprofile(const ARMISARegisters *id)
22
uint64_t pmccfiltr_el0; /* Performance Monitor Filter Register */
25
return FIELD_EX32(id->id_pfr1, ID_PFR1, MPROGMOD) != 0;
23
uint64_t vpidr_el2; /* Virtualization Processor ID Register */
26
}
24
uint64_t vmpidr_el2; /* Virtualization Multiprocessor ID Register */
27
25
+ uint64_t tfsr_el[4]; /* tfsre0_el1 is index 0. */
28
+static inline bool isar_feature_aa32_m_sec_state(const ARMISARegisters *id)
26
+ uint64_t gcr_el1;
27
+ uint64_t rgsr_el1;
28
} cp15;
29
30
struct {
31
@@ -XXX,XX +XXX,XX @@ void pmu_init(ARMCPU *cpu);
32
#define PSTATE_SS (1U << 21)
33
#define PSTATE_PAN (1U << 22)
34
#define PSTATE_UAO (1U << 23)
35
+#define PSTATE_TCO (1U << 25)
36
#define PSTATE_V (1U << 28)
37
#define PSTATE_C (1U << 29)
38
#define PSTATE_Z (1U << 30)
39
diff --git a/target/arm/internals.h b/target/arm/internals.h
40
index XXXXXXX..XXXXXXX 100644
41
--- a/target/arm/internals.h
42
+++ b/target/arm/internals.h
43
@@ -XXX,XX +XXX,XX @@ static inline uint32_t aarch64_pstate_valid_mask(const ARMISARegisters *id)
44
if (isar_feature_aa64_uao(id)) {
45
valid |= PSTATE_UAO;
46
}
47
+ if (isar_feature_aa64_mte(id)) {
48
+ valid |= PSTATE_TCO;
49
+ }
50
51
return valid;
52
}
53
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
54
55
#endif /* !CONFIG_USER_ONLY */
56
57
+/*
58
+ * The log2 of the words in the tag block, for GMID_EL1.BS.
59
+ * The is the maximum, 256 bytes, which manipulates 64-bits of tags.
60
+ */
61
+#define GMID_EL1_BS 6
62
+
63
#endif
64
diff --git a/target/arm/helper.c b/target/arm/helper.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/helper.c
67
+++ b/target/arm/helper.c
68
@@ -XXX,XX +XXX,XX @@ static void define_arm_vh_e2h_redirects_aliases(ARMCPU *cpu)
69
{ K(3, 0, 1, 2, 0), K(3, 4, 1, 2, 0), K(3, 5, 1, 2, 0),
70
"ZCR_EL1", "ZCR_EL2", "ZCR_EL12", isar_feature_aa64_sve },
71
72
+ { K(3, 0, 5, 6, 0), K(3, 4, 5, 6, 0), K(3, 5, 5, 6, 0),
73
+ "TFSR_EL1", "TFSR_EL2", "TFSR_EL12", isar_feature_aa64_mte },
74
+
75
/* TODO: ARMv8.2-SPE -- PMSCR_EL2 */
76
/* TODO: ARMv8.4-Trace -- TRFCR_EL2 */
77
};
78
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo dcpodp_reg[] = {
79
};
80
#endif /*CONFIG_USER_ONLY*/
81
82
+static CPAccessResult access_aa64_tid5(CPUARMState *env, const ARMCPRegInfo *ri,
83
+ bool isread)
84
+{
29
+{
85
+ if ((arm_current_el(env) < 2) && (arm_hcr_el2_eff(env) & HCR_TID5)) {
30
+ /*
86
+ return CP_ACCESS_TRAP_EL2;
31
+ * Return true if M-profile state handling insns
87
+ }
32
+ * (VSCCLRM, CLRM, FPCTX access insns) are implemented
88
+
33
+ */
89
+ return CP_ACCESS_OK;
34
+ return FIELD_EX32(id->id_pfr1, ID_PFR1, SECURITY) >= 3;
90
+}
35
+}
91
+
36
+
92
+static CPAccessResult access_mte(CPUARMState *env, const ARMCPRegInfo *ri,
37
static inline bool isar_feature_aa32_fp16_arith(const ARMISARegisters *id)
93
+ bool isread)
38
{
39
/* Sadly this is encoded differently for A-profile and M-profile */
40
diff --git a/target/arm/m-nocp.decode b/target/arm/m-nocp.decode
41
index XXXXXXX..XXXXXXX 100644
42
--- a/target/arm/m-nocp.decode
43
+++ b/target/arm/m-nocp.decode
44
@@ -XXX,XX +XXX,XX @@
45
# If the coprocessor is not present or disabled then we will generate
46
# the NOCP exception; otherwise we let the insn through to the main decode.
47
48
+%vd_dp 22:1 12:4
49
+%vd_sp 12:4 22:1
50
+
51
&nocp cp
52
53
{
54
# Special cases which do not take an early NOCP: VLLDM and VLSTM
55
VLLDM_VLSTM 1110 1100 001 l:1 rn:4 0000 1010 0000 0000
56
- # TODO: VSCCLRM (new in v8.1M) is similar:
57
- #VSCCLRM 1110 1100 1-01 1111 ---- 1011 ---- ---0
58
+ # VSCCLRM (new in v8.1M) is similar:
59
+ VSCCLRM 1110 1100 1.01 1111 .... 1011 imm:7 0 vd=%vd_dp size=3
60
+ VSCCLRM 1110 1100 1.01 1111 .... 1010 imm:8 vd=%vd_sp size=2
61
62
NOCP 111- 1110 ---- ---- ---- cp:4 ---- ---- &nocp
63
NOCP 111- 110- ---- ---- ---- cp:4 ---- ---- &nocp
64
diff --git a/target/arm/translate.c b/target/arm/translate.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/target/arm/translate.c
67
+++ b/target/arm/translate.c
68
@@ -XXX,XX +XXX,XX @@ void arm_translate_init(void)
69
a64_translate_init();
70
}
71
72
+/* Generate a label used for skipping this instruction */
73
+static void arm_gen_condlabel(DisasContext *s)
94
+{
74
+{
95
+ int el = arm_current_el(env);
75
+ if (!s->condjmp) {
96
+
76
+ s->condlabel = gen_new_label();
97
+ if (el < 2 &&
77
+ s->condjmp = 1;
98
+ arm_feature(env, ARM_FEATURE_EL2) &&
78
+ }
99
+ !(arm_hcr_el2_eff(env) & HCR_ATA)) {
100
+ return CP_ACCESS_TRAP_EL2;
101
+ }
102
+ if (el < 3 &&
103
+ arm_feature(env, ARM_FEATURE_EL3) &&
104
+ !(env->cp15.scr_el3 & SCR_ATA)) {
105
+ return CP_ACCESS_TRAP_EL3;
106
+ }
107
+ return CP_ACCESS_OK;
108
+}
79
+}
109
+
80
+
110
+static uint64_t tco_read(CPUARMState *env, const ARMCPRegInfo *ri)
81
/* Flags for the disas_set_da_iss info argument:
82
* lower bits hold the Rt register number, higher bits are flags.
83
*/
84
@@ -XXX,XX +XXX,XX @@ static void write_neon_element64(TCGv_i64 src, int reg, int ele, MemOp memop)
85
long off = neon_element_offset(reg, ele, memop);
86
87
switch (memop) {
88
+ case MO_32:
89
+ tcg_gen_st32_i64(src, cpu_env, off);
90
+ break;
91
case MO_64:
92
tcg_gen_st_i64(src, cpu_env, off);
93
break;
94
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
95
s->base.is_jmp = DISAS_UPDATE_EXIT;
96
}
97
98
-/* Generate a label used for skipping this instruction */
99
-static void arm_gen_condlabel(DisasContext *s)
100
-{
101
- if (!s->condjmp) {
102
- s->condlabel = gen_new_label();
103
- s->condjmp = 1;
104
- }
105
-}
106
-
107
/* Skip this instruction if the ARM condition is false */
108
static void arm_skip_unless(DisasContext *s, uint32_t cond)
109
{
110
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
111
index XXXXXXX..XXXXXXX 100644
112
--- a/target/arm/translate-vfp.c.inc
113
+++ b/target/arm/translate-vfp.c.inc
114
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
115
return true;
116
}
117
118
+static bool trans_VSCCLRM(DisasContext *s, arg_VSCCLRM *a)
111
+{
119
+{
112
+ return env->pstate & PSTATE_TCO;
120
+ int btmreg, topreg;
121
+ TCGv_i64 zero;
122
+ TCGv_i32 aspen, sfpa;
123
+
124
+ if (!dc_isar_feature(aa32_m_sec_state, s)) {
125
+ /* Before v8.1M, fall through in decode to NOCP check */
126
+ return false;
127
+ }
128
+
129
+ /* Explicitly UNDEF because this takes precedence over NOCP */
130
+ if (!arm_dc_feature(s, ARM_FEATURE_M_MAIN) || !s->v8m_secure) {
131
+ unallocated_encoding(s);
132
+ return true;
133
+ }
134
+
135
+ if (!dc_isar_feature(aa32_vfp_simd, s)) {
136
+ /* NOP if we have neither FP nor MVE */
137
+ return true;
138
+ }
139
+
140
+ /*
141
+ * If FPCCR.ASPEN != 0 && CONTROL_S.SFPA == 0 then there is no
142
+ * active floating point context so we must NOP (without doing
143
+ * any lazy state preservation or the NOCP check).
144
+ */
145
+ aspen = load_cpu_field(v7m.fpccr[M_REG_S]);
146
+ sfpa = load_cpu_field(v7m.control[M_REG_S]);
147
+ tcg_gen_andi_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
148
+ tcg_gen_xori_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
149
+ tcg_gen_andi_i32(sfpa, sfpa, R_V7M_CONTROL_SFPA_MASK);
150
+ tcg_gen_or_i32(sfpa, sfpa, aspen);
151
+ arm_gen_condlabel(s);
152
+ tcg_gen_brcondi_i32(TCG_COND_EQ, sfpa, 0, s->condlabel);
153
+
154
+ if (s->fp_excp_el != 0) {
155
+ gen_exception_insn(s, s->pc_curr, EXCP_NOCP,
156
+ syn_uncategorized(), s->fp_excp_el);
157
+ return true;
158
+ }
159
+
160
+ topreg = a->vd + a->imm - 1;
161
+ btmreg = a->vd;
162
+
163
+ /* Convert to Sreg numbers if the insn specified in Dregs */
164
+ if (a->size == 3) {
165
+ topreg = topreg * 2 + 1;
166
+ btmreg *= 2;
167
+ }
168
+
169
+ if (topreg > 63 || (topreg > 31 && !(topreg & 1))) {
170
+ /* UNPREDICTABLE: we choose to undef */
171
+ unallocated_encoding(s);
172
+ return true;
173
+ }
174
+
175
+ /* Silently ignore requests to clear D16-D31 if they don't exist */
176
+ if (topreg > 31 && !dc_isar_feature(aa32_simd_r32, s)) {
177
+ topreg = 31;
178
+ }
179
+
180
+ if (!vfp_access_check(s)) {
181
+ return true;
182
+ }
183
+
184
+ /* Zero the Sregs from btmreg to topreg inclusive. */
185
+ zero = tcg_const_i64(0);
186
+ if (btmreg & 1) {
187
+ write_neon_element64(zero, btmreg >> 1, 1, MO_32);
188
+ btmreg++;
189
+ }
190
+ for (; btmreg + 1 <= topreg; btmreg += 2) {
191
+ write_neon_element64(zero, btmreg >> 1, 0, MO_64);
192
+ }
193
+ if (btmreg == topreg) {
194
+ write_neon_element64(zero, btmreg >> 1, 0, MO_32);
195
+ btmreg++;
196
+ }
197
+ assert(btmreg == topreg + 1);
198
+ /* TODO: when MVE is implemented, zero VPR here */
199
+ return true;
113
+}
200
+}
114
+
201
+
115
+static void tco_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t val)
202
static bool trans_NOCP(DisasContext *s, arg_nocp *a)
116
+{
203
{
117
+ env->pstate = (env->pstate & ~PSTATE_TCO) | (val & PSTATE_TCO);
204
/*
118
+}
119
+
120
+static const ARMCPRegInfo mte_reginfo[] = {
121
+ { .name = "TFSRE0_EL1", .state = ARM_CP_STATE_AA64,
122
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 1,
123
+ .access = PL1_RW, .accessfn = access_mte,
124
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[0]) },
125
+ { .name = "TFSR_EL1", .state = ARM_CP_STATE_AA64,
126
+ .opc0 = 3, .opc1 = 0, .crn = 5, .crm = 6, .opc2 = 0,
127
+ .access = PL1_RW, .accessfn = access_mte,
128
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[1]) },
129
+ { .name = "TFSR_EL2", .state = ARM_CP_STATE_AA64,
130
+ .opc0 = 3, .opc1 = 4, .crn = 5, .crm = 6, .opc2 = 0,
131
+ .access = PL2_RW, .accessfn = access_mte,
132
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[2]) },
133
+ { .name = "TFSR_EL3", .state = ARM_CP_STATE_AA64,
134
+ .opc0 = 3, .opc1 = 6, .crn = 5, .crm = 6, .opc2 = 0,
135
+ .access = PL3_RW,
136
+ .fieldoffset = offsetof(CPUARMState, cp15.tfsr_el[3]) },
137
+ { .name = "RGSR_EL1", .state = ARM_CP_STATE_AA64,
138
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 5,
139
+ .access = PL1_RW, .accessfn = access_mte,
140
+ .fieldoffset = offsetof(CPUARMState, cp15.rgsr_el1) },
141
+ { .name = "GCR_EL1", .state = ARM_CP_STATE_AA64,
142
+ .opc0 = 3, .opc1 = 0, .crn = 1, .crm = 0, .opc2 = 6,
143
+ .access = PL1_RW, .accessfn = access_mte,
144
+ .fieldoffset = offsetof(CPUARMState, cp15.gcr_el1) },
145
+ { .name = "GMID_EL1", .state = ARM_CP_STATE_AA64,
146
+ .opc0 = 3, .opc1 = 1, .crn = 0, .crm = 0, .opc2 = 4,
147
+ .access = PL1_R, .accessfn = access_aa64_tid5,
148
+ .type = ARM_CP_CONST, .resetvalue = GMID_EL1_BS },
149
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
150
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
151
+ .type = ARM_CP_NO_RAW,
152
+ .access = PL0_RW, .readfn = tco_read, .writefn = tco_write },
153
+ REGINFO_SENTINEL
154
+};
155
+
156
+static const ARMCPRegInfo mte_tco_ro_reginfo[] = {
157
+ { .name = "TCO", .state = ARM_CP_STATE_AA64,
158
+ .opc0 = 3, .opc1 = 3, .crn = 4, .crm = 2, .opc2 = 7,
159
+ .type = ARM_CP_CONST, .access = PL0_RW, },
160
+ REGINFO_SENTINEL
161
+};
162
#endif
163
164
static CPAccessResult access_predinv(CPUARMState *env, const ARMCPRegInfo *ri,
165
@@ -XXX,XX +XXX,XX @@ void register_cp_regs_for_features(ARMCPU *cpu)
166
}
167
}
168
#endif /*CONFIG_USER_ONLY*/
169
+
170
+ /*
171
+ * If full MTE is enabled, add all of the system registers.
172
+ * If only "instructions available at EL0" are enabled,
173
+ * then define only a RAZ/WI version of PSTATE.TCO.
174
+ */
175
+ if (cpu_isar_feature(aa64_mte, cpu)) {
176
+ define_arm_cp_regs(cpu, mte_reginfo);
177
+ } else if (cpu_isar_feature(aa64_mte_insn_reg, cpu)) {
178
+ define_arm_cp_regs(cpu, mte_tco_ro_reginfo);
179
+ }
180
#endif
181
182
if (cpu_isar_feature(any_predinv, cpu)) {
183
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
184
index XXXXXXX..XXXXXXX 100644
185
--- a/target/arm/translate-a64.c
186
+++ b/target/arm/translate-a64.c
187
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
188
s->base.is_jmp = DISAS_UPDATE_EXIT;
189
break;
190
191
+ case 0x1c: /* TCO */
192
+ if (dc_isar_feature(aa64_mte, s)) {
193
+ /* Full MTE is enabled -- set the TCO bit as directed. */
194
+ if (crm & 1) {
195
+ set_pstate_bits(PSTATE_TCO);
196
+ } else {
197
+ clear_pstate_bits(PSTATE_TCO);
198
+ }
199
+ t1 = tcg_const_i32(s->current_el);
200
+ gen_helper_rebuild_hflags_a64(cpu_env, t1);
201
+ tcg_temp_free_i32(t1);
202
+ /* Many factors, including TCO, go into MTE_ACTIVE. */
203
+ s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
204
+ } else if (dc_isar_feature(aa64_mte_insn_reg, s)) {
205
+ /* Only "instructions accessible at EL0" -- PSTATE.TCO is WI. */
206
+ s->base.is_jmp = DISAS_NEXT;
207
+ } else {
208
+ goto do_unallocated;
209
+ }
210
+ break;
211
+
212
default:
213
do_unallocated:
214
unallocated_encoding(s);
215
--
205
--
216
2.20.1
206
2.20.1
217
207
218
208
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.1M the new CLRM instruction allows zeroing an arbitrary set of
2
the general-purpose registers and APSR. Implement this.
2
3
3
Emphasize that the is_jmp option exits to the main loop.
4
The encoding is a subset of the LDMIA T2 encoding, using what would
5
be Rn=0b1111 (which UNDEFs for LDMIA).
4
6
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-6-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
8
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20201119215617.29887-6-peter.maydell@linaro.org
9
---
10
---
10
target/arm/translate.h | 14 ++++++++------
11
target/arm/t32.decode | 6 +++++-
11
target/arm/translate-a64.c | 8 ++++----
12
target/arm/translate.c | 38 ++++++++++++++++++++++++++++++++++++++
12
target/arm/translate-vfp.inc.c | 4 ++--
13
2 files changed, 43 insertions(+), 1 deletion(-)
13
target/arm/translate.c | 12 ++++++------
14
4 files changed, 20 insertions(+), 18 deletions(-)
15
14
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
15
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
17
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.h
17
--- a/target/arm/t32.decode
19
+++ b/target/arm/translate.h
18
+++ b/target/arm/t32.decode
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
19
@@ -XXX,XX +XXX,XX @@ UXTAB 1111 1010 0101 .... 1111 .... 10.. .... @rrr_rot
21
20
22
/* is_jmp field values */
21
STM_t32 1110 1000 10.0 .... ................ @ldstm i=1 b=0
23
#define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
22
STM_t32 1110 1001 00.0 .... ................ @ldstm i=0 b=1
24
-#define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
23
-LDM_t32 1110 1000 10.1 .... ................ @ldstm i=1 b=0
25
+/* CPU state was modified dynamically; exit to main loop for interrupts. */
24
+{
26
+#define DISAS_UPDATE_EXIT DISAS_TARGET_1
25
+ # Rn=15 UNDEFs for LDM; M-profile CLRM uses that encoding
27
/* These instructions trap after executing, so the A32/T32 decoder must
26
+ CLRM 1110 1000 1001 1111 list:16
28
* defer them until after the conditional execution state has been updated.
27
+ LDM_t32 1110 1000 10.1 .... ................ @ldstm i=1 b=0
29
* WFI also needs special handling when single-stepping.
28
+}
30
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
29
LDM_t32 1110 1001 00.1 .... ................ @ldstm i=0 b=1
31
* custom end-of-TB code)
30
32
*/
31
&rfe !extern rn w pu
33
#define DISAS_BX_EXCRET DISAS_TARGET_8
34
-/* For instructions which want an immediate exit to the main loop,
35
- * as opposed to attempting to use lookup_and_goto_ptr. Unlike
36
- * DISAS_UPDATE this doesn't write the PC on exiting the translation
37
- * loop so you need to ensure something (gen_a64_set_pc_im or runtime
38
- * helper) has done so before we reach return from cpu_tb_exec.
39
+/*
40
+ * For instructions which want an immediate exit to the main loop, as opposed
41
+ * to attempting to use lookup_and_goto_ptr. Unlike DISAS_UPDATE_EXIT, this
42
+ * doesn't write the PC on exiting the translation loop so you need to ensure
43
+ * something (gen_a64_set_pc_im or runtime helper) has done so before we reach
44
+ * return from cpu_tb_exec.
45
*/
46
#define DISAS_EXIT DISAS_TARGET_9
47
48
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/target/arm/translate-a64.c
51
+++ b/target/arm/translate-a64.c
52
@@ -XXX,XX +XXX,XX @@ static void handle_msr_i(DisasContext *s, uint32_t insn,
53
gen_helper_msr_i_daifclear(cpu_env, t1);
54
tcg_temp_free_i32(t1);
55
/* For DAIFClear, exit the cpu loop to re-evaluate pending IRQs. */
56
- s->base.is_jmp = DISAS_UPDATE;
57
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
58
break;
59
60
default:
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
63
if ((tb_cflags(s->base.tb) & CF_USE_ICOUNT) && (ri->type & ARM_CP_IO)) {
64
/* I/O operations must end the TB here (whether read or write) */
65
- s->base.is_jmp = DISAS_UPDATE;
66
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
67
}
68
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
69
/*
70
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
71
* but allow this to be suppressed by the register definition
72
* (usually only necessary to work around guest bugs).
73
*/
74
- s->base.is_jmp = DISAS_UPDATE;
75
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
76
}
77
}
78
79
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
80
gen_goto_tb(dc, 1, dc->base.pc_next);
81
break;
82
default:
83
- case DISAS_UPDATE:
84
+ case DISAS_UPDATE_EXIT:
85
gen_a64_set_pc_im(dc->base.pc_next);
86
/* fall through */
87
case DISAS_EXIT:
88
diff --git a/target/arm/translate-vfp.inc.c b/target/arm/translate-vfp.inc.c
89
index XXXXXXX..XXXXXXX 100644
90
--- a/target/arm/translate-vfp.inc.c
91
+++ b/target/arm/translate-vfp.inc.c
92
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
93
* this to be the last insn in the TB).
94
*/
95
if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
96
- s->base.is_jmp = DISAS_UPDATE;
97
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
98
gen_io_start();
99
}
100
gen_helper_v7m_preserve_fp_state(cpu_env);
101
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
102
tcg_temp_free_i32(fptr);
103
104
/* End the TB, because we have updated FP control bits */
105
- s->base.is_jmp = DISAS_UPDATE;
106
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
107
return true;
108
}
109
diff --git a/target/arm/translate.c b/target/arm/translate.c
32
diff --git a/target/arm/translate.c b/target/arm/translate.c
110
index XXXXXXX..XXXXXXX 100644
33
index XXXXXXX..XXXXXXX 100644
111
--- a/target/arm/translate.c
34
--- a/target/arm/translate.c
112
+++ b/target/arm/translate.c
35
+++ b/target/arm/translate.c
113
@@ -XXX,XX +XXX,XX @@ static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
36
@@ -XXX,XX +XXX,XX @@ static bool trans_LDM_t16(DisasContext *s, arg_ldst_block *a)
114
tcg_temp_free_i32(tcg_tgtmode);
37
return do_ldm(s, a, 1);
115
tcg_temp_free_i32(tcg_regno);
116
tcg_temp_free_i32(tcg_reg);
117
- s->base.is_jmp = DISAS_UPDATE;
118
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
119
}
38
}
120
39
121
static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
40
+static bool trans_CLRM(DisasContext *s, arg_CLRM *a)
122
@@ -XXX,XX +XXX,XX @@ static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
41
+{
123
tcg_temp_free_i32(tcg_tgtmode);
42
+ int i;
124
tcg_temp_free_i32(tcg_regno);
43
+ TCGv_i32 zero;
125
store_reg(s, rn, tcg_reg);
44
+
126
- s->base.is_jmp = DISAS_UPDATE;
45
+ if (!dc_isar_feature(aa32_m_sec_state, s)) {
127
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
46
+ return false;
128
}
47
+ }
129
48
+
130
/* Store value to PC as for an exception return (ie don't
49
+ if (extract32(a->list, 13, 1)) {
131
@@ -XXX,XX +XXX,XX @@ static void gen_srs(DisasContext *s,
50
+ return false;
132
tcg_temp_free_i32(tmp);
51
+ }
133
}
52
+
134
tcg_temp_free_i32(addr);
53
+ if (!a->list) {
135
- s->base.is_jmp = DISAS_UPDATE;
54
+ /* UNPREDICTABLE; we choose to UNDEF */
136
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
55
+ return false;
137
}
56
+ }
138
57
+
139
/* Generate a label used for skipping this instruction */
58
+ zero = tcg_const_i32(0);
140
@@ -XXX,XX +XXX,XX @@ static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
59
+ for (i = 0; i < 15; i++) {
141
}
60
+ if (extract32(a->list, i, 1)) {
142
if (a->E != (s->be_data == MO_BE)) {
61
+ /* Clear R[i] */
143
gen_helper_setend(cpu_env);
62
+ tcg_gen_mov_i32(cpu_R[i], zero);
144
- s->base.is_jmp = DISAS_UPDATE;
63
+ }
145
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
64
+ }
146
}
65
+ if (extract32(a->list, 15, 1)) {
147
return true;
66
+ /*
148
}
67
+ * Clear APSR (by calling the MSR helper with the same argument
149
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
68
+ * as for "MSR APSR_nzcvqg, Rn": mask = 0b1100, SYSM=0)
150
break;
69
+ */
151
case DISAS_NEXT:
70
+ TCGv_i32 maskreg = tcg_const_i32(0xc << 8);
152
case DISAS_TOO_MANY:
71
+ gen_helper_v7m_msr(cpu_env, maskreg, zero);
153
- case DISAS_UPDATE:
72
+ tcg_temp_free_i32(maskreg);
154
+ case DISAS_UPDATE_EXIT:
73
+ }
155
gen_set_pc_im(dc, dc->base.pc_next);
74
+ tcg_temp_free_i32(zero);
156
/* fall through */
75
+ return true;
157
default:
76
+}
158
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
77
+
159
case DISAS_JUMP:
78
/*
160
gen_goto_ptr();
79
* Branch, branch with link
161
break;
80
*/
162
- case DISAS_UPDATE:
163
+ case DISAS_UPDATE_EXIT:
164
gen_set_pc_im(dc, dc->base.pc_next);
165
/* fall through */
166
default:
167
--
81
--
168
2.20.1
82
2.20.1
169
83
170
84
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
For M-profile before v8.1M, the only valid register for VMSR/VMRS is
2
the FPSCR. We have a comment that states this, but the actual logic
3
to forbid accesses for any other register value is missing, so we
4
would end up with A-profile style behaviour. Add the missing check.
2
5
3
Replace existing uses of check_data_tbi in translate-a64.c that
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
perform multiple logical memory access. Leave the helper blank
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
for now to reduce the patch size.
8
Message-id: 20201119215617.29887-7-peter.maydell@linaro.org
9
---
10
target/arm/translate-vfp.c.inc | 5 ++++-
11
1 file changed, 4 insertions(+), 1 deletion(-)
6
12
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-25-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/translate-a64.h | 2 ++
14
target/arm/mte_helper.c | 8 +++++
15
target/arm/translate-a64.c | 71 +++++++++++++++++++++++++++++---------
16
4 files changed, 66 insertions(+), 16 deletions(-)
17
18
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
19
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/helper-a64.h
15
--- a/target/arm/translate-vfp.c.inc
21
+++ b/target/arm/helper-a64.h
16
+++ b/target/arm/translate-vfp.c.inc
22
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
17
@@ -XXX,XX +XXX,XX @@ static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
23
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
18
* Accesses to R15 are UNPREDICTABLE; we choose to undef.
24
19
* (FPSCR -> r15 is a special case which writes to the PSR flags.)
25
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
20
*/
26
+DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
21
- if (a->rt == 15 && (!a->l || a->reg != ARM_VFP_FPSCR)) {
27
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+ if (a->reg != ARM_VFP_FPSCR) {
28
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
+ return false;
29
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
24
+ }
30
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
25
+ if (a->rt == 15 && !a->l) {
31
index XXXXXXX..XXXXXXX 100644
26
return false;
32
--- a/target/arm/translate-a64.h
33
+++ b/target/arm/translate-a64.h
34
@@ -XXX,XX +XXX,XX @@ bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
35
bool sve_access_check(DisasContext *s);
36
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
37
bool tag_checked, int log2_size);
38
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
39
+ bool tag_checked, int count, int log2_esize);
40
41
/* We should have at some point before trying to access an FP register
42
* done the necessary access check, so assert that
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
48
{
49
return ptr;
50
}
51
+
52
+/*
53
+ * Perform an MTE checked access for multiple logical accesses.
54
+ */
55
+uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
56
+{
57
+ return ptr;
58
+}
59
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/target/arm/translate-a64.c
62
+++ b/target/arm/translate-a64.c
63
@@ -XXX,XX +XXX,XX @@ TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
64
false, get_mem_index(s));
65
}
66
67
+/*
68
+ * For MTE, check multiple logical sequential accesses.
69
+ */
70
+TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
71
+ bool tag_checked, int log2_esize, int total_size)
72
+{
73
+ if (tag_checked && s->mte_active[0] && total_size != (1 << log2_esize)) {
74
+ TCGv_i32 tcg_desc;
75
+ TCGv_i64 ret;
76
+ int desc = 0;
77
+
78
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
79
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
80
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
81
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
82
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_esize);
83
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, total_size);
84
+ tcg_desc = tcg_const_i32(desc);
85
+
86
+ ret = new_tmp_a64(s);
87
+ gen_helper_mte_checkN(ret, cpu_env, tcg_desc, addr);
88
+ tcg_temp_free_i32(tcg_desc);
89
+
90
+ return ret;
91
+ }
92
+ return gen_mte_check1(s, addr, is_write, tag_checked, log2_esize);
93
+}
94
+
95
typedef struct DisasCompare64 {
96
TCGCond cond;
97
TCGv_i64 value;
98
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
99
}
27
}
100
}
28
}
101
102
- clean_addr = clean_data_tbi(s, dirty_addr);
103
+ clean_addr = gen_mte_checkN(s, dirty_addr, !is_load,
104
+ (wback || rn != 31) && !set_tag,
105
+ size, 2 << size);
106
+
107
if (is_vector) {
108
if (is_load) {
109
do_fp_ld(s, rt, clean_addr, size);
110
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
111
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
112
MemOp endian = s->be_data;
113
114
- int ebytes; /* bytes per element */
115
+ int total; /* total bytes */
116
int elements; /* elements per vector */
117
int rpt; /* num iterations */
118
int selem; /* structure elements */
119
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
120
endian = MO_LE;
121
}
122
123
- /* Consecutive little-endian elements from a single register
124
+ total = rpt * selem * (is_q ? 16 : 8);
125
+ tcg_rn = cpu_reg_sp(s, rn);
126
+
127
+ /*
128
+ * Issue the MTE check vs the logical repeat count, before we
129
+ * promote consecutive little-endian elements below.
130
+ */
131
+ clean_addr = gen_mte_checkN(s, tcg_rn, is_store, is_postidx || rn != 31,
132
+ size, total);
133
+
134
+ /*
135
+ * Consecutive little-endian elements from a single register
136
* can be promoted to a larger little-endian operation.
137
*/
138
if (selem == 1 && endian == MO_LE) {
139
size = 3;
140
}
141
- ebytes = 1 << size;
142
- elements = (is_q ? 16 : 8) / ebytes;
143
-
144
- tcg_rn = cpu_reg_sp(s, rn);
145
- clean_addr = clean_data_tbi(s, tcg_rn);
146
- tcg_ebytes = tcg_const_i64(ebytes);
147
+ elements = (is_q ? 16 : 8) >> size;
148
149
+ tcg_ebytes = tcg_const_i64(1 << size);
150
for (r = 0; r < rpt; r++) {
151
int e;
152
for (e = 0; e < elements; e++) {
153
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
154
155
if (is_postidx) {
156
if (rm == 31) {
157
- tcg_gen_addi_i64(tcg_rn, tcg_rn, rpt * elements * selem * ebytes);
158
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
159
} else {
160
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
161
}
162
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
163
int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
164
bool replicate = false;
165
int index = is_q << 3 | S << 2 | size;
166
- int ebytes, xs;
167
+ int xs, total;
168
TCGv_i64 clean_addr, tcg_rn, tcg_ebytes;
169
170
if (extract32(insn, 31, 1)) {
171
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
172
return;
173
}
174
175
- ebytes = 1 << scale;
176
-
177
if (rn == 31) {
178
gen_check_sp_alignment(s);
179
}
180
181
+ total = selem << scale;
182
tcg_rn = cpu_reg_sp(s, rn);
183
- clean_addr = clean_data_tbi(s, tcg_rn);
184
- tcg_ebytes = tcg_const_i64(ebytes);
185
186
+ clean_addr = gen_mte_checkN(s, tcg_rn, !is_load, is_postidx || rn != 31,
187
+ scale, total);
188
+
189
+ tcg_ebytes = tcg_const_i64(1 << scale);
190
for (xs = 0; xs < selem; xs++) {
191
if (replicate) {
192
/* Load and replicate to all elements */
193
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
194
195
if (is_postidx) {
196
if (rm == 31) {
197
- tcg_gen_addi_i64(tcg_rn, tcg_rn, selem * ebytes);
198
+ tcg_gen_addi_i64(tcg_rn, tcg_rn, total);
199
} else {
200
tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
201
}
202
--
29
--
203
2.20.1
30
2.20.1
204
31
205
32
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
Currently M-profile borrows the A-profile code for VMSR and VMRS
2
2
(access to the FP system registers), because all it needs to support
3
Cache the composite ATA setting.
3
is the FPSCR. In v8.1M things become significantly more complicated
4
4
in two ways:
5
Cache when MTE is fully enabled, i.e. access to tags are enabled
5
6
and tag checks affect the PE. Do this for both the normal context
6
* there are several new FP system registers; some have side effects
7
and the UNPRIV context.
7
on read, and one (FPCXT_NS) needs to avoid the usual
8
8
vfp_access_check() and the "only if FPU implemented" check
9
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
10
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
* all sysregs are now accessible both by VMRS/VMSR (which
11
Message-id: 20200626033144.790098-9-richard.henderson@linaro.org
11
reads/writes a general purpose register) and also by VLDR/VSTR
12
(which reads/writes them directly to memory)
13
14
Refactor the structure of how we handle VMSR/VMRS to cope with this:
15
16
* keep the M-profile code entirely separate from the A-profile code
17
18
* abstract out the "read or write the general purpose register" part
19
of the code into a loadfn or storefn function pointer, so we can
20
reuse it for VLDR/VSTR.
21
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
22
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
23
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
24
Message-id: 20201119215617.29887-8-peter.maydell@linaro.org
13
---
25
---
14
target/arm/cpu.h | 12 ++++++++----
26
target/arm/cpu.h | 3 +
15
target/arm/internals.h | 18 +++++++++++++++++
27
target/arm/translate-vfp.c.inc | 182 ++++++++++++++++++++++++++++++---
16
target/arm/translate.h | 5 +++++
28
2 files changed, 171 insertions(+), 14 deletions(-)
17
target/arm/helper.c | 40 ++++++++++++++++++++++++++++++++++++++
18
target/arm/translate-a64.c | 4 ++++
19
5 files changed, 75 insertions(+), 4 deletions(-)
20
29
21
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
30
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
22
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
23
--- a/target/arm/cpu.h
32
--- a/target/arm/cpu.h
24
+++ b/target/arm/cpu.h
33
+++ b/target/arm/cpu.h
25
@@ -XXX,XX +XXX,XX @@ typedef ARMCPU ArchCPU;
34
@@ -XXX,XX +XXX,XX @@ enum arm_cpu_mode {
26
* | | | TBFLAG_A32 | |
35
#define ARM_VFP_FPINST 9
27
* | | +-----+----------+ TBFLAG_AM32 |
36
#define ARM_VFP_FPINST2 10
28
* | TBFLAG_ANY | |TBFLAG_M32| |
37
29
- * | | +-+----------+--------------|
38
+/* QEMU-internal value meaning "FPSCR, but we care only about NZCV" */
30
- * | | | TBFLAG_A64 |
39
+#define QEMU_VFP_FPSCR_NZCV 0xffff
31
- * +--------------+---------+---------------------------+
40
+
32
- * 31 20 15 0
41
/* iwMMXt coprocessor control registers. */
33
+ * | +-----------+----------+--------------|
42
#define ARM_IWMMXT_wCID 0
34
+ * | | TBFLAG_A64 |
43
#define ARM_IWMMXT_wCon 1
35
+ * +--------------+-------------------------------------+
44
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
36
+ * 31 20 0
37
*
38
* Unless otherwise noted, these bits are cached in env->hflags.
39
*/
40
@@ -XXX,XX +XXX,XX @@ FIELD(TBFLAG_A64, BT, 9, 1)
41
FIELD(TBFLAG_A64, BTYPE, 10, 2) /* Not cached. */
42
FIELD(TBFLAG_A64, TBID, 12, 2)
43
FIELD(TBFLAG_A64, UNPRIV, 14, 1)
44
+FIELD(TBFLAG_A64, ATA, 15, 1)
45
+FIELD(TBFLAG_A64, TCMA, 16, 2)
46
+FIELD(TBFLAG_A64, MTE_ACTIVE, 18, 1)
47
+FIELD(TBFLAG_A64, MTE0_ACTIVE, 19, 1)
48
49
/**
50
* cpu_mmu_index:
51
diff --git a/target/arm/internals.h b/target/arm/internals.h
52
index XXXXXXX..XXXXXXX 100644
45
index XXXXXXX..XXXXXXX 100644
53
--- a/target/arm/internals.h
46
--- a/target/arm/translate-vfp.c.inc
54
+++ b/target/arm/internals.h
47
+++ b/target/arm/translate-vfp.c.inc
55
@@ -XXX,XX +XXX,XX @@ static inline int exception_target_el(CPUARMState *env)
48
@@ -XXX,XX +XXX,XX @@ static bool trans_VDUP(DisasContext *s, arg_VDUP *a)
56
return target_el;
49
return true;
57
}
50
}
58
51
59
+/* Determine if allocation tags are available. */
52
+/*
60
+static inline bool allocation_tag_access_enabled(CPUARMState *env, int el,
53
+ * M-profile provides two different sets of instructions that can
61
+ uint64_t sctlr)
54
+ * access floating point system registers: VMSR/VMRS (which move
62
+{
55
+ * to/from a general purpose register) and VLDR/VSTR sysreg (which
63
+ if (el < 3
56
+ * move directly to/from memory). In some cases there are also side
64
+ && arm_feature(env, ARM_FEATURE_EL3)
57
+ * effects which must happen after any write to memory (which could
65
+ && !(env->cp15.scr_el3 & SCR_ATA)) {
58
+ * cause an exception). So we implement the common logic for the
59
+ * sysreg access in gen_M_fp_sysreg_write() and gen_M_fp_sysreg_read(),
60
+ * which take pointers to callback functions which will perform the
61
+ * actual "read/write general purpose register" and "read/write
62
+ * memory" operations.
63
+ */
64
+
65
+/*
66
+ * Emit code to store the sysreg to its final destination; frees the
67
+ * TCG temp 'value' it is passed.
68
+ */
69
+typedef void fp_sysreg_storefn(DisasContext *s, void *opaque, TCGv_i32 value);
70
+/*
71
+ * Emit code to load the value to be copied to the sysreg; returns
72
+ * a new TCG temporary
73
+ */
74
+typedef TCGv_i32 fp_sysreg_loadfn(DisasContext *s, void *opaque);
75
+
76
+/* Common decode/access checks for fp sysreg read/write */
77
+typedef enum FPSysRegCheckResult {
78
+ FPSysRegCheckFailed, /* caller should return false */
79
+ FPSysRegCheckDone, /* caller should return true */
80
+ FPSysRegCheckContinue, /* caller should continue generating code */
81
+} FPSysRegCheckResult;
82
+
83
+static FPSysRegCheckResult fp_sysreg_checks(DisasContext *s, int regno)
84
+{
85
+ if (!dc_isar_feature(aa32_fpsp_v2, s)) {
86
+ return FPSysRegCheckFailed;
87
+ }
88
+
89
+ switch (regno) {
90
+ case ARM_VFP_FPSCR:
91
+ case QEMU_VFP_FPSCR_NZCV:
92
+ break;
93
+ default:
94
+ return FPSysRegCheckFailed;
95
+ }
96
+
97
+ if (!vfp_access_check(s)) {
98
+ return FPSysRegCheckDone;
99
+ }
100
+
101
+ return FPSysRegCheckContinue;
102
+}
103
+
104
+static bool gen_M_fp_sysreg_write(DisasContext *s, int regno,
105
+
106
+ fp_sysreg_loadfn *loadfn,
107
+ void *opaque)
108
+{
109
+ /* Do a write to an M-profile floating point system register */
110
+ TCGv_i32 tmp;
111
+
112
+ switch (fp_sysreg_checks(s, regno)) {
113
+ case FPSysRegCheckFailed:
66
+ return false;
114
+ return false;
67
+ }
115
+ case FPSysRegCheckDone:
68
+ if (el < 2
116
+ return true;
69
+ && arm_feature(env, ARM_FEATURE_EL2)
117
+ case FPSysRegCheckContinue:
70
+ && !(arm_hcr_el2_eff(env) & HCR_ATA)) {
118
+ break;
119
+ }
120
+
121
+ switch (regno) {
122
+ case ARM_VFP_FPSCR:
123
+ tmp = loadfn(s, opaque);
124
+ gen_helper_vfp_set_fpscr(cpu_env, tmp);
125
+ tcg_temp_free_i32(tmp);
126
+ gen_lookup_tb(s);
127
+ break;
128
+ default:
129
+ g_assert_not_reached();
130
+ }
131
+ return true;
132
+}
133
+
134
+static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
135
+ fp_sysreg_storefn *storefn,
136
+ void *opaque)
137
+{
138
+ /* Do a read from an M-profile floating point system register */
139
+ TCGv_i32 tmp;
140
+
141
+ switch (fp_sysreg_checks(s, regno)) {
142
+ case FPSysRegCheckFailed:
71
+ return false;
143
+ return false;
72
+ }
144
+ case FPSysRegCheckDone:
73
+ sctlr &= (el == 0 ? SCTLR_ATA0 : SCTLR_ATA);
145
+ return true;
74
+ return sctlr != 0;
146
+ case FPSysRegCheckContinue:
75
+}
147
+ break;
76
+
148
+ }
77
#ifndef CONFIG_USER_ONLY
149
+
78
150
+ switch (regno) {
79
/* Security attributes for an address, as returned by v8m_security_lookup. */
151
+ case ARM_VFP_FPSCR:
80
diff --git a/target/arm/translate.h b/target/arm/translate.h
152
+ tmp = tcg_temp_new_i32();
81
index XXXXXXX..XXXXXXX 100644
153
+ gen_helper_vfp_get_fpscr(tmp, cpu_env);
82
--- a/target/arm/translate.h
154
+ storefn(s, opaque, tmp);
83
+++ b/target/arm/translate.h
155
+ break;
84
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
156
+ case QEMU_VFP_FPSCR_NZCV:
85
ARMMMUIdx mmu_idx; /* MMU index to use for normal loads/stores */
157
+ /*
86
uint8_t tbii; /* TBI1|TBI0 for insns */
158
+ * Read just NZCV; this is a special case to avoid the
87
uint8_t tbid; /* TBI1|TBI0 for data */
159
+ * helper call for the "VMRS to CPSR.NZCV" insn.
88
+ uint8_t tcma; /* TCMA1|TCMA0 for MTE */
160
+ */
89
bool ns; /* Use non-secure CPREG bank on access */
161
+ tmp = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
90
int fp_excp_el; /* FP exception EL or 0 if enabled */
162
+ tcg_gen_andi_i32(tmp, tmp, 0xf0000000);
91
int sve_excp_el; /* SVE exception EL or 0 if enabled */
163
+ storefn(s, opaque, tmp);
92
@@ -XXX,XX +XXX,XX @@ typedef struct DisasContext {
164
+ break;
93
bool unpriv;
165
+ default:
94
/* True if v8.3-PAuth is active. */
166
+ g_assert_not_reached();
95
bool pauth_active;
167
+ }
96
+ /* True if v8.5-MTE access to tags is enabled. */
168
+ return true;
97
+ bool ata;
169
+}
98
+ /* True if v8.5-MTE tag checks affect the PE; index with is_unpriv. */
170
+
99
+ bool mte_active[2];
171
+static void fp_sysreg_to_gpr(DisasContext *s, void *opaque, TCGv_i32 value)
100
/* True with v8.5-BTI and SCTLR_ELx.BT* set. */
172
+{
101
bool bt;
173
+ arg_VMSR_VMRS *a = opaque;
102
/* True if any CP15 access is trapped by HSTR_EL2 */
174
+
103
diff --git a/target/arm/helper.c b/target/arm/helper.c
175
+ if (a->rt == 15) {
104
index XXXXXXX..XXXXXXX 100644
176
+ /* Set the 4 flag bits in the CPSR */
105
--- a/target/arm/helper.c
177
+ gen_set_nzcv(value);
106
+++ b/target/arm/helper.c
178
+ tcg_temp_free_i32(value);
107
@@ -XXX,XX +XXX,XX @@ static int aa64_va_parameter_tbid(uint64_t tcr, ARMMMUIdx mmu_idx)
179
+ } else {
180
+ store_reg(s, a->rt, value);
181
+ }
182
+}
183
+
184
+static TCGv_i32 gpr_to_fp_sysreg(DisasContext *s, void *opaque)
185
+{
186
+ arg_VMSR_VMRS *a = opaque;
187
+
188
+ return load_reg(s, a->rt);
189
+}
190
+
191
+static bool gen_M_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
192
+{
193
+ /*
194
+ * Accesses to R15 are UNPREDICTABLE; we choose to undef.
195
+ * FPSCR -> r15 is a special case which writes to the PSR flags;
196
+ * set a->reg to a special value to tell gen_M_fp_sysreg_read()
197
+ * we only care about the top 4 bits of FPSCR there.
198
+ */
199
+ if (a->rt == 15) {
200
+ if (a->l && a->reg == ARM_VFP_FPSCR) {
201
+ a->reg = QEMU_VFP_FPSCR_NZCV;
202
+ } else {
203
+ return false;
204
+ }
205
+ }
206
+
207
+ if (a->l) {
208
+ /* VMRS, move FP system register to gp register */
209
+ return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_gpr, a);
210
+ } else {
211
+ /* VMSR, move gp register to FP system register */
212
+ return gen_M_fp_sysreg_write(s, a->reg, gpr_to_fp_sysreg, a);
213
+ }
214
+}
215
+
216
static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
217
{
218
TCGv_i32 tmp;
219
bool ignore_vfp_enabled = false;
220
221
- if (!dc_isar_feature(aa32_fpsp_v2, s)) {
222
- return false;
223
+ if (arm_dc_feature(s, ARM_FEATURE_M)) {
224
+ return gen_M_VMSR_VMRS(s, a);
108
}
225
}
109
}
226
110
227
- if (arm_dc_feature(s, ARM_FEATURE_M)) {
111
+static int aa64_va_parameter_tcma(uint64_t tcr, ARMMMUIdx mmu_idx)
228
- /*
112
+{
229
- * The only M-profile VFP vmrs/vmsr sysreg is FPSCR.
113
+ if (regime_has_2_ranges(mmu_idx)) {
230
- * Accesses to R15 are UNPREDICTABLE; we choose to undef.
114
+ return extract64(tcr, 57, 2);
231
- * (FPSCR -> r15 is a special case which writes to the PSR flags.)
115
+ } else {
232
- */
116
+ /* Replicate the single TCMA bit so we always have 2 bits. */
233
- if (a->reg != ARM_VFP_FPSCR) {
117
+ return extract32(tcr, 30, 1) * 3;
234
- return false;
118
+ }
235
- }
119
+}
236
- if (a->rt == 15 && !a->l) {
120
+
237
- return false;
121
ARMVAParameters aa64_va_parameters(CPUARMState *env, uint64_t va,
238
- }
122
ARMMMUIdx mmu_idx, bool data)
239
+ if (!dc_isar_feature(aa32_fpsp_v2, s)) {
123
{
240
+ return false;
124
@@ -XXX,XX +XXX,XX @@ static uint32_t rebuild_hflags_a64(CPUARMState *env, int el, int fp_el,
125
}
126
}
241
}
127
242
128
+ if (cpu_isar_feature(aa64_mte, env_archcpu(env))) {
243
switch (a->reg) {
129
+ /*
130
+ * Set MTE_ACTIVE if any access may be Checked, and leave clear
131
+ * if all accesses must be Unchecked:
132
+ * 1) If no TBI, then there are no tags in the address to check,
133
+ * 2) If Tag Check Override, then all accesses are Unchecked,
134
+ * 3) If Tag Check Fail == 0, then Checked access have no effect,
135
+ * 4) If no Allocation Tag Access, then all accesses are Unchecked.
136
+ */
137
+ if (allocation_tag_access_enabled(env, el, sctlr)) {
138
+ flags = FIELD_DP32(flags, TBFLAG_A64, ATA, 1);
139
+ if (tbid
140
+ && !(env->pstate & PSTATE_TCO)
141
+ && (sctlr & (el == 0 ? SCTLR_TCF0 : SCTLR_TCF))) {
142
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE_ACTIVE, 1);
143
+ }
144
+ }
145
+ /* And again for unprivileged accesses, if required. */
146
+ if (FIELD_EX32(flags, TBFLAG_A64, UNPRIV)
147
+ && tbid
148
+ && !(env->pstate & PSTATE_TCO)
149
+ && (sctlr & SCTLR_TCF0)
150
+ && allocation_tag_access_enabled(env, 0, sctlr)) {
151
+ flags = FIELD_DP32(flags, TBFLAG_A64, MTE0_ACTIVE, 1);
152
+ }
153
+ /* Cache TCMA as well as TBI. */
154
+ flags = FIELD_DP32(flags, TBFLAG_A64, TCMA,
155
+ aa64_va_parameter_tcma(tcr, mmu_idx));
156
+ }
157
+
158
return rebuild_hflags_common(env, fp_el, mmu_idx, flags);
159
}
160
161
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
162
index XXXXXXX..XXXXXXX 100644
163
--- a/target/arm/translate-a64.c
164
+++ b/target/arm/translate-a64.c
165
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
166
dc->mmu_idx = core_to_aa64_mmu_idx(core_mmu_idx);
167
dc->tbii = FIELD_EX32(tb_flags, TBFLAG_A64, TBII);
168
dc->tbid = FIELD_EX32(tb_flags, TBFLAG_A64, TBID);
169
+ dc->tcma = FIELD_EX32(tb_flags, TBFLAG_A64, TCMA);
170
dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
171
#if !defined(CONFIG_USER_ONLY)
172
dc->user = (dc->current_el == 0);
173
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
174
dc->bt = FIELD_EX32(tb_flags, TBFLAG_A64, BT);
175
dc->btype = FIELD_EX32(tb_flags, TBFLAG_A64, BTYPE);
176
dc->unpriv = FIELD_EX32(tb_flags, TBFLAG_A64, UNPRIV);
177
+ dc->ata = FIELD_EX32(tb_flags, TBFLAG_A64, ATA);
178
+ dc->mte_active[0] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE_ACTIVE);
179
+ dc->mte_active[1] = FIELD_EX32(tb_flags, TBFLAG_A64, MTE0_ACTIVE);
180
dc->vec_len = 0;
181
dc->vec_stride = 0;
182
dc->cp_regs = arm_cpu->cp_regs;
183
--
244
--
184
2.20.1
245
2.20.1
185
246
186
247
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
The constant-expander functions like negate, plus_2, etc, are
2
generally useful; move them up in translate.c so we can use them in
3
the VFP/Neon decoders as well as in the A32/T32/T16 decoders.
2
4
3
We will shortly need this in mte_helper.c as well.
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201119215617.29887-9-peter.maydell@linaro.org
8
---
9
target/arm/translate.c | 46 +++++++++++++++++++++++-------------------
10
1 file changed, 25 insertions(+), 21 deletions(-)
4
11
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
diff --git a/target/arm/translate.c b/target/arm/translate.c
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-22-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 36 ++++++++++++++++++++++++++++++++++++
11
target/arm/helper.c | 36 ------------------------------------
12
2 files changed, 36 insertions(+), 36 deletions(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
14
--- a/target/arm/translate.c
17
+++ b/target/arm/internals.h
15
+++ b/target/arm/translate.c
18
@@ -XXX,XX +XXX,XX @@ static inline bool regime_is_pan(CPUARMState *env, ARMMMUIdx mmu_idx)
16
@@ -XXX,XX +XXX,XX @@ static void arm_gen_condlabel(DisasContext *s)
19
}
17
}
20
}
18
}
21
19
22
+/* Return the exception level which controls this address translation regime */
20
+/*
23
+static inline uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
21
+ * Constant expanders for the decoders.
22
+ */
23
+
24
+static int negate(DisasContext *s, int x)
24
+{
25
+{
25
+ switch (mmu_idx) {
26
+ return -x;
26
+ case ARMMMUIdx_E20_0:
27
+ case ARMMMUIdx_E20_2:
28
+ case ARMMMUIdx_E20_2_PAN:
29
+ case ARMMMUIdx_Stage2:
30
+ case ARMMMUIdx_E2:
31
+ return 2;
32
+ case ARMMMUIdx_SE3:
33
+ return 3;
34
+ case ARMMMUIdx_SE10_0:
35
+ return arm_el_is_aa64(env, 3) ? 1 : 3;
36
+ case ARMMMUIdx_SE10_1:
37
+ case ARMMMUIdx_SE10_1_PAN:
38
+ case ARMMMUIdx_Stage1_E0:
39
+ case ARMMMUIdx_Stage1_E1:
40
+ case ARMMMUIdx_Stage1_E1_PAN:
41
+ case ARMMMUIdx_E10_0:
42
+ case ARMMMUIdx_E10_1:
43
+ case ARMMMUIdx_E10_1_PAN:
44
+ case ARMMMUIdx_MPrivNegPri:
45
+ case ARMMMUIdx_MUserNegPri:
46
+ case ARMMMUIdx_MPriv:
47
+ case ARMMMUIdx_MUser:
48
+ case ARMMMUIdx_MSPrivNegPri:
49
+ case ARMMMUIdx_MSUserNegPri:
50
+ case ARMMMUIdx_MSPriv:
51
+ case ARMMMUIdx_MSUser:
52
+ return 1;
53
+ default:
54
+ g_assert_not_reached();
55
+ }
56
+}
27
+}
57
+
28
+
58
/* Return the FSR value for a debug exception (watchpoint, hardware
29
+static int plus_2(DisasContext *s, int x)
59
* breakpoint or BKPT insn) targeting the specified exception level.
30
+{
31
+ return x + 2;
32
+}
33
+
34
+static int times_2(DisasContext *s, int x)
35
+{
36
+ return x * 2;
37
+}
38
+
39
+static int times_4(DisasContext *s, int x)
40
+{
41
+ return x * 4;
42
+}
43
+
44
/* Flags for the disas_set_da_iss info argument:
45
* lower bits hold the Rt register number, higher bits are flags.
60
*/
46
*/
61
diff --git a/target/arm/helper.c b/target/arm/helper.c
47
@@ -XXX,XX +XXX,XX @@ static void arm_skip_unless(DisasContext *s, uint32_t cond)
62
index XXXXXXX..XXXXXXX 100644
48
63
--- a/target/arm/helper.c
49
64
+++ b/target/arm/helper.c
50
/*
65
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_interrupt(CPUState *cs)
51
- * Constant expanders for the decoders.
66
}
52
+ * Constant expanders used by T16/T32 decode
67
#endif /* !CONFIG_USER_ONLY */
53
*/
68
54
69
-/* Return the exception level which controls this address translation regime */
55
-static int negate(DisasContext *s, int x)
70
-static uint32_t regime_el(CPUARMState *env, ARMMMUIdx mmu_idx)
71
-{
56
-{
72
- switch (mmu_idx) {
57
- return -x;
73
- case ARMMMUIdx_E20_0:
74
- case ARMMMUIdx_E20_2:
75
- case ARMMMUIdx_E20_2_PAN:
76
- case ARMMMUIdx_Stage2:
77
- case ARMMMUIdx_E2:
78
- return 2;
79
- case ARMMMUIdx_SE3:
80
- return 3;
81
- case ARMMMUIdx_SE10_0:
82
- return arm_el_is_aa64(env, 3) ? 1 : 3;
83
- case ARMMMUIdx_SE10_1:
84
- case ARMMMUIdx_SE10_1_PAN:
85
- case ARMMMUIdx_Stage1_E0:
86
- case ARMMMUIdx_Stage1_E1:
87
- case ARMMMUIdx_Stage1_E1_PAN:
88
- case ARMMMUIdx_E10_0:
89
- case ARMMMUIdx_E10_1:
90
- case ARMMMUIdx_E10_1_PAN:
91
- case ARMMMUIdx_MPrivNegPri:
92
- case ARMMMUIdx_MUserNegPri:
93
- case ARMMMUIdx_MPriv:
94
- case ARMMMUIdx_MUser:
95
- case ARMMMUIdx_MSPrivNegPri:
96
- case ARMMMUIdx_MSUserNegPri:
97
- case ARMMMUIdx_MSPriv:
98
- case ARMMMUIdx_MSUser:
99
- return 1;
100
- default:
101
- g_assert_not_reached();
102
- }
103
-}
58
-}
104
-
59
-
105
uint64_t arm_sctlr(CPUARMState *env, int el)
60
-static int plus_2(DisasContext *s, int x)
61
-{
62
- return x + 2;
63
-}
64
-
65
-static int times_2(DisasContext *s, int x)
66
-{
67
- return x * 2;
68
-}
69
-
70
-static int times_4(DisasContext *s, int x)
71
-{
72
- return x * 4;
73
-}
74
-
75
/* Return only the rotation part of T32ExpandImm. */
76
static int t32_expandimm_rot(DisasContext *s, int x)
106
{
77
{
107
/* Only EL0 needs to be adjusted for EL1&0 or EL2&0. */
108
--
78
--
109
2.20.1
79
2.20.1
110
80
111
81
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
Implement the new-in-v8.1M VLDR/VSTR variants which directly
2
read or write FP system registers to memory.
2
3
3
Fill out the stub that was added earlier.
4
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20201119215617.29887-10-peter.maydell@linaro.org
7
---
8
target/arm/vfp.decode | 14 ++++++
9
target/arm/translate-vfp.c.inc | 91 ++++++++++++++++++++++++++++++++++
10
2 files changed, 105 insertions(+)
4
11
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
diff --git a/target/arm/vfp.decode b/target/arm/vfp.decode
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-26-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 48 +++++++++++++++
11
target/arm/mte_helper.c | 132 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 179 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
14
--- a/target/arm/vfp.decode
17
+++ b/target/arm/internals.h
15
+++ b/target/arm/vfp.decode
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, WRITE, 8, 1)
16
@@ -XXX,XX +XXX,XX @@ VLDR_VSTR_hp ---- 1101 u:1 .0 l:1 rn:4 .... 1001 imm:8 vd=%vd_sp
19
FIELD(MTEDESC, ESIZE, 9, 5)
17
VLDR_VSTR_sp ---- 1101 u:1 .0 l:1 rn:4 .... 1010 imm:8 vd=%vd_sp
20
FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
18
VLDR_VSTR_dp ---- 1101 u:1 .0 l:1 rn:4 .... 1011 imm:8 vd=%vd_dp
21
19
22
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
20
+# M-profile VLDR/VSTR to sysreg
23
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
21
+%vldr_sysreg 22:1 13:3
24
+ uint64_t ptr, uintptr_t ra);
22
+%imm7_0x4 0:7 !function=times_4
25
+
23
+
26
static inline int allocation_tag_from_addr(uint64_t ptr)
24
+&vldr_sysreg rn reg imm a w p
27
{
25
+@vldr_sysreg .... ... . a:1 . . . rn:4 ... . ... .. ....... \
28
return extract64(ptr, 56, 4);
26
+ reg=%vldr_sysreg imm=%imm7_0x4 &vldr_sysreg
29
@@ -XXX,XX +XXX,XX @@ static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
27
+
30
return deposit64(ptr, 56, 4, rtag);
28
+# P=0 W=0 is SEE "Related encodings", so split into two patterns
29
+VLDR_sysreg ---- 110 1 . . w:1 1 .... ... 0 111 11 ....... @vldr_sysreg p=1
30
+VLDR_sysreg ---- 110 0 . . 1 1 .... ... 0 111 11 ....... @vldr_sysreg p=0 w=1
31
+VSTR_sysreg ---- 110 1 . . w:1 0 .... ... 0 111 11 ....... @vldr_sysreg p=1
32
+VSTR_sysreg ---- 110 0 . . 1 0 .... ... 0 111 11 ....... @vldr_sysreg p=0 w=1
33
+
34
# We split the load/store multiple up into two patterns to avoid
35
# overlap with other insns in the "Advanced SIMD load/store and 64-bit move"
36
# grouping:
37
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
38
index XXXXXXX..XXXXXXX 100644
39
--- a/target/arm/translate-vfp.c.inc
40
+++ b/target/arm/translate-vfp.c.inc
41
@@ -XXX,XX +XXX,XX @@ static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
42
return true;
31
}
43
}
32
44
33
+/* Return true if tbi bits mean that the access is checked. */
45
+static void fp_sysreg_to_memory(DisasContext *s, void *opaque, TCGv_i32 value)
34
+static inline bool tbi_check(uint32_t desc, int bit55)
35
+{
46
+{
36
+ return (desc >> (R_MTEDESC_TBI_SHIFT + bit55)) & 1;
47
+ arg_vldr_sysreg *a = opaque;
37
+}
48
+ uint32_t offset = a->imm;
49
+ TCGv_i32 addr;
38
+
50
+
39
+/* Return true if tcma bits mean that the access is unchecked. */
51
+ if (!a->a) {
40
+static inline bool tcma_check(uint32_t desc, int bit55, int ptr_tag)
52
+ offset = - offset;
41
+{
42
+ /*
43
+ * We had extracted bit55 and ptr_tag for other reasons, so fold
44
+ * (ptr<59:55> == 00000 || ptr<59:55> == 11111) into a single test.
45
+ */
46
+ bool match = ((ptr_tag + bit55) & 0xf) == 0;
47
+ bool tcma = (desc >> (R_MTEDESC_TCMA_SHIFT + bit55)) & 1;
48
+ return tcma && match;
49
+}
50
+
51
+/*
52
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
53
+ * for the tag to be present in the FAR_ELx register. But for user-only
54
+ * mode, we do not have a TLB with which to implement this, so we must
55
+ * remove the top byte.
56
+ */
57
+static inline uint64_t useronly_clean_ptr(uint64_t ptr)
58
+{
59
+ /* TBI is known to be enabled. */
60
+#ifdef CONFIG_USER_ONLY
61
+ ptr = sextract64(ptr, 0, 56);
62
+#endif
63
+ return ptr;
64
+}
65
+
66
+static inline uint64_t useronly_maybe_clean_ptr(uint32_t desc, uint64_t ptr)
67
+{
68
+#ifdef CONFIG_USER_ONLY
69
+ int64_t clean_ptr = sextract64(ptr, 0, 56);
70
+ if (tbi_check(desc, clean_ptr < 0)) {
71
+ ptr = clean_ptr;
72
+ }
73
+#endif
74
+ return ptr;
75
+}
76
+
77
#endif
78
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
79
index XXXXXXX..XXXXXXX 100644
80
--- a/target/arm/mte_helper.c
81
+++ b/target/arm/mte_helper.c
82
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
83
}
84
}
85
86
+/* Record a tag check failure. */
87
+static void mte_check_fail(CPUARMState *env, int mmu_idx,
88
+ uint64_t dirty_ptr, uintptr_t ra)
89
+{
90
+ ARMMMUIdx arm_mmu_idx = core_to_aa64_mmu_idx(mmu_idx);
91
+ int el, reg_el, tcf, select;
92
+ uint64_t sctlr;
93
+
94
+ reg_el = regime_el(env, arm_mmu_idx);
95
+ sctlr = env->cp15.sctlr_el[reg_el];
96
+
97
+ switch (arm_mmu_idx) {
98
+ case ARMMMUIdx_E10_0:
99
+ case ARMMMUIdx_E20_0:
100
+ el = 0;
101
+ tcf = extract64(sctlr, 38, 2);
102
+ break;
103
+ default:
104
+ el = reg_el;
105
+ tcf = extract64(sctlr, 40, 2);
106
+ }
53
+ }
107
+
54
+
108
+ switch (tcf) {
55
+ addr = load_reg(s, a->rn);
109
+ case 1:
56
+ if (a->p) {
110
+ /*
57
+ tcg_gen_addi_i32(addr, addr, offset);
111
+ * Tag check fail causes a synchronous exception.
58
+ }
112
+ *
113
+ * In restore_state_to_opc, we set the exception syndrome
114
+ * for the load or store operation. Unwind first so we
115
+ * may overwrite that with the syndrome for the tag check.
116
+ */
117
+ cpu_restore_state(env_cpu(env), ra, true);
118
+ env->exception.vaddress = dirty_ptr;
119
+ raise_exception(env, EXCP_DATA_ABORT,
120
+ syn_data_abort_no_iss(el != 0, 0, 0, 0, 0, 0, 0x11),
121
+ exception_target_el(env));
122
+ /* noreturn, but fall through to the assert anyway */
123
+
59
+
124
+ case 0:
60
+ if (s->v8m_stackcheck && a->rn == 13 && a->w) {
125
+ /*
61
+ gen_helper_v8m_stackcheck(cpu_env, addr);
126
+ * Tag check fail does not affect the PE.
62
+ }
127
+ * We eliminate this case by not setting MTE_ACTIVE
128
+ * in tb_flags, so that we never make this runtime call.
129
+ */
130
+ g_assert_not_reached();
131
+
63
+
132
+ case 2:
64
+ gen_aa32_st_i32(s, value, addr, get_mem_index(s),
133
+ /* Tag check fail causes asynchronous flag set. */
65
+ MO_UL | MO_ALIGN | s->be_data);
134
+ mmu_idx = arm_mmu_idx_el(env, el);
66
+ tcg_temp_free_i32(value);
135
+ if (regime_has_2_ranges(mmu_idx)) {
67
+
136
+ select = extract64(dirty_ptr, 55, 1);
68
+ if (a->w) {
137
+ } else {
69
+ /* writeback */
138
+ select = 0;
70
+ if (!a->p) {
71
+ tcg_gen_addi_i32(addr, addr, offset);
139
+ }
72
+ }
140
+ env->cp15.tfsr_el[el] |= 1 << select;
73
+ store_reg(s, a->rn, addr);
141
+ break;
74
+ } else {
142
+
75
+ tcg_temp_free_i32(addr);
143
+ default:
144
+ /* Case 3: Reserved. */
145
+ qemu_log_mask(LOG_GUEST_ERROR,
146
+ "Tag check failure with SCTLR_EL%d.TCF%s "
147
+ "set to reserved value %d\n",
148
+ reg_el, el ? "" : "0", tcf);
149
+ break;
150
+ }
76
+ }
151
+}
77
+}
152
+
78
+
153
/*
79
+static TCGv_i32 memory_to_fp_sysreg(DisasContext *s, void *opaque)
154
* Perform an MTE checked access for a single logical or atomic access.
155
*/
156
+static bool mte_probe1_int(CPUARMState *env, uint32_t desc, uint64_t ptr,
157
+ uintptr_t ra, int bit55)
158
+{
80
+{
159
+ int mem_tag, mmu_idx, ptr_tag, size;
81
+ arg_vldr_sysreg *a = opaque;
160
+ MMUAccessType type;
82
+ uint32_t offset = a->imm;
161
+ uint8_t *mem;
83
+ TCGv_i32 addr;
84
+ TCGv_i32 value = tcg_temp_new_i32();
162
+
85
+
163
+ ptr_tag = allocation_tag_from_addr(ptr);
86
+ if (!a->a) {
164
+
87
+ offset = - offset;
165
+ if (tcma_check(desc, bit55, ptr_tag)) {
166
+ return true;
167
+ }
88
+ }
168
+
89
+
169
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
90
+ addr = load_reg(s, a->rn);
170
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
91
+ if (a->p) {
171
+ size = FIELD_EX32(desc, MTEDESC, ESIZE);
92
+ tcg_gen_addi_i32(addr, addr, offset);
172
+
173
+ mem = allocation_tag_mem(env, mmu_idx, ptr, type, size,
174
+ MMU_DATA_LOAD, 1, ra);
175
+ if (!mem) {
176
+ return true;
177
+ }
93
+ }
178
+
94
+
179
+ mem_tag = load_tag1(ptr, mem);
95
+ if (s->v8m_stackcheck && a->rn == 13 && a->w) {
180
+ return ptr_tag == mem_tag;
96
+ gen_helper_v8m_stackcheck(cpu_env, addr);
97
+ }
98
+
99
+ gen_aa32_ld_i32(s, value, addr, get_mem_index(s),
100
+ MO_UL | MO_ALIGN | s->be_data);
101
+
102
+ if (a->w) {
103
+ /* writeback */
104
+ if (!a->p) {
105
+ tcg_gen_addi_i32(addr, addr, offset);
106
+ }
107
+ store_reg(s, a->rn, addr);
108
+ } else {
109
+ tcg_temp_free_i32(addr);
110
+ }
111
+ return value;
181
+}
112
+}
182
+
113
+
183
+/*
114
+static bool trans_VLDR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
184
+ * No-fault version of mte_check1, to be used by SVE for MemSingleNF.
185
+ * Returns false if the access is Checked and the check failed. This
186
+ * is only intended to probe the tag -- the validity of the page must
187
+ * be checked beforehand.
188
+ */
189
+bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr)
190
+{
115
+{
191
+ int bit55 = extract64(ptr, 55, 1);
116
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
192
+
117
+ return false;
193
+ /* If TBI is disabled, the access is unchecked. */
194
+ if (unlikely(!tbi_check(desc, bit55))) {
195
+ return true;
196
+ }
118
+ }
197
+
119
+ if (a->rn == 15) {
198
+ return mte_probe1_int(env, desc, ptr, 0, bit55);
120
+ return false;
121
+ }
122
+ return gen_M_fp_sysreg_write(s, a->reg, memory_to_fp_sysreg, a);
199
+}
123
+}
200
+
124
+
201
+uint64_t mte_check1(CPUARMState *env, uint32_t desc,
125
+static bool trans_VSTR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
202
+ uint64_t ptr, uintptr_t ra)
203
+{
126
+{
204
+ int bit55 = extract64(ptr, 55, 1);
127
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
205
+
128
+ return false;
206
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
207
+ if (unlikely(!tbi_check(desc, bit55))) {
208
+ return ptr;
209
+ }
129
+ }
210
+
130
+ if (a->rn == 15) {
211
+ if (unlikely(!mte_probe1_int(env, desc, ptr, ra, bit55))) {
131
+ return false;
212
+ int mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
213
+ mte_check_fail(env, mmu_idx, ptr, ra);
214
+ }
132
+ }
215
+
133
+ return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_memory, a);
216
+ return useronly_clean_ptr(ptr);
217
+}
134
+}
218
+
135
+
219
uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
136
static bool trans_VMOV_half(DisasContext *s, arg_VMOV_single *a)
220
{
137
{
221
- return ptr;
138
TCGv_i32 tmp;
222
+ return mte_check1(env, desc, ptr, GETPC());
223
}
224
225
/*
226
--
139
--
227
2.20.1
140
2.20.1
228
141
229
142
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
v8.1M defines a new FP system register FPSCR_nzcvqc; this behaves
2
like the existing FPSCR, except that it reads and writes only bits
3
[31:27] of the FPSCR (the N, Z, C, V and QC flag bits). (Unlike the
4
FPSCR, the special case for Rt=15 of writing the CPSR.NZCV is not
5
permitted.)
2
6
3
This is DC GVA and DC GZVA, and the tag check for DC ZVA.
7
Implement the register. Since we don't yet implement MVE, we handle
8
the QC bit as RES0, with todo comments for where we will need to add
9
support later.
4
10
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-40-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20201119215617.29887-11-peter.maydell@linaro.org
9
---
14
---
10
target/arm/cpu.h | 4 +++-
15
target/arm/cpu.h | 13 +++++++++++++
11
target/arm/helper.c | 16 ++++++++++++++++
16
target/arm/translate-vfp.c.inc | 27 +++++++++++++++++++++++++++
12
target/arm/translate-a64.c | 39 ++++++++++++++++++++++++++++++++++++++
17
2 files changed, 40 insertions(+)
13
3 files changed, 58 insertions(+), 1 deletion(-)
14
18
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
19
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
16
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.h
21
--- a/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
22
+++ b/target/arm/cpu.h
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t cpreg_to_kvm_id(uint32_t cpregid)
23
@@ -XXX,XX +XXX,XX @@ void vfp_set_fpscr(CPUARMState *env, uint32_t val);
20
#define ARM_CP_NZCV (ARM_CP_SPECIAL | 0x0300)
24
#define FPCR_FZ (1 << 24) /* Flush-to-zero enable bit */
21
#define ARM_CP_CURRENTEL (ARM_CP_SPECIAL | 0x0400)
25
#define FPCR_DN (1 << 25) /* Default NaN enable bit */
22
#define ARM_CP_DC_ZVA (ARM_CP_SPECIAL | 0x0500)
26
#define FPCR_QC (1 << 27) /* Cumulative saturation bit */
23
-#define ARM_LAST_SPECIAL ARM_CP_DC_ZVA
27
+#define FPCR_V (1 << 28) /* FP overflow flag */
24
+#define ARM_CP_DC_GVA (ARM_CP_SPECIAL | 0x0600)
28
+#define FPCR_C (1 << 29) /* FP carry flag */
25
+#define ARM_CP_DC_GZVA (ARM_CP_SPECIAL | 0x0700)
29
+#define FPCR_Z (1 << 30) /* FP zero flag */
26
+#define ARM_LAST_SPECIAL ARM_CP_DC_GZVA
30
+#define FPCR_N (1 << 31) /* FP negative flag */
27
#define ARM_CP_FPU 0x1000
31
+
28
#define ARM_CP_SVE 0x2000
32
+#define FPCR_NZCV_MASK (FPCR_N | FPCR_Z | FPCR_C | FPCR_V)
29
#define ARM_CP_NO_GDB 0x4000
33
+#define FPCR_NZCVQC_MASK (FPCR_NZCV_MASK | FPCR_QC)
30
diff --git a/target/arm/helper.c b/target/arm/helper.c
34
35
static inline uint32_t vfp_get_fpsr(CPUARMState *env)
36
{
37
@@ -XXX,XX +XXX,XX @@ enum arm_cpu_mode {
38
#define ARM_VFP_FPEXC 8
39
#define ARM_VFP_FPINST 9
40
#define ARM_VFP_FPINST2 10
41
+/* These ones are M-profile only */
42
+#define ARM_VFP_FPSCR_NZCVQC 2
43
+#define ARM_VFP_VPR 12
44
+#define ARM_VFP_P0 13
45
+#define ARM_VFP_FPCXT_NS 14
46
+#define ARM_VFP_FPCXT_S 15
47
48
/* QEMU-internal value meaning "FPSCR, but we care only about NZCV" */
49
#define QEMU_VFP_FPSCR_NZCV 0xffff
50
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
31
index XXXXXXX..XXXXXXX 100644
51
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.c
52
--- a/target/arm/translate-vfp.c.inc
33
+++ b/target/arm/helper.c
53
+++ b/target/arm/translate-vfp.c.inc
34
@@ -XXX,XX +XXX,XX @@ static const ARMCPRegInfo mte_el0_cacheop_reginfo[] = {
54
@@ -XXX,XX +XXX,XX @@ static FPSysRegCheckResult fp_sysreg_checks(DisasContext *s, int regno)
35
.opc0 = 1, .opc1 = 3, .crn = 7, .crm = 14, .opc2 = 5,
55
case ARM_VFP_FPSCR:
36
.type = ARM_CP_NOP, .access = PL0_W,
56
case QEMU_VFP_FPSCR_NZCV:
37
.accessfn = aa64_cacheop_poc_access },
57
break;
38
+ { .name = "DC_GVA", .state = ARM_CP_STATE_AA64,
58
+ case ARM_VFP_FPSCR_NZCVQC:
39
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 3,
59
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
40
+ .access = PL0_W, .type = ARM_CP_DC_GVA,
60
+ return false;
41
+#ifndef CONFIG_USER_ONLY
42
+ /* Avoid overhead of an access check that always passes in user-mode */
43
+ .accessfn = aa64_zva_access,
44
+#endif
45
+ },
46
+ { .name = "DC_GZVA", .state = ARM_CP_STATE_AA64,
47
+ .opc0 = 1, .opc1 = 3, .crn = 7, .crm = 4, .opc2 = 4,
48
+ .access = PL0_W, .type = ARM_CP_DC_GZVA,
49
+#ifndef CONFIG_USER_ONLY
50
+ /* Avoid overhead of an access check that always passes in user-mode */
51
+ .accessfn = aa64_zva_access,
52
+#endif
53
+ },
54
REGINFO_SENTINEL
55
};
56
57
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
58
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/translate-a64.c
60
+++ b/target/arm/translate-a64.c
61
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
62
}
63
gen_helper_dc_zva(cpu_env, tcg_rt);
64
return;
65
+ case ARM_CP_DC_GVA:
66
+ {
67
+ TCGv_i64 clean_addr, tag;
68
+
69
+ /*
70
+ * DC_GVA, like DC_ZVA, requires that we supply the original
71
+ * pointer for an invalid page. Probe that address first.
72
+ */
73
+ tcg_rt = cpu_reg(s, rt);
74
+ clean_addr = clean_data_tbi(s, tcg_rt);
75
+ gen_probe_access(s, clean_addr, MMU_DATA_STORE, MO_8);
76
+
77
+ if (s->ata) {
78
+ /* Extract the tag from the register to match STZGM. */
79
+ tag = tcg_temp_new_i64();
80
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
81
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
82
+ tcg_temp_free_i64(tag);
83
+ }
84
+ }
61
+ }
85
+ return;
62
+ break;
86
+ case ARM_CP_DC_GZVA:
87
+ {
88
+ TCGv_i64 clean_addr, tag;
89
+
90
+ /* For DC_GZVA, we can rely on DC_ZVA for the proper fault. */
91
+ tcg_rt = cpu_reg(s, rt);
92
+ clean_addr = clean_data_tbi(s, tcg_rt);
93
+ gen_helper_dc_zva(cpu_env, clean_addr);
94
+
95
+ if (s->ata) {
96
+ /* Extract the tag from the register to match STZGM. */
97
+ tag = tcg_temp_new_i64();
98
+ tcg_gen_shri_i64(tag, tcg_rt, 56);
99
+ gen_helper_stzgm_tags(cpu_env, clean_addr, tag);
100
+ tcg_temp_free_i64(tag);
101
+ }
102
+ }
103
+ return;
104
default:
63
default:
64
return FPSysRegCheckFailed;
65
}
66
@@ -XXX,XX +XXX,XX @@ static bool gen_M_fp_sysreg_write(DisasContext *s, int regno,
67
tcg_temp_free_i32(tmp);
68
gen_lookup_tb(s);
105
break;
69
break;
70
+ case ARM_VFP_FPSCR_NZCVQC:
71
+ {
72
+ TCGv_i32 fpscr;
73
+ tmp = loadfn(s, opaque);
74
+ /*
75
+ * TODO: when we implement MVE, write the QC bit.
76
+ * For non-MVE, QC is RES0.
77
+ */
78
+ tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
79
+ fpscr = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
80
+ tcg_gen_andi_i32(fpscr, fpscr, ~FPCR_NZCV_MASK);
81
+ tcg_gen_or_i32(fpscr, fpscr, tmp);
82
+ store_cpu_field(fpscr, vfp.xregs[ARM_VFP_FPSCR]);
83
+ tcg_temp_free_i32(tmp);
84
+ break;
85
+ }
86
default:
87
g_assert_not_reached();
106
}
88
}
89
@@ -XXX,XX +XXX,XX @@ static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
90
gen_helper_vfp_get_fpscr(tmp, cpu_env);
91
storefn(s, opaque, tmp);
92
break;
93
+ case ARM_VFP_FPSCR_NZCVQC:
94
+ /*
95
+ * TODO: MVE has a QC bit, which we probably won't store
96
+ * in the xregs[] field. For non-MVE, where QC is RES0,
97
+ * we can just fall through to the FPSCR_NZCV case.
98
+ */
99
case QEMU_VFP_FPSCR_NZCV:
100
/*
101
* Read just NZCV; this is a special case to avoid the
107
--
102
--
108
2.20.1
103
2.20.1
109
104
110
105
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
We defined a constant name for the mask of NZCV bits in the FPCR/FPSCR
2
in the previous commit; use it in a couple of places in existing code,
3
where we're masking out everything except NZCV for the "load to Rt=15
4
sets CPSR.NZCV" special case.
2
5
3
There are a number of paths by which the TBI is still intact
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
for user-only in the SVE helpers.
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20201119215617.29887-12-peter.maydell@linaro.org
9
---
10
target/arm/translate-vfp.c.inc | 4 ++--
11
1 file changed, 2 insertions(+), 2 deletions(-)
5
12
6
Because we currently always set TBI for user-only, we do not
13
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
7
need to pass down the actual TBI setting from above, and we
8
can remove the top byte in the inner-most primitives, so that
9
none are forgotten. Moreover, this keeps the "dirty" pointer
10
around at the higher levels, where we need it for any MTE checking.
11
12
Since the normal case, especially for user-only, goes through
13
RAM, this clearing merely adds two insns per page lookup, which
14
will be completely in the noise.
15
16
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
17
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20200626033144.790098-39-richard.henderson@linaro.org
19
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
20
---
21
target/arm/cpu.c | 3 +++
22
target/arm/sve_helper.c | 19 +++++++++++++++++--
23
target/arm/translate-a64.c | 5 +++++
24
3 files changed, 25 insertions(+), 2 deletions(-)
25
26
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
27
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
28
--- a/target/arm/cpu.c
15
--- a/target/arm/translate-vfp.c.inc
29
+++ b/target/arm/cpu.c
16
+++ b/target/arm/translate-vfp.c.inc
30
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
17
@@ -XXX,XX +XXX,XX @@ static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
31
* Enable TBI0 and TBI1. While the real kernel only enables TBI0,
18
* helper call for the "VMRS to CPSR.NZCV" insn.
32
* turning on both here will produce smaller code and otherwise
33
* make no difference to the user-level emulation.
34
+ *
35
+ * In sve_probe_page, we assume that this is set.
36
+ * Do not modify this without other changes.
37
*/
19
*/
38
env->cp15.tcr_el[1].raw_tcr = (3ULL << 37);
20
tmp = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
39
#else
21
- tcg_gen_andi_i32(tmp, tmp, 0xf0000000);
40
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
22
+ tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
41
index XXXXXXX..XXXXXXX 100644
23
storefn(s, opaque, tmp);
42
--- a/target/arm/sve_helper.c
24
break;
43
+++ b/target/arm/sve_helper.c
25
default:
44
@@ -XXX,XX +XXX,XX @@ static void sve_##NAME##_host(void *vd, intptr_t reg_off, void *host) \
26
@@ -XXX,XX +XXX,XX @@ static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
45
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
27
case ARM_VFP_FPSCR:
46
target_ulong addr, uintptr_t ra) \
28
if (a->rt == 15) {
47
{ \
29
tmp = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
48
- *(TYPEE *)(vd + H(reg_off)) = (TYPEM)TLB(env, addr, ra); \
30
- tcg_gen_andi_i32(tmp, tmp, 0xf0000000);
49
+ *(TYPEE *)(vd + H(reg_off)) = \
31
+ tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
50
+ (TYPEM)TLB(env, useronly_clean_ptr(addr), ra); \
32
} else {
51
}
33
tmp = tcg_temp_new_i32();
52
34
gen_helper_vfp_get_fpscr(tmp, cpu_env);
53
#define DO_ST_TLB(NAME, H, TYPEE, TYPEM, TLB) \
54
static void sve_##NAME##_tlb(CPUARMState *env, void *vd, intptr_t reg_off, \
55
target_ulong addr, uintptr_t ra) \
56
{ \
57
- TLB(env, addr, (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
58
+ TLB(env, useronly_clean_ptr(addr), \
59
+ (TYPEM)*(TYPEE *)(vd + H(reg_off)), ra); \
60
}
61
62
#define DO_LD_PRIM_1(NAME, H, TE, TM) \
63
@@ -XXX,XX +XXX,XX @@ static bool sve_probe_page(SVEHostPage *info, bool nofault,
64
int flags;
65
66
addr += mem_off;
67
+
68
+ /*
69
+ * User-only currently always issues with TBI. See the comment
70
+ * above useronly_clean_ptr. Usually we clean this top byte away
71
+ * during translation, but we can't do that for e.g. vector + imm
72
+ * addressing modes.
73
+ *
74
+ * We currently always enable TBI for user-only, and do not provide
75
+ * a way to turn it off. So clean the pointer unconditionally here,
76
+ * rather than look it up here, or pass it down from above.
77
+ */
78
+ addr = useronly_clean_ptr(addr);
79
+
80
flags = probe_access_flags(env, addr, access_type, mmu_idx, nofault,
81
&info->host, retaddr);
82
info->flags = flags;
83
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
84
index XXXXXXX..XXXXXXX 100644
85
--- a/target/arm/translate-a64.c
86
+++ b/target/arm/translate-a64.c
87
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_init_disas_context(DisasContextBase *dcbase,
88
dc->features = env->features;
89
dc->dcz_blocksize = arm_cpu->dcz_blocksize;
90
91
+#ifdef CONFIG_USER_ONLY
92
+ /* In sve_probe_page, we assume TBI is enabled. */
93
+ tcg_debug_assert(dc->tbid & 1);
94
+#endif
95
+
96
/* Single step state. The code-generation logic here is:
97
* SS_ACTIVE == 0:
98
* generate code with no special handling for single-stepping (except
99
--
35
--
100
2.20.1
36
2.20.1
101
37
102
38
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
Factor out the code which handles M-profile lazy FP state preservation
2
from full_vfp_access_check(); accesses to the FPCXT_NS register are
3
a special case which need to do just this part (corresponding in the
4
pseudocode to the PreserveFPState() function), and not the full
5
set of actions matching the pseudocode ExecuteFPCheck() which
6
normal FP instructions need to do.
2
7
3
Add a trivial representation of the PCA9552 GPIOs.
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
10
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Message-id: 20201119215617.29887-13-peter.maydell@linaro.org
12
---
13
target/arm/translate-vfp.c.inc | 45 ++++++++++++++++++++--------------
14
1 file changed, 27 insertions(+), 18 deletions(-)
4
15
5
Example booting obmc-phosphor-image:
16
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_status
8
1592689902.327837:pca955x_gpio_status pca-unspecified GPIOs 0-15 [*...............]
9
1592689902.329934:pca955x_gpio_status pca-unspecified GPIOs 0-15 [**..............]
10
1592689902.330717:pca955x_gpio_status pca-unspecified GPIOs 0-15 [***.............]
11
1592689902.331431:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****............]
12
1592689902.332163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*..]
13
1592689902.332888:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........**.]
14
1592689902.333629:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
15
1592690032.793289:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
16
1592690033.303163:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
17
1592690033.812962:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
18
1592690034.323234:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........***]
19
1592690034.832922:pca955x_gpio_status pca-unspecified GPIOs 0-15 [****.........*.*]
20
21
We notice the GPIO #14 (front-power LED) starts to blink.
22
23
This LED is described in the witherspoon device-tree [*]:
24
25
front-power {
26
retain-state-shutdown;
27
default-state = "keep";
28
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
29
};
30
31
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
32
33
Suggested-by: Cédric Le Goater <clg@kaod.org>
34
Reviewed-by: Cédric Le Goater <clg@kaod.org>
35
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
36
Tested-by: Cédric Le Goater <clg@kaod.org>
37
Message-id: 20200623072723.6324-7-f4bug@amsat.org
38
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
39
---
40
hw/misc/pca9552.c | 39 +++++++++++++++++++++++++++++++++++++++
41
hw/misc/trace-events | 3 +++
42
2 files changed, 42 insertions(+)
43
44
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
45
index XXXXXXX..XXXXXXX 100644
17
index XXXXXXX..XXXXXXX 100644
46
--- a/hw/misc/pca9552.c
18
--- a/target/arm/translate-vfp.c.inc
47
+++ b/hw/misc/pca9552.c
19
+++ b/target/arm/translate-vfp.c.inc
48
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@ static inline long vfp_f16_offset(unsigned reg, bool top)
49
#include "qemu/osdep.h"
21
return offs;
50
#include "qemu/log.h"
51
#include "qemu/module.h"
52
+#include "qemu/bitops.h"
53
#include "hw/qdev-properties.h"
54
#include "hw/misc/pca9552.h"
55
#include "hw/misc/pca9552_regs.h"
56
#include "migration/vmstate.h"
57
#include "qapi/error.h"
58
#include "qapi/visitor.h"
59
+#include "trace.h"
60
61
typedef struct PCA955xClass {
62
/*< private >*/
63
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_pin_get_config(PCA955xState *s, int pin)
64
return extract32(s->regs[reg], shift, 2);
65
}
22
}
66
23
67
+/* Return INPUT status (bit #N belongs to GPIO #N) */
24
+/*
68
+static uint16_t pca955x_pins_get_status(PCA955xState *s)
25
+ * Generate code for M-profile lazy FP state preservation if needed;
26
+ * this corresponds to the pseudocode PreserveFPState() function.
27
+ */
28
+static void gen_preserve_fp_state(DisasContext *s)
69
+{
29
+{
70
+ return (s->regs[PCA9552_INPUT1] << 8) | s->regs[PCA9552_INPUT0];
30
+ if (s->v7m_lspact) {
71
+}
31
+ /*
72
+
32
+ * Lazy state saving affects external memory and also the NVIC,
73
+static void pca955x_display_pins_status(PCA955xState *s,
33
+ * so we must mark it as an IO operation for icount (and cause
74
+ uint16_t previous_pins_status)
34
+ * this to be the last insn in the TB).
75
+{
35
+ */
76
+ PCA955xClass *k = PCA955X_GET_CLASS(s);
36
+ if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
77
+ uint16_t pins_status, pins_changed;
37
+ s->base.is_jmp = DISAS_UPDATE_EXIT;
78
+ int i;
38
+ gen_io_start();
79
+
80
+ pins_status = pca955x_pins_get_status(s);
81
+ pins_changed = previous_pins_status ^ pins_status;
82
+ if (!pins_changed) {
83
+ return;
84
+ }
85
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_STATUS)) {
86
+ char *buf = g_newa(char, k->pin_count + 1);
87
+
88
+ for (i = 0; i < k->pin_count; i++) {
89
+ if (extract32(pins_status, i, 1)) {
90
+ buf[i] = '*';
91
+ } else {
92
+ buf[i] = '.';
93
+ }
94
+ }
39
+ }
95
+ buf[i] = '\0';
40
+ gen_helper_v7m_preserve_fp_state(cpu_env);
96
+ trace_pca955x_gpio_status(s->description, buf);
41
+ /*
42
+ * If the preserve_fp_state helper doesn't throw an exception
43
+ * then it will clear LSPACT; we don't need to repeat this for
44
+ * any further FP insns in this TB.
45
+ */
46
+ s->v7m_lspact = false;
97
+ }
47
+ }
98
+}
48
+}
99
+
49
+
100
static void pca955x_update_pin_input(PCA955xState *s)
50
/*
101
{
51
* Check that VFP access is enabled. If it is, do the necessary
102
PCA955xClass *k = PCA955X_GET_CLASS(s);
52
* M-profile lazy-FP handling and then return true.
103
@@ -XXX,XX +XXX,XX @@ static uint8_t pca955x_read(PCA955xState *s, uint8_t reg)
53
@@ -XXX,XX +XXX,XX @@ static bool full_vfp_access_check(DisasContext *s, bool ignore_vfp_enabled)
104
54
/* Handle M-profile lazy FP state mechanics */
105
static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
55
106
{
56
/* Trigger lazy-state preservation if necessary */
107
+ uint16_t pins_status;
57
- if (s->v7m_lspact) {
108
+
58
- /*
109
switch (reg) {
59
- * Lazy state saving affects external memory and also the NVIC,
110
case PCA9552_PSC0:
60
- * so we must mark it as an IO operation for icount (and cause
111
case PCA9552_PWM0:
61
- * this to be the last insn in the TB).
112
@@ -XXX,XX +XXX,XX @@ static void pca955x_write(PCA955xState *s, uint8_t reg, uint8_t data)
62
- */
113
case PCA9552_LS1:
63
- if (tb_cflags(s->base.tb) & CF_USE_ICOUNT) {
114
case PCA9552_LS2:
64
- s->base.is_jmp = DISAS_UPDATE_EXIT;
115
case PCA9552_LS3:
65
- gen_io_start();
116
+ pins_status = pca955x_pins_get_status(s);
66
- }
117
s->regs[reg] = data;
67
- gen_helper_v7m_preserve_fp_state(cpu_env);
118
pca955x_update_pin_input(s);
68
- /*
119
+ pca955x_display_pins_status(s, pins_status);
69
- * If the preserve_fp_state helper doesn't throw an exception
120
break;
70
- * then it will clear LSPACT; we don't need to repeat this for
121
71
- * any further FP insns in this TB.
122
case PCA9552_INPUT0:
72
- */
123
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
- s->v7m_lspact = false;
124
index XXXXXXX..XXXXXXX 100644
74
- }
125
--- a/hw/misc/trace-events
75
+ gen_preserve_fp_state(s);
126
+++ b/hw/misc/trace-events
76
127
@@ -XXX,XX +XXX,XX @@ via1_adb_poll(uint8_t data, const char *vadbint, int status, int index, int size
77
/* Update ownership of FP context: set FPCCR.S to match current state */
128
# grlib_ahb_apb_pnp.c
78
if (s->v8m_fpccr_s_wrong) {
129
grlib_ahb_pnp_read(uint64_t addr, uint32_t value) "AHB PnP read addr:0x%03"PRIx64" data:0x%08x"
130
grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx64" data:0x%08x"
131
+
132
+# pca9552.c
133
+pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
134
--
79
--
135
2.20.1
80
2.20.1
136
81
137
82
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
Implement the new-in-v8.1M FPCXT_S floating point system register.
2
This is for saving and restoring the secure floating point context,
3
and it reads and writes bits [27:0] from the FPSCR and the
4
CONTROL.SFPA bit in bit [31].
2
5
3
Because the elements are non-sequential, we cannot eliminate many
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
tests straight away like we can for sequential operations. But
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
5
we often have the PTE details handy, so we can test for Tagged.
8
Message-id: 20201119215617.29887-14-peter.maydell@linaro.org
9
---
10
target/arm/translate-vfp.c.inc | 58 ++++++++++++++++++++++++++++++++++
11
1 file changed, 58 insertions(+)
6
12
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
13
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-38-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-sve.h | 285 ++++++++++++++++
13
target/arm/sve_helper.c | 185 +++++++++--
14
target/arm/translate-sve.c | 650 +++++++++++++++++++++++++------------
15
3 files changed, 872 insertions(+), 248 deletions(-)
16
17
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
18
index XXXXXXX..XXXXXXX 100644
14
index XXXXXXX..XXXXXXX 100644
19
--- a/target/arm/helper-sve.h
15
--- a/target/arm/translate-vfp.c.inc
20
+++ b/target/arm/helper-sve.h
16
+++ b/target/arm/translate-vfp.c.inc
21
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldsds_le_zd, TCG_CALL_NO_WG,
17
@@ -XXX,XX +XXX,XX @@ static FPSysRegCheckResult fp_sysreg_checks(DisasContext *s, int regno)
22
DEF_HELPER_FLAGS_6(sve_ldsds_be_zd, TCG_CALL_NO_WG,
18
return false;
23
void, env, ptr, ptr, ptr, tl, i32)
19
}
24
25
+DEF_HELPER_FLAGS_6(sve_ldbsu_zsu_mte, TCG_CALL_NO_WG,
26
+ void, env, ptr, ptr, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu_mte, TCG_CALL_NO_WG,
28
+ void, env, ptr, ptr, ptr, tl, i32)
29
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zsu_mte, TCG_CALL_NO_WG,
30
+ void, env, ptr, ptr, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_6(sve_ldss_le_zsu_mte, TCG_CALL_NO_WG,
32
+ void, env, ptr, ptr, ptr, tl, i32)
33
+DEF_HELPER_FLAGS_6(sve_ldss_be_zsu_mte, TCG_CALL_NO_WG,
34
+ void, env, ptr, ptr, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_6(sve_ldbss_zsu_mte, TCG_CALL_NO_WG,
36
+ void, env, ptr, ptr, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zsu_mte, TCG_CALL_NO_WG,
38
+ void, env, ptr, ptr, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zsu_mte, TCG_CALL_NO_WG,
40
+ void, env, ptr, ptr, ptr, tl, i32)
41
+
42
+DEF_HELPER_FLAGS_6(sve_ldbsu_zss_mte, TCG_CALL_NO_WG,
43
+ void, env, ptr, ptr, ptr, tl, i32)
44
+DEF_HELPER_FLAGS_6(sve_ldhsu_le_zss_mte, TCG_CALL_NO_WG,
45
+ void, env, ptr, ptr, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_6(sve_ldhsu_be_zss_mte, TCG_CALL_NO_WG,
47
+ void, env, ptr, ptr, ptr, tl, i32)
48
+DEF_HELPER_FLAGS_6(sve_ldss_le_zss_mte, TCG_CALL_NO_WG,
49
+ void, env, ptr, ptr, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_6(sve_ldss_be_zss_mte, TCG_CALL_NO_WG,
51
+ void, env, ptr, ptr, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_6(sve_ldbss_zss_mte, TCG_CALL_NO_WG,
53
+ void, env, ptr, ptr, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_6(sve_ldhss_le_zss_mte, TCG_CALL_NO_WG,
55
+ void, env, ptr, ptr, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_6(sve_ldhss_be_zss_mte, TCG_CALL_NO_WG,
57
+ void, env, ptr, ptr, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_6(sve_ldbdu_zsu_mte, TCG_CALL_NO_WG,
60
+ void, env, ptr, ptr, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zsu_mte, TCG_CALL_NO_WG,
62
+ void, env, ptr, ptr, ptr, tl, i32)
63
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zsu_mte, TCG_CALL_NO_WG,
64
+ void, env, ptr, ptr, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zsu_mte, TCG_CALL_NO_WG,
66
+ void, env, ptr, ptr, ptr, tl, i32)
67
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zsu_mte, TCG_CALL_NO_WG,
68
+ void, env, ptr, ptr, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_6(sve_lddd_le_zsu_mte, TCG_CALL_NO_WG,
70
+ void, env, ptr, ptr, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_6(sve_lddd_be_zsu_mte, TCG_CALL_NO_WG,
72
+ void, env, ptr, ptr, ptr, tl, i32)
73
+DEF_HELPER_FLAGS_6(sve_ldbds_zsu_mte, TCG_CALL_NO_WG,
74
+ void, env, ptr, ptr, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zsu_mte, TCG_CALL_NO_WG,
76
+ void, env, ptr, ptr, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zsu_mte, TCG_CALL_NO_WG,
78
+ void, env, ptr, ptr, ptr, tl, i32)
79
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zsu_mte, TCG_CALL_NO_WG,
80
+ void, env, ptr, ptr, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zsu_mte, TCG_CALL_NO_WG,
82
+ void, env, ptr, ptr, ptr, tl, i32)
83
+
84
+DEF_HELPER_FLAGS_6(sve_ldbdu_zss_mte, TCG_CALL_NO_WG,
85
+ void, env, ptr, ptr, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zss_mte, TCG_CALL_NO_WG,
87
+ void, env, ptr, ptr, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zss_mte, TCG_CALL_NO_WG,
89
+ void, env, ptr, ptr, ptr, tl, i32)
90
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zss_mte, TCG_CALL_NO_WG,
91
+ void, env, ptr, ptr, ptr, tl, i32)
92
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zss_mte, TCG_CALL_NO_WG,
93
+ void, env, ptr, ptr, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_6(sve_lddd_le_zss_mte, TCG_CALL_NO_WG,
95
+ void, env, ptr, ptr, ptr, tl, i32)
96
+DEF_HELPER_FLAGS_6(sve_lddd_be_zss_mte, TCG_CALL_NO_WG,
97
+ void, env, ptr, ptr, ptr, tl, i32)
98
+DEF_HELPER_FLAGS_6(sve_ldbds_zss_mte, TCG_CALL_NO_WG,
99
+ void, env, ptr, ptr, ptr, tl, i32)
100
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zss_mte, TCG_CALL_NO_WG,
101
+ void, env, ptr, ptr, ptr, tl, i32)
102
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zss_mte, TCG_CALL_NO_WG,
103
+ void, env, ptr, ptr, ptr, tl, i32)
104
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zss_mte, TCG_CALL_NO_WG,
105
+ void, env, ptr, ptr, ptr, tl, i32)
106
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zss_mte, TCG_CALL_NO_WG,
107
+ void, env, ptr, ptr, ptr, tl, i32)
108
+
109
+DEF_HELPER_FLAGS_6(sve_ldbdu_zd_mte, TCG_CALL_NO_WG,
110
+ void, env, ptr, ptr, ptr, tl, i32)
111
+DEF_HELPER_FLAGS_6(sve_ldhdu_le_zd_mte, TCG_CALL_NO_WG,
112
+ void, env, ptr, ptr, ptr, tl, i32)
113
+DEF_HELPER_FLAGS_6(sve_ldhdu_be_zd_mte, TCG_CALL_NO_WG,
114
+ void, env, ptr, ptr, ptr, tl, i32)
115
+DEF_HELPER_FLAGS_6(sve_ldsdu_le_zd_mte, TCG_CALL_NO_WG,
116
+ void, env, ptr, ptr, ptr, tl, i32)
117
+DEF_HELPER_FLAGS_6(sve_ldsdu_be_zd_mte, TCG_CALL_NO_WG,
118
+ void, env, ptr, ptr, ptr, tl, i32)
119
+DEF_HELPER_FLAGS_6(sve_lddd_le_zd_mte, TCG_CALL_NO_WG,
120
+ void, env, ptr, ptr, ptr, tl, i32)
121
+DEF_HELPER_FLAGS_6(sve_lddd_be_zd_mte, TCG_CALL_NO_WG,
122
+ void, env, ptr, ptr, ptr, tl, i32)
123
+DEF_HELPER_FLAGS_6(sve_ldbds_zd_mte, TCG_CALL_NO_WG,
124
+ void, env, ptr, ptr, ptr, tl, i32)
125
+DEF_HELPER_FLAGS_6(sve_ldhds_le_zd_mte, TCG_CALL_NO_WG,
126
+ void, env, ptr, ptr, ptr, tl, i32)
127
+DEF_HELPER_FLAGS_6(sve_ldhds_be_zd_mte, TCG_CALL_NO_WG,
128
+ void, env, ptr, ptr, ptr, tl, i32)
129
+DEF_HELPER_FLAGS_6(sve_ldsds_le_zd_mte, TCG_CALL_NO_WG,
130
+ void, env, ptr, ptr, ptr, tl, i32)
131
+DEF_HELPER_FLAGS_6(sve_ldsds_be_zd_mte, TCG_CALL_NO_WG,
132
+ void, env, ptr, ptr, ptr, tl, i32)
133
+
134
DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu, TCG_CALL_NO_WG,
135
void, env, ptr, ptr, ptr, tl, i32)
136
DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu, TCG_CALL_NO_WG,
137
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd, TCG_CALL_NO_WG,
138
DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd, TCG_CALL_NO_WG,
139
void, env, ptr, ptr, ptr, tl, i32)
140
141
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zsu_mte, TCG_CALL_NO_WG,
142
+ void, env, ptr, ptr, ptr, tl, i32)
143
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zsu_mte, TCG_CALL_NO_WG,
144
+ void, env, ptr, ptr, ptr, tl, i32)
145
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zsu_mte, TCG_CALL_NO_WG,
146
+ void, env, ptr, ptr, ptr, tl, i32)
147
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zsu_mte, TCG_CALL_NO_WG,
148
+ void, env, ptr, ptr, ptr, tl, i32)
149
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zsu_mte, TCG_CALL_NO_WG,
150
+ void, env, ptr, ptr, ptr, tl, i32)
151
+DEF_HELPER_FLAGS_6(sve_ldffbss_zsu_mte, TCG_CALL_NO_WG,
152
+ void, env, ptr, ptr, ptr, tl, i32)
153
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zsu_mte, TCG_CALL_NO_WG,
154
+ void, env, ptr, ptr, ptr, tl, i32)
155
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zsu_mte, TCG_CALL_NO_WG,
156
+ void, env, ptr, ptr, ptr, tl, i32)
157
+
158
+DEF_HELPER_FLAGS_6(sve_ldffbsu_zss_mte, TCG_CALL_NO_WG,
159
+ void, env, ptr, ptr, ptr, tl, i32)
160
+DEF_HELPER_FLAGS_6(sve_ldffhsu_le_zss_mte, TCG_CALL_NO_WG,
161
+ void, env, ptr, ptr, ptr, tl, i32)
162
+DEF_HELPER_FLAGS_6(sve_ldffhsu_be_zss_mte, TCG_CALL_NO_WG,
163
+ void, env, ptr, ptr, ptr, tl, i32)
164
+DEF_HELPER_FLAGS_6(sve_ldffss_le_zss_mte, TCG_CALL_NO_WG,
165
+ void, env, ptr, ptr, ptr, tl, i32)
166
+DEF_HELPER_FLAGS_6(sve_ldffss_be_zss_mte, TCG_CALL_NO_WG,
167
+ void, env, ptr, ptr, ptr, tl, i32)
168
+DEF_HELPER_FLAGS_6(sve_ldffbss_zss_mte, TCG_CALL_NO_WG,
169
+ void, env, ptr, ptr, ptr, tl, i32)
170
+DEF_HELPER_FLAGS_6(sve_ldffhss_le_zss_mte, TCG_CALL_NO_WG,
171
+ void, env, ptr, ptr, ptr, tl, i32)
172
+DEF_HELPER_FLAGS_6(sve_ldffhss_be_zss_mte, TCG_CALL_NO_WG,
173
+ void, env, ptr, ptr, ptr, tl, i32)
174
+
175
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zsu_mte, TCG_CALL_NO_WG,
176
+ void, env, ptr, ptr, ptr, tl, i32)
177
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zsu_mte, TCG_CALL_NO_WG,
178
+ void, env, ptr, ptr, ptr, tl, i32)
179
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zsu_mte, TCG_CALL_NO_WG,
180
+ void, env, ptr, ptr, ptr, tl, i32)
181
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zsu_mte, TCG_CALL_NO_WG,
182
+ void, env, ptr, ptr, ptr, tl, i32)
183
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zsu_mte, TCG_CALL_NO_WG,
184
+ void, env, ptr, ptr, ptr, tl, i32)
185
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zsu_mte, TCG_CALL_NO_WG,
186
+ void, env, ptr, ptr, ptr, tl, i32)
187
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zsu_mte, TCG_CALL_NO_WG,
188
+ void, env, ptr, ptr, ptr, tl, i32)
189
+DEF_HELPER_FLAGS_6(sve_ldffbds_zsu_mte, TCG_CALL_NO_WG,
190
+ void, env, ptr, ptr, ptr, tl, i32)
191
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zsu_mte, TCG_CALL_NO_WG,
192
+ void, env, ptr, ptr, ptr, tl, i32)
193
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zsu_mte, TCG_CALL_NO_WG,
194
+ void, env, ptr, ptr, ptr, tl, i32)
195
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zsu_mte, TCG_CALL_NO_WG,
196
+ void, env, ptr, ptr, ptr, tl, i32)
197
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zsu_mte, TCG_CALL_NO_WG,
198
+ void, env, ptr, ptr, ptr, tl, i32)
199
+
200
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zss_mte, TCG_CALL_NO_WG,
201
+ void, env, ptr, ptr, ptr, tl, i32)
202
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zss_mte, TCG_CALL_NO_WG,
203
+ void, env, ptr, ptr, ptr, tl, i32)
204
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zss_mte, TCG_CALL_NO_WG,
205
+ void, env, ptr, ptr, ptr, tl, i32)
206
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zss_mte, TCG_CALL_NO_WG,
207
+ void, env, ptr, ptr, ptr, tl, i32)
208
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zss_mte, TCG_CALL_NO_WG,
209
+ void, env, ptr, ptr, ptr, tl, i32)
210
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zss_mte, TCG_CALL_NO_WG,
211
+ void, env, ptr, ptr, ptr, tl, i32)
212
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zss_mte, TCG_CALL_NO_WG,
213
+ void, env, ptr, ptr, ptr, tl, i32)
214
+DEF_HELPER_FLAGS_6(sve_ldffbds_zss_mte, TCG_CALL_NO_WG,
215
+ void, env, ptr, ptr, ptr, tl, i32)
216
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zss_mte, TCG_CALL_NO_WG,
217
+ void, env, ptr, ptr, ptr, tl, i32)
218
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zss_mte, TCG_CALL_NO_WG,
219
+ void, env, ptr, ptr, ptr, tl, i32)
220
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zss_mte, TCG_CALL_NO_WG,
221
+ void, env, ptr, ptr, ptr, tl, i32)
222
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zss_mte, TCG_CALL_NO_WG,
223
+ void, env, ptr, ptr, ptr, tl, i32)
224
+
225
+DEF_HELPER_FLAGS_6(sve_ldffbdu_zd_mte, TCG_CALL_NO_WG,
226
+ void, env, ptr, ptr, ptr, tl, i32)
227
+DEF_HELPER_FLAGS_6(sve_ldffhdu_le_zd_mte, TCG_CALL_NO_WG,
228
+ void, env, ptr, ptr, ptr, tl, i32)
229
+DEF_HELPER_FLAGS_6(sve_ldffhdu_be_zd_mte, TCG_CALL_NO_WG,
230
+ void, env, ptr, ptr, ptr, tl, i32)
231
+DEF_HELPER_FLAGS_6(sve_ldffsdu_le_zd_mte, TCG_CALL_NO_WG,
232
+ void, env, ptr, ptr, ptr, tl, i32)
233
+DEF_HELPER_FLAGS_6(sve_ldffsdu_be_zd_mte, TCG_CALL_NO_WG,
234
+ void, env, ptr, ptr, ptr, tl, i32)
235
+DEF_HELPER_FLAGS_6(sve_ldffdd_le_zd_mte, TCG_CALL_NO_WG,
236
+ void, env, ptr, ptr, ptr, tl, i32)
237
+DEF_HELPER_FLAGS_6(sve_ldffdd_be_zd_mte, TCG_CALL_NO_WG,
238
+ void, env, ptr, ptr, ptr, tl, i32)
239
+DEF_HELPER_FLAGS_6(sve_ldffbds_zd_mte, TCG_CALL_NO_WG,
240
+ void, env, ptr, ptr, ptr, tl, i32)
241
+DEF_HELPER_FLAGS_6(sve_ldffhds_le_zd_mte, TCG_CALL_NO_WG,
242
+ void, env, ptr, ptr, ptr, tl, i32)
243
+DEF_HELPER_FLAGS_6(sve_ldffhds_be_zd_mte, TCG_CALL_NO_WG,
244
+ void, env, ptr, ptr, ptr, tl, i32)
245
+DEF_HELPER_FLAGS_6(sve_ldffsds_le_zd_mte, TCG_CALL_NO_WG,
246
+ void, env, ptr, ptr, ptr, tl, i32)
247
+DEF_HELPER_FLAGS_6(sve_ldffsds_be_zd_mte, TCG_CALL_NO_WG,
248
+ void, env, ptr, ptr, ptr, tl, i32)
249
+
250
DEF_HELPER_FLAGS_6(sve_stbs_zsu, TCG_CALL_NO_WG,
251
void, env, ptr, ptr, ptr, tl, i32)
252
DEF_HELPER_FLAGS_6(sve_sths_le_zsu, TCG_CALL_NO_WG,
253
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_6(sve_stdd_le_zd, TCG_CALL_NO_WG,
254
DEF_HELPER_FLAGS_6(sve_stdd_be_zd, TCG_CALL_NO_WG,
255
void, env, ptr, ptr, ptr, tl, i32)
256
257
+DEF_HELPER_FLAGS_6(sve_stbs_zsu_mte, TCG_CALL_NO_WG,
258
+ void, env, ptr, ptr, ptr, tl, i32)
259
+DEF_HELPER_FLAGS_6(sve_sths_le_zsu_mte, TCG_CALL_NO_WG,
260
+ void, env, ptr, ptr, ptr, tl, i32)
261
+DEF_HELPER_FLAGS_6(sve_sths_be_zsu_mte, TCG_CALL_NO_WG,
262
+ void, env, ptr, ptr, ptr, tl, i32)
263
+DEF_HELPER_FLAGS_6(sve_stss_le_zsu_mte, TCG_CALL_NO_WG,
264
+ void, env, ptr, ptr, ptr, tl, i32)
265
+DEF_HELPER_FLAGS_6(sve_stss_be_zsu_mte, TCG_CALL_NO_WG,
266
+ void, env, ptr, ptr, ptr, tl, i32)
267
+
268
+DEF_HELPER_FLAGS_6(sve_stbs_zss_mte, TCG_CALL_NO_WG,
269
+ void, env, ptr, ptr, ptr, tl, i32)
270
+DEF_HELPER_FLAGS_6(sve_sths_le_zss_mte, TCG_CALL_NO_WG,
271
+ void, env, ptr, ptr, ptr, tl, i32)
272
+DEF_HELPER_FLAGS_6(sve_sths_be_zss_mte, TCG_CALL_NO_WG,
273
+ void, env, ptr, ptr, ptr, tl, i32)
274
+DEF_HELPER_FLAGS_6(sve_stss_le_zss_mte, TCG_CALL_NO_WG,
275
+ void, env, ptr, ptr, ptr, tl, i32)
276
+DEF_HELPER_FLAGS_6(sve_stss_be_zss_mte, TCG_CALL_NO_WG,
277
+ void, env, ptr, ptr, ptr, tl, i32)
278
+
279
+DEF_HELPER_FLAGS_6(sve_stbd_zsu_mte, TCG_CALL_NO_WG,
280
+ void, env, ptr, ptr, ptr, tl, i32)
281
+DEF_HELPER_FLAGS_6(sve_sthd_le_zsu_mte, TCG_CALL_NO_WG,
282
+ void, env, ptr, ptr, ptr, tl, i32)
283
+DEF_HELPER_FLAGS_6(sve_sthd_be_zsu_mte, TCG_CALL_NO_WG,
284
+ void, env, ptr, ptr, ptr, tl, i32)
285
+DEF_HELPER_FLAGS_6(sve_stsd_le_zsu_mte, TCG_CALL_NO_WG,
286
+ void, env, ptr, ptr, ptr, tl, i32)
287
+DEF_HELPER_FLAGS_6(sve_stsd_be_zsu_mte, TCG_CALL_NO_WG,
288
+ void, env, ptr, ptr, ptr, tl, i32)
289
+DEF_HELPER_FLAGS_6(sve_stdd_le_zsu_mte, TCG_CALL_NO_WG,
290
+ void, env, ptr, ptr, ptr, tl, i32)
291
+DEF_HELPER_FLAGS_6(sve_stdd_be_zsu_mte, TCG_CALL_NO_WG,
292
+ void, env, ptr, ptr, ptr, tl, i32)
293
+
294
+DEF_HELPER_FLAGS_6(sve_stbd_zss_mte, TCG_CALL_NO_WG,
295
+ void, env, ptr, ptr, ptr, tl, i32)
296
+DEF_HELPER_FLAGS_6(sve_sthd_le_zss_mte, TCG_CALL_NO_WG,
297
+ void, env, ptr, ptr, ptr, tl, i32)
298
+DEF_HELPER_FLAGS_6(sve_sthd_be_zss_mte, TCG_CALL_NO_WG,
299
+ void, env, ptr, ptr, ptr, tl, i32)
300
+DEF_HELPER_FLAGS_6(sve_stsd_le_zss_mte, TCG_CALL_NO_WG,
301
+ void, env, ptr, ptr, ptr, tl, i32)
302
+DEF_HELPER_FLAGS_6(sve_stsd_be_zss_mte, TCG_CALL_NO_WG,
303
+ void, env, ptr, ptr, ptr, tl, i32)
304
+DEF_HELPER_FLAGS_6(sve_stdd_le_zss_mte, TCG_CALL_NO_WG,
305
+ void, env, ptr, ptr, ptr, tl, i32)
306
+DEF_HELPER_FLAGS_6(sve_stdd_be_zss_mte, TCG_CALL_NO_WG,
307
+ void, env, ptr, ptr, ptr, tl, i32)
308
+
309
+DEF_HELPER_FLAGS_6(sve_stbd_zd_mte, TCG_CALL_NO_WG,
310
+ void, env, ptr, ptr, ptr, tl, i32)
311
+DEF_HELPER_FLAGS_6(sve_sthd_le_zd_mte, TCG_CALL_NO_WG,
312
+ void, env, ptr, ptr, ptr, tl, i32)
313
+DEF_HELPER_FLAGS_6(sve_sthd_be_zd_mte, TCG_CALL_NO_WG,
314
+ void, env, ptr, ptr, ptr, tl, i32)
315
+DEF_HELPER_FLAGS_6(sve_stsd_le_zd_mte, TCG_CALL_NO_WG,
316
+ void, env, ptr, ptr, ptr, tl, i32)
317
+DEF_HELPER_FLAGS_6(sve_stsd_be_zd_mte, TCG_CALL_NO_WG,
318
+ void, env, ptr, ptr, ptr, tl, i32)
319
+DEF_HELPER_FLAGS_6(sve_stdd_le_zd_mte, TCG_CALL_NO_WG,
320
+ void, env, ptr, ptr, ptr, tl, i32)
321
+DEF_HELPER_FLAGS_6(sve_stdd_be_zd_mte, TCG_CALL_NO_WG,
322
+ void, env, ptr, ptr, ptr, tl, i32)
323
+
324
DEF_HELPER_FLAGS_4(sve2_pmull_h, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32)
325
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
326
index XXXXXXX..XXXXXXX 100644
327
--- a/target/arm/sve_helper.c
328
+++ b/target/arm/sve_helper.c
329
@@ -XXX,XX +XXX,XX @@ static target_ulong off_zd_d(void *reg, intptr_t reg_ofs)
330
static inline QEMU_ALWAYS_INLINE
331
void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
332
target_ulong base, uint32_t desc, uintptr_t retaddr,
333
- int esize, int msize, zreg_off_fn *off_fn,
334
+ uint32_t mtedesc, int esize, int msize,
335
+ zreg_off_fn *off_fn,
336
sve_ldst1_host_fn *host_fn,
337
sve_ldst1_tlb_fn *tlb_fn)
338
{
339
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
340
cpu_check_watchpoint(env_cpu(env), addr, msize,
341
info.attrs, BP_MEM_READ, retaddr);
342
}
343
- /* TODO: MTE check */
344
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
345
+ mte_check1(env, mtedesc, addr, retaddr);
346
+ }
347
host_fn(&scratch, reg_off, info.host);
348
} else {
349
/* Element crosses the page boundary. */
350
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
351
msize, info.attrs,
352
BP_MEM_READ, retaddr);
353
}
354
- /* TODO: MTE check */
355
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
356
+ mte_check1(env, mtedesc, addr, retaddr);
357
+ }
358
tlb_fn(env, &scratch, reg_off, addr, retaddr);
359
}
360
}
361
@@ -XXX,XX +XXX,XX @@ void sve_ld1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
362
memcpy(vd, &scratch, reg_max);
363
}
364
365
+static inline QEMU_ALWAYS_INLINE
366
+void sve_ld1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
367
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
368
+ int esize, int msize, zreg_off_fn *off_fn,
369
+ sve_ldst1_host_fn *host_fn,
370
+ sve_ldst1_tlb_fn *tlb_fn)
371
+{
372
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
373
+ /* Remove mtedesc from the normal sve descriptor. */
374
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
375
+
376
+ /*
377
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
378
+ * offset base entirely over the address space hole to change the
379
+ * pointer tag, or change the bit55 selector. So we could here
380
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
381
+ */
382
+ sve_ld1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
383
+ esize, msize, off_fn, host_fn, tlb_fn);
384
+}
385
+
386
#define DO_LD1_ZPZ_S(MEM, OFS, MSZ) \
387
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
388
void *vm, target_ulong base, uint32_t desc) \
389
{ \
390
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
391
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
392
off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
393
+} \
394
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
395
+ void *vm, target_ulong base, uint32_t desc) \
396
+{ \
397
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
398
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
399
}
400
401
#define DO_LD1_ZPZ_D(MEM, OFS, MSZ) \
402
void HELPER(sve_ld##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
403
void *vm, target_ulong base, uint32_t desc) \
404
{ \
405
- sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
406
+ sve_ld1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
407
off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
408
+} \
409
+void HELPER(sve_ld##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
410
+ void *vm, target_ulong base, uint32_t desc) \
411
+{ \
412
+ sve_ld1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
413
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
414
}
415
416
DO_LD1_ZPZ_S(bsu, zsu, MO_8)
417
@@ -XXX,XX +XXX,XX @@ DO_LD1_ZPZ_D(dd_be, zd, MO_64)
418
static inline QEMU_ALWAYS_INLINE
419
void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
420
target_ulong base, uint32_t desc, uintptr_t retaddr,
421
- const int esz, const int msz, zreg_off_fn *off_fn,
422
+ uint32_t mtedesc, const int esz, const int msz,
423
+ zreg_off_fn *off_fn,
424
sve_ldst1_host_fn *host_fn,
425
sve_ldst1_tlb_fn *tlb_fn)
426
{
427
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
428
* Probe the first element, allowing faults.
429
*/
430
addr = base + (off_fn(vm, reg_off) << scale);
431
+ if (mtedesc) {
432
+ mte_check1(env, mtedesc, addr, retaddr);
433
+ }
434
tlb_fn(env, vd, reg_off, addr, retaddr);
435
436
/* After any fault, zero the other elements. */
437
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
438
(env_cpu(env), addr, msize) & BP_MEM_READ)) {
439
goto fault;
440
}
441
- /* TODO: MTE check. */
442
+ if (mtedesc &&
443
+ arm_tlb_mte_tagged(&info.attrs) &&
444
+ !mte_probe1(env, mtedesc, addr)) {
445
+ goto fault;
446
+ }
447
448
host_fn(vd, reg_off, info.host);
449
}
450
@@ -XXX,XX +XXX,XX @@ void sve_ldff1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
451
record_fault(env, reg_off, reg_max);
452
}
453
454
-#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
455
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
456
- void *vm, target_ulong base, uint32_t desc) \
457
-{ \
458
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
459
- off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
460
+static inline QEMU_ALWAYS_INLINE
461
+void sve_ldff1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
462
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
463
+ const int esz, const int msz,
464
+ zreg_off_fn *off_fn,
465
+ sve_ldst1_host_fn *host_fn,
466
+ sve_ldst1_tlb_fn *tlb_fn)
467
+{
468
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
469
+ /* Remove mtedesc from the normal sve descriptor. */
470
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
471
+
472
+ /*
473
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
474
+ * offset base entirely over the address space hole to change the
475
+ * pointer tag, or change the bit55 selector. So we could here
476
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
477
+ */
478
+ sve_ldff1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
479
+ esz, msz, off_fn, host_fn, tlb_fn);
480
}
481
482
-#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
483
-void HELPER(sve_ldff##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
484
- void *vm, target_ulong base, uint32_t desc) \
485
-{ \
486
- sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
487
- off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
488
+#define DO_LDFF1_ZPZ_S(MEM, OFS, MSZ) \
489
+void HELPER(sve_ldff##MEM##_##OFS) \
490
+ (CPUARMState *env, void *vd, void *vg, \
491
+ void *vm, target_ulong base, uint32_t desc) \
492
+{ \
493
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_32, MSZ, \
494
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
495
+} \
496
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
497
+ (CPUARMState *env, void *vd, void *vg, \
498
+ void *vm, target_ulong base, uint32_t desc) \
499
+{ \
500
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_32, MSZ, \
501
+ off_##OFS##_s, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
502
+}
503
+
504
+#define DO_LDFF1_ZPZ_D(MEM, OFS, MSZ) \
505
+void HELPER(sve_ldff##MEM##_##OFS) \
506
+ (CPUARMState *env, void *vd, void *vg, \
507
+ void *vm, target_ulong base, uint32_t desc) \
508
+{ \
509
+ sve_ldff1_z(env, vd, vg, vm, base, desc, GETPC(), 0, MO_64, MSZ, \
510
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
511
+} \
512
+void HELPER(sve_ldff##MEM##_##OFS##_mte) \
513
+ (CPUARMState *env, void *vd, void *vg, \
514
+ void *vm, target_ulong base, uint32_t desc) \
515
+{ \
516
+ sve_ldff1_z_mte(env, vd, vg, vm, base, desc, GETPC(), MO_64, MSZ, \
517
+ off_##OFS##_d, sve_ld1##MEM##_host, sve_ld1##MEM##_tlb); \
518
}
519
520
DO_LDFF1_ZPZ_S(bsu, zsu, MO_8)
521
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_ZPZ_D(dd_be, zd, MO_64)
522
static inline QEMU_ALWAYS_INLINE
523
void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
524
target_ulong base, uint32_t desc, uintptr_t retaddr,
525
- int esize, int msize, zreg_off_fn *off_fn,
526
+ uint32_t mtedesc, int esize, int msize,
527
+ zreg_off_fn *off_fn,
528
sve_ldst1_host_fn *host_fn,
529
sve_ldst1_tlb_fn *tlb_fn)
530
{
531
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
532
cpu_check_watchpoint(env_cpu(env), addr, msize,
533
info.attrs, BP_MEM_WRITE, retaddr);
534
}
535
- /* TODO: MTE check. */
536
+
537
+ if (mtedesc && arm_tlb_mte_tagged(&info.attrs)) {
538
+ mte_check1(env, mtedesc, addr, retaddr);
539
+ }
540
}
541
i += 1;
542
reg_off += esize;
543
@@ -XXX,XX +XXX,XX @@ void sve_st1_z(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
544
} while (reg_off < reg_max);
545
}
546
547
-#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
548
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
549
- void *vm, target_ulong base, uint32_t desc) \
550
-{ \
551
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
552
- off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
553
+static inline QEMU_ALWAYS_INLINE
554
+void sve_st1_z_mte(CPUARMState *env, void *vd, uint64_t *vg, void *vm,
555
+ target_ulong base, uint32_t desc, uintptr_t retaddr,
556
+ int esize, int msize, zreg_off_fn *off_fn,
557
+ sve_ldst1_host_fn *host_fn,
558
+ sve_ldst1_tlb_fn *tlb_fn)
559
+{
560
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
561
+ /* Remove mtedesc from the normal sve descriptor. */
562
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
563
+
564
+ /*
565
+ * ??? TODO: For the 32-bit offset extractions, base + ofs cannot
566
+ * offset base entirely over the address space hole to change the
567
+ * pointer tag, or change the bit55 selector. So we could here
568
+ * examine TBI + TCMA like we do for sve_ldN_r_mte().
569
+ */
570
+ sve_st1_z(env, vd, vg, vm, base, desc, retaddr, mtedesc,
571
+ esize, msize, off_fn, host_fn, tlb_fn);
572
}
573
574
-#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
575
-void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
576
+#define DO_ST1_ZPZ_S(MEM, OFS, MSZ) \
577
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
578
void *vm, target_ulong base, uint32_t desc) \
579
-{ \
580
- sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
581
- off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
582
+{ \
583
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 4, 1 << MSZ, \
584
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
585
+} \
586
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
587
+ void *vm, target_ulong base, uint32_t desc) \
588
+{ \
589
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 4, 1 << MSZ, \
590
+ off_##OFS##_s, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
591
+}
592
+
593
+#define DO_ST1_ZPZ_D(MEM, OFS, MSZ) \
594
+void HELPER(sve_st##MEM##_##OFS)(CPUARMState *env, void *vd, void *vg, \
595
+ void *vm, target_ulong base, uint32_t desc) \
596
+{ \
597
+ sve_st1_z(env, vd, vg, vm, base, desc, GETPC(), 0, 8, 1 << MSZ, \
598
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
599
+} \
600
+void HELPER(sve_st##MEM##_##OFS##_mte)(CPUARMState *env, void *vd, void *vg, \
601
+ void *vm, target_ulong base, uint32_t desc) \
602
+{ \
603
+ sve_st1_z_mte(env, vd, vg, vm, base, desc, GETPC(), 8, 1 << MSZ, \
604
+ off_##OFS##_d, sve_st1##MEM##_host, sve_st1##MEM##_tlb); \
605
}
606
607
DO_ST1_ZPZ_S(bs, zsu, MO_8)
608
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
609
index XXXXXXX..XXXXXXX 100644
610
--- a/target/arm/translate-sve.c
611
+++ b/target/arm/translate-sve.c
612
@@ -XXX,XX +XXX,XX @@ static bool trans_ST_zpri(DisasContext *s, arg_rpri_store *a)
613
*/
614
615
static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
616
- int scale, TCGv_i64 scalar, int msz,
617
+ int scale, TCGv_i64 scalar, int msz, bool is_write,
618
gen_helper_gvec_mem_scatter *fn)
619
{
620
unsigned vsz = vec_full_reg_size(s);
621
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
622
TCGv_ptr t_pg = tcg_temp_new_ptr();
623
TCGv_ptr t_zt = tcg_temp_new_ptr();
624
TCGv_i32 t_desc;
625
- int desc;
626
+ int desc = 0;
627
628
+ if (s->mte_active[0]) {
629
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
630
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
631
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
632
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
633
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
634
+ desc <<= SVE_MTEDESC_SHIFT;
635
+ }
636
desc = simd_desc(vsz, vsz, scale);
637
t_desc = tcg_const_i32(desc);
638
639
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpz(DisasContext *s, int zt, int pg, int zm,
640
tcg_temp_free_i32(t_desc);
641
}
642
643
-/* Indexed by [be][ff][xs][u][msz]. */
644
-static gen_helper_gvec_mem_scatter * const gather_load_fn32[2][2][2][2][3] = {
645
- /* Little-endian */
646
- { { { { gen_helper_sve_ldbss_zsu,
647
- gen_helper_sve_ldhss_le_zsu,
648
- NULL, },
649
- { gen_helper_sve_ldbsu_zsu,
650
- gen_helper_sve_ldhsu_le_zsu,
651
- gen_helper_sve_ldss_le_zsu, } },
652
- { { gen_helper_sve_ldbss_zss,
653
- gen_helper_sve_ldhss_le_zss,
654
- NULL, },
655
- { gen_helper_sve_ldbsu_zss,
656
- gen_helper_sve_ldhsu_le_zss,
657
- gen_helper_sve_ldss_le_zss, } } },
658
+/* Indexed by [mte][be][ff][xs][u][msz]. */
659
+static gen_helper_gvec_mem_scatter * const
660
+gather_load_fn32[2][2][2][2][2][3] = {
661
+ { /* MTE Inactive */
662
+ { /* Little-endian */
663
+ { { { gen_helper_sve_ldbss_zsu,
664
+ gen_helper_sve_ldhss_le_zsu,
665
+ NULL, },
666
+ { gen_helper_sve_ldbsu_zsu,
667
+ gen_helper_sve_ldhsu_le_zsu,
668
+ gen_helper_sve_ldss_le_zsu, } },
669
+ { { gen_helper_sve_ldbss_zss,
670
+ gen_helper_sve_ldhss_le_zss,
671
+ NULL, },
672
+ { gen_helper_sve_ldbsu_zss,
673
+ gen_helper_sve_ldhsu_le_zss,
674
+ gen_helper_sve_ldss_le_zss, } } },
675
676
- /* First-fault */
677
- { { { gen_helper_sve_ldffbss_zsu,
678
- gen_helper_sve_ldffhss_le_zsu,
679
- NULL, },
680
- { gen_helper_sve_ldffbsu_zsu,
681
- gen_helper_sve_ldffhsu_le_zsu,
682
- gen_helper_sve_ldffss_le_zsu, } },
683
- { { gen_helper_sve_ldffbss_zss,
684
- gen_helper_sve_ldffhss_le_zss,
685
- NULL, },
686
- { gen_helper_sve_ldffbsu_zss,
687
- gen_helper_sve_ldffhsu_le_zss,
688
- gen_helper_sve_ldffss_le_zss, } } } },
689
+ /* First-fault */
690
+ { { { gen_helper_sve_ldffbss_zsu,
691
+ gen_helper_sve_ldffhss_le_zsu,
692
+ NULL, },
693
+ { gen_helper_sve_ldffbsu_zsu,
694
+ gen_helper_sve_ldffhsu_le_zsu,
695
+ gen_helper_sve_ldffss_le_zsu, } },
696
+ { { gen_helper_sve_ldffbss_zss,
697
+ gen_helper_sve_ldffhss_le_zss,
698
+ NULL, },
699
+ { gen_helper_sve_ldffbsu_zss,
700
+ gen_helper_sve_ldffhsu_le_zss,
701
+ gen_helper_sve_ldffss_le_zss, } } } },
702
703
- /* Big-endian */
704
- { { { { gen_helper_sve_ldbss_zsu,
705
- gen_helper_sve_ldhss_be_zsu,
706
- NULL, },
707
- { gen_helper_sve_ldbsu_zsu,
708
- gen_helper_sve_ldhsu_be_zsu,
709
- gen_helper_sve_ldss_be_zsu, } },
710
- { { gen_helper_sve_ldbss_zss,
711
- gen_helper_sve_ldhss_be_zss,
712
- NULL, },
713
- { gen_helper_sve_ldbsu_zss,
714
- gen_helper_sve_ldhsu_be_zss,
715
- gen_helper_sve_ldss_be_zss, } } },
716
+ { /* Big-endian */
717
+ { { { gen_helper_sve_ldbss_zsu,
718
+ gen_helper_sve_ldhss_be_zsu,
719
+ NULL, },
720
+ { gen_helper_sve_ldbsu_zsu,
721
+ gen_helper_sve_ldhsu_be_zsu,
722
+ gen_helper_sve_ldss_be_zsu, } },
723
+ { { gen_helper_sve_ldbss_zss,
724
+ gen_helper_sve_ldhss_be_zss,
725
+ NULL, },
726
+ { gen_helper_sve_ldbsu_zss,
727
+ gen_helper_sve_ldhsu_be_zss,
728
+ gen_helper_sve_ldss_be_zss, } } },
729
730
- /* First-fault */
731
- { { { gen_helper_sve_ldffbss_zsu,
732
- gen_helper_sve_ldffhss_be_zsu,
733
- NULL, },
734
- { gen_helper_sve_ldffbsu_zsu,
735
- gen_helper_sve_ldffhsu_be_zsu,
736
- gen_helper_sve_ldffss_be_zsu, } },
737
- { { gen_helper_sve_ldffbss_zss,
738
- gen_helper_sve_ldffhss_be_zss,
739
- NULL, },
740
- { gen_helper_sve_ldffbsu_zss,
741
- gen_helper_sve_ldffhsu_be_zss,
742
- gen_helper_sve_ldffss_be_zss, } } } },
743
+ /* First-fault */
744
+ { { { gen_helper_sve_ldffbss_zsu,
745
+ gen_helper_sve_ldffhss_be_zsu,
746
+ NULL, },
747
+ { gen_helper_sve_ldffbsu_zsu,
748
+ gen_helper_sve_ldffhsu_be_zsu,
749
+ gen_helper_sve_ldffss_be_zsu, } },
750
+ { { gen_helper_sve_ldffbss_zss,
751
+ gen_helper_sve_ldffhss_be_zss,
752
+ NULL, },
753
+ { gen_helper_sve_ldffbsu_zss,
754
+ gen_helper_sve_ldffhsu_be_zss,
755
+ gen_helper_sve_ldffss_be_zss, } } } } },
756
+ { /* MTE Active */
757
+ { /* Little-endian */
758
+ { { { gen_helper_sve_ldbss_zsu_mte,
759
+ gen_helper_sve_ldhss_le_zsu_mte,
760
+ NULL, },
761
+ { gen_helper_sve_ldbsu_zsu_mte,
762
+ gen_helper_sve_ldhsu_le_zsu_mte,
763
+ gen_helper_sve_ldss_le_zsu_mte, } },
764
+ { { gen_helper_sve_ldbss_zss_mte,
765
+ gen_helper_sve_ldhss_le_zss_mte,
766
+ NULL, },
767
+ { gen_helper_sve_ldbsu_zss_mte,
768
+ gen_helper_sve_ldhsu_le_zss_mte,
769
+ gen_helper_sve_ldss_le_zss_mte, } } },
770
+
771
+ /* First-fault */
772
+ { { { gen_helper_sve_ldffbss_zsu_mte,
773
+ gen_helper_sve_ldffhss_le_zsu_mte,
774
+ NULL, },
775
+ { gen_helper_sve_ldffbsu_zsu_mte,
776
+ gen_helper_sve_ldffhsu_le_zsu_mte,
777
+ gen_helper_sve_ldffss_le_zsu_mte, } },
778
+ { { gen_helper_sve_ldffbss_zss_mte,
779
+ gen_helper_sve_ldffhss_le_zss_mte,
780
+ NULL, },
781
+ { gen_helper_sve_ldffbsu_zss_mte,
782
+ gen_helper_sve_ldffhsu_le_zss_mte,
783
+ gen_helper_sve_ldffss_le_zss_mte, } } } },
784
+
785
+ { /* Big-endian */
786
+ { { { gen_helper_sve_ldbss_zsu_mte,
787
+ gen_helper_sve_ldhss_be_zsu_mte,
788
+ NULL, },
789
+ { gen_helper_sve_ldbsu_zsu_mte,
790
+ gen_helper_sve_ldhsu_be_zsu_mte,
791
+ gen_helper_sve_ldss_be_zsu_mte, } },
792
+ { { gen_helper_sve_ldbss_zss_mte,
793
+ gen_helper_sve_ldhss_be_zss_mte,
794
+ NULL, },
795
+ { gen_helper_sve_ldbsu_zss_mte,
796
+ gen_helper_sve_ldhsu_be_zss_mte,
797
+ gen_helper_sve_ldss_be_zss_mte, } } },
798
+
799
+ /* First-fault */
800
+ { { { gen_helper_sve_ldffbss_zsu_mte,
801
+ gen_helper_sve_ldffhss_be_zsu_mte,
802
+ NULL, },
803
+ { gen_helper_sve_ldffbsu_zsu_mte,
804
+ gen_helper_sve_ldffhsu_be_zsu_mte,
805
+ gen_helper_sve_ldffss_be_zsu_mte, } },
806
+ { { gen_helper_sve_ldffbss_zss_mte,
807
+ gen_helper_sve_ldffhss_be_zss_mte,
808
+ NULL, },
809
+ { gen_helper_sve_ldffbsu_zss_mte,
810
+ gen_helper_sve_ldffhsu_be_zss_mte,
811
+ gen_helper_sve_ldffss_be_zss_mte, } } } } },
812
};
813
814
/* Note that we overload xs=2 to indicate 64-bit offset. */
815
-static gen_helper_gvec_mem_scatter * const gather_load_fn64[2][2][3][2][4] = {
816
- /* Little-endian */
817
- { { { { gen_helper_sve_ldbds_zsu,
818
- gen_helper_sve_ldhds_le_zsu,
819
- gen_helper_sve_ldsds_le_zsu,
820
- NULL, },
821
- { gen_helper_sve_ldbdu_zsu,
822
- gen_helper_sve_ldhdu_le_zsu,
823
- gen_helper_sve_ldsdu_le_zsu,
824
- gen_helper_sve_lddd_le_zsu, } },
825
- { { gen_helper_sve_ldbds_zss,
826
- gen_helper_sve_ldhds_le_zss,
827
- gen_helper_sve_ldsds_le_zss,
828
- NULL, },
829
- { gen_helper_sve_ldbdu_zss,
830
- gen_helper_sve_ldhdu_le_zss,
831
- gen_helper_sve_ldsdu_le_zss,
832
- gen_helper_sve_lddd_le_zss, } },
833
- { { gen_helper_sve_ldbds_zd,
834
- gen_helper_sve_ldhds_le_zd,
835
- gen_helper_sve_ldsds_le_zd,
836
- NULL, },
837
- { gen_helper_sve_ldbdu_zd,
838
- gen_helper_sve_ldhdu_le_zd,
839
- gen_helper_sve_ldsdu_le_zd,
840
- gen_helper_sve_lddd_le_zd, } } },
841
+static gen_helper_gvec_mem_scatter * const
842
+gather_load_fn64[2][2][2][3][2][4] = {
843
+ { /* MTE Inactive */
844
+ { /* Little-endian */
845
+ { { { gen_helper_sve_ldbds_zsu,
846
+ gen_helper_sve_ldhds_le_zsu,
847
+ gen_helper_sve_ldsds_le_zsu,
848
+ NULL, },
849
+ { gen_helper_sve_ldbdu_zsu,
850
+ gen_helper_sve_ldhdu_le_zsu,
851
+ gen_helper_sve_ldsdu_le_zsu,
852
+ gen_helper_sve_lddd_le_zsu, } },
853
+ { { gen_helper_sve_ldbds_zss,
854
+ gen_helper_sve_ldhds_le_zss,
855
+ gen_helper_sve_ldsds_le_zss,
856
+ NULL, },
857
+ { gen_helper_sve_ldbdu_zss,
858
+ gen_helper_sve_ldhdu_le_zss,
859
+ gen_helper_sve_ldsdu_le_zss,
860
+ gen_helper_sve_lddd_le_zss, } },
861
+ { { gen_helper_sve_ldbds_zd,
862
+ gen_helper_sve_ldhds_le_zd,
863
+ gen_helper_sve_ldsds_le_zd,
864
+ NULL, },
865
+ { gen_helper_sve_ldbdu_zd,
866
+ gen_helper_sve_ldhdu_le_zd,
867
+ gen_helper_sve_ldsdu_le_zd,
868
+ gen_helper_sve_lddd_le_zd, } } },
869
870
- /* First-fault */
871
- { { { gen_helper_sve_ldffbds_zsu,
872
- gen_helper_sve_ldffhds_le_zsu,
873
- gen_helper_sve_ldffsds_le_zsu,
874
- NULL, },
875
- { gen_helper_sve_ldffbdu_zsu,
876
- gen_helper_sve_ldffhdu_le_zsu,
877
- gen_helper_sve_ldffsdu_le_zsu,
878
- gen_helper_sve_ldffdd_le_zsu, } },
879
- { { gen_helper_sve_ldffbds_zss,
880
- gen_helper_sve_ldffhds_le_zss,
881
- gen_helper_sve_ldffsds_le_zss,
882
- NULL, },
883
- { gen_helper_sve_ldffbdu_zss,
884
- gen_helper_sve_ldffhdu_le_zss,
885
- gen_helper_sve_ldffsdu_le_zss,
886
- gen_helper_sve_ldffdd_le_zss, } },
887
- { { gen_helper_sve_ldffbds_zd,
888
- gen_helper_sve_ldffhds_le_zd,
889
- gen_helper_sve_ldffsds_le_zd,
890
- NULL, },
891
- { gen_helper_sve_ldffbdu_zd,
892
- gen_helper_sve_ldffhdu_le_zd,
893
- gen_helper_sve_ldffsdu_le_zd,
894
- gen_helper_sve_ldffdd_le_zd, } } } },
895
+ /* First-fault */
896
+ { { { gen_helper_sve_ldffbds_zsu,
897
+ gen_helper_sve_ldffhds_le_zsu,
898
+ gen_helper_sve_ldffsds_le_zsu,
899
+ NULL, },
900
+ { gen_helper_sve_ldffbdu_zsu,
901
+ gen_helper_sve_ldffhdu_le_zsu,
902
+ gen_helper_sve_ldffsdu_le_zsu,
903
+ gen_helper_sve_ldffdd_le_zsu, } },
904
+ { { gen_helper_sve_ldffbds_zss,
905
+ gen_helper_sve_ldffhds_le_zss,
906
+ gen_helper_sve_ldffsds_le_zss,
907
+ NULL, },
908
+ { gen_helper_sve_ldffbdu_zss,
909
+ gen_helper_sve_ldffhdu_le_zss,
910
+ gen_helper_sve_ldffsdu_le_zss,
911
+ gen_helper_sve_ldffdd_le_zss, } },
912
+ { { gen_helper_sve_ldffbds_zd,
913
+ gen_helper_sve_ldffhds_le_zd,
914
+ gen_helper_sve_ldffsds_le_zd,
915
+ NULL, },
916
+ { gen_helper_sve_ldffbdu_zd,
917
+ gen_helper_sve_ldffhdu_le_zd,
918
+ gen_helper_sve_ldffsdu_le_zd,
919
+ gen_helper_sve_ldffdd_le_zd, } } } },
920
+ { /* Big-endian */
921
+ { { { gen_helper_sve_ldbds_zsu,
922
+ gen_helper_sve_ldhds_be_zsu,
923
+ gen_helper_sve_ldsds_be_zsu,
924
+ NULL, },
925
+ { gen_helper_sve_ldbdu_zsu,
926
+ gen_helper_sve_ldhdu_be_zsu,
927
+ gen_helper_sve_ldsdu_be_zsu,
928
+ gen_helper_sve_lddd_be_zsu, } },
929
+ { { gen_helper_sve_ldbds_zss,
930
+ gen_helper_sve_ldhds_be_zss,
931
+ gen_helper_sve_ldsds_be_zss,
932
+ NULL, },
933
+ { gen_helper_sve_ldbdu_zss,
934
+ gen_helper_sve_ldhdu_be_zss,
935
+ gen_helper_sve_ldsdu_be_zss,
936
+ gen_helper_sve_lddd_be_zss, } },
937
+ { { gen_helper_sve_ldbds_zd,
938
+ gen_helper_sve_ldhds_be_zd,
939
+ gen_helper_sve_ldsds_be_zd,
940
+ NULL, },
941
+ { gen_helper_sve_ldbdu_zd,
942
+ gen_helper_sve_ldhdu_be_zd,
943
+ gen_helper_sve_ldsdu_be_zd,
944
+ gen_helper_sve_lddd_be_zd, } } },
945
946
- /* Big-endian */
947
- { { { { gen_helper_sve_ldbds_zsu,
948
- gen_helper_sve_ldhds_be_zsu,
949
- gen_helper_sve_ldsds_be_zsu,
950
- NULL, },
951
- { gen_helper_sve_ldbdu_zsu,
952
- gen_helper_sve_ldhdu_be_zsu,
953
- gen_helper_sve_ldsdu_be_zsu,
954
- gen_helper_sve_lddd_be_zsu, } },
955
- { { gen_helper_sve_ldbds_zss,
956
- gen_helper_sve_ldhds_be_zss,
957
- gen_helper_sve_ldsds_be_zss,
958
- NULL, },
959
- { gen_helper_sve_ldbdu_zss,
960
- gen_helper_sve_ldhdu_be_zss,
961
- gen_helper_sve_ldsdu_be_zss,
962
- gen_helper_sve_lddd_be_zss, } },
963
- { { gen_helper_sve_ldbds_zd,
964
- gen_helper_sve_ldhds_be_zd,
965
- gen_helper_sve_ldsds_be_zd,
966
- NULL, },
967
- { gen_helper_sve_ldbdu_zd,
968
- gen_helper_sve_ldhdu_be_zd,
969
- gen_helper_sve_ldsdu_be_zd,
970
- gen_helper_sve_lddd_be_zd, } } },
971
+ /* First-fault */
972
+ { { { gen_helper_sve_ldffbds_zsu,
973
+ gen_helper_sve_ldffhds_be_zsu,
974
+ gen_helper_sve_ldffsds_be_zsu,
975
+ NULL, },
976
+ { gen_helper_sve_ldffbdu_zsu,
977
+ gen_helper_sve_ldffhdu_be_zsu,
978
+ gen_helper_sve_ldffsdu_be_zsu,
979
+ gen_helper_sve_ldffdd_be_zsu, } },
980
+ { { gen_helper_sve_ldffbds_zss,
981
+ gen_helper_sve_ldffhds_be_zss,
982
+ gen_helper_sve_ldffsds_be_zss,
983
+ NULL, },
984
+ { gen_helper_sve_ldffbdu_zss,
985
+ gen_helper_sve_ldffhdu_be_zss,
986
+ gen_helper_sve_ldffsdu_be_zss,
987
+ gen_helper_sve_ldffdd_be_zss, } },
988
+ { { gen_helper_sve_ldffbds_zd,
989
+ gen_helper_sve_ldffhds_be_zd,
990
+ gen_helper_sve_ldffsds_be_zd,
991
+ NULL, },
992
+ { gen_helper_sve_ldffbdu_zd,
993
+ gen_helper_sve_ldffhdu_be_zd,
994
+ gen_helper_sve_ldffsdu_be_zd,
995
+ gen_helper_sve_ldffdd_be_zd, } } } } },
996
+ { /* MTE Active */
997
+ { /* Little-endian */
998
+ { { { gen_helper_sve_ldbds_zsu_mte,
999
+ gen_helper_sve_ldhds_le_zsu_mte,
1000
+ gen_helper_sve_ldsds_le_zsu_mte,
1001
+ NULL, },
1002
+ { gen_helper_sve_ldbdu_zsu_mte,
1003
+ gen_helper_sve_ldhdu_le_zsu_mte,
1004
+ gen_helper_sve_ldsdu_le_zsu_mte,
1005
+ gen_helper_sve_lddd_le_zsu_mte, } },
1006
+ { { gen_helper_sve_ldbds_zss_mte,
1007
+ gen_helper_sve_ldhds_le_zss_mte,
1008
+ gen_helper_sve_ldsds_le_zss_mte,
1009
+ NULL, },
1010
+ { gen_helper_sve_ldbdu_zss_mte,
1011
+ gen_helper_sve_ldhdu_le_zss_mte,
1012
+ gen_helper_sve_ldsdu_le_zss_mte,
1013
+ gen_helper_sve_lddd_le_zss_mte, } },
1014
+ { { gen_helper_sve_ldbds_zd_mte,
1015
+ gen_helper_sve_ldhds_le_zd_mte,
1016
+ gen_helper_sve_ldsds_le_zd_mte,
1017
+ NULL, },
1018
+ { gen_helper_sve_ldbdu_zd_mte,
1019
+ gen_helper_sve_ldhdu_le_zd_mte,
1020
+ gen_helper_sve_ldsdu_le_zd_mte,
1021
+ gen_helper_sve_lddd_le_zd_mte, } } },
1022
1023
- /* First-fault */
1024
- { { { gen_helper_sve_ldffbds_zsu,
1025
- gen_helper_sve_ldffhds_be_zsu,
1026
- gen_helper_sve_ldffsds_be_zsu,
1027
- NULL, },
1028
- { gen_helper_sve_ldffbdu_zsu,
1029
- gen_helper_sve_ldffhdu_be_zsu,
1030
- gen_helper_sve_ldffsdu_be_zsu,
1031
- gen_helper_sve_ldffdd_be_zsu, } },
1032
- { { gen_helper_sve_ldffbds_zss,
1033
- gen_helper_sve_ldffhds_be_zss,
1034
- gen_helper_sve_ldffsds_be_zss,
1035
- NULL, },
1036
- { gen_helper_sve_ldffbdu_zss,
1037
- gen_helper_sve_ldffhdu_be_zss,
1038
- gen_helper_sve_ldffsdu_be_zss,
1039
- gen_helper_sve_ldffdd_be_zss, } },
1040
- { { gen_helper_sve_ldffbds_zd,
1041
- gen_helper_sve_ldffhds_be_zd,
1042
- gen_helper_sve_ldffsds_be_zd,
1043
- NULL, },
1044
- { gen_helper_sve_ldffbdu_zd,
1045
- gen_helper_sve_ldffhdu_be_zd,
1046
- gen_helper_sve_ldffsdu_be_zd,
1047
- gen_helper_sve_ldffdd_be_zd, } } } },
1048
+ /* First-fault */
1049
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1050
+ gen_helper_sve_ldffhds_le_zsu_mte,
1051
+ gen_helper_sve_ldffsds_le_zsu_mte,
1052
+ NULL, },
1053
+ { gen_helper_sve_ldffbdu_zsu_mte,
1054
+ gen_helper_sve_ldffhdu_le_zsu_mte,
1055
+ gen_helper_sve_ldffsdu_le_zsu_mte,
1056
+ gen_helper_sve_ldffdd_le_zsu_mte, } },
1057
+ { { gen_helper_sve_ldffbds_zss_mte,
1058
+ gen_helper_sve_ldffhds_le_zss_mte,
1059
+ gen_helper_sve_ldffsds_le_zss_mte,
1060
+ NULL, },
1061
+ { gen_helper_sve_ldffbdu_zss_mte,
1062
+ gen_helper_sve_ldffhdu_le_zss_mte,
1063
+ gen_helper_sve_ldffsdu_le_zss_mte,
1064
+ gen_helper_sve_ldffdd_le_zss_mte, } },
1065
+ { { gen_helper_sve_ldffbds_zd_mte,
1066
+ gen_helper_sve_ldffhds_le_zd_mte,
1067
+ gen_helper_sve_ldffsds_le_zd_mte,
1068
+ NULL, },
1069
+ { gen_helper_sve_ldffbdu_zd_mte,
1070
+ gen_helper_sve_ldffhdu_le_zd_mte,
1071
+ gen_helper_sve_ldffsdu_le_zd_mte,
1072
+ gen_helper_sve_ldffdd_le_zd_mte, } } } },
1073
+ { /* Big-endian */
1074
+ { { { gen_helper_sve_ldbds_zsu_mte,
1075
+ gen_helper_sve_ldhds_be_zsu_mte,
1076
+ gen_helper_sve_ldsds_be_zsu_mte,
1077
+ NULL, },
1078
+ { gen_helper_sve_ldbdu_zsu_mte,
1079
+ gen_helper_sve_ldhdu_be_zsu_mte,
1080
+ gen_helper_sve_ldsdu_be_zsu_mte,
1081
+ gen_helper_sve_lddd_be_zsu_mte, } },
1082
+ { { gen_helper_sve_ldbds_zss_mte,
1083
+ gen_helper_sve_ldhds_be_zss_mte,
1084
+ gen_helper_sve_ldsds_be_zss_mte,
1085
+ NULL, },
1086
+ { gen_helper_sve_ldbdu_zss_mte,
1087
+ gen_helper_sve_ldhdu_be_zss_mte,
1088
+ gen_helper_sve_ldsdu_be_zss_mte,
1089
+ gen_helper_sve_lddd_be_zss_mte, } },
1090
+ { { gen_helper_sve_ldbds_zd_mte,
1091
+ gen_helper_sve_ldhds_be_zd_mte,
1092
+ gen_helper_sve_ldsds_be_zd_mte,
1093
+ NULL, },
1094
+ { gen_helper_sve_ldbdu_zd_mte,
1095
+ gen_helper_sve_ldhdu_be_zd_mte,
1096
+ gen_helper_sve_ldsdu_be_zd_mte,
1097
+ gen_helper_sve_lddd_be_zd_mte, } } },
1098
+
1099
+ /* First-fault */
1100
+ { { { gen_helper_sve_ldffbds_zsu_mte,
1101
+ gen_helper_sve_ldffhds_be_zsu_mte,
1102
+ gen_helper_sve_ldffsds_be_zsu_mte,
1103
+ NULL, },
1104
+ { gen_helper_sve_ldffbdu_zsu_mte,
1105
+ gen_helper_sve_ldffhdu_be_zsu_mte,
1106
+ gen_helper_sve_ldffsdu_be_zsu_mte,
1107
+ gen_helper_sve_ldffdd_be_zsu_mte, } },
1108
+ { { gen_helper_sve_ldffbds_zss_mte,
1109
+ gen_helper_sve_ldffhds_be_zss_mte,
1110
+ gen_helper_sve_ldffsds_be_zss_mte,
1111
+ NULL, },
1112
+ { gen_helper_sve_ldffbdu_zss_mte,
1113
+ gen_helper_sve_ldffhdu_be_zss_mte,
1114
+ gen_helper_sve_ldffsdu_be_zss_mte,
1115
+ gen_helper_sve_ldffdd_be_zss_mte, } },
1116
+ { { gen_helper_sve_ldffbds_zd_mte,
1117
+ gen_helper_sve_ldffhds_be_zd_mte,
1118
+ gen_helper_sve_ldffsds_be_zd_mte,
1119
+ NULL, },
1120
+ { gen_helper_sve_ldffbdu_zd_mte,
1121
+ gen_helper_sve_ldffhdu_be_zd_mte,
1122
+ gen_helper_sve_ldffsdu_be_zd_mte,
1123
+ gen_helper_sve_ldffdd_be_zd_mte, } } } } },
1124
};
1125
1126
static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1127
{
1128
gen_helper_gvec_mem_scatter *fn = NULL;
1129
- int be = s->be_data == MO_BE;
1130
+ bool be = s->be_data == MO_BE;
1131
+ bool mte = s->mte_active[0];
1132
1133
if (!sve_access_check(s)) {
1134
return true;
1135
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zprz(DisasContext *s, arg_LD1_zprz *a)
1136
1137
switch (a->esz) {
1138
case MO_32:
1139
- fn = gather_load_fn32[be][a->ff][a->xs][a->u][a->msz];
1140
+ fn = gather_load_fn32[mte][be][a->ff][a->xs][a->u][a->msz];
1141
break;
20
break;
1142
case MO_64:
21
+ case ARM_VFP_FPCXT_S:
1143
- fn = gather_load_fn64[be][a->ff][a->xs][a->u][a->msz];
22
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
1144
+ fn = gather_load_fn64[mte][be][a->ff][a->xs][a->u][a->msz];
23
+ return false;
24
+ }
25
+ if (!s->v8m_secure) {
26
+ return false;
27
+ }
28
+ break;
29
default:
30
return FPSysRegCheckFailed;
31
}
32
@@ -XXX,XX +XXX,XX @@ static bool gen_M_fp_sysreg_write(DisasContext *s, int regno,
33
tcg_temp_free_i32(tmp);
1145
break;
34
break;
1146
}
35
}
1147
assert(fn != NULL);
36
+ case ARM_VFP_FPCXT_S:
1148
37
+ {
1149
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
38
+ TCGv_i32 sfpa, control, fpscr;
1150
- cpu_reg_sp(s, a->rn), a->msz, fn);
39
+ /* Set FPSCR[27:0] and CONTROL.SFPA from value */
1151
+ cpu_reg_sp(s, a->rn), a->msz, false, fn);
40
+ tmp = loadfn(s, opaque);
1152
return true;
41
+ sfpa = tcg_temp_new_i32();
1153
}
42
+ tcg_gen_shri_i32(sfpa, tmp, 31);
1154
43
+ control = load_cpu_field(v7m.control[M_REG_S]);
1155
static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
44
+ tcg_gen_deposit_i32(control, control, sfpa,
1156
{
45
+ R_V7M_CONTROL_SFPA_SHIFT, 1);
1157
gen_helper_gvec_mem_scatter *fn = NULL;
46
+ store_cpu_field(control, v7m.control[M_REG_S]);
1158
- int be = s->be_data == MO_BE;
47
+ fpscr = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
1159
+ bool be = s->be_data == MO_BE;
48
+ tcg_gen_andi_i32(fpscr, fpscr, FPCR_NZCV_MASK);
1160
+ bool mte = s->mte_active[0];
49
+ tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
1161
TCGv_i64 imm;
50
+ tcg_gen_or_i32(fpscr, fpscr, tmp);
1162
51
+ store_cpu_field(fpscr, vfp.xregs[ARM_VFP_FPSCR]);
1163
if (a->esz < a->msz || (a->esz == a->msz && !a->u)) {
52
+ tcg_temp_free_i32(tmp);
1164
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
53
+ tcg_temp_free_i32(sfpa);
1165
54
+ break;
1166
switch (a->esz) {
55
+ }
1167
case MO_32:
1168
- fn = gather_load_fn32[be][a->ff][0][a->u][a->msz];
1169
+ fn = gather_load_fn32[mte][be][a->ff][0][a->u][a->msz];
1170
break;
1171
case MO_64:
1172
- fn = gather_load_fn64[be][a->ff][2][a->u][a->msz];
1173
+ fn = gather_load_fn64[mte][be][a->ff][2][a->u][a->msz];
1174
break;
1175
}
1176
assert(fn != NULL);
1177
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1_zpiz(DisasContext *s, arg_LD1_zpiz *a)
1178
* by loading the immediate into the scalar parameter.
1179
*/
1180
imm = tcg_const_i64(a->imm << a->msz);
1181
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1182
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, false, fn);
1183
tcg_temp_free_i64(imm);
1184
return true;
1185
}
1186
1187
-/* Indexed by [be][xs][msz]. */
1188
-static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][3] = {
1189
- /* Little-endian */
1190
- { { gen_helper_sve_stbs_zsu,
1191
- gen_helper_sve_sths_le_zsu,
1192
- gen_helper_sve_stss_le_zsu, },
1193
- { gen_helper_sve_stbs_zss,
1194
- gen_helper_sve_sths_le_zss,
1195
- gen_helper_sve_stss_le_zss, } },
1196
- /* Big-endian */
1197
- { { gen_helper_sve_stbs_zsu,
1198
- gen_helper_sve_sths_be_zsu,
1199
- gen_helper_sve_stss_be_zsu, },
1200
- { gen_helper_sve_stbs_zss,
1201
- gen_helper_sve_sths_be_zss,
1202
- gen_helper_sve_stss_be_zss, } },
1203
+/* Indexed by [mte][be][xs][msz]. */
1204
+static gen_helper_gvec_mem_scatter * const scatter_store_fn32[2][2][2][3] = {
1205
+ { /* MTE Inactive */
1206
+ { /* Little-endian */
1207
+ { gen_helper_sve_stbs_zsu,
1208
+ gen_helper_sve_sths_le_zsu,
1209
+ gen_helper_sve_stss_le_zsu, },
1210
+ { gen_helper_sve_stbs_zss,
1211
+ gen_helper_sve_sths_le_zss,
1212
+ gen_helper_sve_stss_le_zss, } },
1213
+ { /* Big-endian */
1214
+ { gen_helper_sve_stbs_zsu,
1215
+ gen_helper_sve_sths_be_zsu,
1216
+ gen_helper_sve_stss_be_zsu, },
1217
+ { gen_helper_sve_stbs_zss,
1218
+ gen_helper_sve_sths_be_zss,
1219
+ gen_helper_sve_stss_be_zss, } } },
1220
+ { /* MTE Active */
1221
+ { /* Little-endian */
1222
+ { gen_helper_sve_stbs_zsu_mte,
1223
+ gen_helper_sve_sths_le_zsu_mte,
1224
+ gen_helper_sve_stss_le_zsu_mte, },
1225
+ { gen_helper_sve_stbs_zss_mte,
1226
+ gen_helper_sve_sths_le_zss_mte,
1227
+ gen_helper_sve_stss_le_zss_mte, } },
1228
+ { /* Big-endian */
1229
+ { gen_helper_sve_stbs_zsu_mte,
1230
+ gen_helper_sve_sths_be_zsu_mte,
1231
+ gen_helper_sve_stss_be_zsu_mte, },
1232
+ { gen_helper_sve_stbs_zss_mte,
1233
+ gen_helper_sve_sths_be_zss_mte,
1234
+ gen_helper_sve_stss_be_zss_mte, } } },
1235
};
1236
1237
/* Note that we overload xs=2 to indicate 64-bit offset. */
1238
-static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][3][4] = {
1239
- /* Little-endian */
1240
- { { gen_helper_sve_stbd_zsu,
1241
- gen_helper_sve_sthd_le_zsu,
1242
- gen_helper_sve_stsd_le_zsu,
1243
- gen_helper_sve_stdd_le_zsu, },
1244
- { gen_helper_sve_stbd_zss,
1245
- gen_helper_sve_sthd_le_zss,
1246
- gen_helper_sve_stsd_le_zss,
1247
- gen_helper_sve_stdd_le_zss, },
1248
- { gen_helper_sve_stbd_zd,
1249
- gen_helper_sve_sthd_le_zd,
1250
- gen_helper_sve_stsd_le_zd,
1251
- gen_helper_sve_stdd_le_zd, } },
1252
- /* Big-endian */
1253
- { { gen_helper_sve_stbd_zsu,
1254
- gen_helper_sve_sthd_be_zsu,
1255
- gen_helper_sve_stsd_be_zsu,
1256
- gen_helper_sve_stdd_be_zsu, },
1257
- { gen_helper_sve_stbd_zss,
1258
- gen_helper_sve_sthd_be_zss,
1259
- gen_helper_sve_stsd_be_zss,
1260
- gen_helper_sve_stdd_be_zss, },
1261
- { gen_helper_sve_stbd_zd,
1262
- gen_helper_sve_sthd_be_zd,
1263
- gen_helper_sve_stsd_be_zd,
1264
- gen_helper_sve_stdd_be_zd, } },
1265
+static gen_helper_gvec_mem_scatter * const scatter_store_fn64[2][2][3][4] = {
1266
+ { /* MTE Inactive */
1267
+ { /* Little-endian */
1268
+ { gen_helper_sve_stbd_zsu,
1269
+ gen_helper_sve_sthd_le_zsu,
1270
+ gen_helper_sve_stsd_le_zsu,
1271
+ gen_helper_sve_stdd_le_zsu, },
1272
+ { gen_helper_sve_stbd_zss,
1273
+ gen_helper_sve_sthd_le_zss,
1274
+ gen_helper_sve_stsd_le_zss,
1275
+ gen_helper_sve_stdd_le_zss, },
1276
+ { gen_helper_sve_stbd_zd,
1277
+ gen_helper_sve_sthd_le_zd,
1278
+ gen_helper_sve_stsd_le_zd,
1279
+ gen_helper_sve_stdd_le_zd, } },
1280
+ { /* Big-endian */
1281
+ { gen_helper_sve_stbd_zsu,
1282
+ gen_helper_sve_sthd_be_zsu,
1283
+ gen_helper_sve_stsd_be_zsu,
1284
+ gen_helper_sve_stdd_be_zsu, },
1285
+ { gen_helper_sve_stbd_zss,
1286
+ gen_helper_sve_sthd_be_zss,
1287
+ gen_helper_sve_stsd_be_zss,
1288
+ gen_helper_sve_stdd_be_zss, },
1289
+ { gen_helper_sve_stbd_zd,
1290
+ gen_helper_sve_sthd_be_zd,
1291
+ gen_helper_sve_stsd_be_zd,
1292
+ gen_helper_sve_stdd_be_zd, } } },
1293
+ { /* MTE Inactive */
1294
+ { /* Little-endian */
1295
+ { gen_helper_sve_stbd_zsu_mte,
1296
+ gen_helper_sve_sthd_le_zsu_mte,
1297
+ gen_helper_sve_stsd_le_zsu_mte,
1298
+ gen_helper_sve_stdd_le_zsu_mte, },
1299
+ { gen_helper_sve_stbd_zss_mte,
1300
+ gen_helper_sve_sthd_le_zss_mte,
1301
+ gen_helper_sve_stsd_le_zss_mte,
1302
+ gen_helper_sve_stdd_le_zss_mte, },
1303
+ { gen_helper_sve_stbd_zd_mte,
1304
+ gen_helper_sve_sthd_le_zd_mte,
1305
+ gen_helper_sve_stsd_le_zd_mte,
1306
+ gen_helper_sve_stdd_le_zd_mte, } },
1307
+ { /* Big-endian */
1308
+ { gen_helper_sve_stbd_zsu_mte,
1309
+ gen_helper_sve_sthd_be_zsu_mte,
1310
+ gen_helper_sve_stsd_be_zsu_mte,
1311
+ gen_helper_sve_stdd_be_zsu_mte, },
1312
+ { gen_helper_sve_stbd_zss_mte,
1313
+ gen_helper_sve_sthd_be_zss_mte,
1314
+ gen_helper_sve_stsd_be_zss_mte,
1315
+ gen_helper_sve_stdd_be_zss_mte, },
1316
+ { gen_helper_sve_stbd_zd_mte,
1317
+ gen_helper_sve_sthd_be_zd_mte,
1318
+ gen_helper_sve_stsd_be_zd_mte,
1319
+ gen_helper_sve_stdd_be_zd_mte, } } },
1320
};
1321
1322
static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1323
{
1324
gen_helper_gvec_mem_scatter *fn;
1325
- int be = s->be_data == MO_BE;
1326
+ bool be = s->be_data == MO_BE;
1327
+ bool mte = s->mte_active[0];
1328
1329
if (a->esz < a->msz || (a->msz == 0 && a->scale)) {
1330
return false;
1331
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zprz(DisasContext *s, arg_ST1_zprz *a)
1332
}
1333
switch (a->esz) {
1334
case MO_32:
1335
- fn = scatter_store_fn32[be][a->xs][a->msz];
1336
+ fn = scatter_store_fn32[mte][be][a->xs][a->msz];
1337
break;
1338
case MO_64:
1339
- fn = scatter_store_fn64[be][a->xs][a->msz];
1340
+ fn = scatter_store_fn64[mte][be][a->xs][a->msz];
1341
break;
1342
default:
56
default:
1343
g_assert_not_reached();
57
g_assert_not_reached();
1344
}
58
}
1345
do_mem_zpz(s, a->rd, a->pg, a->rm, a->scale * a->msz,
59
@@ -XXX,XX +XXX,XX @@ static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
1346
- cpu_reg_sp(s, a->rn), a->msz, fn);
60
tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
1347
+ cpu_reg_sp(s, a->rn), a->msz, true, fn);
61
storefn(s, opaque, tmp);
1348
return true;
1349
}
1350
1351
static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1352
{
1353
gen_helper_gvec_mem_scatter *fn = NULL;
1354
- int be = s->be_data == MO_BE;
1355
+ bool be = s->be_data == MO_BE;
1356
+ bool mte = s->mte_active[0];
1357
TCGv_i64 imm;
1358
1359
if (a->esz < a->msz) {
1360
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1361
1362
switch (a->esz) {
1363
case MO_32:
1364
- fn = scatter_store_fn32[be][0][a->msz];
1365
+ fn = scatter_store_fn32[mte][be][0][a->msz];
1366
break;
62
break;
1367
case MO_64:
63
+ case ARM_VFP_FPCXT_S:
1368
- fn = scatter_store_fn64[be][2][a->msz];
64
+ {
1369
+ fn = scatter_store_fn64[mte][be][2][a->msz];
65
+ TCGv_i32 control, sfpa, fpscr;
1370
break;
66
+ /* Bits [27:0] from FPSCR, bit [31] from CONTROL.SFPA */
67
+ tmp = tcg_temp_new_i32();
68
+ sfpa = tcg_temp_new_i32();
69
+ gen_helper_vfp_get_fpscr(tmp, cpu_env);
70
+ tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
71
+ control = load_cpu_field(v7m.control[M_REG_S]);
72
+ tcg_gen_andi_i32(sfpa, control, R_V7M_CONTROL_SFPA_MASK);
73
+ tcg_gen_shli_i32(sfpa, sfpa, 31 - R_V7M_CONTROL_SFPA_SHIFT);
74
+ tcg_gen_or_i32(tmp, tmp, sfpa);
75
+ tcg_temp_free_i32(sfpa);
76
+ /*
77
+ * Store result before updating FPSCR etc, in case
78
+ * it is a memory write which causes an exception.
79
+ */
80
+ storefn(s, opaque, tmp);
81
+ /*
82
+ * Now we must reset FPSCR from FPDSCR_NS, and clear
83
+ * CONTROL.SFPA; so we'll end the TB here.
84
+ */
85
+ tcg_gen_andi_i32(control, control, ~R_V7M_CONTROL_SFPA_MASK);
86
+ store_cpu_field(control, v7m.control[M_REG_S]);
87
+ fpscr = load_cpu_field(v7m.fpdscr[M_REG_NS]);
88
+ gen_helper_vfp_set_fpscr(cpu_env, fpscr);
89
+ tcg_temp_free_i32(fpscr);
90
+ gen_lookup_tb(s);
91
+ break;
92
+ }
93
default:
94
g_assert_not_reached();
1371
}
95
}
1372
assert(fn != NULL);
1373
@@ -XXX,XX +XXX,XX @@ static bool trans_ST1_zpiz(DisasContext *s, arg_ST1_zpiz *a)
1374
* by loading the immediate into the scalar parameter.
1375
*/
1376
imm = tcg_const_i64(a->imm << a->msz);
1377
- do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, fn);
1378
+ do_mem_zpz(s, a->rd, a->pg, a->rn, 0, imm, a->msz, true, fn);
1379
tcg_temp_free_i64(imm);
1380
return true;
1381
}
1382
--
96
--
1383
2.20.1
97
2.20.1
1384
98
1385
99
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
The FPDSCR register has a similar layout to the FPSCR. In v8.1M it
2
gains new fields FZ16 (if half-precision floating point is supported)
3
and LTPSIZE (always reads as 4). Update the reset value and the code
4
that handles writes to this register accordingly.
2
5
3
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
4
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
5
Message-id: 20200626033144.790098-44-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20201119215617.29887-16-peter.maydell@linaro.org
7
---
9
---
8
target/arm/cpu.h | 6 ++++++
10
target/arm/cpu.h | 5 +++++
9
hw/arm/virt.c | 55 ++++++++++++++++++++++++++++++++++++++++++++++--
11
hw/intc/armv7m_nvic.c | 9 ++++++++-
10
target/arm/cpu.c | 52 +++++++++++++++++++++++++++++++++++++++++----
12
target/arm/cpu.c | 3 +++
11
3 files changed, 107 insertions(+), 6 deletions(-)
13
3 files changed, 16 insertions(+), 1 deletion(-)
12
14
13
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
15
--- a/target/arm/cpu.h
17
--- a/target/arm/cpu.h
16
+++ b/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
17
@@ -XXX,XX +XXX,XX @@ struct ARMCPU {
19
@@ -XXX,XX +XXX,XX @@ void vfp_set_fpscr(CPUARMState *env, uint32_t val);
18
/* MemoryRegion to use for secure physical accesses */
20
#define FPCR_IXE (1 << 12) /* Inexact exception trap enable */
19
MemoryRegion *secure_memory;
21
#define FPCR_IDE (1 << 15) /* Input Denormal exception trap enable */
20
22
#define FPCR_FZ16 (1 << 19) /* ARMv8.2+, FP16 flush-to-zero */
21
+ /* MemoryRegion to use for allocation tag accesses */
23
+#define FPCR_RMODE_MASK (3 << 22) /* Rounding mode */
22
+ MemoryRegion *tag_memory;
24
#define FPCR_FZ (1 << 24) /* Flush-to-zero enable bit */
23
+ MemoryRegion *secure_tag_memory;
25
#define FPCR_DN (1 << 25) /* Default NaN enable bit */
26
+#define FPCR_AHP (1 << 26) /* Alternative half-precision */
27
#define FPCR_QC (1 << 27) /* Cumulative saturation bit */
28
#define FPCR_V (1 << 28) /* FP overflow flag */
29
#define FPCR_C (1 << 29) /* FP carry flag */
30
#define FPCR_Z (1 << 30) /* FP zero flag */
31
#define FPCR_N (1 << 31) /* FP negative flag */
32
33
+#define FPCR_LTPSIZE_SHIFT 16 /* LTPSIZE, M-profile only */
34
+#define FPCR_LTPSIZE_MASK (7 << FPCR_LTPSIZE_SHIFT)
24
+
35
+
25
/* For v8M, pointer to the IDAU interface provided by board/SoC */
36
#define FPCR_NZCV_MASK (FPCR_N | FPCR_Z | FPCR_C | FPCR_V)
26
Object *idau;
37
#define FPCR_NZCVQC_MASK (FPCR_NZCV_MASK | FPCR_QC)
27
38
28
@@ -XXX,XX +XXX,XX @@ typedef enum ARMMMUIdxBit {
39
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
29
typedef enum ARMASIdx {
30
ARMASIdx_NS = 0,
31
ARMASIdx_S = 1,
32
+ ARMASIdx_TagNS = 2,
33
+ ARMASIdx_TagS = 3,
34
} ARMASIdx;
35
36
/* Return the Exception Level targeted by debug exceptions. */
37
diff --git a/hw/arm/virt.c b/hw/arm/virt.c
38
index XXXXXXX..XXXXXXX 100644
40
index XXXXXXX..XXXXXXX 100644
39
--- a/hw/arm/virt.c
41
--- a/hw/intc/armv7m_nvic.c
40
+++ b/hw/arm/virt.c
42
+++ b/hw/intc/armv7m_nvic.c
41
@@ -XXX,XX +XXX,XX @@ static void create_platform_bus(VirtMachineState *vms)
43
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
42
sysbus_mmio_get_region(s, 0));
44
break;
43
}
45
case 0xf3c: /* FPDSCR */
44
46
if (cpu_isar_feature(aa32_vfp_simd, cpu)) {
45
+static void create_tag_ram(MemoryRegion *tag_sysmem,
47
- value &= 0x07c00000;
46
+ hwaddr base, hwaddr size,
48
+ uint32_t mask = FPCR_AHP | FPCR_DN | FPCR_FZ | FPCR_RMODE_MASK;
47
+ const char *name)
49
+ if (cpu_isar_feature(any_fp16, cpu)) {
48
+{
50
+ mask |= FPCR_FZ16;
49
+ MemoryRegion *tagram = g_new(MemoryRegion, 1);
51
+ }
50
+
52
+ value &= mask;
51
+ memory_region_init_ram(tagram, NULL, name, size / 32, &error_fatal);
53
+ if (cpu_isar_feature(aa32_lob, cpu)) {
52
+ memory_region_add_subregion(tag_sysmem, base / 32, tagram);
54
+ value |= 4 << FPCR_LTPSIZE_SHIFT;
53
+}
55
+ }
54
+
56
cpu->env.v7m.fpdscr[attrs.secure] = value;
55
static void create_secure_ram(VirtMachineState *vms,
56
- MemoryRegion *secure_sysmem)
57
+ MemoryRegion *secure_sysmem,
58
+ MemoryRegion *secure_tag_sysmem)
59
{
60
MemoryRegion *secram = g_new(MemoryRegion, 1);
61
char *nodename;
62
@@ -XXX,XX +XXX,XX @@ static void create_secure_ram(VirtMachineState *vms,
63
qemu_fdt_setprop_string(vms->fdt, nodename, "status", "disabled");
64
qemu_fdt_setprop_string(vms->fdt, nodename, "secure-status", "okay");
65
66
+ if (secure_tag_sysmem) {
67
+ create_tag_ram(secure_tag_sysmem, base, size, "mach-virt.secure-tag");
68
+ }
69
+
70
g_free(nodename);
71
}
72
73
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
74
const CPUArchIdList *possible_cpus;
75
MemoryRegion *sysmem = get_system_memory();
76
MemoryRegion *secure_sysmem = NULL;
77
+ MemoryRegion *tag_sysmem = NULL;
78
+ MemoryRegion *secure_tag_sysmem = NULL;
79
int n, virt_max_cpus;
80
bool firmware_loaded;
81
bool aarch64 = true;
82
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
83
"secure-memory", &error_abort);
84
}
57
}
85
58
break;
86
+ /*
87
+ * The cpu adds the property if and only if MemTag is supported.
88
+ * If it is, we must allocate the ram to back that up.
89
+ */
90
+ if (object_property_find(cpuobj, "tag-memory", NULL)) {
91
+ if (!tag_sysmem) {
92
+ tag_sysmem = g_new(MemoryRegion, 1);
93
+ memory_region_init(tag_sysmem, OBJECT(machine),
94
+ "tag-memory", UINT64_MAX / 32);
95
+
96
+ if (vms->secure) {
97
+ secure_tag_sysmem = g_new(MemoryRegion, 1);
98
+ memory_region_init(secure_tag_sysmem, OBJECT(machine),
99
+ "secure-tag-memory", UINT64_MAX / 32);
100
+
101
+ /* As with ram, secure-tag takes precedence over tag. */
102
+ memory_region_add_subregion_overlap(secure_tag_sysmem, 0,
103
+ tag_sysmem, -1);
104
+ }
105
+ }
106
+
107
+ object_property_set_link(cpuobj, OBJECT(tag_sysmem),
108
+ "tag-memory", &error_abort);
109
+ if (vms->secure) {
110
+ object_property_set_link(cpuobj, OBJECT(secure_tag_sysmem),
111
+ "secure-tag-memory", &error_abort);
112
+ }
113
+ }
114
+
115
qdev_realize(DEVICE(cpuobj), NULL, &error_fatal);
116
object_unref(cpuobj);
117
}
118
@@ -XXX,XX +XXX,XX @@ static void machvirt_init(MachineState *machine)
119
create_uart(vms, VIRT_UART, sysmem, serial_hd(0));
120
121
if (vms->secure) {
122
- create_secure_ram(vms, secure_sysmem);
123
+ create_secure_ram(vms, secure_sysmem, secure_tag_sysmem);
124
create_uart(vms, VIRT_SECURE_UART, secure_sysmem, serial_hd(1));
125
}
126
127
+ if (tag_sysmem) {
128
+ create_tag_ram(tag_sysmem, vms->memmap[VIRT_MEM].base,
129
+ machine->ram_size, "mach-virt.tag");
130
+ }
131
+
132
vms->highmem_ecam &= vms->highmem && (!firmware_loaded || aarch64);
133
134
create_rtc(vms);
135
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
59
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
136
index XXXXXXX..XXXXXXX 100644
60
index XXXXXXX..XXXXXXX 100644
137
--- a/target/arm/cpu.c
61
--- a/target/arm/cpu.c
138
+++ b/target/arm/cpu.c
62
+++ b/target/arm/cpu.c
139
@@ -XXX,XX +XXX,XX @@ void arm_cpu_post_init(Object *obj)
63
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_reset(DeviceState *dev)
140
if (kvm_enabled()) {
64
* always reset to 4.
141
kvm_arm_add_vcpu_properties(obj);
65
*/
142
}
66
env->v7m.ltpsize = 4;
143
+
67
+ /* The LTPSIZE field in FPDSCR is constant and reads as 4. */
144
+#ifndef CONFIG_USER_ONLY
68
+ env->v7m.fpdscr[M_REG_NS] = 4 << FPCR_LTPSIZE_SHIFT;
145
+ if (arm_feature(&cpu->env, ARM_FEATURE_AARCH64) &&
69
+ env->v7m.fpdscr[M_REG_S] = 4 << FPCR_LTPSIZE_SHIFT;
146
+ cpu_isar_feature(aa64_mte, cpu)) {
147
+ object_property_add_link(obj, "tag-memory",
148
+ TYPE_MEMORY_REGION,
149
+ (Object **)&cpu->tag_memory,
150
+ qdev_prop_allow_set_link_before_realize,
151
+ OBJ_PROP_LINK_STRONG);
152
+
153
+ if (arm_feature(&cpu->env, ARM_FEATURE_EL3)) {
154
+ object_property_add_link(obj, "secure-tag-memory",
155
+ TYPE_MEMORY_REGION,
156
+ (Object **)&cpu->secure_tag_memory,
157
+ qdev_prop_allow_set_link_before_realize,
158
+ OBJ_PROP_LINK_STRONG);
159
+ }
160
+ }
161
+#endif
162
}
163
164
static void arm_cpu_finalizefn(Object *obj)
165
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_realizefn(DeviceState *dev, Error **errp)
166
#ifndef CONFIG_USER_ONLY
167
MachineState *ms = MACHINE(qdev_get_machine());
168
unsigned int smp_cpus = ms->smp.cpus;
169
+ bool has_secure = cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY);
170
171
- if (cpu->has_el3 || arm_feature(env, ARM_FEATURE_M_SECURITY)) {
172
- cs->num_ases = 2;
173
+ /*
174
+ * We must set cs->num_ases to the final value before
175
+ * the first call to cpu_address_space_init.
176
+ */
177
+ if (cpu->tag_memory != NULL) {
178
+ cs->num_ases = 3 + has_secure;
179
+ } else {
180
+ cs->num_ases = 1 + has_secure;
181
+ }
182
183
+ if (has_secure) {
184
if (!cpu->secure_memory) {
185
cpu->secure_memory = cs->memory;
186
}
70
}
187
cpu_address_space_init(cs, ARMASIdx_S, "cpu-secure-memory",
71
188
cpu->secure_memory);
72
if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
189
- } else {
190
- cs->num_ases = 1;
191
}
192
+
193
+ if (cpu->tag_memory != NULL) {
194
+ cpu_address_space_init(cs, ARMASIdx_TagNS, "cpu-tag-memory",
195
+ cpu->tag_memory);
196
+ if (has_secure) {
197
+ cpu_address_space_init(cs, ARMASIdx_TagS, "cpu-tag-memory",
198
+ cpu->secure_tag_memory);
199
+ }
200
+ } else if (cpu_isar_feature(aa64_mte, cpu)) {
201
+ /*
202
+ * Since there is no tag memory, we can't meaningfully support MTE
203
+ * to its fullest. To avoid problems later, when we would come to
204
+ * use the tag memory, downgrade support to insns only.
205
+ */
206
+ cpu->isar.id_aa64pfr1 =
207
+ FIELD_DP64(cpu->isar.id_aa64pfr1, ID_AA64PFR1, MTE, 1);
208
+ }
209
+
210
cpu_address_space_init(cs, ARMASIdx_NS, "cpu-memory", cs->memory);
211
212
/* No core_count specified, default to smp_cpus. */
213
--
73
--
214
2.20.1
74
2.20.1
215
75
216
76
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.0M, on exception entry the registers R0-R3, R12, APSR and EPSR
2
are zeroed for an exception taken to Non-secure state; for an
3
exception taken to Secure state they become UNKNOWN, and we chose to
4
leave them at their previous values.
2
5
3
We still need to handle tbi for user-only when mte is inactive.
6
In v8.1M the behaviour is specified more tightly and these registers
7
are always zeroed regardless of the security state that the exception
8
targets (see rule R_KPZV). Implement this.
4
9
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-37-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20201119215617.29887-17-peter.maydell@linaro.org
9
---
13
---
10
target/arm/translate-a64.h | 1 +
14
target/arm/m_helper.c | 16 ++++++++++++----
11
target/arm/translate-a64.c | 2 +-
15
1 file changed, 12 insertions(+), 4 deletions(-)
12
target/arm/translate-sve.c | 6 ++++--
13
3 files changed, 6 insertions(+), 3 deletions(-)
14
16
15
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
17
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
16
index XXXXXXX..XXXXXXX 100644
18
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/translate-a64.h
19
--- a/target/arm/m_helper.c
18
+++ b/target/arm/translate-a64.h
20
+++ b/target/arm/m_helper.c
19
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
21
@@ -XXX,XX +XXX,XX @@ static void v7m_exception_taken(ARMCPU *cpu, uint32_t lr, bool dotailchain,
20
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
22
* Clear registers if necessary to prevent non-secure exception
21
unsigned int imms, unsigned int immr);
23
* code being able to see register values from secure code.
22
bool sve_access_check(DisasContext *s);
24
* Where register values become architecturally UNKNOWN we leave
23
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr);
25
- * them with their previous values.
24
TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
26
+ * them with their previous values. v8.1M is tighter than v8.0M
25
bool tag_checked, int log2_size);
27
+ * here and always zeroes the caller-saved registers regardless
26
TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
28
+ * of the security state the exception is targeting.
27
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
29
*/
28
index XXXXXXX..XXXXXXX 100644
30
if (arm_feature(env, ARM_FEATURE_M_SECURITY)) {
29
--- a/target/arm/translate-a64.c
31
- if (!targets_secure) {
30
+++ b/target/arm/translate-a64.c
32
+ if (!targets_secure || arm_feature(env, ARM_FEATURE_V8_1M)) {
31
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
33
/*
32
* of the write-back address.
34
* Always clear the caller-saved registers (they have been
33
*/
35
* pushed to the stack earlier in v7m_push_stack()).
34
36
@@ -XXX,XX +XXX,XX @@ static void v7m_exception_taken(ARMCPU *cpu, uint32_t lr, bool dotailchain,
35
-static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
37
* v7m_push_callee_stack()).
36
+TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
38
*/
37
{
39
int i;
38
TCGv_i64 clean = new_tmp_a64(s);
40
+ /*
39
#ifdef CONFIG_USER_ONLY
41
+ * r4..r11 are callee-saves, zero only if background
40
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
42
+ * state was Secure (EXCRET.S == 1) and exception
41
index XXXXXXX..XXXXXXX 100644
43
+ * targets Non-secure state
42
--- a/target/arm/translate-sve.c
44
+ */
43
+++ b/target/arm/translate-sve.c
45
+ bool zero_callee_saves = !targets_secure &&
44
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
46
+ (lr & R_V7M_EXCRET_S_MASK);
45
* For e.g. LD4, there are not enough arguments to pass all 4
47
46
* registers as pointers, so encode the regno into the data field.
48
for (i = 0; i < 13; i++) {
47
* For consistency, do this even for LD1.
49
- /* r4..r11 are callee-saves, zero only if EXCRET.S == 1 */
48
- * TODO: mte_n check here while callers are updated.
50
- if (i < 4 || i > 11 || (lr & R_V7M_EXCRET_S_MASK)) {
49
*/
51
+ if (i < 4 || i > 11 || zero_callee_saves) {
50
- if (mte_n && s->mte_active[0]) {
52
env->regs[i] = 0;
51
+ if (s->mte_active[0]) {
53
}
52
int msz = dtype_msz(dtype);
54
}
53
54
desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
55
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
56
desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
57
desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
58
desc <<= SVE_MTEDESC_SHIFT;
59
+ } else {
60
+ addr = clean_data_tbi(s, addr);
61
}
62
+
63
desc = simd_desc(vsz, vsz, zt | desc);
64
t_desc = tcg_const_i32(desc);
65
t_pg = tcg_temp_new_ptr();
66
--
55
--
67
2.20.1
56
2.20.1
68
57
69
58
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.1M, vector table fetch failures don't set HFSR.FORCED (see rule
2
R_LLRP). (In previous versions of the architecture this was either
3
required or IMPDEF.)
2
4
3
Now that we know that the operation is on a single page,
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
we need not loop over pages while probing.
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201119215617.29887-18-peter.maydell@linaro.org
8
---
9
target/arm/m_helper.c | 6 +++++-
10
1 file changed, 5 insertions(+), 1 deletion(-)
5
11
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-19-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-a64.c | 94 +++++++++++------------------------------
12
1 file changed, 25 insertions(+), 69 deletions(-)
13
14
diff --git a/target/arm/helper-a64.c b/target/arm/helper-a64.c
15
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.c
14
--- a/target/arm/m_helper.c
17
+++ b/target/arm/helper-a64.c
15
+++ b/target/arm/m_helper.c
18
@@ -XXX,XX +XXX,XX @@ void HELPER(dc_zva)(CPUARMState *env, uint64_t vaddr_in)
16
@@ -XXX,XX +XXX,XX @@ load_fail:
19
* (which matches the usual QEMU behaviour of not implementing either
17
* The HardFault is Secure if BFHFNMINS is 0 (meaning that all HFs are
20
* alignment faults or any memory attribute handling).
18
* secure); otherwise it targets the same security state as the
19
* underlying exception.
20
+ * In v8.1M HardFaults from vector table fetch fails don't set FORCED.
21
*/
21
*/
22
-
22
if (!(cpu->env.v7m.aircr & R_V7M_AIRCR_BFHFNMINS_MASK)) {
23
- ARMCPU *cpu = env_archcpu(env);
23
exc_secure = true;
24
- uint64_t blocklen = 4 << cpu->dcz_blocksize;
25
+ int blocklen = 4 << env_archcpu(env)->dcz_blocksize;
26
uint64_t vaddr = vaddr_in & ~(blocklen - 1);
27
+ int mmu_idx = cpu_mmu_index(env, false);
28
+ void *mem;
29
+
30
+ /*
31
+ * Trapless lookup. In addition to actual invalid page, may
32
+ * return NULL for I/O, watchpoints, clean pages, etc.
33
+ */
34
+ mem = tlb_vaddr_to_host(env, vaddr, MMU_DATA_STORE, mmu_idx);
35
36
#ifndef CONFIG_USER_ONLY
37
- {
38
+ if (unlikely(!mem)) {
39
+ uintptr_t ra = GETPC();
40
+
41
/*
42
- * Slightly awkwardly, QEMU's TARGET_PAGE_SIZE may be less than
43
- * the block size so we might have to do more than one TLB lookup.
44
- * We know that in fact for any v8 CPU the page size is at least 4K
45
- * and the block size must be 2K or less, but TARGET_PAGE_SIZE is only
46
- * 1K as an artefact of legacy v5 subpage support being present in the
47
- * same QEMU executable. So in practice the hostaddr[] array has
48
- * two entries, given the current setting of TARGET_PAGE_BITS_MIN.
49
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
50
+ * the original pointer for an invalid page. But watchpoints require
51
+ * that we probe the actual space. So do both.
52
*/
53
- int maxidx = DIV_ROUND_UP(blocklen, TARGET_PAGE_SIZE);
54
- void *hostaddr[DIV_ROUND_UP(2 * KiB, 1 << TARGET_PAGE_BITS_MIN)];
55
- int try, i;
56
- unsigned mmu_idx = cpu_mmu_index(env, false);
57
- TCGMemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
58
+ (void) probe_write(env, vaddr_in, 1, mmu_idx, ra);
59
+ mem = probe_write(env, vaddr, blocklen, mmu_idx, ra);
60
61
- assert(maxidx <= ARRAY_SIZE(hostaddr));
62
-
63
- for (try = 0; try < 2; try++) {
64
-
65
- for (i = 0; i < maxidx; i++) {
66
- hostaddr[i] = tlb_vaddr_to_host(env,
67
- vaddr + TARGET_PAGE_SIZE * i,
68
- 1, mmu_idx);
69
- if (!hostaddr[i]) {
70
- break;
71
- }
72
- }
73
- if (i == maxidx) {
74
- /*
75
- * If it's all in the TLB it's fair game for just writing to;
76
- * we know we don't need to update dirty status, etc.
77
- */
78
- for (i = 0; i < maxidx - 1; i++) {
79
- memset(hostaddr[i], 0, TARGET_PAGE_SIZE);
80
- }
81
- memset(hostaddr[i], 0, blocklen - (i * TARGET_PAGE_SIZE));
82
- return;
83
- }
84
+ if (unlikely(!mem)) {
85
/*
86
- * OK, try a store and see if we can populate the tlb. This
87
- * might cause an exception if the memory isn't writable,
88
- * in which case we will longjmp out of here. We must for
89
- * this purpose use the actual register value passed to us
90
- * so that we get the fault address right.
91
+ * The only remaining reason for mem == NULL is I/O.
92
+ * Just do a series of byte writes as the architecture demands.
93
*/
94
- helper_ret_stb_mmu(env, vaddr_in, 0, oi, GETPC());
95
- /* Now we can populate the other TLB entries, if any */
96
- for (i = 0; i < maxidx; i++) {
97
- uint64_t va = vaddr + TARGET_PAGE_SIZE * i;
98
- if (va != (vaddr_in & TARGET_PAGE_MASK)) {
99
- helper_ret_stb_mmu(env, va, 0, oi, GETPC());
100
- }
101
+ for (int i = 0; i < blocklen; i++) {
102
+ cpu_stb_mmuidx_ra(env, vaddr + i, 0, mmu_idx, ra);
103
}
104
- }
105
-
106
- /*
107
- * Slow path (probably attempt to do this to an I/O device or
108
- * similar, or clearing of a block of code we have translations
109
- * cached for). Just do a series of byte writes as the architecture
110
- * demands. It's not worth trying to use a cpu_physical_memory_map(),
111
- * memset(), unmap() sequence here because:
112
- * + we'd need to account for the blocksize being larger than a page
113
- * + the direct-RAM access case is almost always going to be dealt
114
- * with in the fastpath code above, so there's no speed benefit
115
- * + we would have to deal with the map returning NULL because the
116
- * bounce buffer was in use
117
- */
118
- for (i = 0; i < blocklen; i++) {
119
- helper_ret_stb_mmu(env, vaddr + i, 0, oi, GETPC());
120
+ return;
121
}
122
}
24
}
123
-#else
25
- env->v7m.hfsr |= R_V7M_HFSR_VECTTBL_MASK | R_V7M_HFSR_FORCED_MASK;
124
- memset(g2h(vaddr), 0, blocklen);
26
+ env->v7m.hfsr |= R_V7M_HFSR_VECTTBL_MASK;
125
#endif
27
+ if (!arm_feature(env, ARM_FEATURE_V8_1M)) {
126
+
28
+ env->v7m.hfsr |= R_V7M_HFSR_FORCED_MASK;
127
+ memset(mem, 0, blocklen);
29
+ }
30
armv7m_nvic_set_pending_derived(env->nvic, ARMV7M_EXCP_HARD, exc_secure);
31
return false;
128
}
32
}
129
--
33
--
130
2.20.1
34
2.20.1
131
35
132
36
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.1M a REVIDR register is defined, which is at address 0xe00ecfc
2
and is a read-only IMPDEF register providing implementation specific
3
minor revision information, like the v8A REVIDR_EL1. Implement this.
2
4
3
We now implement all of the components of MTE, without actually
4
supporting any tagged memory. All MTE instructions will work,
5
trivially, so we can enable support.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-46-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201119215617.29887-19-peter.maydell@linaro.org
11
---
8
---
12
target/arm/cpu64.c | 5 +++++
9
hw/intc/armv7m_nvic.c | 5 +++++
13
1 file changed, 5 insertions(+)
10
1 file changed, 5 insertions(+)
14
11
15
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
12
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
16
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu64.c
14
--- a/hw/intc/armv7m_nvic.c
18
+++ b/target/arm/cpu64.c
15
+++ b/hw/intc/armv7m_nvic.c
19
@@ -XXX,XX +XXX,XX @@ static void aarch64_max_initfn(Object *obj)
16
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
20
17
}
21
t = cpu->isar.id_aa64pfr1;
18
return val;
22
t = FIELD_DP64(t, ID_AA64PFR1, BT, 1);
19
}
23
+ /*
20
+ case 0xcfc:
24
+ * Begin with full support for MTE; will be downgraded to MTE=1
21
+ if (!arm_feature(&cpu->env, ARM_FEATURE_V8_1M)) {
25
+ * during realize if the board provides no tag memory.
22
+ goto bad_offset;
26
+ */
23
+ }
27
+ t = FIELD_DP64(t, ID_AA64PFR1, MTE, 2);
24
+ return cpu->revidr;
28
cpu->isar.id_aa64pfr1 = t;
25
case 0xd00: /* CPUID Base. */
29
26
return cpu->midr;
30
t = cpu->isar.id_aa64mmfr1;
27
case 0xd04: /* Interrupt Control State (ICSR) */
31
--
28
--
32
2.20.1
29
2.20.1
33
30
34
31
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In v8.1M a new exception return check is added which may cause a NOCP
2
UsageFault (see rule R_XLTP): before we clear s0..s15 and the FPSCR
3
we must check whether access to CP10 from the Security state of the
4
returning exception is disabled; if it is then we must take a fault.
2
5
3
We need to check the memattr of a page in order to determine
6
(Note that for our implementation CPPWR is always RAZ/WI and so can
4
whether it is Tagged for MTE. Between Stage1 and Stage2,
7
never cause CP10 accesses to fail.)
5
this becomes simpler if we always collect this data, instead
6
of occasionally being presented with NULL.
7
8
8
Use the nonnull attribute to allow the compiler to check that
9
The other v8.1M change to this register-clearing code is that if MVE
9
all pointer arguments are non-null.
10
is implemented VPR must also be cleared, so add a TODO comment to
11
that effect.
10
12
11
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Message-id: 20200626033144.790098-42-richard.henderson@linaro.org
14
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 20201119215617.29887-20-peter.maydell@linaro.org
15
---
16
---
16
target/arm/internals.h | 3 ++-
17
target/arm/m_helper.c | 22 +++++++++++++++++++++-
17
target/arm/helper.c | 60 ++++++++++++++++++++---------------------
18
1 file changed, 21 insertions(+), 1 deletion(-)
18
target/arm/m_helper.c | 11 +++++---
19
target/arm/tlb_helper.c | 4 ++-
20
4 files changed, 42 insertions(+), 36 deletions(-)
21
19
22
diff --git a/target/arm/internals.h b/target/arm/internals.h
23
index XXXXXXX..XXXXXXX 100644
24
--- a/target/arm/internals.h
25
+++ b/target/arm/internals.h
26
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
27
MMUAccessType access_type, ARMMMUIdx mmu_idx,
28
hwaddr *phys_ptr, MemTxAttrs *attrs, int *prot,
29
target_ulong *page_size,
30
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
31
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
32
+ __attribute__((nonnull));
33
34
void arm_log_exception(int idx);
35
36
diff --git a/target/arm/helper.c b/target/arm/helper.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/target/arm/helper.c
39
+++ b/target/arm/helper.c
40
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
41
bool s1_is_el0,
42
hwaddr *phys_ptr, MemTxAttrs *txattrs, int *prot,
43
target_ulong *page_size_ptr,
44
- ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs);
45
+ ARMMMUFaultInfo *fi, ARMCacheAttrs *cacheattrs)
46
+ __attribute__((nonnull));
47
#endif
48
49
static void switch_mode(CPUARMState *env, int mode);
50
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
51
arm_tlb_bti_gp(txattrs) = true;
52
}
53
54
- if (cacheattrs != NULL) {
55
- if (mmu_idx == ARMMMUIdx_Stage2) {
56
- cacheattrs->attrs = convert_stage2_attrs(env,
57
- extract32(attrs, 0, 4));
58
- } else {
59
- /* Index into MAIR registers for cache attributes */
60
- uint8_t attrindx = extract32(attrs, 0, 3);
61
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
62
- assert(attrindx <= 7);
63
- cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
64
- }
65
- cacheattrs->shareability = extract32(attrs, 6, 2);
66
+ if (mmu_idx == ARMMMUIdx_Stage2) {
67
+ cacheattrs->attrs = convert_stage2_attrs(env, extract32(attrs, 0, 4));
68
+ } else {
69
+ /* Index into MAIR registers for cache attributes */
70
+ uint8_t attrindx = extract32(attrs, 0, 3);
71
+ uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
72
+ assert(attrindx <= 7);
73
+ cacheattrs->attrs = extract64(mair, attrindx * 8, 8);
74
}
75
+ cacheattrs->shareability = extract32(attrs, 6, 2);
76
77
*phys_ptr = descaddr;
78
*page_size_ptr = page_size;
79
@@ -XXX,XX +XXX,XX @@ bool get_phys_addr(CPUARMState *env, target_ulong address,
80
ret = get_phys_addr_lpae(env, ipa, access_type, ARMMMUIdx_Stage2,
81
mmu_idx == ARMMMUIdx_E10_0,
82
phys_ptr, attrs, &s2_prot,
83
- page_size, fi,
84
- cacheattrs != NULL ? &cacheattrs2 : NULL);
85
+ page_size, fi, &cacheattrs2);
86
fi->s2addr = ipa;
87
/* Combine the S1 and S2 perms. */
88
*prot &= s2_prot;
89
90
- /* Combine the S1 and S2 cache attributes, if needed */
91
- if (!ret && cacheattrs != NULL) {
92
- if (env->cp15.hcr_el2 & HCR_DC) {
93
- /*
94
- * HCR.DC forces the first stage attributes to
95
- * Normal Non-Shareable,
96
- * Inner Write-Back Read-Allocate Write-Allocate,
97
- * Outer Write-Back Read-Allocate Write-Allocate.
98
- */
99
- cacheattrs->attrs = 0xff;
100
- cacheattrs->shareability = 0;
101
- }
102
- *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
103
+ /* If S2 fails, return early. */
104
+ if (ret) {
105
+ return ret;
106
}
107
108
- return ret;
109
+ /* Combine the S1 and S2 cache attributes. */
110
+ if (env->cp15.hcr_el2 & HCR_DC) {
111
+ /*
112
+ * HCR.DC forces the first stage attributes to
113
+ * Normal Non-Shareable,
114
+ * Inner Write-Back Read-Allocate Write-Allocate,
115
+ * Outer Write-Back Read-Allocate Write-Allocate.
116
+ */
117
+ cacheattrs->attrs = 0xff;
118
+ cacheattrs->shareability = 0;
119
+ }
120
+ *cacheattrs = combine_cacheattrs(*cacheattrs, cacheattrs2);
121
+ return 0;
122
} else {
123
/*
124
* For non-EL2 CPUs a stage1+stage2 translation is just stage 1.
125
@@ -XXX,XX +XXX,XX @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
126
bool ret;
127
ARMMMUFaultInfo fi = {};
128
ARMMMUIdx mmu_idx = arm_mmu_idx(env);
129
+ ARMCacheAttrs cacheattrs = {};
130
131
*attrs = (MemTxAttrs) {};
132
133
ret = get_phys_addr(env, addr, 0, mmu_idx, &phys_addr,
134
- attrs, &prot, &page_size, &fi, NULL);
135
+ attrs, &prot, &page_size, &fi, &cacheattrs);
136
137
if (ret) {
138
return -1;
139
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
20
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
140
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
141
--- a/target/arm/m_helper.c
22
--- a/target/arm/m_helper.c
142
+++ b/target/arm/m_helper.c
23
+++ b/target/arm/m_helper.c
143
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
24
@@ -XXX,XX +XXX,XX @@ static void do_v7m_exception_exit(ARMCPU *cpu)
144
hwaddr physaddr;
25
v7m_exception_taken(cpu, excret, true, false);
145
int prot;
26
return;
146
ARMMMUFaultInfo fi = {};
27
} else {
147
+ ARMCacheAttrs cacheattrs = {};
28
- /* Clear s0..s15 and FPSCR */
148
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
29
+ if (arm_feature(env, ARM_FEATURE_V8_1M)) {
149
int exc;
30
+ /* v8.1M adds this NOCP check */
150
bool exc_secure;
31
+ bool nsacr_pass = exc_secure ||
151
32
+ extract32(env->v7m.nsacr, 10, 1);
152
if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &physaddr,
33
+ bool cpacr_pass = v7m_cpacr_pass(env, exc_secure, true);
153
- &attrs, &prot, &page_size, &fi, NULL)) {
34
+ if (!nsacr_pass) {
154
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
35
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE, true);
155
/* MPU/SAU lookup failed */
36
+ env->v7m.cfsr[M_REG_S] |= R_V7M_CFSR_NOCP_MASK;
156
if (fi.type == ARMFault_QEMU_SFault) {
37
+ qemu_log_mask(CPU_LOG_INT, "...taking UsageFault on existing "
157
if (mode == STACK_LAZYFP) {
38
+ "stackframe: NSACR prevents clearing FPU registers\n");
158
@@ -XXX,XX +XXX,XX @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
39
+ v7m_exception_taken(cpu, excret, true, false);
159
hwaddr physaddr;
40
+ } else if (!cpacr_pass) {
160
int prot;
41
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_USAGE,
161
ARMMMUFaultInfo fi = {};
42
+ exc_secure);
162
+ ARMCacheAttrs cacheattrs = {};
43
+ env->v7m.cfsr[exc_secure] |= R_V7M_CFSR_NOCP_MASK;
163
bool secure = mmu_idx & ARM_MMU_IDX_M_S;
44
+ qemu_log_mask(CPU_LOG_INT, "...taking UsageFault on existing "
164
int exc;
45
+ "stackframe: CPACR prevents clearing FPU registers\n");
165
bool exc_secure;
46
+ v7m_exception_taken(cpu, excret, true, false);
166
uint32_t value;
47
+ }
167
48
+ }
168
if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
49
+ /* Clear s0..s15 and FPSCR; TODO also VPR when MVE is implemented */
169
- &attrs, &prot, &page_size, &fi, NULL)) {
50
int i;
170
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
51
171
/* MPU/SAU lookup failed */
52
for (i = 0; i < 16; i += 2) {
172
if (fi.type == ARMFault_QEMU_SFault) {
173
qemu_log_mask(CPU_LOG_INT,
174
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
175
V8M_SAttributes sattrs = {};
176
MemTxAttrs attrs = {};
177
ARMMMUFaultInfo fi = {};
178
+ ARMCacheAttrs cacheattrs = {};
179
MemTxResult txres;
180
target_ulong page_size;
181
hwaddr physaddr;
182
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
183
"...really SecureFault with SFSR.INVEP\n");
184
return false;
185
}
186
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx,
187
- &physaddr, &attrs, &prot, &page_size, &fi, NULL)) {
188
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &physaddr,
189
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
190
/* the MPU lookup failed */
191
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
192
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
193
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
194
index XXXXXXX..XXXXXXX 100644
195
--- a/target/arm/tlb_helper.c
196
+++ b/target/arm/tlb_helper.c
197
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
198
int prot, ret;
199
MemTxAttrs attrs = {};
200
ARMMMUFaultInfo fi = {};
201
+ ARMCacheAttrs cacheattrs = {};
202
203
/*
204
* Walk the page table and (if the mapping exists) add the page
205
@@ -XXX,XX +XXX,XX @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
206
*/
207
ret = get_phys_addr(&cpu->env, address, access_type,
208
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
209
- &phys_addr, &attrs, &prot, &page_size, &fi, NULL);
210
+ &phys_addr, &attrs, &prot, &page_size,
211
+ &fi, &cacheattrs);
212
if (likely(!ret)) {
213
/*
214
* Map a single [sub]page. Regions smaller than our declared
215
--
53
--
216
2.20.1
54
2.20.1
217
55
218
56
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
v8.1M adds new encodings of VLLDM and VLSTM (where bit 7 is set).
2
The only difference is that:
3
* the old T1 encodings UNDEF if the implementation implements 32
4
Dregs (this is currently architecturally impossible for M-profile)
5
* the new T2 encodings have the implementation-defined option to
6
read from memory (discarding the data) or write UNKNOWN values to
7
memory for the stack slots that would be D16-D31
2
8
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
We choose not to make those accesses, so for us the two
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
instructions behave identically assuming they don't UNDEF.
5
Message-id: 20200626033144.790098-16-richard.henderson@linaro.org
11
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
14
Message-id: 20201119215617.29887-21-peter.maydell@linaro.org
7
---
15
---
8
target/arm/helper-a64.h | 7 ++
16
target/arm/m-nocp.decode | 2 +-
9
target/arm/helper.h | 2 +
17
target/arm/translate-vfp.c.inc | 25 +++++++++++++++++++++++++
10
target/arm/mte_helper.c | 194 +++++++++++++++++++++++++++++++++++++
18
2 files changed, 26 insertions(+), 1 deletion(-)
11
target/arm/op_helper.c | 16 +++
12
target/arm/translate-a64.c | 172 +++++++++++++++++++++++++++++++-
13
5 files changed, 386 insertions(+), 5 deletions(-)
14
19
15
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
diff --git a/target/arm/m-nocp.decode b/target/arm/m-nocp.decode
16
index XXXXXXX..XXXXXXX 100644
21
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper-a64.h
22
--- a/target/arm/m-nocp.decode
18
+++ b/target/arm/helper-a64.h
23
+++ b/target/arm/m-nocp.decode
19
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
24
@@ -XXX,XX +XXX,XX @@
20
25
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
26
{
22
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
27
# Special cases which do not take an early NOCP: VLLDM and VLSTM
23
+DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
28
- VLLDM_VLSTM 1110 1100 001 l:1 rn:4 0000 1010 0000 0000
24
+DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
29
+ VLLDM_VLSTM 1110 1100 001 l:1 rn:4 0000 1010 op:1 000 0000
25
+DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
30
# VSCCLRM (new in v8.1M) is similar:
26
+DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
31
VSCCLRM 1110 1100 1.01 1111 .... 1011 imm:7 0 vd=%vd_dp size=3
27
+DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
32
VSCCLRM 1110 1100 1.01 1111 .... 1010 imm:8 vd=%vd_sp size=2
28
+DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
33
diff --git a/target/arm/translate-vfp.c.inc b/target/arm/translate-vfp.c.inc
29
+DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
30
diff --git a/target/arm/helper.h b/target/arm/helper.h
31
index XXXXXXX..XXXXXXX 100644
34
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper.h
35
--- a/target/arm/translate-vfp.c.inc
33
+++ b/target/arm/helper.h
36
+++ b/target/arm/translate-vfp.c.inc
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
37
@@ -XXX,XX +XXX,XX @@ static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
35
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
38
!arm_dc_feature(s, ARM_FEATURE_V8)) {
36
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
39
return false;
37
40
}
38
+DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
39
+
41
+
40
DEF_HELPER_1(vfp_get_fpscr, i32, env)
42
+ if (a->op) {
41
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
43
+ /*
42
44
+ * T2 encoding ({D0-D31} reglist): v8.1M and up. We choose not
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
45
+ * to take the IMPDEF option to make memory accesses to the stack
44
index XXXXXXX..XXXXXXX 100644
46
+ * slots that correspond to the D16-D31 registers (discarding
45
--- a/target/arm/mte_helper.c
47
+ * read data and writing UNKNOWN values), so for us the T2
46
+++ b/target/arm/mte_helper.c
48
+ * encoding behaves identically to the T1 encoding.
47
@@ -XXX,XX +XXX,XX @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
49
+ */
48
return tag;
50
+ if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
49
}
51
+ return false;
50
51
+/**
52
+ * allocation_tag_mem:
53
+ * @env: the cpu environment
54
+ * @ptr_mmu_idx: the addressing regime to use for the virtual address
55
+ * @ptr: the virtual address for which to look up tag memory
56
+ * @ptr_access: the access to use for the virtual address
57
+ * @ptr_size: the number of bytes in the normal memory access
58
+ * @tag_access: the access to use for the tag memory
59
+ * @tag_size: the number of bytes in the tag memory access
60
+ * @ra: the return address for exception handling
61
+ *
62
+ * Our tag memory is formatted as a sequence of little-endian nibbles.
63
+ * That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
64
+ * tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
65
+ * for the higher addr.
66
+ *
67
+ * Here, resolve the physical address from the virtual address, and return
68
+ * a pointer to the corresponding tag byte. Exit with exception if the
69
+ * virtual address is not accessible for @ptr_access.
70
+ *
71
+ * The @ptr_size and @tag_size values may not have an obvious relation
72
+ * due to the alignment of @ptr, and the number of tag checks required.
73
+ *
74
+ * If there is no tag storage corresponding to @ptr, return NULL.
75
+ */
76
+static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
77
+ uint64_t ptr, MMUAccessType ptr_access,
78
+ int ptr_size, MMUAccessType tag_access,
79
+ int tag_size, uintptr_t ra)
80
+{
81
+ /* Tag storage not implemented. */
82
+ return NULL;
83
+}
84
+
85
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
86
{
87
int rtag;
88
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
89
90
return address_with_allocation_tag(ptr + offset, rtag);
91
}
92
+
93
+static int load_tag1(uint64_t ptr, uint8_t *mem)
94
+{
95
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
96
+ return extract32(*mem, ofs, 4);
97
+}
98
+
99
+uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
100
+{
101
+ int mmu_idx = cpu_mmu_index(env, false);
102
+ uint8_t *mem;
103
+ int rtag = 0;
104
+
105
+ /* Trap if accessing an invalid page. */
106
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
107
+ MMU_DATA_LOAD, 1, GETPC());
108
+
109
+ /* Load if page supports tags. */
110
+ if (mem) {
111
+ rtag = load_tag1(ptr, mem);
112
+ }
113
+
114
+ return address_with_allocation_tag(xt, rtag);
115
+}
116
+
117
+static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
118
+{
119
+ if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
120
+ arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
121
+ cpu_mmu_index(env, false), ra);
122
+ g_assert_not_reached();
123
+ }
124
+}
125
+
126
+/* For use in a non-parallel context, store to the given nibble. */
127
+static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
128
+{
129
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
130
+ *mem = deposit32(*mem, ofs, 4, tag);
131
+}
132
+
133
+/* For use in a parallel context, atomically store to the given nibble. */
134
+static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
135
+{
136
+ int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
137
+ uint8_t old = atomic_read(mem);
138
+
139
+ while (1) {
140
+ uint8_t new = deposit32(old, ofs, 4, tag);
141
+ uint8_t cmp = atomic_cmpxchg(mem, old, new);
142
+ if (likely(cmp == old)) {
143
+ return;
144
+ }
145
+ old = cmp;
146
+ }
147
+}
148
+
149
+typedef void stg_store1(uint64_t, uint8_t *, int);
150
+
151
+static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
152
+ uintptr_t ra, stg_store1 store1)
153
+{
154
+ int mmu_idx = cpu_mmu_index(env, false);
155
+ uint8_t *mem;
156
+
157
+ check_tag_aligned(env, ptr, ra);
158
+
159
+ /* Trap if accessing an invalid page. */
160
+ mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
161
+ MMU_DATA_STORE, 1, ra);
162
+
163
+ /* Store if page supports tags. */
164
+ if (mem) {
165
+ store1(ptr, mem, allocation_tag_from_addr(xt));
166
+ }
167
+}
168
+
169
+void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
170
+{
171
+ do_stg(env, ptr, xt, GETPC(), store_tag1);
172
+}
173
+
174
+void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
175
+{
176
+ do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
177
+}
178
+
179
+void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
180
+{
181
+ int mmu_idx = cpu_mmu_index(env, false);
182
+ uintptr_t ra = GETPC();
183
+
184
+ check_tag_aligned(env, ptr, ra);
185
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
186
+}
187
+
188
+static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
189
+ uintptr_t ra, stg_store1 store1)
190
+{
191
+ int mmu_idx = cpu_mmu_index(env, false);
192
+ int tag = allocation_tag_from_addr(xt);
193
+ uint8_t *mem1, *mem2;
194
+
195
+ check_tag_aligned(env, ptr, ra);
196
+
197
+ /*
198
+ * Trap if accessing an invalid page(s).
199
+ * This takes priority over !allocation_tag_access_enabled.
200
+ */
201
+ if (ptr & TAG_GRANULE) {
202
+ /* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
203
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
204
+ TAG_GRANULE, MMU_DATA_STORE, 1, ra);
205
+ mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
206
+ MMU_DATA_STORE, TAG_GRANULE,
207
+ MMU_DATA_STORE, 1, ra);
208
+
209
+ /* Store if page(s) support tags. */
210
+ if (mem1) {
211
+ store1(TAG_GRANULE, mem1, tag);
212
+ }
213
+ if (mem2) {
214
+ store1(0, mem2, tag);
215
+ }
52
+ }
216
+ } else {
53
+ } else {
217
+ /* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
54
+ /*
218
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
55
+ * T1 encoding ({D0-D15} reglist); undef if we have 32 Dregs.
219
+ 2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
56
+ * This is currently architecturally impossible, but we add the
220
+ if (mem1) {
57
+ * check to stay in line with the pseudocode. Note that we must
221
+ tag |= tag << 4;
58
+ * emit code for the UNDEF so it takes precedence over the NOCP.
222
+ atomic_set(mem1, tag);
59
+ */
223
+ }
60
+ if (dc_isar_feature(aa32_simd_r32, s)) {
224
+ }
61
+ unallocated_encoding(s);
225
+}
62
+ return true;
226
+
227
+void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
228
+{
229
+ do_st2g(env, ptr, xt, GETPC(), store_tag1);
230
+}
231
+
232
+void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
233
+{
234
+ do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
235
+}
236
+
237
+void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
238
+{
239
+ int mmu_idx = cpu_mmu_index(env, false);
240
+ uintptr_t ra = GETPC();
241
+ int in_page = -(ptr | TARGET_PAGE_MASK);
242
+
243
+ check_tag_aligned(env, ptr, ra);
244
+
245
+ if (likely(in_page >= 2 * TAG_GRANULE)) {
246
+ probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
247
+ } else {
248
+ probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
249
+ probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
250
+ }
251
+}
252
diff --git a/target/arm/op_helper.c b/target/arm/op_helper.c
253
index XXXXXXX..XXXXXXX 100644
254
--- a/target/arm/op_helper.c
255
+++ b/target/arm/op_helper.c
256
@@ -XXX,XX +XXX,XX @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
257
return ((uint32_t)x >> shift) | (x << (32 - shift));
258
}
259
}
260
+
261
+void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
262
+ uint32_t access_type, uint32_t mmu_idx,
263
+ uint32_t size)
264
+{
265
+ uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
266
+ uintptr_t ra = GETPC();
267
+
268
+ if (likely(size <= in_page)) {
269
+ probe_access(env, ptr, size, access_type, mmu_idx, ra);
270
+ } else {
271
+ probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
272
+ probe_access(env, ptr + in_page, size - in_page,
273
+ access_type, mmu_idx, ra);
274
+ }
275
+}
276
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
277
index XXXXXXX..XXXXXXX 100644
278
--- a/target/arm/translate-a64.c
279
+++ b/target/arm/translate-a64.c
280
@@ -XXX,XX +XXX,XX @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
281
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
282
}
283
284
+static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
285
+ MMUAccessType acc, int log2_size)
286
+{
287
+ TCGv_i32 t_acc = tcg_const_i32(acc);
288
+ TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
289
+ TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
290
+
291
+ gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
292
+ tcg_temp_free_i32(t_acc);
293
+ tcg_temp_free_i32(t_idx);
294
+ tcg_temp_free_i32(t_size);
295
+}
296
+
297
typedef struct DisasCompare64 {
298
TCGCond cond;
299
TCGv_i64 value;
300
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
301
}
302
}
303
304
+/*
305
+ * Load/Store memory tags
306
+ *
307
+ * 31 30 29 24 22 21 12 10 5 0
308
+ * +-----+-------------+-----+---+------+-----+------+------+
309
+ * | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
310
+ * +-----+-------------+-----+---+------+-----+------+------+
311
+ */
312
+static void disas_ldst_tag(DisasContext *s, uint32_t insn)
313
+{
314
+ int rt = extract32(insn, 0, 5);
315
+ int rn = extract32(insn, 5, 5);
316
+ uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
317
+ int op2 = extract32(insn, 10, 2);
318
+ int op1 = extract32(insn, 22, 2);
319
+ bool is_load = false, is_pair = false, is_zero = false;
320
+ int index = 0;
321
+ TCGv_i64 addr, clean_addr, tcg_rt;
322
+
323
+ /* We checked insn bits [29:24,21] in the caller. */
324
+ if (extract32(insn, 30, 2) != 3) {
325
+ goto do_unallocated;
326
+ }
327
+
328
+ /*
329
+ * @index is a tri-state variable which has 3 states:
330
+ * < 0 : post-index, writeback
331
+ * = 0 : signed offset
332
+ * > 0 : pre-index, writeback
333
+ */
334
+ switch (op1) {
335
+ case 0:
336
+ if (op2 != 0) {
337
+ /* STG */
338
+ index = op2 - 2;
339
+ break;
340
+ }
341
+ goto do_unallocated;
342
+ case 1:
343
+ if (op2 != 0) {
344
+ /* STZG */
345
+ is_zero = true;
346
+ index = op2 - 2;
347
+ } else {
348
+ /* LDG */
349
+ is_load = true;
350
+ }
351
+ break;
352
+ case 2:
353
+ if (op2 != 0) {
354
+ /* ST2G */
355
+ is_pair = true;
356
+ index = op2 - 2;
357
+ break;
358
+ }
359
+ goto do_unallocated;
360
+ case 3:
361
+ if (op2 != 0) {
362
+ /* STZ2G */
363
+ is_pair = is_zero = true;
364
+ index = op2 - 2;
365
+ break;
366
+ }
367
+ goto do_unallocated;
368
+
369
+ default:
370
+ do_unallocated:
371
+ unallocated_encoding(s);
372
+ return;
373
+ }
374
+
375
+ if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
376
+ goto do_unallocated;
377
+ }
378
+
379
+ if (rn == 31) {
380
+ gen_check_sp_alignment(s);
381
+ }
382
+
383
+ addr = read_cpu_reg_sp(s, rn, true);
384
+ if (index >= 0) {
385
+ /* pre-index or signed offset */
386
+ tcg_gen_addi_i64(addr, addr, offset);
387
+ }
388
+
389
+ if (is_load) {
390
+ tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
391
+ tcg_rt = cpu_reg(s, rt);
392
+ if (s->ata) {
393
+ gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
394
+ } else {
395
+ clean_addr = clean_data_tbi(s, addr);
396
+ gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
397
+ gen_address_with_allocation_tag0(tcg_rt, addr);
398
+ }
399
+ } else {
400
+ tcg_rt = cpu_reg_sp(s, rt);
401
+ if (!s->ata) {
402
+ /*
403
+ * For STG and ST2G, we need to check alignment and probe memory.
404
+ * TODO: For STZG and STZ2G, we could rely on the stores below,
405
+ * at least for system mode; user-only won't enforce alignment.
406
+ */
407
+ if (is_pair) {
408
+ gen_helper_st2g_stub(cpu_env, addr);
409
+ } else {
410
+ gen_helper_stg_stub(cpu_env, addr);
411
+ }
412
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
413
+ if (is_pair) {
414
+ gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
415
+ } else {
416
+ gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
417
+ }
418
+ } else {
419
+ if (is_pair) {
420
+ gen_helper_st2g(cpu_env, addr, tcg_rt);
421
+ } else {
422
+ gen_helper_stg(cpu_env, addr, tcg_rt);
423
+ }
424
+ }
63
+ }
425
+ }
64
+ }
426
+
65
+
427
+ if (is_zero) {
66
/*
428
+ TCGv_i64 clean_addr = clean_data_tbi(s, addr);
67
* If not secure, UNDEF. We must emit code for this
429
+ TCGv_i64 tcg_zero = tcg_const_i64(0);
68
* rather than returning false so that this takes
430
+ int mem_index = get_mem_index(s);
431
+ int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
432
+
433
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
434
+ MO_Q | MO_ALIGN_16);
435
+ for (i = 8; i < n; i += 8) {
436
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
437
+ tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
438
+ }
439
+ tcg_temp_free_i64(tcg_zero);
440
+ }
441
+
442
+ if (index != 0) {
443
+ /* pre-index or post-index */
444
+ if (index < 0) {
445
+ /* post-index */
446
+ tcg_gen_addi_i64(addr, addr, offset);
447
+ }
448
+ tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
449
+ }
450
+}
451
+
452
/* Loads and stores */
453
static void disas_ldst(DisasContext *s, uint32_t insn)
454
{
455
@@ -XXX,XX +XXX,XX @@ static void disas_ldst(DisasContext *s, uint32_t insn)
456
case 0x0d: /* AdvSIMD load/store single structure */
457
disas_ldst_single_struct(s, insn);
458
break;
459
- case 0x19: /* LDAPR/STLR (unscaled immediate) */
460
- if (extract32(insn, 10, 2) != 0 ||
461
- extract32(insn, 21, 1) != 0) {
462
+ case 0x19:
463
+ if (extract32(insn, 21, 1) != 0) {
464
+ disas_ldst_tag(s, insn);
465
+ } else if (extract32(insn, 10, 2) == 0) {
466
+ disas_ldst_ldapr_stlr(s, insn);
467
+ } else {
468
unallocated_encoding(s);
469
- break;
470
}
471
- disas_ldst_ldapr_stlr(s, insn);
472
break;
473
default:
474
unallocated_encoding(s);
475
--
69
--
476
2.20.1
70
2.20.1
477
71
478
72
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
v8.1M introduces a new TRD flag in the CCR register, which enables
2
checking for stack frame integrity signatures on SG instructions.
3
This bit is not banked, and is always RAZ/WI to Non-secure code.
4
Adjust the code for handling CCR reads and writes to handle this.
2
5
3
Introduce an lvalue macro to wrap target_tlb_bit0.
4
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-33-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20201119215617.29887-23-peter.maydell@linaro.org
9
---
9
---
10
target/arm/cpu.h | 13 +++++++++++++
10
target/arm/cpu.h | 2 ++
11
target/arm/helper.c | 2 +-
11
hw/intc/armv7m_nvic.c | 26 ++++++++++++++++++--------
12
target/arm/translate-a64.c | 2 +-
12
2 files changed, 20 insertions(+), 8 deletions(-)
13
3 files changed, 15 insertions(+), 2 deletions(-)
14
13
15
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
14
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
16
index XXXXXXX..XXXXXXX 100644
15
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/cpu.h
16
--- a/target/arm/cpu.h
18
+++ b/target/arm/cpu.h
17
+++ b/target/arm/cpu.h
19
@@ -XXX,XX +XXX,XX @@ static inline uint64_t *aa64_vfp_qreg(CPUARMState *env, unsigned regno)
18
@@ -XXX,XX +XXX,XX @@ FIELD(V7M_CCR, STKOFHFNMIGN, 10, 1)
20
/* Shared between translate-sve.c and sve_helper.c. */
19
FIELD(V7M_CCR, DC, 16, 1)
21
extern const uint64_t pred_esz_masks[4];
20
FIELD(V7M_CCR, IC, 17, 1)
22
21
FIELD(V7M_CCR, BP, 18, 1)
23
+/* Helper for the macros below, validating the argument type. */
22
+FIELD(V7M_CCR, LOB, 19, 1)
24
+static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
23
+FIELD(V7M_CCR, TRD, 20, 1)
25
+{
24
26
+ return x;
25
/* V7M SCR bits */
27
+}
26
FIELD(V7M_SCR, SLEEPONEXIT, 1, 1)
27
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
28
index XXXXXXX..XXXXXXX 100644
29
--- a/hw/intc/armv7m_nvic.c
30
+++ b/hw/intc/armv7m_nvic.c
31
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
32
}
33
return cpu->env.v7m.scr[attrs.secure];
34
case 0xd14: /* Configuration Control. */
35
- /* The BFHFNMIGN bit is the only non-banked bit; we
36
- * keep it in the non-secure copy of the register.
37
+ /*
38
+ * Non-banked bits: BFHFNMIGN (stored in the NS copy of the register)
39
+ * and TRD (stored in the S copy of the register)
40
*/
41
val = cpu->env.v7m.ccr[attrs.secure];
42
val |= cpu->env.v7m.ccr[M_REG_NS] & R_V7M_CCR_BFHFNMIGN_MASK;
43
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
44
cpu->env.v7m.scr[attrs.secure] = value;
45
break;
46
case 0xd14: /* Configuration Control. */
47
+ {
48
+ uint32_t mask;
28
+
49
+
29
+/*
50
if (!arm_feature(&cpu->env, ARM_FEATURE_M_MAIN)) {
30
+ * Lvalue macros for ARM TLB bits that we must cache in the TCG TLB.
51
goto bad_offset;
31
+ * Using these should be a bit more self-documenting than using the
52
}
32
+ * generic target bits directly.
53
33
+ */
54
/* Enforce RAZ/WI on reserved and must-RAZ/WI bits */
34
+#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
55
- value &= (R_V7M_CCR_STKALIGN_MASK |
35
+
56
- R_V7M_CCR_BFHFNMIGN_MASK |
36
/*
57
- R_V7M_CCR_DIV_0_TRP_MASK |
37
* Naming convention for isar_feature functions:
58
- R_V7M_CCR_UNALIGN_TRP_MASK |
38
* Functions which test 32-bit ID registers should have _aa32_ in
59
- R_V7M_CCR_USERSETMPEND_MASK |
39
diff --git a/target/arm/helper.c b/target/arm/helper.c
60
- R_V7M_CCR_NONBASETHRDENA_MASK);
40
index XXXXXXX..XXXXXXX 100644
61
+ mask = R_V7M_CCR_STKALIGN_MASK |
41
--- a/target/arm/helper.c
62
+ R_V7M_CCR_BFHFNMIGN_MASK |
42
+++ b/target/arm/helper.c
63
+ R_V7M_CCR_DIV_0_TRP_MASK |
43
@@ -XXX,XX +XXX,XX @@ static bool get_phys_addr_lpae(CPUARMState *env, target_ulong address,
64
+ R_V7M_CCR_UNALIGN_TRP_MASK |
44
}
65
+ R_V7M_CCR_USERSETMPEND_MASK |
45
/* When in aarch64 mode, and BTI is enabled, remember GP in the IOTLB. */
66
+ R_V7M_CCR_NONBASETHRDENA_MASK;
46
if (aarch64 && guarded && cpu_isar_feature(aa64_bti, cpu)) {
67
+ if (arm_feature(&cpu->env, ARM_FEATURE_V8_1M) && attrs.secure) {
47
- txattrs->target_tlb_bit0 = true;
68
+ /* TRD is always RAZ/WI from NS */
48
+ arm_tlb_bti_gp(txattrs) = true;
69
+ mask |= R_V7M_CCR_TRD_MASK;
49
}
70
+ }
50
71
+ value &= mask;
51
if (cacheattrs != NULL) {
72
52
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
73
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
53
index XXXXXXX..XXXXXXX 100644
74
/* v8M makes NONBASETHRDENA and STKALIGN be RES1 */
54
--- a/target/arm/translate-a64.c
75
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
55
+++ b/target/arm/translate-a64.c
76
56
@@ -XXX,XX +XXX,XX @@ static bool is_guarded_page(CPUARMState *env, DisasContext *s)
77
cpu->env.v7m.ccr[attrs.secure] = value;
57
* table entry even for that case.
78
break;
58
*/
79
+ }
59
return (tlb_hit(entry->addr_code, addr) &&
80
case 0xd24: /* System Handler Control and State (SHCSR) */
60
- env_tlb(env)->d[mmu_idx].iotlb[index].attrs.target_tlb_bit0);
81
if (!arm_feature(&cpu->env, ARM_FEATURE_V7)) {
61
+ arm_tlb_bti_gp(&env_tlb(env)->d[mmu_idx].iotlb[index].attrs));
82
goto bad_offset;
62
#endif
63
}
64
65
--
83
--
66
2.20.1
84
2.20.1
67
85
68
86
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
v8.1M introduces a new TRD flag in the CCR register, which enables
2
checking for stack frame integrity signatures on SG instructions.
3
Add the code in the SG insn implementation for the new behaviour.
2
4
3
Fill out the stub that was added earlier.
5
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
6
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20201119215617.29887-24-peter.maydell@linaro.org
8
---
9
target/arm/m_helper.c | 86 +++++++++++++++++++++++++++++++++++++++++++
10
1 file changed, 86 insertions(+)
4
11
5
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
12
diff --git a/target/arm/m_helper.c b/target/arm/m_helper.c
6
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
7
Message-id: 20200626033144.790098-27-richard.henderson@linaro.org
8
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
9
---
10
target/arm/internals.h | 2 +
11
target/arm/mte_helper.c | 165 +++++++++++++++++++++++++++++++++++++++-
12
2 files changed, 166 insertions(+), 1 deletion(-)
13
14
diff --git a/target/arm/internals.h b/target/arm/internals.h
15
index XXXXXXX..XXXXXXX 100644
13
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/internals.h
14
--- a/target/arm/m_helper.c
17
+++ b/target/arm/internals.h
15
+++ b/target/arm/m_helper.c
18
@@ -XXX,XX +XXX,XX @@ FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
16
@@ -XXX,XX +XXX,XX @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx,
19
bool mte_probe1(CPUARMState *env, uint32_t desc, uint64_t ptr);
17
return true;
20
uint64_t mte_check1(CPUARMState *env, uint32_t desc,
18
}
21
uint64_t ptr, uintptr_t ra);
19
22
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
20
+static bool v7m_read_sg_stack_word(ARMCPU *cpu, ARMMMUIdx mmu_idx,
23
+ uint64_t ptr, uintptr_t ra);
21
+ uint32_t addr, uint32_t *spdata)
24
22
+{
25
static inline int allocation_tag_from_addr(uint64_t ptr)
23
+ /*
26
{
24
+ * Read a word of data from the stack for the SG instruction,
27
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
25
+ * writing the value into *spdata. If the load succeeds, return
28
index XXXXXXX..XXXXXXX 100644
26
+ * true; otherwise pend an appropriate exception and return false.
29
--- a/target/arm/mte_helper.c
27
+ * (We can't use data load helpers here that throw an exception
30
+++ b/target/arm/mte_helper.c
28
+ * because of the context we're called in, which is halfway through
31
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
29
+ * arm_v7m_cpu_do_interrupt().)
32
/*
30
+ */
33
* Perform an MTE checked access for multiple logical accesses.
31
+ CPUState *cs = CPU(cpu);
34
*/
32
+ CPUARMState *env = &cpu->env;
33
+ MemTxAttrs attrs = {};
34
+ MemTxResult txres;
35
+ target_ulong page_size;
36
+ hwaddr physaddr;
37
+ int prot;
38
+ ARMMMUFaultInfo fi = {};
39
+ ARMCacheAttrs cacheattrs = {};
40
+ uint32_t value;
35
+
41
+
36
+/**
42
+ if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &physaddr,
37
+ * checkN:
43
+ &attrs, &prot, &page_size, &fi, &cacheattrs)) {
38
+ * @tag: tag memory to test
44
+ /* MPU/SAU lookup failed */
39
+ * @odd: true to begin testing at tags at odd nibble
45
+ if (fi.type == ARMFault_QEMU_SFault) {
40
+ * @cmp: the tag to compare against
46
+ qemu_log_mask(CPU_LOG_INT,
41
+ * @count: number of tags to test
47
+ "...SecureFault during stack word read\n");
42
+ *
48
+ env->v7m.sfsr |= R_V7M_SFSR_AUVIOL_MASK | R_V7M_SFSR_SFARVALID_MASK;
43
+ * Return the number of successful tests.
49
+ env->v7m.sfar = addr;
44
+ * Thus a return value < @count indicates a failure.
50
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_SECURE, false);
45
+ *
51
+ } else {
46
+ * A note about sizes: count is expected to be small.
52
+ qemu_log_mask(CPU_LOG_INT,
47
+ *
53
+ "...MemManageFault during stack word read\n");
48
+ * The most common use will be LDP/STP of two integer registers,
54
+ env->v7m.cfsr[M_REG_S] |= R_V7M_CFSR_DACCVIOL_MASK |
49
+ * which means 16 bytes of memory touching at most 2 tags, but
55
+ R_V7M_CFSR_MMARVALID_MASK;
50
+ * often the access is aligned and thus just 1 tag.
56
+ env->v7m.mmfar[M_REG_S] = addr;
51
+ *
57
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, false);
52
+ * Using AdvSIMD LD/ST (multiple), one can access 64 bytes of memory,
58
+ }
53
+ * touching at most 5 tags. SVE LDR/STR (vector) with the default
59
+ return false;
54
+ * vector length is also 64 bytes; the maximum architectural length
60
+ }
55
+ * is 256 bytes touching at most 9 tags.
61
+ value = address_space_ldl(arm_addressspace(cs, attrs), physaddr,
56
+ *
62
+ attrs, &txres);
57
+ * The loop below uses 7 logical operations and 1 memory operation
63
+ if (txres != MEMTX_OK) {
58
+ * per tag pair. An implementation that loads an aligned word and
64
+ /* BusFault trying to read the data */
59
+ * uses masking to ignore adjacent tags requires 18 logical operations
65
+ qemu_log_mask(CPU_LOG_INT,
60
+ * and thus does not begin to pay off until 6 tags.
66
+ "...BusFault during stack word read\n");
61
+ * Which, according to the survey above, is unlikely to be common.
67
+ env->v7m.cfsr[M_REG_NS] |=
62
+ */
68
+ (R_V7M_CFSR_PRECISERR_MASK | R_V7M_CFSR_BFARVALID_MASK);
63
+static int checkN(uint8_t *mem, int odd, int cmp, int count)
69
+ env->v7m.bfar = addr;
64
+{
70
+ armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_BUS, false);
65
+ int n = 0, diff;
71
+ return false;
66
+
67
+ /* Replicate the test tag and compare. */
68
+ cmp *= 0x11;
69
+ diff = *mem++ ^ cmp;
70
+
71
+ if (odd) {
72
+ goto start_odd;
73
+ }
72
+ }
74
+
73
+
75
+ while (1) {
74
+ *spdata = value;
76
+ /* Test even tag. */
75
+ return true;
77
+ if (unlikely((diff) & 0x0f)) {
76
+}
78
+ break;
77
+
79
+ }
78
static bool v7m_handle_execute_nsc(ARMCPU *cpu)
80
+ if (++n == count) {
79
{
81
+ break;
80
/*
81
@@ -XXX,XX +XXX,XX @@ static bool v7m_handle_execute_nsc(ARMCPU *cpu)
82
*/
83
qemu_log_mask(CPU_LOG_INT, "...really an SG instruction at 0x%08" PRIx32
84
", executing it\n", env->regs[15]);
85
+
86
+ if (cpu_isar_feature(aa32_m_sec_state, cpu) &&
87
+ !arm_v7m_is_handler_mode(env)) {
88
+ /*
89
+ * v8.1M exception stack frame integrity check. Note that we
90
+ * must perform the memory access even if CCR_S.TRD is zero
91
+ * and we aren't going to check what the data loaded is.
92
+ */
93
+ uint32_t spdata, sp;
94
+
95
+ /*
96
+ * We know we are currently NS, so the S stack pointers must be
97
+ * in other_ss_{psp,msp}, not in regs[13]/other_sp.
98
+ */
99
+ sp = v7m_using_psp(env) ? env->v7m.other_ss_psp : env->v7m.other_ss_msp;
100
+ if (!v7m_read_sg_stack_word(cpu, mmu_idx, sp, &spdata)) {
101
+ /* Stack access failed and an exception has been pended */
102
+ return false;
82
+ }
103
+ }
83
+
104
+
84
+ start_odd:
105
+ if (env->v7m.ccr[M_REG_S] & R_V7M_CCR_TRD_MASK) {
85
+ /* Test odd tag. */
106
+ if (((spdata & ~1) == 0xfefa125a) ||
86
+ if (unlikely((diff) & 0xf0)) {
107
+ !(env->v7m.control[M_REG_S] & 1)) {
87
+ break;
108
+ goto gen_invep;
88
+ }
89
+ if (++n == count) {
90
+ break;
91
+ }
92
+
93
+ diff = *mem++ ^ cmp;
94
+ }
95
+ return n;
96
+}
97
+
98
+uint64_t mte_checkN(CPUARMState *env, uint32_t desc,
99
+ uint64_t ptr, uintptr_t ra)
100
+{
101
+ int mmu_idx, ptr_tag, bit55;
102
+ uint64_t ptr_last, ptr_end, prev_page, next_page;
103
+ uint64_t tag_first, tag_end;
104
+ uint64_t tag_byte_first, tag_byte_end;
105
+ uint32_t esize, total, tag_count, tag_size, n, c;
106
+ uint8_t *mem1, *mem2;
107
+ MMUAccessType type;
108
+
109
+ bit55 = extract64(ptr, 55, 1);
110
+
111
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
112
+ if (unlikely(!tbi_check(desc, bit55))) {
113
+ return ptr;
114
+ }
115
+
116
+ ptr_tag = allocation_tag_from_addr(ptr);
117
+
118
+ if (tcma_check(desc, bit55, ptr_tag)) {
119
+ goto done;
120
+ }
121
+
122
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
123
+ type = FIELD_EX32(desc, MTEDESC, WRITE) ? MMU_DATA_STORE : MMU_DATA_LOAD;
124
+ esize = FIELD_EX32(desc, MTEDESC, ESIZE);
125
+ total = FIELD_EX32(desc, MTEDESC, TSIZE);
126
+
127
+ /* Find the addr of the end of the access, and of the last element. */
128
+ ptr_end = ptr + total;
129
+ ptr_last = ptr_end - esize;
130
+
131
+ /* Round the bounds to the tag granule, and compute the number of tags. */
132
+ tag_first = QEMU_ALIGN_DOWN(ptr, TAG_GRANULE);
133
+ tag_end = QEMU_ALIGN_UP(ptr_last, TAG_GRANULE);
134
+ tag_count = (tag_end - tag_first) / TAG_GRANULE;
135
+
136
+ /* Round the bounds to twice the tag granule, and compute the bytes. */
137
+ tag_byte_first = QEMU_ALIGN_DOWN(ptr, 2 * TAG_GRANULE);
138
+ tag_byte_end = QEMU_ALIGN_UP(ptr_last, 2 * TAG_GRANULE);
139
+
140
+ /* Locate the page boundaries. */
141
+ prev_page = ptr & TARGET_PAGE_MASK;
142
+ next_page = prev_page + TARGET_PAGE_SIZE;
143
+
144
+ if (likely(tag_end - prev_page <= TARGET_PAGE_SIZE)) {
145
+ /* Memory access stays on one page. */
146
+ tag_size = (tag_byte_end - tag_byte_first) / (2 * TAG_GRANULE);
147
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, total,
148
+ MMU_DATA_LOAD, tag_size, ra);
149
+ if (!mem1) {
150
+ goto done;
151
+ }
152
+ /* Perform all of the comparisons. */
153
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, tag_count);
154
+ } else {
155
+ /* Memory access crosses to next page. */
156
+ tag_size = (next_page - tag_byte_first) / (2 * TAG_GRANULE);
157
+ mem1 = allocation_tag_mem(env, mmu_idx, ptr, type, next_page - ptr,
158
+ MMU_DATA_LOAD, tag_size, ra);
159
+
160
+ tag_size = (tag_byte_end - next_page) / (2 * TAG_GRANULE);
161
+ mem2 = allocation_tag_mem(env, mmu_idx, next_page, type,
162
+ ptr_end - next_page,
163
+ MMU_DATA_LOAD, tag_size, ra);
164
+
165
+ /*
166
+ * Perform all of the comparisons.
167
+ * Note the possible but unlikely case of the operation spanning
168
+ * two pages that do not both have tagging enabled.
169
+ */
170
+ n = c = (next_page - tag_first) / TAG_GRANULE;
171
+ if (mem1) {
172
+ n = checkN(mem1, ptr & TAG_GRANULE, ptr_tag, c);
173
+ }
174
+ if (n == c) {
175
+ if (!mem2) {
176
+ goto done;
177
+ }
109
+ }
178
+ n += checkN(mem2, 0, ptr_tag, tag_count - c);
179
+ }
110
+ }
180
+ }
111
+ }
181
+
112
+
182
+ /*
113
env->regs[14] &= ~1;
183
+ * If we failed, we know which granule. Compute the element that
114
env->v7m.control[M_REG_S] &= ~R_V7M_CONTROL_SFPA_MASK;
184
+ * is first in that granule, and signal failure on that element.
115
switch_v7m_security_state(env, true);
185
+ */
186
+ if (unlikely(n < tag_count)) {
187
+ uint64_t fail_ofs;
188
+
189
+ fail_ofs = tag_first + n * TAG_GRANULE - ptr;
190
+ fail_ofs = ROUND_UP(fail_ofs, esize);
191
+ mte_check_fail(env, mmu_idx, ptr + fail_ofs, ra);
192
+ }
193
+
194
+ done:
195
+ return useronly_clean_ptr(ptr);
196
+}
197
+
198
uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
199
{
200
- return ptr;
201
+ return mte_checkN(env, desc, ptr, GETPC());
202
}
203
--
116
--
204
2.20.1
117
2.20.1
205
118
206
119
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
In commit 077d7449100d824a4 we added code to handle the v8M
2
requirement that returns from NMI or HardFault forcibly deactivate
3
those exceptions regardless of what interrupt the guest is trying to
4
deactivate. Unfortunately this broke the handling of the "illegal
5
exception return because the returning exception number is not
6
active" check for those cases. In the pseudocode this test is done
7
on the exception the guest asks to return from, but because our
8
implementation was doing this in armv7m_nvic_complete_irq() after the
9
new "deactivate NMI/HardFault regardless" code we ended up doing the
10
test on the VecInfo for that exception instead, which usually meant
11
failing to raise the illegal exception return fault.
2
12
3
This does not attempt to rectify all of the res0 bits, but does
13
In the case for "configurable exception targeting the opposite
4
clear the mte bits when not enabled. Since there is no high-part
14
security state" we detected the illegal-return case but went ahead
5
mapping of SCTLR, aa32 mode cannot write to these bits.
15
and deactivated the VecInfo anyway, which is wrong because that is
16
the VecInfo for the other security state.
6
17
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
18
Rearrange the code so that we first identify the illegal return
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
19
cases, then see if we really need to deactivate NMI or HardFault
9
Message-id: 20200626033144.790098-4-richard.henderson@linaro.org
20
instead, and finally do the deactivation.
21
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
22
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
23
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
24
Message-id: 20201119215617.29887-25-peter.maydell@linaro.org
11
---
25
---
12
target/arm/helper.c | 23 +++++++++++++++++------
26
hw/intc/armv7m_nvic.c | 59 +++++++++++++++++++++++--------------------
13
1 file changed, 17 insertions(+), 6 deletions(-)
27
1 file changed, 32 insertions(+), 27 deletions(-)
14
28
15
diff --git a/target/arm/helper.c b/target/arm/helper.c
29
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
16
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
17
--- a/target/arm/helper.c
31
--- a/hw/intc/armv7m_nvic.c
18
+++ b/target/arm/helper.c
32
+++ b/hw/intc/armv7m_nvic.c
19
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
33
@@ -XXX,XX +XXX,XX @@ int armv7m_nvic_complete_irq(void *opaque, int irq, bool secure)
20
{
34
{
21
ARMCPU *cpu = env_archcpu(env);
35
NVICState *s = (NVICState *)opaque;
22
36
VecInfo *vec = NULL;
23
+ if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
37
- int ret;
24
+ /* M bit is RAZ/WI for PMSA with no MPU implemented */
38
+ int ret = 0;
25
+ value &= ~SCTLR_M;
39
40
assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
41
42
+ trace_nvic_complete_irq(irq, secure);
43
+
44
+ if (secure && exc_is_banked(irq)) {
45
+ vec = &s->sec_vectors[irq];
46
+ } else {
47
+ vec = &s->vectors[irq];
26
+ }
48
+ }
27
+
49
+
28
+ /* ??? Lots of these bits are not implemented. */
50
+ /*
29
+
51
+ * Identify illegal exception return cases. We can't immediately
30
+ if (ri->state == ARM_CP_STATE_AA64 && !cpu_isar_feature(aa64_mte, cpu)) {
52
+ * return at this point because we still need to deactivate
31
+ if (ri->opc1 == 6) { /* SCTLR_EL3 */
53
+ * (either this exception or NMI/HardFault) first.
32
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF | SCTLR_ATA);
54
+ */
33
+ } else {
55
+ if (!exc_is_banked(irq) && exc_targets_secure(s, irq) != secure) {
34
+ value &= ~(SCTLR_ITFSB | SCTLR_TCF0 | SCTLR_TCF |
56
+ /*
35
+ SCTLR_ATA0 | SCTLR_ATA);
57
+ * Return from a configurable exception targeting the opposite
36
+ }
58
+ * security state from the one we're trying to complete it for.
59
+ * Clear vec because it's not really the VecInfo for this
60
+ * (irq, secstate) so we mustn't deactivate it.
61
+ */
62
+ ret = -1;
63
+ vec = NULL;
64
+ } else if (!vec->active) {
65
+ /* Return from an inactive interrupt */
66
+ ret = -1;
67
+ } else {
68
+ /* Legal return, we will return the RETTOBASE bit value to the caller */
69
+ ret = nvic_rettobase(s);
37
+ }
70
+ }
38
+
71
+
39
if (raw_read(env, ri) == value) {
72
/*
40
/* Skip the TLB flush if nothing actually changed; Linux likes
73
* For negative priorities, v8M will forcibly deactivate the appropriate
41
* to do a lot of pointless SCTLR writes.
74
* NMI or HardFault regardless of what interrupt we're being asked to
42
@@ -XXX,XX +XXX,XX @@ static void sctlr_write(CPUARMState *env, const ARMCPRegInfo *ri,
75
@@ -XXX,XX +XXX,XX @@ int armv7m_nvic_complete_irq(void *opaque, int irq, bool secure)
43
return;
44
}
76
}
45
77
46
- if (arm_feature(env, ARM_FEATURE_PMSA) && !cpu->has_mpu) {
78
if (!vec) {
47
- /* M bit is RAZ/WI for PMSA with no MPU implemented */
79
- if (secure && exc_is_banked(irq)) {
48
- value &= ~SCTLR_M;
80
- vec = &s->sec_vectors[irq];
81
- } else {
82
- vec = &s->vectors[irq];
83
- }
49
- }
84
- }
50
-
85
-
51
raw_write(env, ri, value);
86
- trace_nvic_complete_irq(irq, secure);
52
- /* ??? Lots of these bits are not implemented. */
87
-
53
+
88
- if (!vec->active) {
54
/* This may enable/disable the MMU, so do a TLB flush. */
89
- /* Tell the caller this was an illegal exception return */
55
tlb_flush(CPU(cpu));
90
- return -1;
56
91
- }
92
-
93
- /*
94
- * If this is a configurable exception and it is currently
95
- * targeting the opposite security state from the one we're trying
96
- * to complete it for, this counts as an illegal exception return.
97
- * We still need to deactivate whatever vector the logic above has
98
- * selected, though, as it might not be the same as the one for the
99
- * requested exception number.
100
- */
101
- if (!exc_is_banked(irq) && exc_targets_secure(s, irq) != secure) {
102
- ret = -1;
103
- } else {
104
- ret = nvic_rettobase(s);
105
+ return ret;
106
}
107
108
vec->active = 0;
57
--
109
--
58
2.20.1
110
2.20.1
59
111
60
112
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
For v8.1M the architecture mandates that CPUs must provide at
2
least the "minimal RAS implementation" from the Reliability,
3
Availability and Serviceability extension. This consists of:
4
* an ESB instruction which is a NOP
5
-- since it is in the HINT space we need only add a comment
6
* an RFSR register which will RAZ/WI
7
* a RAZ/WI AIRCR.IESB bit
8
-- the code which handles writes to AIRCR does not allow setting
9
of RES0 bits, so we already treat this as RAZ/WI; add a comment
10
noting that this is deliberate
11
* minimal implementation of the RAS register block at 0xe0005000
12
-- this will be in a subsequent commit
13
* setting the ID_PFR0.RAS field to 0b0010
14
-- we will do this when we add the Cortex-M55 CPU model
2
15
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-2-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
17
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
18
Message-id: 20201119215617.29887-26-peter.maydell@linaro.org
7
---
19
---
8
target/arm/cpu.h | 10 ++++++++++
20
target/arm/cpu.h | 14 ++++++++++++++
9
1 file changed, 10 insertions(+)
21
target/arm/t32.decode | 4 ++++
22
hw/intc/armv7m_nvic.c | 13 +++++++++++++
23
3 files changed, 31 insertions(+)
10
24
11
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
25
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
12
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/cpu.h
27
--- a/target/arm/cpu.h
14
+++ b/target/arm/cpu.h
28
+++ b/target/arm/cpu.h
15
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa64_bti(const ARMISARegisters *id)
29
@@ -XXX,XX +XXX,XX @@ FIELD(ID_MMFR4, LSM, 20, 4)
16
return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, BT) != 0;
30
FIELD(ID_MMFR4, CCIDX, 24, 4)
31
FIELD(ID_MMFR4, EVT, 28, 4)
32
33
+FIELD(ID_PFR0, STATE0, 0, 4)
34
+FIELD(ID_PFR0, STATE1, 4, 4)
35
+FIELD(ID_PFR0, STATE2, 8, 4)
36
+FIELD(ID_PFR0, STATE3, 12, 4)
37
+FIELD(ID_PFR0, CSV2, 16, 4)
38
+FIELD(ID_PFR0, AMU, 20, 4)
39
+FIELD(ID_PFR0, DIT, 24, 4)
40
+FIELD(ID_PFR0, RAS, 28, 4)
41
+
42
FIELD(ID_PFR1, PROGMOD, 0, 4)
43
FIELD(ID_PFR1, SECURITY, 4, 4)
44
FIELD(ID_PFR1, MPROGMOD, 8, 4)
45
@@ -XXX,XX +XXX,XX @@ static inline bool isar_feature_aa32_predinv(const ARMISARegisters *id)
46
return FIELD_EX32(id->id_isar6, ID_ISAR6, SPECRES) != 0;
17
}
47
}
18
48
19
+static inline bool isar_feature_aa64_mte_insn_reg(const ARMISARegisters *id)
49
+static inline bool isar_feature_aa32_ras(const ARMISARegisters *id)
20
+{
50
+{
21
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) != 0;
51
+ return FIELD_EX32(id->id_pfr0, ID_PFR0, RAS) != 0;
22
+}
52
+}
23
+
53
+
24
+static inline bool isar_feature_aa64_mte(const ARMISARegisters *id)
54
static inline bool isar_feature_aa32_mprofile(const ARMISARegisters *id)
25
+{
55
{
26
+ return FIELD_EX64(id->id_aa64pfr1, ID_AA64PFR1, MTE) >= 2;
56
return FIELD_EX32(id->id_pfr1, ID_PFR1, MPROGMOD) != 0;
27
+}
57
diff --git a/target/arm/t32.decode b/target/arm/t32.decode
58
index XXXXXXX..XXXXXXX 100644
59
--- a/target/arm/t32.decode
60
+++ b/target/arm/t32.decode
61
@@ -XXX,XX +XXX,XX @@ CLZ 1111 1010 1011 ---- 1111 .... 1000 .... @rdm
62
# SEV 1111 0011 1010 1111 1000 0000 0000 0100
63
# SEVL 1111 0011 1010 1111 1000 0000 0000 0101
64
65
+ # For M-profile minimal-RAS ESB can be a NOP, which is the
66
+ # default behaviour since it is in the hint space.
67
+ # ESB 1111 0011 1010 1111 1000 0000 0001 0000
28
+
68
+
29
static inline bool isar_feature_aa64_pmu_8_1(const ARMISARegisters *id)
69
# The canonical nop ends in 0000 0000, but the whole rest
30
{
70
# of the space is "reserved hint, behaves as nop".
31
return FIELD_EX64(id->id_aa64dfr0, ID_AA64DFR0, PMUVER) >= 4 &&
71
NOP 1111 0011 1010 1111 1000 0000 ---- ----
72
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/intc/armv7m_nvic.c
75
+++ b/hw/intc/armv7m_nvic.c
76
@@ -XXX,XX +XXX,XX @@ static uint32_t nvic_readl(NVICState *s, uint32_t offset, MemTxAttrs attrs)
77
return 0;
78
}
79
return cpu->env.v7m.sfar;
80
+ case 0xf04: /* RFSR */
81
+ if (!cpu_isar_feature(aa32_ras, cpu)) {
82
+ goto bad_offset;
83
+ }
84
+ /* We provide minimal-RAS only: RFSR is RAZ/WI */
85
+ return 0;
86
case 0xf34: /* FPCCR */
87
if (!cpu_isar_feature(aa32_vfp_simd, cpu)) {
88
return 0;
89
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
90
R_V7M_AIRCR_PRIGROUP_SHIFT,
91
R_V7M_AIRCR_PRIGROUP_LENGTH);
92
}
93
+ /* AIRCR.IESB is RAZ/WI because we implement only minimal RAS */
94
if (attrs.secure) {
95
/* These bits are only writable by secure */
96
cpu->env.v7m.aircr = value &
97
@@ -XXX,XX +XXX,XX @@ static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value,
98
}
99
break;
100
}
101
+ case 0xf04: /* RFSR */
102
+ if (!cpu_isar_feature(aa32_ras, cpu)) {
103
+ goto bad_offset;
104
+ }
105
+ /* We provide minimal-RAS only: RFSR is RAZ/WI */
106
+ break;
107
case 0xf34: /* FPCCR */
108
if (cpu_isar_feature(aa32_vfp_simd, cpu)) {
109
/* Not all bits here are banked. */
32
--
110
--
33
2.20.1
111
2.20.1
34
112
35
113
diff view generated by jsdifflib
1
From: Richard Henderson <richard.henderson@linaro.org>
1
The RAS feature has a block of memory-mapped registers at offset
2
0x5000 within the PPB. For a "minimal RAS" implementation we provide
3
no error records and so the only registers that exist in the block
4
are ERRIIDR and ERRDEVID.
2
5
3
Because the elements are sequential, we can eliminate many tests all
6
The "RAZ/WI for privileged, BusFault for nonprivileged" behaviour
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
7
of the "nvic-default" region is actually valid for minimal-RAS,
8
so the main benefit of providing an explicit implementation of
9
the register block is more accurate LOG_UNIMP messages, and a
10
framework for where we could add a real RAS implementation later
11
if necessary.
5
12
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-34-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
15
Message-id: 20201119215617.29887-27-peter.maydell@linaro.org
10
---
16
---
11
target/arm/cpu.h | 1 +
17
include/hw/intc/armv7m_nvic.h | 1 +
12
target/arm/helper-sve.h | 58 ++++++++++
18
hw/intc/armv7m_nvic.c | 56 +++++++++++++++++++++++++++++++++++
13
target/arm/internals.h | 6 +
19
2 files changed, 57 insertions(+)
14
target/arm/sve_helper.c | 218 ++++++++++++++++++++++++++++++-------
15
target/arm/translate-sve.c | 186 ++++++++++++++++++++++---------
16
5 files changed, 378 insertions(+), 91 deletions(-)
17
20
18
diff --git a/target/arm/cpu.h b/target/arm/cpu.h
21
diff --git a/include/hw/intc/armv7m_nvic.h b/include/hw/intc/armv7m_nvic.h
19
index XXXXXXX..XXXXXXX 100644
22
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/cpu.h
23
--- a/include/hw/intc/armv7m_nvic.h
21
+++ b/target/arm/cpu.h
24
+++ b/include/hw/intc/armv7m_nvic.h
22
@@ -XXX,XX +XXX,XX @@ static inline MemTxAttrs *typecheck_memtxattrs(MemTxAttrs *x)
25
@@ -XXX,XX +XXX,XX @@ struct NVICState {
23
* generic target bits directly.
26
MemoryRegion sysreg_ns_mem;
24
*/
27
MemoryRegion systickmem;
25
#define arm_tlb_bti_gp(x) (typecheck_memtxattrs(x)->target_tlb_bit0)
28
MemoryRegion systick_ns_mem;
26
+#define arm_tlb_mte_tagged(x) (typecheck_memtxattrs(x)->target_tlb_bit1)
29
+ MemoryRegion ras_mem;
27
30
MemoryRegion container;
28
/*
31
MemoryRegion defaultmem;
29
* Naming convention for isar_feature functions:
32
30
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
33
diff --git a/hw/intc/armv7m_nvic.c b/hw/intc/armv7m_nvic.c
31
index XXXXXXX..XXXXXXX 100644
34
index XXXXXXX..XXXXXXX 100644
32
--- a/target/arm/helper-sve.h
35
--- a/hw/intc/armv7m_nvic.c
33
+++ b/target/arm/helper-sve.h
36
+++ b/hw/intc/armv7m_nvic.c
34
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_ld1sds_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
@@ -XXX,XX +XXX,XX @@ static const MemoryRegionOps nvic_systick_ops = {
35
DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
.endianness = DEVICE_NATIVE_ENDIAN,
36
DEF_HELPER_FLAGS_4(sve_ld1sds_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
};
37
40
38
+DEF_HELPER_FLAGS_4(sve_ld1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
39
+DEF_HELPER_FLAGS_4(sve_ld2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_ld3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_ld4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+
41
+
43
+DEF_HELPER_FLAGS_4(sve_ld1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+static MemTxResult ras_read(void *opaque, hwaddr addr,
44
+DEF_HELPER_FLAGS_4(sve_ld2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
+ uint64_t *data, unsigned size,
45
+DEF_HELPER_FLAGS_4(sve_ld3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
44
+ MemTxAttrs attrs)
46
+DEF_HELPER_FLAGS_4(sve_ld4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+
48
+DEF_HELPER_FLAGS_4(sve_ld1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
49
+DEF_HELPER_FLAGS_4(sve_ld2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_ld3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_ld4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+
53
+DEF_HELPER_FLAGS_4(sve_ld1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
54
+DEF_HELPER_FLAGS_4(sve_ld2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_ld3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_ld4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+
58
+DEF_HELPER_FLAGS_4(sve_ld1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
59
+DEF_HELPER_FLAGS_4(sve_ld2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_ld3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_ld4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_ld1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_ld2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_ld3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_ld4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_ld1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_ld2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+DEF_HELPER_FLAGS_4(sve_ld3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
71
+DEF_HELPER_FLAGS_4(sve_ld4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
72
+
73
+DEF_HELPER_FLAGS_4(sve_ld1bhu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
74
+DEF_HELPER_FLAGS_4(sve_ld1bsu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
75
+DEF_HELPER_FLAGS_4(sve_ld1bdu_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
76
+DEF_HELPER_FLAGS_4(sve_ld1bhs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
77
+DEF_HELPER_FLAGS_4(sve_ld1bss_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
78
+DEF_HELPER_FLAGS_4(sve_ld1bds_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
79
+
80
+DEF_HELPER_FLAGS_4(sve_ld1hsu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
81
+DEF_HELPER_FLAGS_4(sve_ld1hdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
82
+DEF_HELPER_FLAGS_4(sve_ld1hss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
83
+DEF_HELPER_FLAGS_4(sve_ld1hds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
84
+
85
+DEF_HELPER_FLAGS_4(sve_ld1hsu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
86
+DEF_HELPER_FLAGS_4(sve_ld1hdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
87
+DEF_HELPER_FLAGS_4(sve_ld1hss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
88
+DEF_HELPER_FLAGS_4(sve_ld1hds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
89
+
90
+DEF_HELPER_FLAGS_4(sve_ld1sdu_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
91
+DEF_HELPER_FLAGS_4(sve_ld1sds_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
92
+
93
+DEF_HELPER_FLAGS_4(sve_ld1sdu_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
94
+DEF_HELPER_FLAGS_4(sve_ld1sds_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
95
+
96
DEF_HELPER_FLAGS_4(sve_ldff1bb_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
97
DEF_HELPER_FLAGS_4(sve_ldff1bhu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
98
DEF_HELPER_FLAGS_4(sve_ldff1bsu_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
99
diff --git a/target/arm/internals.h b/target/arm/internals.h
100
index XXXXXXX..XXXXXXX 100644
101
--- a/target/arm/internals.h
102
+++ b/target/arm/internals.h
103
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
104
#define LOG2_TAG_GRANULE 4
105
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
106
107
+/*
108
+ * The SVE simd_data field, for memory ops, contains either
109
+ * rd (5 bits) or a shift count (2 bits).
110
+ */
111
+#define SVE_MTEDESC_SHIFT 5
112
+
113
/* Bits within a descriptor passed to the helper_mte_check* functions. */
114
FIELD(MTEDESC, MIDX, 0, 4)
115
FIELD(MTEDESC, TBI, 4, 2)
116
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
117
index XXXXXXX..XXXXXXX 100644
118
--- a/target/arm/sve_helper.c
119
+++ b/target/arm/sve_helper.c
120
@@ -XXX,XX +XXX,XX @@ static void sve_cont_ldst_watchpoints(SVEContLdSt *info, CPUARMState *env,
121
#endif
122
}
123
124
+typedef uint64_t mte_check_fn(CPUARMState *, uint32_t, uint64_t, uintptr_t);
125
+
126
+static inline QEMU_ALWAYS_INLINE
127
+void sve_cont_ldst_mte_check_int(SVEContLdSt *info, CPUARMState *env,
128
+ uint64_t *vg, target_ulong addr, int esize,
129
+ int msize, uint32_t mtedesc, uintptr_t ra,
130
+ mte_check_fn *check)
131
+{
45
+{
132
+ intptr_t mem_off, reg_off, reg_last;
46
+ if (attrs.user) {
133
+
47
+ return MEMTX_ERROR;
134
+ /* Process the page only if MemAttr == Tagged. */
135
+ if (arm_tlb_mte_tagged(&info->page[0].attrs)) {
136
+ mem_off = info->mem_off_first[0];
137
+ reg_off = info->reg_off_first[0];
138
+ reg_last = info->reg_off_split;
139
+ if (reg_last < 0) {
140
+ reg_last = info->reg_off_last[0];
141
+ }
142
+
143
+ do {
144
+ uint64_t pg = vg[reg_off >> 6];
145
+ do {
146
+ if ((pg >> (reg_off & 63)) & 1) {
147
+ check(env, mtedesc, addr, ra);
148
+ }
149
+ reg_off += esize;
150
+ mem_off += msize;
151
+ } while (reg_off <= reg_last && (reg_off & 63));
152
+ } while (reg_off <= reg_last);
153
+ }
48
+ }
154
+
49
+
155
+ mem_off = info->mem_off_first[1];
50
+ switch (addr) {
156
+ if (mem_off >= 0 && arm_tlb_mte_tagged(&info->page[1].attrs)) {
51
+ case 0xe10: /* ERRIIDR */
157
+ reg_off = info->reg_off_first[1];
52
+ /* architect field = Arm; product/variant/revision 0 */
158
+ reg_last = info->reg_off_last[1];
53
+ *data = 0x43b;
159
+
54
+ break;
160
+ do {
55
+ case 0xfc8: /* ERRDEVID */
161
+ uint64_t pg = vg[reg_off >> 6];
56
+ /* Minimal RAS: we implement 0 error record indexes */
162
+ do {
57
+ *data = 0;
163
+ if ((pg >> (reg_off & 63)) & 1) {
58
+ break;
164
+ check(env, mtedesc, addr, ra);
59
+ default:
165
+ }
60
+ qemu_log_mask(LOG_UNIMP, "Read RAS register offset 0x%x\n",
166
+ reg_off += esize;
61
+ (uint32_t)addr);
167
+ mem_off += msize;
62
+ *data = 0;
168
+ } while (reg_off & 63);
63
+ break;
169
+ } while (reg_off <= reg_last);
170
+ }
64
+ }
65
+ return MEMTX_OK;
171
+}
66
+}
172
+
67
+
173
+typedef void sve_cont_ldst_mte_check_fn(SVEContLdSt *info, CPUARMState *env,
68
+static MemTxResult ras_write(void *opaque, hwaddr addr,
174
+ uint64_t *vg, target_ulong addr,
69
+ uint64_t value, unsigned size,
175
+ int esize, int msize, uint32_t mtedesc,
70
+ MemTxAttrs attrs)
176
+ uintptr_t ra);
71
+{
72
+ if (attrs.user) {
73
+ return MEMTX_ERROR;
74
+ }
177
+
75
+
178
+static void sve_cont_ldst_mte_check1(SVEContLdSt *info, CPUARMState *env,
76
+ switch (addr) {
179
+ uint64_t *vg, target_ulong addr,
77
+ default:
180
+ int esize, int msize, uint32_t mtedesc,
78
+ qemu_log_mask(LOG_UNIMP, "Write to RAS register offset 0x%x\n",
181
+ uintptr_t ra)
79
+ (uint32_t)addr);
182
+{
80
+ break;
183
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
81
+ }
184
+ mtedesc, ra, mte_check1);
82
+ return MEMTX_OK;
185
+}
83
+}
186
+
84
+
187
+static void sve_cont_ldst_mte_checkN(SVEContLdSt *info, CPUARMState *env,
85
+static const MemoryRegionOps ras_ops = {
188
+ uint64_t *vg, target_ulong addr,
86
+ .read_with_attrs = ras_read,
189
+ int esize, int msize, uint32_t mtedesc,
87
+ .write_with_attrs = ras_write,
190
+ uintptr_t ra)
88
+ .endianness = DEVICE_NATIVE_ENDIAN,
191
+{
89
+};
192
+ sve_cont_ldst_mte_check_int(info, env, vg, addr, esize, msize,
193
+ mtedesc, ra, mte_checkN);
194
+}
195
+
196
+
90
+
197
/*
91
/*
198
* Common helper for all contiguous 1,2,3,4-register predicated stores.
92
* Unassigned portions of the PPB space are RAZ/WI for privileged
199
*/
93
* accesses, and fault for non-privileged accesses.
200
static inline QEMU_ALWAYS_INLINE
94
@@ -XXX,XX +XXX,XX @@ static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
201
void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
95
&s->systick_ns_mem, 1);
202
uint32_t desc, const uintptr_t retaddr,
203
- const int esz, const int msz, const int N,
204
+ const int esz, const int msz, const int N, uint32_t mtedesc,
205
sve_ldst1_host_fn *host_fn,
206
- sve_ldst1_tlb_fn *tlb_fn)
207
+ sve_ldst1_tlb_fn *tlb_fn,
208
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
209
{
210
const unsigned rd = simd_data(desc);
211
const intptr_t reg_max = simd_oprsz(desc);
212
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
213
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
214
BP_MEM_READ, retaddr);
215
216
- /* TODO: MTE check. */
217
+ /*
218
+ * Handle mte checks for all active elements.
219
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
220
+ */
221
+ if (mte_check_fn && mtedesc) {
222
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
223
+ mtedesc, retaddr);
224
+ }
225
226
flags = info.page[0].flags | info.page[1].flags;
227
if (unlikely(flags != 0)) {
228
@@ -XXX,XX +XXX,XX @@ void sve_ldN_r(CPUARMState *env, uint64_t *vg, const target_ulong addr,
229
}
96
}
230
}
97
231
98
+ if (cpu_isar_feature(aa32_ras, s->cpu)) {
232
-#define DO_LD1_1(NAME, ESZ) \
99
+ memory_region_init_io(&s->ras_mem, OBJECT(s),
233
-void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
100
+ &ras_ops, s, "nvic_ras", 0x1000);
234
- target_ulong addr, uint32_t desc) \
101
+ memory_region_add_subregion(&s->container, 0x5000, &s->ras_mem);
235
-{ \
236
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
237
- sve_##NAME##_host, sve_##NAME##_tlb); \
238
+static inline QEMU_ALWAYS_INLINE
239
+void sve_ldN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
240
+ uint32_t desc, const uintptr_t ra,
241
+ const int esz, const int msz, const int N,
242
+ sve_ldst1_host_fn *host_fn,
243
+ sve_ldst1_tlb_fn *tlb_fn)
244
+{
245
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
246
+ int bit55 = extract64(addr, 55, 1);
247
+
248
+ /* Remove mtedesc from the normal sve descriptor. */
249
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
250
+
251
+ /* Perform gross MTE suppression early. */
252
+ if (!tbi_check(desc, bit55) ||
253
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
254
+ mtedesc = 0;
255
+ }
102
+ }
256
+
103
+
257
+ sve_ldN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
104
sysbus_init_mmio(SYS_BUS_DEVICE(dev), &s->container);
258
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
259
}
105
}
260
106
261
-#define DO_LD1_2(NAME, ESZ, MSZ) \
262
-void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
263
- target_ulong addr, uint32_t desc) \
264
-{ \
265
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
266
- sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
267
-} \
268
-void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
269
- target_ulong addr, uint32_t desc) \
270
-{ \
271
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
272
- sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
273
+#define DO_LD1_1(NAME, ESZ) \
274
+void HELPER(sve_##NAME##_r)(CPUARMState *env, void *vg, \
275
+ target_ulong addr, uint32_t desc) \
276
+{ \
277
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, 0, \
278
+ sve_##NAME##_host, sve_##NAME##_tlb, NULL); \
279
+} \
280
+void HELPER(sve_##NAME##_r_mte)(CPUARMState *env, void *vg, \
281
+ target_ulong addr, uint32_t desc) \
282
+{ \
283
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, 1, \
284
+ sve_##NAME##_host, sve_##NAME##_tlb); \
285
+}
286
+
287
+#define DO_LD1_2(NAME, ESZ, MSZ) \
288
+void HELPER(sve_##NAME##_le_r)(CPUARMState *env, void *vg, \
289
+ target_ulong addr, uint32_t desc) \
290
+{ \
291
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
292
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb, NULL); \
293
+} \
294
+void HELPER(sve_##NAME##_be_r)(CPUARMState *env, void *vg, \
295
+ target_ulong addr, uint32_t desc) \
296
+{ \
297
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, 0, \
298
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb, NULL); \
299
+} \
300
+void HELPER(sve_##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
301
+ target_ulong addr, uint32_t desc) \
302
+{ \
303
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
304
+ sve_##NAME##_le_host, sve_##NAME##_le_tlb); \
305
+} \
306
+void HELPER(sve_##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
307
+ target_ulong addr, uint32_t desc) \
308
+{ \
309
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, 1, \
310
+ sve_##NAME##_be_host, sve_##NAME##_be_tlb); \
311
}
312
313
DO_LD1_1(ld1bb, MO_8)
314
@@ -XXX,XX +XXX,XX @@ DO_LD1_2(ld1dd, MO_64, MO_64)
315
#undef DO_LD1_1
316
#undef DO_LD1_2
317
318
-#define DO_LDN_1(N) \
319
-void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
320
- target_ulong addr, uint32_t desc) \
321
-{ \
322
- sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
323
- sve_ld1bb_host, sve_ld1bb_tlb); \
324
+#define DO_LDN_1(N) \
325
+void HELPER(sve_ld##N##bb_r)(CPUARMState *env, void *vg, \
326
+ target_ulong addr, uint32_t desc) \
327
+{ \
328
+ sve_ldN_r(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, 0, \
329
+ sve_ld1bb_host, sve_ld1bb_tlb, NULL); \
330
+} \
331
+void HELPER(sve_ld##N##bb_r_mte)(CPUARMState *env, void *vg, \
332
+ target_ulong addr, uint32_t desc) \
333
+{ \
334
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), MO_8, MO_8, N, \
335
+ sve_ld1bb_host, sve_ld1bb_tlb); \
336
}
337
338
-#define DO_LDN_2(N, SUFF, ESZ) \
339
-void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
340
- target_ulong addr, uint32_t desc) \
341
-{ \
342
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
343
- sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
344
-} \
345
-void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
346
- target_ulong addr, uint32_t desc) \
347
-{ \
348
- sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
349
- sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
350
+#define DO_LDN_2(N, SUFF, ESZ) \
351
+void HELPER(sve_ld##N##SUFF##_le_r)(CPUARMState *env, void *vg, \
352
+ target_ulong addr, uint32_t desc) \
353
+{ \
354
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
355
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb, NULL); \
356
+} \
357
+void HELPER(sve_ld##N##SUFF##_be_r)(CPUARMState *env, void *vg, \
358
+ target_ulong addr, uint32_t desc) \
359
+{ \
360
+ sve_ldN_r(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, 0, \
361
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb, NULL); \
362
+} \
363
+void HELPER(sve_ld##N##SUFF##_le_r_mte)(CPUARMState *env, void *vg, \
364
+ target_ulong addr, uint32_t desc) \
365
+{ \
366
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
367
+ sve_ld1##SUFF##_le_host, sve_ld1##SUFF##_le_tlb); \
368
+} \
369
+void HELPER(sve_ld##N##SUFF##_be_r_mte)(CPUARMState *env, void *vg, \
370
+ target_ulong addr, uint32_t desc) \
371
+{ \
372
+ sve_ldN_r_mte(env, vg, addr, desc, GETPC(), ESZ, ESZ, N, \
373
+ sve_ld1##SUFF##_be_host, sve_ld1##SUFF##_be_tlb); \
374
}
375
376
DO_LDN_1(2)
377
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
378
index XXXXXXX..XXXXXXX 100644
379
--- a/target/arm/translate-sve.c
380
+++ b/target/arm/translate-sve.c
381
@@ -XXX,XX +XXX,XX @@ static const uint8_t dtype_esz[16] = {
382
};
383
384
static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
385
- int dtype, gen_helper_gvec_mem *fn)
386
+ int dtype, uint32_t mte_n, bool is_write,
387
+ gen_helper_gvec_mem *fn)
388
{
389
unsigned vsz = vec_full_reg_size(s);
390
TCGv_ptr t_pg;
391
TCGv_i32 t_desc;
392
- int desc;
393
+ int desc = 0;
394
395
- /* For e.g. LD4, there are not enough arguments to pass all 4
396
+ /*
397
+ * For e.g. LD4, there are not enough arguments to pass all 4
398
* registers as pointers, so encode the regno into the data field.
399
* For consistency, do this even for LD1.
400
+ * TODO: mte_n check here while callers are updated.
401
*/
402
- desc = simd_desc(vsz, vsz, zt);
403
+ if (mte_n && s->mte_active[0]) {
404
+ int msz = dtype_msz(dtype);
405
+
406
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
407
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
408
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
409
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
410
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << msz);
411
+ desc = FIELD_DP32(desc, MTEDESC, TSIZE, mte_n << msz);
412
+ desc <<= SVE_MTEDESC_SHIFT;
413
+ }
414
+ desc = simd_desc(vsz, vsz, zt | desc);
415
t_desc = tcg_const_i32(desc);
416
t_pg = tcg_temp_new_ptr();
417
418
@@ -XXX,XX +XXX,XX @@ static void do_mem_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
419
static void do_ld_zpa(DisasContext *s, int zt, int pg,
420
TCGv_i64 addr, int dtype, int nreg)
421
{
422
- static gen_helper_gvec_mem * const fns[2][16][4] = {
423
- /* Little-endian */
424
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
425
+ static gen_helper_gvec_mem * const fns[2][2][16][4] = {
426
+ { /* mte inactive, little-endian */
427
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
428
gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
429
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
430
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
431
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
432
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
433
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
434
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
435
436
- { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
437
- { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
438
- gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
439
- { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
440
- { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
441
+ { gen_helper_sve_ld1sds_le_r, NULL, NULL, NULL },
442
+ { gen_helper_sve_ld1hh_le_r, gen_helper_sve_ld2hh_le_r,
443
+ gen_helper_sve_ld3hh_le_r, gen_helper_sve_ld4hh_le_r },
444
+ { gen_helper_sve_ld1hsu_le_r, NULL, NULL, NULL },
445
+ { gen_helper_sve_ld1hdu_le_r, NULL, NULL, NULL },
446
447
- { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
448
- { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
449
- { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
450
- gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
451
- { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
452
+ { gen_helper_sve_ld1hds_le_r, NULL, NULL, NULL },
453
+ { gen_helper_sve_ld1hss_le_r, NULL, NULL, NULL },
454
+ { gen_helper_sve_ld1ss_le_r, gen_helper_sve_ld2ss_le_r,
455
+ gen_helper_sve_ld3ss_le_r, gen_helper_sve_ld4ss_le_r },
456
+ { gen_helper_sve_ld1sdu_le_r, NULL, NULL, NULL },
457
458
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
459
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
460
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
461
- { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
462
- gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
463
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
464
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
465
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
466
+ { gen_helper_sve_ld1dd_le_r, gen_helper_sve_ld2dd_le_r,
467
+ gen_helper_sve_ld3dd_le_r, gen_helper_sve_ld4dd_le_r } },
468
469
- /* Big-endian */
470
- { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
471
- gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
472
- { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
473
- { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
474
- { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
475
+ /* mte inactive, big-endian */
476
+ { { gen_helper_sve_ld1bb_r, gen_helper_sve_ld2bb_r,
477
+ gen_helper_sve_ld3bb_r, gen_helper_sve_ld4bb_r },
478
+ { gen_helper_sve_ld1bhu_r, NULL, NULL, NULL },
479
+ { gen_helper_sve_ld1bsu_r, NULL, NULL, NULL },
480
+ { gen_helper_sve_ld1bdu_r, NULL, NULL, NULL },
481
482
- { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
483
- { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
484
- gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
485
- { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
486
- { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
487
+ { gen_helper_sve_ld1sds_be_r, NULL, NULL, NULL },
488
+ { gen_helper_sve_ld1hh_be_r, gen_helper_sve_ld2hh_be_r,
489
+ gen_helper_sve_ld3hh_be_r, gen_helper_sve_ld4hh_be_r },
490
+ { gen_helper_sve_ld1hsu_be_r, NULL, NULL, NULL },
491
+ { gen_helper_sve_ld1hdu_be_r, NULL, NULL, NULL },
492
493
- { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
494
- { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
495
- { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
496
- gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
497
- { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
498
+ { gen_helper_sve_ld1hds_be_r, NULL, NULL, NULL },
499
+ { gen_helper_sve_ld1hss_be_r, NULL, NULL, NULL },
500
+ { gen_helper_sve_ld1ss_be_r, gen_helper_sve_ld2ss_be_r,
501
+ gen_helper_sve_ld3ss_be_r, gen_helper_sve_ld4ss_be_r },
502
+ { gen_helper_sve_ld1sdu_be_r, NULL, NULL, NULL },
503
504
- { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
505
- { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
506
- { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
507
- { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
508
- gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } }
509
+ { gen_helper_sve_ld1bds_r, NULL, NULL, NULL },
510
+ { gen_helper_sve_ld1bss_r, NULL, NULL, NULL },
511
+ { gen_helper_sve_ld1bhs_r, NULL, NULL, NULL },
512
+ { gen_helper_sve_ld1dd_be_r, gen_helper_sve_ld2dd_be_r,
513
+ gen_helper_sve_ld3dd_be_r, gen_helper_sve_ld4dd_be_r } } },
514
+
515
+ { /* mte active, little-endian */
516
+ { { gen_helper_sve_ld1bb_r_mte,
517
+ gen_helper_sve_ld2bb_r_mte,
518
+ gen_helper_sve_ld3bb_r_mte,
519
+ gen_helper_sve_ld4bb_r_mte },
520
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
521
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
522
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
523
+
524
+ { gen_helper_sve_ld1sds_le_r_mte, NULL, NULL, NULL },
525
+ { gen_helper_sve_ld1hh_le_r_mte,
526
+ gen_helper_sve_ld2hh_le_r_mte,
527
+ gen_helper_sve_ld3hh_le_r_mte,
528
+ gen_helper_sve_ld4hh_le_r_mte },
529
+ { gen_helper_sve_ld1hsu_le_r_mte, NULL, NULL, NULL },
530
+ { gen_helper_sve_ld1hdu_le_r_mte, NULL, NULL, NULL },
531
+
532
+ { gen_helper_sve_ld1hds_le_r_mte, NULL, NULL, NULL },
533
+ { gen_helper_sve_ld1hss_le_r_mte, NULL, NULL, NULL },
534
+ { gen_helper_sve_ld1ss_le_r_mte,
535
+ gen_helper_sve_ld2ss_le_r_mte,
536
+ gen_helper_sve_ld3ss_le_r_mte,
537
+ gen_helper_sve_ld4ss_le_r_mte },
538
+ { gen_helper_sve_ld1sdu_le_r_mte, NULL, NULL, NULL },
539
+
540
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
541
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
542
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
543
+ { gen_helper_sve_ld1dd_le_r_mte,
544
+ gen_helper_sve_ld2dd_le_r_mte,
545
+ gen_helper_sve_ld3dd_le_r_mte,
546
+ gen_helper_sve_ld4dd_le_r_mte } },
547
+
548
+ /* mte active, big-endian */
549
+ { { gen_helper_sve_ld1bb_r_mte,
550
+ gen_helper_sve_ld2bb_r_mte,
551
+ gen_helper_sve_ld3bb_r_mte,
552
+ gen_helper_sve_ld4bb_r_mte },
553
+ { gen_helper_sve_ld1bhu_r_mte, NULL, NULL, NULL },
554
+ { gen_helper_sve_ld1bsu_r_mte, NULL, NULL, NULL },
555
+ { gen_helper_sve_ld1bdu_r_mte, NULL, NULL, NULL },
556
+
557
+ { gen_helper_sve_ld1sds_be_r_mte, NULL, NULL, NULL },
558
+ { gen_helper_sve_ld1hh_be_r_mte,
559
+ gen_helper_sve_ld2hh_be_r_mte,
560
+ gen_helper_sve_ld3hh_be_r_mte,
561
+ gen_helper_sve_ld4hh_be_r_mte },
562
+ { gen_helper_sve_ld1hsu_be_r_mte, NULL, NULL, NULL },
563
+ { gen_helper_sve_ld1hdu_be_r_mte, NULL, NULL, NULL },
564
+
565
+ { gen_helper_sve_ld1hds_be_r_mte, NULL, NULL, NULL },
566
+ { gen_helper_sve_ld1hss_be_r_mte, NULL, NULL, NULL },
567
+ { gen_helper_sve_ld1ss_be_r_mte,
568
+ gen_helper_sve_ld2ss_be_r_mte,
569
+ gen_helper_sve_ld3ss_be_r_mte,
570
+ gen_helper_sve_ld4ss_be_r_mte },
571
+ { gen_helper_sve_ld1sdu_be_r_mte, NULL, NULL, NULL },
572
+
573
+ { gen_helper_sve_ld1bds_r_mte, NULL, NULL, NULL },
574
+ { gen_helper_sve_ld1bss_r_mte, NULL, NULL, NULL },
575
+ { gen_helper_sve_ld1bhs_r_mte, NULL, NULL, NULL },
576
+ { gen_helper_sve_ld1dd_be_r_mte,
577
+ gen_helper_sve_ld2dd_be_r_mte,
578
+ gen_helper_sve_ld3dd_be_r_mte,
579
+ gen_helper_sve_ld4dd_be_r_mte } } },
580
};
581
- gen_helper_gvec_mem *fn = fns[s->be_data == MO_BE][dtype][nreg];
582
+ gen_helper_gvec_mem *fn
583
+ = fns[s->mte_active[0]][s->be_data == MO_BE][dtype][nreg];
584
585
- /* While there are holes in the table, they are not
586
+ /*
587
+ * While there are holes in the table, they are not
588
* accessible via the instruction encoding.
589
*/
590
assert(fn != NULL);
591
- do_mem_zpa(s, zt, pg, addr, dtype, fn);
592
+ do_mem_zpa(s, zt, pg, addr, dtype, nreg, false, fn);
593
}
594
595
static bool trans_LD_zprr(DisasContext *s, arg_rprr_load *a)
596
@@ -XXX,XX +XXX,XX @@ static bool trans_LDFF1_zprr(DisasContext *s, arg_rprr_load *a)
597
TCGv_i64 addr = new_tmp_a64(s);
598
tcg_gen_shli_i64(addr, cpu_reg(s, a->rm), dtype_msz(a->dtype));
599
tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, a->rn));
600
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
601
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
602
fns[s->be_data == MO_BE][a->dtype]);
603
}
604
return true;
605
@@ -XXX,XX +XXX,XX @@ static bool trans_LDNF1_zpri(DisasContext *s, arg_rpri_load *a)
606
TCGv_i64 addr = new_tmp_a64(s);
607
608
tcg_gen_addi_i64(addr, cpu_reg_sp(s, a->rn), off);
609
- do_mem_zpa(s, a->rd, a->pg, addr, a->dtype,
610
+ do_mem_zpa(s, a->rd, a->pg, addr, a->dtype, 0, false,
611
fns[s->be_data == MO_BE][a->dtype]);
612
}
613
return true;
614
@@ -XXX,XX +XXX,XX @@ static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
615
fn = fn_multiple[be][nreg - 1][msz];
616
}
617
assert(fn != NULL);
618
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), fn);
619
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
620
}
621
622
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
623
--
107
--
624
2.20.1
108
2.20.1
625
109
626
110
diff view generated by jsdifflib
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
1
Correct a typo in the name we give the NVIC object.
2
2
3
We have 2 distinct PCA9552 devices. Set their description
3
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
4
to distinguish them when looking at the trace events.
4
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
5
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
6
Message-id: 20201119215617.29887-28-peter.maydell@linaro.org
7
---
8
hw/arm/armv7m.c | 2 +-
9
1 file changed, 1 insertion(+), 1 deletion(-)
5
10
6
Description name taken from:
11
diff --git a/hw/arm/armv7m.c b/hw/arm/armv7m.c
7
https://github.com/open-power/witherspoon-xml/blob/master/witherspoon.xml
8
9
Reviewed-by: Cédric Le Goater <clg@kaod.org>
10
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
11
Reviewed-by: Corey Minyard <cminyard@mvista.com>
12
Reviewed-by: Markus Armbruster <armbru@redhat.com>
13
Tested-by: Cédric Le Goater <clg@kaod.org>
14
Message-id: 20200623072723.6324-8-f4bug@amsat.org
15
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
16
---
17
hw/arm/aspeed.c | 13 +++++++++----
18
1 file changed, 9 insertions(+), 4 deletions(-)
19
20
diff --git a/hw/arm/aspeed.c b/hw/arm/aspeed.c
21
index XXXXXXX..XXXXXXX 100644
12
index XXXXXXX..XXXXXXX 100644
22
--- a/hw/arm/aspeed.c
13
--- a/hw/arm/armv7m.c
23
+++ b/hw/arm/aspeed.c
14
+++ b/hw/arm/armv7m.c
24
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
15
@@ -XXX,XX +XXX,XX @@ static void armv7m_instance_init(Object *obj)
25
{
16
26
AspeedSoCState *soc = &bmc->soc;
17
memory_region_init(&s->container, obj, "armv7m-container", UINT64_MAX);
27
uint8_t *eeprom_buf = g_malloc0(8 * 1024);
18
28
+ DeviceState *dev;
19
- object_initialize_child(obj, "nvnic", &s->nvic, TYPE_NVIC);
29
20
+ object_initialize_child(obj, "nvic", &s->nvic, TYPE_NVIC);
30
/* Bus 3: TODO bmp280@77 */
21
object_property_add_alias(obj, "num-irq",
31
/* Bus 3: TODO max31785@52 */
22
OBJECT(&s->nvic), "num-irq");
32
/* Bus 3: TODO dps310@76 */
33
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3), TYPE_PCA9552,
34
- 0x60);
35
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
36
+ qdev_prop_set_string(dev, "description", "pca1");
37
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 3),
38
+ &error_fatal);
39
40
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 4), "tmp423", 0x4c);
41
i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 5), "tmp423", 0x4c);
42
@@ -XXX,XX +XXX,XX @@ static void witherspoon_bmc_i2c_init(AspeedMachineState *bmc)
43
44
smbus_eeprom_init_one(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), 0x51,
45
eeprom_buf);
46
- i2c_create_slave(aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11), TYPE_PCA9552,
47
- 0x60);
48
+ dev = i2c_try_create_slave(TYPE_PCA9552, 0x60);
49
+ qdev_prop_set_string(dev, "description", "pca0");
50
+ i2c_realize_and_unref(dev, aspeed_i2c_get_bus(DEVICE(&soc->i2c), 11),
51
+ &error_fatal);
52
/* Bus 11: TODO ucd90160@64 */
53
}
54
23
55
--
24
--
56
2.20.1
25
2.20.1
57
26
58
27
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
Emit a trace event when a GPIO change its state.
4
5
Example booting obmc-phosphor-image:
6
7
$ qemu-system-arm -M witherspoon-bmc -trace pca955x_gpio_change
8
1592690552.687372:pca955x_gpio_change pca1 GPIO id:0 status: 0 -> 1
9
1592690552.690169:pca955x_gpio_change pca1 GPIO id:1 status: 0 -> 1
10
1592690552.691673:pca955x_gpio_change pca1 GPIO id:2 status: 0 -> 1
11
1592690552.696886:pca955x_gpio_change pca1 GPIO id:3 status: 0 -> 1
12
1592690552.698614:pca955x_gpio_change pca1 GPIO id:13 status: 0 -> 1
13
1592690552.699833:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
14
1592690552.700842:pca955x_gpio_change pca1 GPIO id:15 status: 0 -> 1
15
1592690683.841921:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
16
1592690683.861660:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
17
1592690684.371460:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
18
1592690684.882115:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
19
1592690685.391411:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
20
1592690685.901391:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
21
1592690686.411678:pca955x_gpio_change pca1 GPIO id:14 status: 1 -> 0
22
1592690686.921279:pca955x_gpio_change pca1 GPIO id:14 status: 0 -> 1
23
24
We notice the GPIO #14 (front-power LED) starts to blink.
25
26
This LED is described in the witherspoon device-tree [*]:
27
28
front-power {
29
retain-state-shutdown;
30
default-state = "keep";
31
gpios = <&pca0 14 GPIO_ACTIVE_LOW>;
32
};
33
34
[*] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/arm/boot/dts/aspeed-bmc-opp-witherspoon.dts?id=b1f9be9392f0#n140
35
36
Reviewed-by: Cédric Le Goater <clg@kaod.org>
37
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
38
Tested-by: Cédric Le Goater <clg@kaod.org>
39
Message-id: 20200623072723.6324-9-f4bug@amsat.org
40
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
41
---
42
hw/misc/pca9552.c | 15 +++++++++++++++
43
hw/misc/trace-events | 1 +
44
2 files changed, 16 insertions(+)
45
46
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
47
index XXXXXXX..XXXXXXX 100644
48
--- a/hw/misc/pca9552.c
49
+++ b/hw/misc/pca9552.c
50
@@ -XXX,XX +XXX,XX @@ static void pca955x_display_pins_status(PCA955xState *s,
51
buf[i] = '\0';
52
trace_pca955x_gpio_status(s->description, buf);
53
}
54
+ if (trace_event_get_state_backends(TRACE_PCA955X_GPIO_CHANGE)) {
55
+ for (i = 0; i < k->pin_count; i++) {
56
+ if (extract32(pins_changed, i, 1)) {
57
+ unsigned new_state = extract32(pins_status, i, 1);
58
+
59
+ /*
60
+ * We display the state using the PCA logic ("active-high").
61
+ * This is not the state of the LED, which signal might be
62
+ * wired "active-low" on the board.
63
+ */
64
+ trace_pca955x_gpio_change(s->description, i,
65
+ !new_state, new_state);
66
+ }
67
+ }
68
+ }
69
}
70
71
static void pca955x_update_pin_input(PCA955xState *s)
72
diff --git a/hw/misc/trace-events b/hw/misc/trace-events
73
index XXXXXXX..XXXXXXX 100644
74
--- a/hw/misc/trace-events
75
+++ b/hw/misc/trace-events
76
@@ -XXX,XX +XXX,XX @@ grlib_apb_pnp_read(uint64_t addr, uint32_t value) "APB PnP read addr:0x%03"PRIx6
77
78
# pca9552.c
79
pca955x_gpio_status(const char *description, const char *buf) "%s GPIOs 0-15 [%s]"
80
+pca955x_gpio_change(const char *description, unsigned id, unsigned prev_state, unsigned current_state) "%s GPIO id:%u status: %u -> %u"
81
--
82
2.20.1
83
84
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <f4bug@amsat.org>
2
1
3
The PCA9552 has 16 GPIOs which can be used as input,
4
output or PWM mode. QEMU models the output GPIO with
5
the qemu_irq type. Let the device expose the 16 GPIOs
6
to allow us to later connect LEDs to these outputs.
7
8
Reviewed-by: Cédric Le Goater <clg@kaod.org>
9
Signed-off-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
10
Tested-by: Cédric Le Goater <clg@kaod.org>
11
Message-id: 20200623072723.6324-10-f4bug@amsat.org
12
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
---
14
include/hw/misc/pca9552.h | 1 +
15
hw/misc/pca9552.c | 6 ++++++
16
2 files changed, 7 insertions(+)
17
18
diff --git a/include/hw/misc/pca9552.h b/include/hw/misc/pca9552.h
19
index XXXXXXX..XXXXXXX 100644
20
--- a/include/hw/misc/pca9552.h
21
+++ b/include/hw/misc/pca9552.h
22
@@ -XXX,XX +XXX,XX @@ typedef struct PCA955xState {
23
uint8_t pointer;
24
25
uint8_t regs[PCA955X_NR_REGS];
26
+ qemu_irq gpio[PCA955X_PIN_COUNT_MAX];
27
char *description; /* For debugging purpose only */
28
} PCA955xState;
29
30
diff --git a/hw/misc/pca9552.c b/hw/misc/pca9552.c
31
index XXXXXXX..XXXXXXX 100644
32
--- a/hw/misc/pca9552.c
33
+++ b/hw/misc/pca9552.c
34
@@ -XXX,XX +XXX,XX @@
35
#include "hw/qdev-properties.h"
36
#include "hw/misc/pca9552.h"
37
#include "hw/misc/pca9552_regs.h"
38
+#include "hw/irq.h"
39
#include "migration/vmstate.h"
40
#include "qapi/error.h"
41
#include "qapi/visitor.h"
42
@@ -XXX,XX +XXX,XX @@ static void pca955x_update_pin_input(PCA955xState *s)
43
44
switch (config) {
45
case PCA9552_LED_ON:
46
+ qemu_set_irq(s->gpio[i], 1);
47
s->regs[input_reg] |= 1 << input_shift;
48
break;
49
case PCA9552_LED_OFF:
50
+ qemu_set_irq(s->gpio[i], 0);
51
s->regs[input_reg] &= ~(1 << input_shift);
52
break;
53
case PCA9552_LED_PWM0:
54
@@ -XXX,XX +XXX,XX @@ static void pca955x_initfn(Object *obj)
55
56
static void pca955x_realize(DeviceState *dev, Error **errp)
57
{
58
+ PCA955xClass *k = PCA955X_GET_CLASS(dev);
59
PCA955xState *s = PCA955X(dev);
60
61
if (!s->description) {
62
s->description = g_strdup("pca-unspecified");
63
}
64
+
65
+ qdev_init_gpio_out(dev, s->gpio, k->pin_count);
66
}
67
68
static Property pca955x_properties[] = {
69
--
70
2.20.1
71
72
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-5-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper.c | 14 +++++++++++---
9
1 file changed, 11 insertions(+), 3 deletions(-)
10
11
diff --git a/target/arm/helper.c b/target/arm/helper.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/helper.c
14
+++ b/target/arm/helper.c
15
@@ -XXX,XX +XXX,XX @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
16
if (cpu_isar_feature(aa64_pauth, cpu)) {
17
valid_mask |= SCR_API | SCR_APK;
18
}
19
+ if (cpu_isar_feature(aa64_mte, cpu)) {
20
+ valid_mask |= SCR_ATA;
21
+ }
22
} else {
23
valid_mask &= ~(SCR_RW | SCR_ST);
24
}
25
@@ -XXX,XX +XXX,XX @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask)
26
if (cpu_isar_feature(aa64_pauth, cpu)) {
27
valid_mask |= HCR_API | HCR_APK;
28
}
29
+ if (cpu_isar_feature(aa64_mte, cpu)) {
30
+ valid_mask |= HCR_ATA | HCR_DCT | HCR_TID5;
31
+ }
32
}
33
34
/* Clear RES0 bits. */
35
value &= valid_mask;
36
37
- /* These bits change the MMU setup:
38
+ /*
39
+ * These bits change the MMU setup:
40
* HCR_VM enables stage 2 translation
41
* HCR_PTW forbids certain page-table setups
42
- * HCR_DC Disables stage1 and enables stage2 translation
43
+ * HCR_DC disables stage1 and enables stage2 translation
44
+ * HCR_DCT enables tagging on (disabled) stage1 translation
45
*/
46
- if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC)) {
47
+ if ((env->cp15.hcr_el2 ^ value) & (HCR_VM | HCR_PTW | HCR_DC | HCR_DCT)) {
48
tlb_flush(CPU(cpu));
49
}
50
env->cp15.hcr_el2 = value;
51
--
52
2.20.1
53
54
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Add an option that writes back the PC, like DISAS_UPDATE_EXIT,
4
but does not exit back to the main loop.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-7-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/translate.h | 2 ++
12
target/arm/translate-a64.c | 3 +++
13
target/arm/translate.c | 4 ++++
14
3 files changed, 9 insertions(+)
15
16
diff --git a/target/arm/translate.h b/target/arm/translate.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/translate.h
19
+++ b/target/arm/translate.h
20
@@ -XXX,XX +XXX,XX @@ static inline void disas_set_insn_syndrome(DisasContext *s, uint32_t syn)
21
* return from cpu_tb_exec.
22
*/
23
#define DISAS_EXIT DISAS_TARGET_9
24
+/* CPU state was modified dynamically; no need to exit, but do not chain. */
25
+#define DISAS_UPDATE_NOCHAIN DISAS_TARGET_10
26
27
#ifdef TARGET_AARCH64
28
void a64_translate_init(void);
29
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
30
index XXXXXXX..XXXXXXX 100644
31
--- a/target/arm/translate-a64.c
32
+++ b/target/arm/translate-a64.c
33
@@ -XXX,XX +XXX,XX @@ static void aarch64_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
34
case DISAS_EXIT:
35
tcg_gen_exit_tb(NULL, 0);
36
break;
37
+ case DISAS_UPDATE_NOCHAIN:
38
+ gen_a64_set_pc_im(dc->base.pc_next);
39
+ /* fall through */
40
case DISAS_JUMP:
41
tcg_gen_lookup_and_goto_ptr();
42
break;
43
diff --git a/target/arm/translate.c b/target/arm/translate.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/translate.c
46
+++ b/target/arm/translate.c
47
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
48
case DISAS_NEXT:
49
case DISAS_TOO_MANY:
50
case DISAS_UPDATE_EXIT:
51
+ case DISAS_UPDATE_NOCHAIN:
52
gen_set_pc_im(dc, dc->base.pc_next);
53
/* fall through */
54
default:
55
@@ -XXX,XX +XXX,XX @@ static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
56
case DISAS_TOO_MANY:
57
gen_goto_tb(dc, 1, dc->base.pc_next);
58
break;
59
+ case DISAS_UPDATE_NOCHAIN:
60
+ gen_set_pc_im(dc, dc->base.pc_next);
61
+ /* fall through */
62
case DISAS_JUMP:
63
gen_goto_ptr();
64
break;
65
--
66
2.20.1
67
68
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
The current Arm ARM has adjusted the official decode of
4
"Add/subtract (immediate)" so that the shift field is only bit 22,
5
and bit 23 is part of the op1 field of the parent category
6
"Data processing - immediate".
7
8
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
9
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
10
Message-id: 20200626033144.790098-11-richard.henderson@linaro.org
11
Suggested-by: Peter Maydell <peter.maydell@linaro.org>
12
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/translate-a64.c | 23 ++++++++---------------
16
1 file changed, 8 insertions(+), 15 deletions(-)
17
18
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
19
index XXXXXXX..XXXXXXX 100644
20
--- a/target/arm/translate-a64.c
21
+++ b/target/arm/translate-a64.c
22
@@ -XXX,XX +XXX,XX @@ static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
23
/*
24
* Add/subtract (immediate)
25
*
26
- * 31 30 29 28 24 23 22 21 10 9 5 4 0
27
- * +--+--+--+-----------+-----+-------------+-----+-----+
28
- * |sf|op| S| 1 0 0 0 1 |shift| imm12 | Rn | Rd |
29
- * +--+--+--+-----------+-----+-------------+-----+-----+
30
+ * 31 30 29 28 23 22 21 10 9 5 4 0
31
+ * +--+--+--+-------------+--+-------------+-----+-----+
32
+ * |sf|op| S| 1 0 0 0 1 0 |sh| imm12 | Rn | Rd |
33
+ * +--+--+--+-------------+--+-------------+-----+-----+
34
*
35
* sf: 0 -> 32bit, 1 -> 64bit
36
* op: 0 -> add , 1 -> sub
37
* S: 1 -> set flags
38
- * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
39
+ * sh: 1 -> LSL imm by 12
40
*/
41
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
42
{
43
int rd = extract32(insn, 0, 5);
44
int rn = extract32(insn, 5, 5);
45
uint64_t imm = extract32(insn, 10, 12);
46
- int shift = extract32(insn, 22, 2);
47
+ bool shift = extract32(insn, 22, 1);
48
bool setflags = extract32(insn, 29, 1);
49
bool sub_op = extract32(insn, 30, 1);
50
bool is_64bit = extract32(insn, 31, 1);
51
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
52
TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
53
TCGv_i64 tcg_result;
54
55
- switch (shift) {
56
- case 0x0:
57
- break;
58
- case 0x1:
59
+ if (shift) {
60
imm <<= 12;
61
- break;
62
- default:
63
- unallocated_encoding(s);
64
- return;
65
}
66
67
tcg_result = tcg_temp_new_i64();
68
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
69
case 0x20: case 0x21: /* PC-rel. addressing */
70
disas_pc_rel_adr(s, insn);
71
break;
72
- case 0x22: case 0x23: /* Add/subtract (immediate) */
73
+ case 0x22: /* Add/subtract (immediate) */
74
disas_add_sub_imm(s, insn);
75
break;
76
case 0x24: /* Logical (immediate) */
77
--
78
2.20.1
79
80
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-12-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/helper-a64.h | 1 +
9
target/arm/internals.h | 9 +++++++
10
target/arm/mte_helper.c | 10 ++++++++
11
target/arm/translate-a64.c | 51 ++++++++++++++++++++++++++++++++++++++
12
4 files changed, 71 insertions(+)
13
14
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/helper-a64.h
17
+++ b/target/arm/helper-a64.h
18
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
19
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
20
21
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
22
+DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
23
diff --git a/target/arm/internals.h b/target/arm/internals.h
24
index XXXXXXX..XXXXXXX 100644
25
--- a/target/arm/internals.h
26
+++ b/target/arm/internals.h
27
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
28
*/
29
#define GMID_EL1_BS 6
30
31
+/* We associate one allocation tag per 16 bytes, the minimum. */
32
+#define LOG2_TAG_GRANULE 4
33
+#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
34
+
35
+static inline int allocation_tag_from_addr(uint64_t ptr)
36
+{
37
+ return extract64(ptr, 56, 4);
38
+}
39
+
40
static inline uint64_t address_with_allocation_tag(uint64_t ptr, int rtag)
41
{
42
return deposit64(ptr, 56, 4, rtag);
43
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
44
index XXXXXXX..XXXXXXX 100644
45
--- a/target/arm/mte_helper.c
46
+++ b/target/arm/mte_helper.c
47
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
48
49
return address_with_allocation_tag(rn, rtag);
50
}
51
+
52
+uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
53
+ int32_t offset, uint32_t tag_offset)
54
+{
55
+ int start_tag = allocation_tag_from_addr(ptr);
56
+ uint16_t exclude = extract32(env->cp15.gcr_el1, 0, 16);
57
+ int rtag = choose_nonexcluded_tag(start_tag, tag_offset, exclude);
58
+
59
+ return address_with_allocation_tag(ptr + offset, rtag);
60
+}
61
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
62
index XXXXXXX..XXXXXXX 100644
63
--- a/target/arm/translate-a64.c
64
+++ b/target/arm/translate-a64.c
65
@@ -XXX,XX +XXX,XX @@ static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
66
tcg_temp_free_i64(tcg_result);
67
}
68
69
+/*
70
+ * Add/subtract (immediate, with tags)
71
+ *
72
+ * 31 30 29 28 23 22 21 16 14 10 9 5 4 0
73
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
74
+ * |sf|op| S| 1 0 0 0 1 1 |o2| uimm6 |o3| uimm4 | Rn | Rd |
75
+ * +--+--+--+-------------+--+---------+--+-------+-----+-----+
76
+ *
77
+ * op: 0 -> add, 1 -> sub
78
+ */
79
+static void disas_add_sub_imm_with_tags(DisasContext *s, uint32_t insn)
80
+{
81
+ int rd = extract32(insn, 0, 5);
82
+ int rn = extract32(insn, 5, 5);
83
+ int uimm4 = extract32(insn, 10, 4);
84
+ int uimm6 = extract32(insn, 16, 6);
85
+ bool sub_op = extract32(insn, 30, 1);
86
+ TCGv_i64 tcg_rn, tcg_rd;
87
+ int imm;
88
+
89
+ /* Test all of sf=1, S=0, o2=0, o3=0. */
90
+ if ((insn & 0xa040c000u) != 0x80000000u ||
91
+ !dc_isar_feature(aa64_mte_insn_reg, s)) {
92
+ unallocated_encoding(s);
93
+ return;
94
+ }
95
+
96
+ imm = uimm6 << LOG2_TAG_GRANULE;
97
+ if (sub_op) {
98
+ imm = -imm;
99
+ }
100
+
101
+ tcg_rn = cpu_reg_sp(s, rn);
102
+ tcg_rd = cpu_reg_sp(s, rd);
103
+
104
+ if (s->ata) {
105
+ TCGv_i32 offset = tcg_const_i32(imm);
106
+ TCGv_i32 tag_offset = tcg_const_i32(uimm4);
107
+
108
+ gen_helper_addsubg(tcg_rd, cpu_env, tcg_rn, offset, tag_offset);
109
+ tcg_temp_free_i32(tag_offset);
110
+ tcg_temp_free_i32(offset);
111
+ } else {
112
+ tcg_gen_addi_i64(tcg_rd, tcg_rn, imm);
113
+ gen_address_with_allocation_tag0(tcg_rd, tcg_rd);
114
+ }
115
+}
116
+
117
/* The input should be a value in the bottom e bits (with higher
118
* bits zero); returns that value replicated into every element
119
* of size e in a 64 bit integer.
120
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
121
case 0x22: /* Add/subtract (immediate) */
122
disas_add_sub_imm(s, insn);
123
break;
124
+ case 0x23: /* Add/subtract (immediate, with tags) */
125
+ disas_add_sub_imm_with_tags(s, insn);
126
+ break;
127
case 0x24: /* Logical (immediate) */
128
disas_logic_imm(s, insn);
129
break;
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-13-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 15 +++++++++++++++
9
1 file changed, 15 insertions(+)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
16
cpu_reg_sp(s, rn));
17
}
18
break;
19
+ case 5: /* GMI */
20
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
21
+ goto do_unallocated;
22
+ } else {
23
+ TCGv_i64 t1 = tcg_const_i64(1);
24
+ TCGv_i64 t2 = tcg_temp_new_i64();
25
+
26
+ tcg_gen_extract_i64(t2, cpu_reg_sp(s, rn), 56, 4);
27
+ tcg_gen_shl_i64(t1, t1, t2);
28
+ tcg_gen_or_i64(cpu_reg(s, rd), cpu_reg(s, rm), t1);
29
+
30
+ tcg_temp_free_i64(t1);
31
+ tcg_temp_free_i64(t2);
32
+ }
33
+ break;
34
case 8: /* LSLV */
35
handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
36
break;
37
--
38
2.20.1
39
40
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-14-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 24 ++++++++++++++++++++++--
9
1 file changed, 22 insertions(+), 2 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void handle_crc32(DisasContext *s,
16
*/
17
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
18
{
19
- unsigned int sf, rm, opcode, rn, rd;
20
+ unsigned int sf, rm, opcode, rn, rd, setflag;
21
sf = extract32(insn, 31, 1);
22
+ setflag = extract32(insn, 29, 1);
23
rm = extract32(insn, 16, 5);
24
opcode = extract32(insn, 10, 6);
25
rn = extract32(insn, 5, 5);
26
rd = extract32(insn, 0, 5);
27
28
- if (extract32(insn, 29, 1)) {
29
+ if (setflag && opcode != 0) {
30
unallocated_encoding(s);
31
return;
32
}
33
34
switch (opcode) {
35
+ case 0: /* SUBP(S) */
36
+ if (sf == 0 || !dc_isar_feature(aa64_mte_insn_reg, s)) {
37
+ goto do_unallocated;
38
+ } else {
39
+ TCGv_i64 tcg_n, tcg_m, tcg_d;
40
+
41
+ tcg_n = read_cpu_reg_sp(s, rn, true);
42
+ tcg_m = read_cpu_reg_sp(s, rm, true);
43
+ tcg_gen_sextract_i64(tcg_n, tcg_n, 0, 56);
44
+ tcg_gen_sextract_i64(tcg_m, tcg_m, 0, 56);
45
+ tcg_d = cpu_reg(s, rd);
46
+
47
+ if (setflag) {
48
+ gen_sub_CC(true, tcg_d, tcg_n, tcg_m);
49
+ } else {
50
+ tcg_gen_sub_i64(tcg_d, tcg_n, tcg_m);
51
+ }
52
+ }
53
+ break;
54
case 2: /* UDIV */
55
handle_div(s, false, sf, rm, rn, rd);
56
break;
57
--
58
2.20.1
59
60
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Use the same code as system mode, so that we generate the same
4
exception + syndrome for the unaligned access.
5
6
For the moment, if MTE is enabled so that this path is reachable,
7
this would generate a SIGSEGV in the user-only cpu_loop. Decoding
8
the syndrome to produce the proper SIGBUS will be done later.
9
10
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
11
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
12
Message-id: 20200626033144.790098-15-richard.henderson@linaro.org
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
---
15
target/arm/cpu.c | 2 +-
16
target/arm/tlb_helper.c | 4 ++--
17
2 files changed, 3 insertions(+), 3 deletions(-)
18
19
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/cpu.c
22
+++ b/target/arm/cpu.c
23
@@ -XXX,XX +XXX,XX @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
24
cc->tlb_fill = arm_cpu_tlb_fill;
25
cc->debug_excp_handler = arm_debug_excp_handler;
26
cc->debug_check_watchpoint = arm_debug_check_watchpoint;
27
-#if !defined(CONFIG_USER_ONLY)
28
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
29
+#if !defined(CONFIG_USER_ONLY)
30
cc->do_transaction_failed = arm_cpu_do_transaction_failed;
31
cc->adjust_watchpoint_address = arm_adjust_watchpoint_address;
32
#endif /* CONFIG_TCG && !CONFIG_USER_ONLY */
33
diff --git a/target/arm/tlb_helper.c b/target/arm/tlb_helper.c
34
index XXXXXXX..XXXXXXX 100644
35
--- a/target/arm/tlb_helper.c
36
+++ b/target/arm/tlb_helper.c
37
@@ -XXX,XX +XXX,XX @@
38
#include "internals.h"
39
#include "exec/exec-all.h"
40
41
-#if !defined(CONFIG_USER_ONLY)
42
-
43
static inline uint32_t merge_syn_data_abort(uint32_t template_syn,
44
unsigned int target_el,
45
bool same_el, bool ea,
46
@@ -XXX,XX +XXX,XX @@ void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr,
47
arm_deliver_fault(cpu, vaddr, access_type, mmu_idx, &fi);
48
}
49
50
+#if !defined(CONFIG_USER_ONLY)
51
+
52
/*
53
* arm_cpu_do_transaction_failed: handle a memory system error response
54
* (eg "no device/memory present at address") by raising an external abort
55
--
56
2.20.1
57
58
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-17-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-a64.c | 29 ++++++++++++++++++++++++++---
9
1 file changed, 26 insertions(+), 3 deletions(-)
10
11
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-a64.c
14
+++ b/target/arm/translate-a64.c
15
@@ -XXX,XX +XXX,XX @@ static void disas_ld_lit(DisasContext *s, uint32_t insn)
16
* +-----+-------+---+---+-------+---+-------+-------+------+------+
17
*
18
* opc: LDP/STP/LDNP/STNP 00 -> 32 bit, 10 -> 64 bit
19
- * LDPSW 01
20
+ * LDPSW/STGP 01
21
* LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
22
* V: 0 -> GPR, 1 -> Vector
23
* idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
24
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
25
bool is_signed = false;
26
bool postindex = false;
27
bool wback = false;
28
+ bool set_tag = false;
29
30
TCGv_i64 clean_addr, dirty_addr;
31
32
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
33
34
if (is_vector) {
35
size = 2 + opc;
36
+ } else if (opc == 1 && !is_load) {
37
+ /* STGP */
38
+ if (!dc_isar_feature(aa64_mte_insn_reg, s) || index == 0) {
39
+ unallocated_encoding(s);
40
+ return;
41
+ }
42
+ size = 3;
43
+ set_tag = true;
44
} else {
45
size = 2 + extract32(opc, 1, 1);
46
is_signed = extract32(opc, 0, 1);
47
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
48
return;
49
}
50
51
- offset <<= size;
52
+ offset <<= (set_tag ? LOG2_TAG_GRANULE : size);
53
54
if (rn == 31) {
55
gen_check_sp_alignment(s);
56
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pair(DisasContext *s, uint32_t insn)
57
if (!postindex) {
58
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
59
}
60
- clean_addr = clean_data_tbi(s, dirty_addr);
61
62
+ if (set_tag) {
63
+ if (!s->ata) {
64
+ /*
65
+ * TODO: We could rely on the stores below, at least for
66
+ * system mode, if we arrange to add MO_ALIGN_16.
67
+ */
68
+ gen_helper_stg_stub(cpu_env, dirty_addr);
69
+ } else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
70
+ gen_helper_stg_parallel(cpu_env, dirty_addr, dirty_addr);
71
+ } else {
72
+ gen_helper_stg(cpu_env, dirty_addr, dirty_addr);
73
+ }
74
+ }
75
+
76
+ clean_addr = clean_data_tbi(s, dirty_addr);
77
if (is_vector) {
78
if (is_load) {
79
do_fp_ld(s, rt, clean_addr, size);
80
--
81
2.20.1
82
83
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Replace existing uses of check_data_tbi in translate-a64.c that
4
perform a single logical memory access. Leave the helper blank
5
for now to reduce the patch size.
6
7
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
8
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
9
Message-id: 20200626033144.790098-24-richard.henderson@linaro.org
10
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
11
---
12
target/arm/helper-a64.h | 1 +
13
target/arm/internals.h | 8 +++
14
target/arm/translate-a64.h | 2 +
15
target/arm/mte_helper.c | 8 +++
16
target/arm/translate-a64.c | 100 ++++++++++++++++++++++++++++---------
17
5 files changed, 95 insertions(+), 24 deletions(-)
18
19
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
20
index XXXXXXX..XXXXXXX 100644
21
--- a/target/arm/helper-a64.h
22
+++ b/target/arm/helper-a64.h
23
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_3(autdb, TCG_CALL_NO_WG, i64, env, i64, i64)
24
DEF_HELPER_FLAGS_2(xpaci, TCG_CALL_NO_RWG_SE, i64, env, i64)
25
DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
26
27
+DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
28
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
29
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
30
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
31
diff --git a/target/arm/internals.h b/target/arm/internals.h
32
index XXXXXXX..XXXXXXX 100644
33
--- a/target/arm/internals.h
34
+++ b/target/arm/internals.h
35
@@ -XXX,XX +XXX,XX @@ void arm_log_exception(int idx);
36
#define LOG2_TAG_GRANULE 4
37
#define TAG_GRANULE (1 << LOG2_TAG_GRANULE)
38
39
+/* Bits within a descriptor passed to the helper_mte_check* functions. */
40
+FIELD(MTEDESC, MIDX, 0, 4)
41
+FIELD(MTEDESC, TBI, 4, 2)
42
+FIELD(MTEDESC, TCMA, 6, 2)
43
+FIELD(MTEDESC, WRITE, 8, 1)
44
+FIELD(MTEDESC, ESIZE, 9, 5)
45
+FIELD(MTEDESC, TSIZE, 14, 10) /* mte_checkN only */
46
+
47
static inline int allocation_tag_from_addr(uint64_t ptr)
48
{
49
return extract64(ptr, 56, 4);
50
diff --git a/target/arm/translate-a64.h b/target/arm/translate-a64.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/target/arm/translate-a64.h
53
+++ b/target/arm/translate-a64.h
54
@@ -XXX,XX +XXX,XX @@ TCGv_ptr get_fpstatus_ptr(bool);
55
bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
56
unsigned int imms, unsigned int immr);
57
bool sve_access_check(DisasContext *s);
58
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
59
+ bool tag_checked, int log2_size);
60
61
/* We should have at some point before trying to access an FP register
62
* done the necessary access check, so assert that
63
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
64
index XXXXXXX..XXXXXXX 100644
65
--- a/target/arm/mte_helper.c
66
+++ b/target/arm/mte_helper.c
67
@@ -XXX,XX +XXX,XX @@ void HELPER(stzgm_tags)(CPUARMState *env, uint64_t ptr, uint64_t val)
68
memset(mem, tag_pair, tag_bytes);
69
}
70
}
71
+
72
+/*
73
+ * Perform an MTE checked access for a single logical or atomic access.
74
+ */
75
+uint64_t HELPER(mte_check1)(CPUARMState *env, uint32_t desc, uint64_t ptr)
76
+{
77
+ return ptr;
78
+}
79
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
80
index XXXXXXX..XXXXXXX 100644
81
--- a/target/arm/translate-a64.c
82
+++ b/target/arm/translate-a64.c
83
@@ -XXX,XX +XXX,XX @@ static void gen_a64_set_pc(DisasContext *s, TCGv_i64 src)
84
}
85
86
/*
87
- * Return a "clean" address for ADDR according to TBID.
88
- * This is always a fresh temporary, as we need to be able to
89
- * increment this independently of a dirty write-back address.
90
+ * Handle MTE and/or TBI.
91
+ *
92
+ * For TBI, ideally, we would do nothing. Proper behaviour on fault is
93
+ * for the tag to be present in the FAR_ELx register. But for user-only
94
+ * mode we do not have a TLB with which to implement this, so we must
95
+ * remove the top byte now.
96
+ *
97
+ * Always return a fresh temporary that we can increment independently
98
+ * of the write-back address.
99
*/
100
+
101
static TCGv_i64 clean_data_tbi(DisasContext *s, TCGv_i64 addr)
102
{
103
TCGv_i64 clean = new_tmp_a64(s);
104
- /*
105
- * In order to get the correct value in the FAR_ELx register,
106
- * we must present the memory subsystem with the "dirty" address
107
- * including the TBI. In system mode we can make this work via
108
- * the TLB, dropping the TBI during translation. But for user-only
109
- * mode we don't have that option, and must remove the top byte now.
110
- */
111
#ifdef CONFIG_USER_ONLY
112
gen_top_byte_ignore(s, clean, addr, s->tbid);
113
#else
114
@@ -XXX,XX +XXX,XX @@ static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
115
tcg_temp_free_i32(t_size);
116
}
117
118
+/*
119
+ * For MTE, check a single logical or atomic access. This probes a single
120
+ * address, the exact one specified. The size and alignment of the access
121
+ * is not relevant to MTE, per se, but watchpoints do require the size,
122
+ * and we want to recognize those before making any other changes to state.
123
+ */
124
+static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
125
+ bool is_write, bool tag_checked,
126
+ int log2_size, bool is_unpriv,
127
+ int core_idx)
128
+{
129
+ if (tag_checked && s->mte_active[is_unpriv]) {
130
+ TCGv_i32 tcg_desc;
131
+ TCGv_i64 ret;
132
+ int desc = 0;
133
+
134
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, core_idx);
135
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
136
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
137
+ desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
138
+ desc = FIELD_DP32(desc, MTEDESC, ESIZE, 1 << log2_size);
139
+ tcg_desc = tcg_const_i32(desc);
140
+
141
+ ret = new_tmp_a64(s);
142
+ gen_helper_mte_check1(ret, cpu_env, tcg_desc, addr);
143
+ tcg_temp_free_i32(tcg_desc);
144
+
145
+ return ret;
146
+ }
147
+ return clean_data_tbi(s, addr);
148
+}
149
+
150
+TCGv_i64 gen_mte_check1(DisasContext *s, TCGv_i64 addr, bool is_write,
151
+ bool tag_checked, int log2_size)
152
+{
153
+ return gen_mte_check1_mmuidx(s, addr, is_write, tag_checked, log2_size,
154
+ false, get_mem_index(s));
155
+}
156
+
157
typedef struct DisasCompare64 {
158
TCGCond cond;
159
TCGv_i64 value;
160
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap(DisasContext *s, int rs, int rt,
161
if (rn == 31) {
162
gen_check_sp_alignment(s);
163
}
164
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
165
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size);
166
tcg_gen_atomic_cmpxchg_i64(tcg_rs, clean_addr, tcg_rs, tcg_rt, memidx,
167
size | MO_ALIGN | s->be_data);
168
}
169
@@ -XXX,XX +XXX,XX @@ static void gen_compare_and_swap_pair(DisasContext *s, int rs, int rt,
170
if (rn == 31) {
171
gen_check_sp_alignment(s);
172
}
173
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
174
+
175
+ /* This is a single atomic access, despite the "pair". */
176
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), true, rn != 31, size + 1);
177
178
if (size == 2) {
179
TCGv_i64 cmp = tcg_temp_new_i64();
180
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
181
if (is_lasr) {
182
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
183
}
184
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
185
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
186
+ true, rn != 31, size);
187
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, false);
188
return;
189
190
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
191
if (rn == 31) {
192
gen_check_sp_alignment(s);
193
}
194
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
195
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
196
+ false, rn != 31, size);
197
s->is_ldex = true;
198
gen_load_exclusive(s, rt, rt2, clean_addr, size, false);
199
if (is_lasr) {
200
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
201
gen_check_sp_alignment(s);
202
}
203
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
204
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
205
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
206
+ true, rn != 31, size);
207
do_gpr_st(s, cpu_reg(s, rt), clean_addr, size, true, rt,
208
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
209
return;
210
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
211
if (rn == 31) {
212
gen_check_sp_alignment(s);
213
}
214
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
215
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
216
+ false, rn != 31, size);
217
do_gpr_ld(s, cpu_reg(s, rt), clean_addr, size, false, false, true, rt,
218
disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
219
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
220
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
221
if (is_lasr) {
222
tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
223
}
224
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
225
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
226
+ true, rn != 31, size);
227
gen_store_exclusive(s, rs, rt, rt2, clean_addr, size, true);
228
return;
229
}
230
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)
231
if (rn == 31) {
232
gen_check_sp_alignment(s);
233
}
234
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
235
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn),
236
+ false, rn != 31, size);
237
s->is_ldex = true;
238
gen_load_exclusive(s, rt, rt2, clean_addr, size, true);
239
if (is_lasr) {
240
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
241
bool iss_valid = !is_vector;
242
bool post_index;
243
bool writeback;
244
+ int memidx;
245
246
TCGv_i64 clean_addr, dirty_addr;
247
248
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
249
if (!post_index) {
250
tcg_gen_addi_i64(dirty_addr, dirty_addr, imm9);
251
}
252
- clean_addr = clean_data_tbi(s, dirty_addr);
253
+
254
+ memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
255
+ clean_addr = gen_mte_check1_mmuidx(s, dirty_addr, is_store,
256
+ writeback || rn != 31,
257
+ size, is_unpriv, memidx);
258
259
if (is_vector) {
260
if (is_store) {
261
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn,
262
}
263
} else {
264
TCGv_i64 tcg_rt = cpu_reg(s, rt);
265
- int memidx = is_unpriv ? get_a64_user_mem_index(s) : get_mem_index(s);
266
bool iss_sf = disas_ldst_compute_iss_sf(size, is_signed, opc);
267
268
if (is_store) {
269
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn,
270
ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
271
272
tcg_gen_add_i64(dirty_addr, dirty_addr, tcg_rm);
273
- clean_addr = clean_data_tbi(s, dirty_addr);
274
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, true, size);
275
276
if (is_vector) {
277
if (is_store) {
278
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn,
279
dirty_addr = read_cpu_reg_sp(s, rn, 1);
280
offset = imm12 << size;
281
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
282
- clean_addr = clean_data_tbi(s, dirty_addr);
283
+ clean_addr = gen_mte_check1(s, dirty_addr, is_store, rn != 31, size);
284
285
if (is_vector) {
286
if (is_store) {
287
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_atomic(DisasContext *s, uint32_t insn,
288
if (rn == 31) {
289
gen_check_sp_alignment(s);
290
}
291
- clean_addr = clean_data_tbi(s, cpu_reg_sp(s, rn));
292
+ clean_addr = gen_mte_check1(s, cpu_reg_sp(s, rn), false, rn != 31, size);
293
294
if (o3_opc == 014) {
295
/*
296
@@ -XXX,XX +XXX,XX @@ static void disas_ldst_pac(DisasContext *s, uint32_t insn,
297
tcg_gen_addi_i64(dirty_addr, dirty_addr, offset);
298
299
/* Note that "clean" and "dirty" here refer to TBI not PAC. */
300
- clean_addr = clean_data_tbi(s, dirty_addr);
301
+ clean_addr = gen_mte_check1(s, dirty_addr, false,
302
+ is_wback || rn != 31, size);
303
304
tcg_rt = cpu_reg(s, rt);
305
do_gpr_ld(s, tcg_rt, clean_addr, size, /* is_signed */ false,
306
--
307
2.20.1
308
309
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Use a special helper for DC_ZVA, rather than the more
4
general mte_checkN.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-28-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-a64.h | 1 +
12
target/arm/mte_helper.c | 106 +++++++++++++++++++++++++++++++++++++
13
target/arm/translate-a64.c | 16 +++++-
14
3 files changed, 122 insertions(+), 1 deletion(-)
15
16
diff --git a/target/arm/helper-a64.h b/target/arm/helper-a64.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-a64.h
19
+++ b/target/arm/helper-a64.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
21
22
DEF_HELPER_FLAGS_3(mte_check1, TCG_CALL_NO_WG, i64, env, i32, i64)
23
DEF_HELPER_FLAGS_3(mte_checkN, TCG_CALL_NO_WG, i64, env, i32, i64)
24
+DEF_HELPER_FLAGS_3(mte_check_zva, TCG_CALL_NO_WG, i64, env, i32, i64)
25
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
26
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
27
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
28
diff --git a/target/arm/mte_helper.c b/target/arm/mte_helper.c
29
index XXXXXXX..XXXXXXX 100644
30
--- a/target/arm/mte_helper.c
31
+++ b/target/arm/mte_helper.c
32
@@ -XXX,XX +XXX,XX @@ uint64_t HELPER(mte_checkN)(CPUARMState *env, uint32_t desc, uint64_t ptr)
33
{
34
return mte_checkN(env, desc, ptr, GETPC());
35
}
36
+
37
+/*
38
+ * Perform an MTE checked access for DC_ZVA.
39
+ */
40
+uint64_t HELPER(mte_check_zva)(CPUARMState *env, uint32_t desc, uint64_t ptr)
41
+{
42
+ uintptr_t ra = GETPC();
43
+ int log2_dcz_bytes, log2_tag_bytes;
44
+ int mmu_idx, bit55;
45
+ intptr_t dcz_bytes, tag_bytes, i;
46
+ void *mem;
47
+ uint64_t ptr_tag, mem_tag, align_ptr;
48
+
49
+ bit55 = extract64(ptr, 55, 1);
50
+
51
+ /* If TBI is disabled, the access is unchecked, and ptr is not dirty. */
52
+ if (unlikely(!tbi_check(desc, bit55))) {
53
+ return ptr;
54
+ }
55
+
56
+ ptr_tag = allocation_tag_from_addr(ptr);
57
+
58
+ if (tcma_check(desc, bit55, ptr_tag)) {
59
+ goto done;
60
+ }
61
+
62
+ /*
63
+ * In arm_cpu_realizefn, we asserted that dcz > LOG2_TAG_GRANULE+1,
64
+ * i.e. 32 bytes, which is an unreasonably small dcz anyway, to make
65
+ * sure that we can access one complete tag byte here.
66
+ */
67
+ log2_dcz_bytes = env_archcpu(env)->dcz_blocksize + 2;
68
+ log2_tag_bytes = log2_dcz_bytes - (LOG2_TAG_GRANULE + 1);
69
+ dcz_bytes = (intptr_t)1 << log2_dcz_bytes;
70
+ tag_bytes = (intptr_t)1 << log2_tag_bytes;
71
+ align_ptr = ptr & -dcz_bytes;
72
+
73
+ /*
74
+ * Trap if accessing an invalid page. DC_ZVA requires that we supply
75
+ * the original pointer for an invalid page. But watchpoints require
76
+ * that we probe the actual space. So do both.
77
+ */
78
+ mmu_idx = FIELD_EX32(desc, MTEDESC, MIDX);
79
+ (void) probe_write(env, ptr, 1, mmu_idx, ra);
80
+ mem = allocation_tag_mem(env, mmu_idx, align_ptr, MMU_DATA_STORE,
81
+ dcz_bytes, MMU_DATA_LOAD, tag_bytes, ra);
82
+ if (!mem) {
83
+ goto done;
84
+ }
85
+
86
+ /*
87
+ * Unlike the reasoning for checkN, DC_ZVA is always aligned, and thus
88
+ * it is quite easy to perform all of the comparisons at once without
89
+ * any extra masking.
90
+ *
91
+ * The most common zva block size is 64; some of the thunderx cpus use
92
+ * a block size of 128. For user-only, aarch64_max_initfn will set the
93
+ * block size to 512. Fill out the other cases for future-proofing.
94
+ *
95
+ * In order to be able to find the first miscompare later, we want the
96
+ * tag bytes to be in little-endian order.
97
+ */
98
+ switch (log2_tag_bytes) {
99
+ case 0: /* zva_blocksize 32 */
100
+ mem_tag = *(uint8_t *)mem;
101
+ ptr_tag *= 0x11u;
102
+ break;
103
+ case 1: /* zva_blocksize 64 */
104
+ mem_tag = cpu_to_le16(*(uint16_t *)mem);
105
+ ptr_tag *= 0x1111u;
106
+ break;
107
+ case 2: /* zva_blocksize 128 */
108
+ mem_tag = cpu_to_le32(*(uint32_t *)mem);
109
+ ptr_tag *= 0x11111111u;
110
+ break;
111
+ case 3: /* zva_blocksize 256 */
112
+ mem_tag = cpu_to_le64(*(uint64_t *)mem);
113
+ ptr_tag *= 0x1111111111111111ull;
114
+ break;
115
+
116
+ default: /* zva_blocksize 512, 1024, 2048 */
117
+ ptr_tag *= 0x1111111111111111ull;
118
+ i = 0;
119
+ do {
120
+ mem_tag = cpu_to_le64(*(uint64_t *)(mem + i));
121
+ if (unlikely(mem_tag != ptr_tag)) {
122
+ goto fail;
123
+ }
124
+ i += 8;
125
+ align_ptr += 16 * TAG_GRANULE;
126
+ } while (i < tag_bytes);
127
+ goto done;
128
+ }
129
+
130
+ if (likely(mem_tag == ptr_tag)) {
131
+ goto done;
132
+ }
133
+
134
+ fail:
135
+ /* Locate the first nibble that differs. */
136
+ i = ctz64(mem_tag ^ ptr_tag) >> 4;
137
+ mte_check_fail(env, mmu_idx, align_ptr + i * TAG_GRANULE, ra);
138
+
139
+ done:
140
+ return useronly_clean_ptr(ptr);
141
+}
142
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
143
index XXXXXXX..XXXXXXX 100644
144
--- a/target/arm/translate-a64.c
145
+++ b/target/arm/translate-a64.c
146
@@ -XXX,XX +XXX,XX @@ static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
147
return;
148
case ARM_CP_DC_ZVA:
149
/* Writes clear the aligned block of memory which rt points into. */
150
- tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
151
+ if (s->mte_active[0]) {
152
+ TCGv_i32 t_desc;
153
+ int desc = 0;
154
+
155
+ desc = FIELD_DP32(desc, MTEDESC, MIDX, get_mem_index(s));
156
+ desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
157
+ desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
158
+ t_desc = tcg_const_i32(desc);
159
+
160
+ tcg_rt = new_tmp_a64(s);
161
+ gen_helper_mte_check_zva(tcg_rt, cpu_env, t_desc, cpu_reg(s, rt));
162
+ tcg_temp_free_i32(t_desc);
163
+ } else {
164
+ tcg_rt = clean_data_tbi(s, cpu_reg(s, rt));
165
+ }
166
gen_helper_dc_zva(cpu_env, tcg_rt);
167
return;
168
default:
169
--
170
2.20.1
171
172
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-29-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
9
1 file changed, 33 insertions(+), 28 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
16
int len_remain = len % 8;
17
int nparts = len / 8 + ctpop8(len_remain);
18
int midx = get_mem_index(s);
19
- TCGv_i64 addr, t0, t1;
20
+ TCGv_i64 dirty_addr, clean_addr, t0, t1;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
- /* Note that unpredicated load/store of vector/predicate registers
30
+ /*
31
+ * Note that unpredicated load/store of vector/predicate registers
32
* are defined as a stream of bytes, which equates to little-endian
33
- * operations on larger quantities. There is no nice way to force
34
- * a little-endian load for aarch64_be-linux-user out of line.
35
- *
36
+ * operations on larger quantities.
37
* Attempt to keep code expansion to a minimum by limiting the
38
* amount of unrolling done.
39
*/
40
if (nparts <= 4) {
41
int i;
42
43
+ t0 = tcg_temp_new_i64();
44
for (i = 0; i < len_align; i += 8) {
45
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
46
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
47
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
48
tcg_gen_st_i64(t0, cpu_env, vofs + i);
49
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
50
}
51
+ tcg_temp_free_i64(t0);
52
} else {
53
TCGLabel *loop = gen_new_label();
54
TCGv_ptr tp, i = tcg_const_local_ptr(0);
55
56
+ /* Copy the clean address into a local temp, live across the loop. */
57
+ t0 = clean_addr;
58
+ clean_addr = tcg_temp_local_new_i64();
59
+ tcg_gen_mov_i64(clean_addr, t0);
60
+ tcg_temp_free_i64(t0);
61
+
62
gen_set_label(loop);
63
64
- /* Minimize the number of local temps that must be re-read from
65
- * the stack each iteration. Instead, re-compute values other
66
- * than the loop counter.
67
- */
68
+ t0 = tcg_temp_new_i64();
69
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEQ);
70
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
71
+
72
tp = tcg_temp_new_ptr();
73
- tcg_gen_addi_ptr(tp, i, imm);
74
- tcg_gen_extu_ptr_i64(addr, tp);
75
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
76
-
77
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ);
78
-
79
tcg_gen_add_ptr(tp, cpu_env, i);
80
tcg_gen_addi_ptr(i, i, 8);
81
tcg_gen_st_i64(t0, tp, vofs);
82
tcg_temp_free_ptr(tp);
83
+ tcg_temp_free_i64(t0);
84
85
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
86
tcg_temp_free_ptr(i);
87
}
88
89
- /* Predicate register loads can be any multiple of 2.
90
+ /*
91
+ * Predicate register loads can be any multiple of 2.
92
* Note that we still store the entire 64-bit unit into cpu_env.
93
*/
94
if (len_remain) {
95
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
96
-
97
+ t0 = tcg_temp_new_i64();
98
switch (len_remain) {
99
case 2:
100
case 4:
101
case 8:
102
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
103
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx,
104
+ MO_LE | ctz32(len_remain));
105
break;
106
107
case 6:
108
t1 = tcg_temp_new_i64();
109
- tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL);
110
- tcg_gen_addi_i64(addr, addr, 4);
111
- tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW);
112
+ tcg_gen_qemu_ld_i64(t0, clean_addr, midx, MO_LEUL);
113
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
114
+ tcg_gen_qemu_ld_i64(t1, clean_addr, midx, MO_LEUW);
115
tcg_gen_deposit_i64(t0, t0, t1, 32, 32);
116
tcg_temp_free_i64(t1);
117
break;
118
@@ -XXX,XX +XXX,XX @@ static void do_ldr(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
119
g_assert_not_reached();
120
}
121
tcg_gen_st_i64(t0, cpu_env, vofs + len_align);
122
+ tcg_temp_free_i64(t0);
123
}
124
- tcg_temp_free_i64(addr);
125
- tcg_temp_free_i64(t0);
126
+ tcg_temp_free_i64(clean_addr);
127
}
128
129
/* Similarly for stores. */
130
--
131
2.20.1
132
133
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-30-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 61 +++++++++++++++++++++-----------------
9
1 file changed, 33 insertions(+), 28 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
16
int len_remain = len % 8;
17
int nparts = len / 8 + ctpop8(len_remain);
18
int midx = get_mem_index(s);
19
- TCGv_i64 addr, t0;
20
+ TCGv_i64 dirty_addr, clean_addr, t0;
21
22
- addr = tcg_temp_new_i64();
23
- t0 = tcg_temp_new_i64();
24
+ dirty_addr = tcg_temp_new_i64();
25
+ tcg_gen_addi_i64(dirty_addr, cpu_reg_sp(s, rn), imm);
26
+ clean_addr = gen_mte_checkN(s, dirty_addr, false, rn != 31, len, MO_8);
27
+ tcg_temp_free_i64(dirty_addr);
28
29
/* Note that unpredicated load/store of vector/predicate registers
30
* are defined as a stream of bytes, which equates to little-endian
31
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
32
if (nparts <= 4) {
33
int i;
34
35
+ t0 = tcg_temp_new_i64();
36
for (i = 0; i < len_align; i += 8) {
37
tcg_gen_ld_i64(t0, cpu_env, vofs + i);
38
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i);
39
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
40
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
41
+ tcg_gen_addi_i64(clean_addr, cpu_reg_sp(s, rn), 8);
42
}
43
+ tcg_temp_free_i64(t0);
44
} else {
45
TCGLabel *loop = gen_new_label();
46
- TCGv_ptr t2, i = tcg_const_local_ptr(0);
47
+ TCGv_ptr tp, i = tcg_const_local_ptr(0);
48
+
49
+ /* Copy the clean address into a local temp, live across the loop. */
50
+ t0 = clean_addr;
51
+ clean_addr = tcg_temp_local_new_i64();
52
+ tcg_gen_mov_i64(clean_addr, t0);
53
+ tcg_temp_free_i64(t0);
54
55
gen_set_label(loop);
56
57
- t2 = tcg_temp_new_ptr();
58
- tcg_gen_add_ptr(t2, cpu_env, i);
59
- tcg_gen_ld_i64(t0, t2, vofs);
60
-
61
- /* Minimize the number of local temps that must be re-read from
62
- * the stack each iteration. Instead, re-compute values other
63
- * than the loop counter.
64
- */
65
- tcg_gen_addi_ptr(t2, i, imm);
66
- tcg_gen_extu_ptr_i64(addr, t2);
67
- tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn));
68
- tcg_temp_free_ptr(t2);
69
-
70
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEQ);
71
-
72
+ t0 = tcg_temp_new_i64();
73
+ tp = tcg_temp_new_ptr();
74
+ tcg_gen_add_ptr(tp, cpu_env, i);
75
+ tcg_gen_ld_i64(t0, tp, vofs);
76
tcg_gen_addi_ptr(i, i, 8);
77
+ tcg_temp_free_ptr(tp);
78
+
79
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEQ);
80
+ tcg_gen_addi_i64(clean_addr, clean_addr, 8);
81
+ tcg_temp_free_i64(t0);
82
83
tcg_gen_brcondi_ptr(TCG_COND_LTU, i, len_align, loop);
84
tcg_temp_free_ptr(i);
85
@@ -XXX,XX +XXX,XX @@ static void do_str(DisasContext *s, uint32_t vofs, int len, int rn, int imm)
86
87
/* Predicate register stores can be any multiple of 2. */
88
if (len_remain) {
89
+ t0 = tcg_temp_new_i64();
90
tcg_gen_ld_i64(t0, cpu_env, vofs + len_align);
91
- tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align);
92
93
switch (len_remain) {
94
case 2:
95
case 4:
96
case 8:
97
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LE | ctz32(len_remain));
98
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx,
99
+ MO_LE | ctz32(len_remain));
100
break;
101
102
case 6:
103
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUL);
104
- tcg_gen_addi_i64(addr, addr, 4);
105
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUL);
106
+ tcg_gen_addi_i64(clean_addr, clean_addr, 4);
107
tcg_gen_shri_i64(t0, t0, 32);
108
- tcg_gen_qemu_st_i64(t0, addr, midx, MO_LEUW);
109
+ tcg_gen_qemu_st_i64(t0, clean_addr, midx, MO_LEUW);
110
break;
111
112
default:
113
g_assert_not_reached();
114
}
115
+ tcg_temp_free_i64(t0);
116
}
117
- tcg_temp_free_i64(addr);
118
- tcg_temp_free_i64(t0);
119
+ tcg_temp_free_i64(clean_addr);
120
}
121
122
static bool trans_LDR_zri(DisasContext *s, arg_rri *a)
123
--
124
2.20.1
125
126
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
4
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
5
Message-id: 20200626033144.790098-31-richard.henderson@linaro.org
6
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
7
---
8
target/arm/translate-sve.c | 6 ++++--
9
1 file changed, 4 insertions(+), 2 deletions(-)
10
11
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/target/arm/translate-sve.c
14
+++ b/target/arm/translate-sve.c
15
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
16
unsigned esz = dtype_esz[a->dtype];
17
unsigned msz = dtype_msz(a->dtype);
18
TCGLabel *over = gen_new_label();
19
- TCGv_i64 temp;
20
+ TCGv_i64 temp, clean_addr;
21
22
/* If the guarding predicate has no bits set, no load occurs. */
23
if (psz <= 8) {
24
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
25
/* Load the data. */
26
temp = tcg_temp_new_i64();
27
tcg_gen_addi_i64(temp, cpu_reg_sp(s, a->rn), a->imm << msz);
28
- tcg_gen_qemu_ld_i64(temp, temp, get_mem_index(s),
29
+ clean_addr = gen_mte_check1(s, temp, false, true, msz);
30
+
31
+ tcg_gen_qemu_ld_i64(temp, clean_addr, get_mem_index(s),
32
s->be_data | dtype_mop[a->dtype]);
33
34
/* Broadcast to *all* elements. */
35
--
36
2.20.1
37
38
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Move the variable declarations to the top of the function,
4
but do not create a new label before sve_access_check.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-32-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/translate-sve.c | 12 +++++++-----
12
1 file changed, 7 insertions(+), 5 deletions(-)
13
14
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/target/arm/translate-sve.c
17
+++ b/target/arm/translate-sve.c
18
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1RQ_zpri(DisasContext *s, arg_rpri_load *a)
19
/* Load and broadcast element. */
20
static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
21
{
22
- if (!sve_access_check(s)) {
23
- return true;
24
- }
25
-
26
unsigned vsz = vec_full_reg_size(s);
27
unsigned psz = pred_full_reg_size(s);
28
unsigned esz = dtype_esz[a->dtype];
29
unsigned msz = dtype_msz(a->dtype);
30
- TCGLabel *over = gen_new_label();
31
+ TCGLabel *over;
32
TCGv_i64 temp, clean_addr;
33
34
+ if (!sve_access_check(s)) {
35
+ return true;
36
+ }
37
+
38
+ over = gen_new_label();
39
+
40
/* If the guarding predicate has no bits set, no load occurs. */
41
if (psz <= 8) {
42
/* Reduce the pred_esz_masks value simply to reduce the
43
--
44
2.20.1
45
46
diff view generated by jsdifflib
Deleted patch
1
From: Richard Henderson <richard.henderson@linaro.org>
2
1
3
Because the elements are sequential, we can eliminate many tests all
4
at once when the tag hits TCMA, or if the page(s) are not Tagged.
5
6
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
7
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
8
Message-id: 20200626033144.790098-35-richard.henderson@linaro.org
9
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
10
---
11
target/arm/helper-sve.h | 47 +++++++++++
12
target/arm/sve_helper.c | 95 ++++++++++++++++------
13
target/arm/translate-sve.c | 162 ++++++++++++++++++++++++-------------
14
3 files changed, 226 insertions(+), 78 deletions(-)
15
16
diff --git a/target/arm/helper-sve.h b/target/arm/helper-sve.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/target/arm/helper-sve.h
19
+++ b/target/arm/helper-sve.h
20
@@ -XXX,XX +XXX,XX @@ DEF_HELPER_FLAGS_4(sve_st1hd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
21
DEF_HELPER_FLAGS_4(sve_st1sd_le_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
22
DEF_HELPER_FLAGS_4(sve_st1sd_be_r, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
23
24
+DEF_HELPER_FLAGS_4(sve_st1bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
25
+DEF_HELPER_FLAGS_4(sve_st2bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
26
+DEF_HELPER_FLAGS_4(sve_st3bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
27
+DEF_HELPER_FLAGS_4(sve_st4bb_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
28
+
29
+DEF_HELPER_FLAGS_4(sve_st1hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
30
+DEF_HELPER_FLAGS_4(sve_st2hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
31
+DEF_HELPER_FLAGS_4(sve_st3hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
32
+DEF_HELPER_FLAGS_4(sve_st4hh_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
33
+
34
+DEF_HELPER_FLAGS_4(sve_st1hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
35
+DEF_HELPER_FLAGS_4(sve_st2hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
36
+DEF_HELPER_FLAGS_4(sve_st3hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
37
+DEF_HELPER_FLAGS_4(sve_st4hh_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
38
+
39
+DEF_HELPER_FLAGS_4(sve_st1ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
40
+DEF_HELPER_FLAGS_4(sve_st2ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
41
+DEF_HELPER_FLAGS_4(sve_st3ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
42
+DEF_HELPER_FLAGS_4(sve_st4ss_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
43
+
44
+DEF_HELPER_FLAGS_4(sve_st1ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
45
+DEF_HELPER_FLAGS_4(sve_st2ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
46
+DEF_HELPER_FLAGS_4(sve_st3ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
47
+DEF_HELPER_FLAGS_4(sve_st4ss_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
48
+
49
+DEF_HELPER_FLAGS_4(sve_st1dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
50
+DEF_HELPER_FLAGS_4(sve_st2dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
51
+DEF_HELPER_FLAGS_4(sve_st3dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
52
+DEF_HELPER_FLAGS_4(sve_st4dd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
53
+
54
+DEF_HELPER_FLAGS_4(sve_st1dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
55
+DEF_HELPER_FLAGS_4(sve_st2dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
56
+DEF_HELPER_FLAGS_4(sve_st3dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
57
+DEF_HELPER_FLAGS_4(sve_st4dd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
58
+
59
+DEF_HELPER_FLAGS_4(sve_st1bh_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
60
+DEF_HELPER_FLAGS_4(sve_st1bs_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
61
+DEF_HELPER_FLAGS_4(sve_st1bd_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
62
+
63
+DEF_HELPER_FLAGS_4(sve_st1hs_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
64
+DEF_HELPER_FLAGS_4(sve_st1hd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
65
+DEF_HELPER_FLAGS_4(sve_st1hs_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
66
+DEF_HELPER_FLAGS_4(sve_st1hd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
67
+
68
+DEF_HELPER_FLAGS_4(sve_st1sd_le_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
69
+DEF_HELPER_FLAGS_4(sve_st1sd_be_r_mte, TCG_CALL_NO_WG, void, env, ptr, tl, i32)
70
+
71
DEF_HELPER_FLAGS_6(sve_ldbsu_zsu, TCG_CALL_NO_WG,
72
void, env, ptr, ptr, ptr, tl, i32)
73
DEF_HELPER_FLAGS_6(sve_ldhsu_le_zsu, TCG_CALL_NO_WG,
74
diff --git a/target/arm/sve_helper.c b/target/arm/sve_helper.c
75
index XXXXXXX..XXXXXXX 100644
76
--- a/target/arm/sve_helper.c
77
+++ b/target/arm/sve_helper.c
78
@@ -XXX,XX +XXX,XX @@ DO_LDFF1_LDNF1_2(dd, MO_64, MO_64)
79
*/
80
81
static inline QEMU_ALWAYS_INLINE
82
-void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
83
- const uintptr_t retaddr, const int esz,
84
- const int msz, const int N,
85
+void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr,
86
+ uint32_t desc, const uintptr_t retaddr,
87
+ const int esz, const int msz, const int N, uint32_t mtedesc,
88
sve_ldst1_host_fn *host_fn,
89
- sve_ldst1_tlb_fn *tlb_fn)
90
+ sve_ldst1_tlb_fn *tlb_fn,
91
+ sve_cont_ldst_mte_check_fn *mte_check_fn)
92
{
93
const unsigned rd = simd_data(desc);
94
const intptr_t reg_max = simd_oprsz(desc);
95
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
96
sve_cont_ldst_watchpoints(&info, env, vg, addr, 1 << esz, N << msz,
97
BP_MEM_WRITE, retaddr);
98
99
- /* TODO: MTE check. */
100
+ /*
101
+ * Handle mte checks for all active elements.
102
+ * Since TBI must be set for MTE, !mtedesc => !mte_active.
103
+ */
104
+ if (mte_check_fn && mtedesc) {
105
+ mte_check_fn(&info, env, vg, addr, 1 << esz, N << msz,
106
+ mtedesc, retaddr);
107
+ }
108
109
flags = info.page[0].flags | info.page[1].flags;
110
if (unlikely(flags != 0)) {
111
@@ -XXX,XX +XXX,XX @@ void sve_stN_r(CPUARMState *env, uint64_t *vg, target_ulong addr, uint32_t desc,
112
}
113
}
114
115
-#define DO_STN_1(N, NAME, ESZ) \
116
-void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
117
- target_ulong addr, uint32_t desc) \
118
-{ \
119
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
120
- sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
121
+static inline QEMU_ALWAYS_INLINE
122
+void sve_stN_r_mte(CPUARMState *env, uint64_t *vg, target_ulong addr,
123
+ uint32_t desc, const uintptr_t ra,
124
+ const int esz, const int msz, const int N,
125
+ sve_ldst1_host_fn *host_fn,
126
+ sve_ldst1_tlb_fn *tlb_fn)
127
+{
128
+ uint32_t mtedesc = desc >> (SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
129
+ int bit55 = extract64(addr, 55, 1);
130
+
131
+ /* Remove mtedesc from the normal sve descriptor. */
132
+ desc = extract32(desc, 0, SIMD_DATA_SHIFT + SVE_MTEDESC_SHIFT);
133
+
134
+ /* Perform gross MTE suppression early. */
135
+ if (!tbi_check(desc, bit55) ||
136
+ tcma_check(desc, bit55, allocation_tag_from_addr(addr))) {
137
+ mtedesc = 0;
138
+ }
139
+
140
+ sve_stN_r(env, vg, addr, desc, ra, esz, msz, N, mtedesc, host_fn, tlb_fn,
141
+ N == 1 ? sve_cont_ldst_mte_check1 : sve_cont_ldst_mte_checkN);
142
}
143
144
-#define DO_STN_2(N, NAME, ESZ, MSZ) \
145
-void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
146
- target_ulong addr, uint32_t desc) \
147
-{ \
148
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
149
- sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
150
-} \
151
-void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
152
- target_ulong addr, uint32_t desc) \
153
-{ \
154
- sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
155
- sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
156
+#define DO_STN_1(N, NAME, ESZ) \
157
+void HELPER(sve_st##N##NAME##_r)(CPUARMState *env, void *vg, \
158
+ target_ulong addr, uint32_t desc) \
159
+{ \
160
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, 0, \
161
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb, NULL); \
162
+} \
163
+void HELPER(sve_st##N##NAME##_r_mte)(CPUARMState *env, void *vg, \
164
+ target_ulong addr, uint32_t desc) \
165
+{ \
166
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MO_8, N, \
167
+ sve_st1##NAME##_host, sve_st1##NAME##_tlb); \
168
+}
169
+
170
+#define DO_STN_2(N, NAME, ESZ, MSZ) \
171
+void HELPER(sve_st##N##NAME##_le_r)(CPUARMState *env, void *vg, \
172
+ target_ulong addr, uint32_t desc) \
173
+{ \
174
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
175
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb, NULL); \
176
+} \
177
+void HELPER(sve_st##N##NAME##_be_r)(CPUARMState *env, void *vg, \
178
+ target_ulong addr, uint32_t desc) \
179
+{ \
180
+ sve_stN_r(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, 0, \
181
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb, NULL); \
182
+} \
183
+void HELPER(sve_st##N##NAME##_le_r_mte)(CPUARMState *env, void *vg, \
184
+ target_ulong addr, uint32_t desc) \
185
+{ \
186
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
187
+ sve_st1##NAME##_le_host, sve_st1##NAME##_le_tlb); \
188
+} \
189
+void HELPER(sve_st##N##NAME##_be_r_mte)(CPUARMState *env, void *vg, \
190
+ target_ulong addr, uint32_t desc) \
191
+{ \
192
+ sve_stN_r_mte(env, vg, addr, desc, GETPC(), ESZ, MSZ, N, \
193
+ sve_st1##NAME##_be_host, sve_st1##NAME##_be_tlb); \
194
}
195
196
DO_STN_1(1, bb, MO_8)
197
diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
198
index XXXXXXX..XXXXXXX 100644
199
--- a/target/arm/translate-sve.c
200
+++ b/target/arm/translate-sve.c
201
@@ -XXX,XX +XXX,XX @@ static bool trans_LD1R_zpri(DisasContext *s, arg_rpri_load *a)
202
static void do_st_zpa(DisasContext *s, int zt, int pg, TCGv_i64 addr,
203
int msz, int esz, int nreg)
204
{
205
- static gen_helper_gvec_mem * const fn_single[2][4][4] = {
206
- { { gen_helper_sve_st1bb_r,
207
- gen_helper_sve_st1bh_r,
208
- gen_helper_sve_st1bs_r,
209
- gen_helper_sve_st1bd_r },
210
- { NULL,
211
- gen_helper_sve_st1hh_le_r,
212
- gen_helper_sve_st1hs_le_r,
213
- gen_helper_sve_st1hd_le_r },
214
- { NULL, NULL,
215
- gen_helper_sve_st1ss_le_r,
216
- gen_helper_sve_st1sd_le_r },
217
- { NULL, NULL, NULL,
218
- gen_helper_sve_st1dd_le_r } },
219
- { { gen_helper_sve_st1bb_r,
220
- gen_helper_sve_st1bh_r,
221
- gen_helper_sve_st1bs_r,
222
- gen_helper_sve_st1bd_r },
223
- { NULL,
224
- gen_helper_sve_st1hh_be_r,
225
- gen_helper_sve_st1hs_be_r,
226
- gen_helper_sve_st1hd_be_r },
227
- { NULL, NULL,
228
- gen_helper_sve_st1ss_be_r,
229
- gen_helper_sve_st1sd_be_r },
230
- { NULL, NULL, NULL,
231
- gen_helper_sve_st1dd_be_r } },
232
+ static gen_helper_gvec_mem * const fn_single[2][2][4][4] = {
233
+ { { { gen_helper_sve_st1bb_r,
234
+ gen_helper_sve_st1bh_r,
235
+ gen_helper_sve_st1bs_r,
236
+ gen_helper_sve_st1bd_r },
237
+ { NULL,
238
+ gen_helper_sve_st1hh_le_r,
239
+ gen_helper_sve_st1hs_le_r,
240
+ gen_helper_sve_st1hd_le_r },
241
+ { NULL, NULL,
242
+ gen_helper_sve_st1ss_le_r,
243
+ gen_helper_sve_st1sd_le_r },
244
+ { NULL, NULL, NULL,
245
+ gen_helper_sve_st1dd_le_r } },
246
+ { { gen_helper_sve_st1bb_r,
247
+ gen_helper_sve_st1bh_r,
248
+ gen_helper_sve_st1bs_r,
249
+ gen_helper_sve_st1bd_r },
250
+ { NULL,
251
+ gen_helper_sve_st1hh_be_r,
252
+ gen_helper_sve_st1hs_be_r,
253
+ gen_helper_sve_st1hd_be_r },
254
+ { NULL, NULL,
255
+ gen_helper_sve_st1ss_be_r,
256
+ gen_helper_sve_st1sd_be_r },
257
+ { NULL, NULL, NULL,
258
+ gen_helper_sve_st1dd_be_r } } },
259
+
260
+ { { { gen_helper_sve_st1bb_r_mte,
261
+ gen_helper_sve_st1bh_r_mte,
262
+ gen_helper_sve_st1bs_r_mte,
263
+ gen_helper_sve_st1bd_r_mte },
264
+ { NULL,
265
+ gen_helper_sve_st1hh_le_r_mte,
266
+ gen_helper_sve_st1hs_le_r_mte,
267
+ gen_helper_sve_st1hd_le_r_mte },
268
+ { NULL, NULL,
269
+ gen_helper_sve_st1ss_le_r_mte,
270
+ gen_helper_sve_st1sd_le_r_mte },
271
+ { NULL, NULL, NULL,
272
+ gen_helper_sve_st1dd_le_r_mte } },
273
+ { { gen_helper_sve_st1bb_r_mte,
274
+ gen_helper_sve_st1bh_r_mte,
275
+ gen_helper_sve_st1bs_r_mte,
276
+ gen_helper_sve_st1bd_r_mte },
277
+ { NULL,
278
+ gen_helper_sve_st1hh_be_r_mte,
279
+ gen_helper_sve_st1hs_be_r_mte,
280
+ gen_helper_sve_st1hd_be_r_mte },
281
+ { NULL, NULL,
282
+ gen_helper_sve_st1ss_be_r_mte,
283
+ gen_helper_sve_st1sd_be_r_mte },
284
+ { NULL, NULL, NULL,
285
+ gen_helper_sve_st1dd_be_r_mte } } },
286
};
287
- static gen_helper_gvec_mem * const fn_multiple[2][3][4] = {
288
- { { gen_helper_sve_st2bb_r,
289
- gen_helper_sve_st2hh_le_r,
290
- gen_helper_sve_st2ss_le_r,
291
- gen_helper_sve_st2dd_le_r },
292
- { gen_helper_sve_st3bb_r,
293
- gen_helper_sve_st3hh_le_r,
294
- gen_helper_sve_st3ss_le_r,
295
- gen_helper_sve_st3dd_le_r },
296
- { gen_helper_sve_st4bb_r,
297
- gen_helper_sve_st4hh_le_r,
298
- gen_helper_sve_st4ss_le_r,
299
- gen_helper_sve_st4dd_le_r } },
300
- { { gen_helper_sve_st2bb_r,
301
- gen_helper_sve_st2hh_be_r,
302
- gen_helper_sve_st2ss_be_r,
303
- gen_helper_sve_st2dd_be_r },
304
- { gen_helper_sve_st3bb_r,
305
- gen_helper_sve_st3hh_be_r,
306
- gen_helper_sve_st3ss_be_r,
307
- gen_helper_sve_st3dd_be_r },
308
- { gen_helper_sve_st4bb_r,
309
- gen_helper_sve_st4hh_be_r,
310
- gen_helper_sve_st4ss_be_r,
311
- gen_helper_sve_st4dd_be_r } },
312
+ static gen_helper_gvec_mem * const fn_multiple[2][2][3][4] = {
313
+ { { { gen_helper_sve_st2bb_r,
314
+ gen_helper_sve_st2hh_le_r,
315
+ gen_helper_sve_st2ss_le_r,
316
+ gen_helper_sve_st2dd_le_r },
317
+ { gen_helper_sve_st3bb_r,
318
+ gen_helper_sve_st3hh_le_r,
319
+ gen_helper_sve_st3ss_le_r,
320
+ gen_helper_sve_st3dd_le_r },
321
+ { gen_helper_sve_st4bb_r,
322
+ gen_helper_sve_st4hh_le_r,
323
+ gen_helper_sve_st4ss_le_r,
324
+ gen_helper_sve_st4dd_le_r } },
325
+ { { gen_helper_sve_st2bb_r,
326
+ gen_helper_sve_st2hh_be_r,
327
+ gen_helper_sve_st2ss_be_r,
328
+ gen_helper_sve_st2dd_be_r },
329
+ { gen_helper_sve_st3bb_r,
330
+ gen_helper_sve_st3hh_be_r,
331
+ gen_helper_sve_st3ss_be_r,
332
+ gen_helper_sve_st3dd_be_r },
333
+ { gen_helper_sve_st4bb_r,
334
+ gen_helper_sve_st4hh_be_r,
335
+ gen_helper_sve_st4ss_be_r,
336
+ gen_helper_sve_st4dd_be_r } } },
337
+ { { { gen_helper_sve_st2bb_r_mte,
338
+ gen_helper_sve_st2hh_le_r_mte,
339
+ gen_helper_sve_st2ss_le_r_mte,
340
+ gen_helper_sve_st2dd_le_r_mte },
341
+ { gen_helper_sve_st3bb_r_mte,
342
+ gen_helper_sve_st3hh_le_r_mte,
343
+ gen_helper_sve_st3ss_le_r_mte,
344
+ gen_helper_sve_st3dd_le_r_mte },
345
+ { gen_helper_sve_st4bb_r_mte,
346
+ gen_helper_sve_st4hh_le_r_mte,
347
+ gen_helper_sve_st4ss_le_r_mte,
348
+ gen_helper_sve_st4dd_le_r_mte } },
349
+ { { gen_helper_sve_st2bb_r_mte,
350
+ gen_helper_sve_st2hh_be_r_mte,
351
+ gen_helper_sve_st2ss_be_r_mte,
352
+ gen_helper_sve_st2dd_be_r_mte },
353
+ { gen_helper_sve_st3bb_r_mte,
354
+ gen_helper_sve_st3hh_be_r_mte,
355
+ gen_helper_sve_st3ss_be_r_mte,
356
+ gen_helper_sve_st3dd_be_r_mte },
357
+ { gen_helper_sve_st4bb_r_mte,
358
+ gen_helper_sve_st4hh_be_r_mte,
359
+ gen_helper_sve_st4ss_be_r_mte,
360
+ gen_helper_sve_st4dd_be_r_mte } } },
361
};
362
gen_helper_gvec_mem *fn;
363
int be = s->be_data == MO_BE;
364
365
if (nreg == 0) {
366
/* ST1 */
367
- fn = fn_single[be][msz][esz];
368
+ fn = fn_single[s->mte_active[0]][be][msz][esz];
369
+ nreg = 1;
370
} else {
371
/* ST2, ST3, ST4 -- msz == esz, enforced by encoding */
372
assert(msz == esz);
373
- fn = fn_multiple[be][nreg - 1][msz];
374
+ fn = fn_multiple[s->mte_active[0]][be][nreg - 1][msz];
375
}
376
assert(fn != NULL);
377
- do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), 0, true, fn);
378
+ do_mem_zpa(s, zt, pg, addr, msz_dtype(s, msz), nreg, true, fn);
379
}
380
381
static bool trans_ST_zprr(DisasContext *s, arg_rprr_store *a)
382
--
383
2.20.1
384
385
diff view generated by jsdifflib